scispace - formally typeset
Search or ask a question

Showing papers in "Online Information Review in 2005"


Journal ArticleDOI
TL;DR: There are massive content omissions presently but it feels that Google Scholar will become an excellent free tool for scholarly information discovery and retrieval with future changes in its structure.
Abstract: Purpose – To identify the pros and the cons of Google Scholar.Design/methodology/approach – Chronicles the recent history of the Google Scholar search engine from its inception in November 2004 and critiques it with regard to its merits and demerits.Findings – Feels that there are massive content omissions presently but that, with future changes in its structure, Google Scholar will become an excellent free tool for scholarly information discovery and retrieval.Originality/value – Presents a useful analysis for potential users of the Google Scholar site.

296 citations


Journal ArticleDOI
TL;DR: This chapter discusses the development of data semantics in heterogeneous database federations GIS and Spatiotemporal Databases, and approaches to the document versioning issue in digital libraries.
Abstract: Active Issues in Databases 6 Business Rules in Databases 60 Managing Inconsistent Databases Using Active Integrity Constraints 87 Active services for federated databases 93 Active Database Management Systems 131 Real-Time Databases Databases and Digital Libraries Technologies 1 Text Categorization 42 Approaches to the document versioning issue in digital libraries 130 An XML Multi-tier Pattern Dissemination System Document Databases 7 Text Databases 95 Signature Files and Signature File Construction 108 On the Path-Oriented Query Evaluation in Document Databases Distributed Databases Distributed databases and Replication issues 21 Checking Integrity Constraints in a Distributed Database 26 Evaluation of the Most Important Replication Methods 50 Replication Mechanisms Over a Set of Distributed UDDI Registries 51 An Overview of Database Replication Approaches 111 Querical Data Networks Interoperability Concerns 20 Ontology-based Data Integration: Different Approaches and Common Features 76 Collective Knowledge Composition in a P2P Network 106 Semantic Information Management with A Common Business Language 110 Integration of data semantics in heterogeneous database federations GIS and Spatiotemporal Databases

291 citations


Journal ArticleDOI
TL;DR: In this paper, the authors study the development of eGovernment initiatives at the regional and local level in the EU through the opinion of those agents directly involved in the projects, finding that eGovernment is not likely to remodel governance in the short term, since democracy initiatives are not on the present agenda of most EU countries.
Abstract: Purpose – The objective of this paper is to study the development of e‐government initiatives at the regional and local level in the EU through the opinion of those agents directly involved in the projects.Design/methodology/approach – The authors wrote a questionnaire that was sent to the regions and the largest cities of EU countries, in order to find out their degree of involvement in e‐government initiatives. Responses were received from 47 regional and local governments.Findings – At present, most regional and local governments have a web site, although in most cases it is little more than a governmental billboard. E‐government is not likely to remodel governance in the short term, since democracy initiatives are not on the present agenda of most EU countries. The survey findings show that e‐government initiatives are still predominantly non‐interactive and non‐deliberative. They tend to reflect present service delivery patterns, not transform them.Research limitations/implications – The construction...

186 citations


Journal ArticleDOI
TL;DR: To classify the most important metrics proposed for web information systems, it was found that about 44 percent of metrics are related to “presentation” and that most metrics (48 percent) are usability metrics.
Abstract: Purpose – The purpose of this paper is to classify the most important metrics proposed for web information systems, with the aim of offering the user a global vision of the state of the research within this area.Design/methodology/approach – WQM distinguishes three dimensions related to web features, lifecycle processes and quality characteristics. A range of recently published (1992‐2004) works that include web metrics definitions have been studied and classified within this model.Findings – In this work, a global vision of web metrics is provided. Concretely, it was found that about 44 percent of metrics are related to “presentation” and that most metrics (48 percent) are usability metrics. Regarding the life cycle, the majority of metrics are related to operation and maintenance processes. Nevertheless, focusing on metrics validation, it was found that there is not too much work done, with only 3 percent of metrics validated theoretically and 37 percent of metrics validated empirically.Practical implic...

143 citations


Journal ArticleDOI
TL;DR: The paper analyses the impact of the open access movement, which came to fruition after the OAI Metadata Harvesting Protocol was established, as it creates the potential for interoperability between e‐print repositories and outlines the challenges for information managers in developing the full potential of open access.
Abstract: Purpose – The purpose of this paper is to provide an overview of the continuous evolution of scholarly publishing and knowledge communication as a result of the internet revolution.Design/methodology/approach – Information was obtained from a literature review of the main contributions on “self‐archiving” – the broad term often applied to electronic publishing of author‐supplied documents on the web without commercial publisher mediation. The paper analyses the impact of the open access movement, which came to fruition after the OAI Metadata Harvesting Protocol was established, as it creates the potential for interoperability between e‐print repositories. It concludes by outlining the challenges for information managers in developing the full potential of open access.Findings – With regard to the future of self‐archiving, particularly in relation to peer‐reviewed journals, information managers have a very important role to perform within their organization.Originality/value – The paper highlights the bene...

76 citations


Journal ArticleDOI
TL;DR: Results include the findings that web users tend to input short queries when searching for visual or textual information, and that image requests have more zero hits and higher specificity, and contain more refined queries.
Abstract: Purpose – This study aims to examine the differences between web image and textual queries.Design/methodology/approach – A large number of web queries from image and textual search engines were analysed and compared based on their factual characteristics, query types, and search interests.Findings – Useful results include the findings that web users tend to input short queries when searching for visual or textual information; that image requests have more zero hits and higher specificity, and contain more refined queries; that web image requests are more focused than textual requests on some popular search interests, and that the variety of textual queries is greater than that of image requests.Originality/value – This study provides results that may enhance one's understanding of web‐searching behaviour and the inherent implications for the improvement of current web image retrieval systems.

48 citations


Journal ArticleDOI
TL;DR: The challenges and problems underlying existing work in invisible web search interfaces are identified, including automated discovery of invisible web site search interfaces, automated classification of invisibleweb sites, label assignment and form filling, and information extraction from the resulting pages.
Abstract: Purpose – The existence and continued growth of the invisible web creates a major challenge for search engines that are attempting to organize all of the material on the web into a form that is easily retrieved by all users. The purpose of this paper is to identify the challenges and problems underlying existing work in this area.Design/methodology/approach – A discussion based on a short survey of prior work, including automated discovery of invisible web site search interfaces, automated classification of invisible web sites, label assignment and form filling, information extraction from the resulting pages, learning the query language of the search interface, building content summary for an invisible web site, selecting proper databases, integrating invisible web‐search interfaces, and accessing the performance of an invisible web site.Findings – Existing technologies and tools for indexing the invisible web follow one of two strategies: indexing the web site interface or examining a portion of the con...

45 citations


Journal ArticleDOI
TL;DR: The differentiation of the marketplace includes unique ways to collect user‐based information and customized, proprietary algorithms that generate what are believed to be the best matches, based on user and matchmaking service criteria.
Abstract: Purpose – To explore and add insight to the online‐dating services phenomena which is the next product and beneficiary of the internet revolution that offers customers a convenient and affordable alternate to traditional methods of dating.Design/methodology/approach – Empirically investigated through conceptual models and statistical methods was the value proposition of online matchmaking services, which boils down to the ability to provide appropriate matches through successfully business‐to‐customer (B2C) customer service enhanced by the web and based on sound customer relations management practices.Findings – The differentiation of the marketplace includes unique ways to collect user‐based information and customized, proprietary algorithms that generate what are believed to be the best matches, based on user and matchmaking service criteria. Online dating services use statistics, data mining, and activity monitoring to provide appropriate matches; thus, differentiating their services and understanding ...

39 citations


Journal ArticleDOI
TL;DR: The study found that there were strong similarities between justifications and reported motives and noted that justifications concerning financial, work‐life and organisational‐mission issues were relatively minor.
Abstract: Purpose – This research seeks to investigate the introduction of new information and communication technology systems and to describe the development of a conceptual model of enterprise resource‐planning systems adoption based on the published rationales organizations use to justify their adoption.Design/methodology/approach – The study uses a grounded theory approach to building the conceptual model from electronically distributed documents. These documents were selected from a sample of universities which adopted enterprise resource‐planning systems.Findings – This paper reports on the use of grounded theory in the internet context. The study found that there were strong similarities between justifications and reported motives. The study noted that justifications concerning financial, work‐life and organisational‐mission issues were relatively minor.Research limitations/implications – This model is built on published justifications, which should not be confused with motives. This picture may distort rea...

36 citations


Journal ArticleDOI
TL;DR: Two user types have been derived from the data analysis of library users' information behaviour on both sides of the information coin – information usage and information production/publishing.
Abstract: Purpose – Aims to study human information behaviour as part of the research project on the interaction of man and the information environment (project VEGA 1/9236/02) and to analyse library users' information behaviour on both sides of the information coin – information usage and information production/publishing.Design/methodology/approach – The methodology was composed of the following steps: pilot studies of students' information‐seeking behaviour; a large‐scale questionnaire survey of users of academic and research libraries; data analyses and interpretations; verification of hypotheses; multiple data sorting; and modelling of user groups.Findings – Library users appreciate easy access and well‐organised forms of information, with an emphasis on electronic sources. In their capacity as authors of professional papers, only few subjects considered print and electronic publishing to be equal. Two user types have been derived from the data analysis. Type S manifests pragmatic ways of information seeking a...

33 citations


Journal ArticleDOI
TL;DR: An artificial neural network is proposed to identify automatically topic changes in a user session by using the statistical characteristics of queries, such as time intervals and query reformulation patterns to identify topic changes within a user search session.
Abstract: Purpose – This study aims to propose an artificial neural network to identify automatically topic changes in a user session by using the statistical characteristics of queries, such as time intervals and query reformulation patterns.Design/methodology/approach – A sample data log from the Norwegian search engine FAST (currently owned by Overture) is selected to train the neural network and then the neural network is used to identify topic changes in the data log.Findings – A total of 98.4 percent of topic shifts and 86.6 percent of topic continuations were estimated correctly.Originality/value – Content analysis of search engine user queries is an important task, since successful exploitation of the content of queries can result in the design of efficient information retrieval algorithms for search engines, which can offer custom‐tailored services to the web user. Identification of topic changes within a user search session is a key issue in the content analysis of search engine user queries.

Journal ArticleDOI
TL;DR: The current paper is centered on the assumption that I‐voting is the next logical step in applying online information‐gathering and retrieval technologies to the field of e‐government.
Abstract: Purpose – To provide practitioners of information management with an overview and framework to explore the various controversies associated with the methods of traditional methods of voting with electronically enhanced voting via the internet (I‐voting). The current paper is centered on the assumption that I‐voting is the next logical step in applying online information‐gathering and retrieval technologies to the field of e‐government.Design/methodology/approach – A review of the applied literature on electronically enhanced voting methods with potentials associated with internet voting, as well as from practical experience, resulted in a basic model for discussion of the emergent nature of I‐voting.Findings – I‐voting would reduce the cost for staffing polling stations, and also the funds needed to pay for voting machines. I‐voting could also reduce the number of errors made by both the voters and the electoral administrators, and allow for easier adoption of uniform standards in the ballot format, since...

Journal ArticleDOI
TL;DR: It was found that many students were initially unfamiliar with many of the databases important to them and that the familiarity developed during the year contributed importantly to their development of information search expertise.
Abstract: Purpose – This paper sets out to report on a study of 12 postgraduate research students' development of information search expertise. It focuses on students' perceived importance and growth of knowledge of different databases as they progress through their studies.Design/methodology/approach – Taking an in‐depth longitudinal approach, this study makes use of surveys, direct observations of students' searching of various databases, students' verbalization of their thoughts while searching databases and interviews during a one‐year period. The paper focuses on analyzing the data from surveying and interviewing the students with regard to their use of databases.Findings – It was found that many students were initially unfamiliar with many of the databases important to them and that the familiarity developed during the year contributed importantly to their development of information search expertise. This suggests that much work has to be done to inform and instruct students about these “treasures”.Research l...

Journal ArticleDOI
TL;DR: The author's experience as a wiki participant on several occasions has been far from positive, and the author's advice to users of wikis: use with caution is provided.
Abstract: Purpose – To provide a brief viewpoint on wikis. A wiki is a web site that allows – and encourages – users to share information by freely writing new content, adding to existing content, and editing or commenting on content. Design/methodology/approach – An opinion piece based on the author's own experiences. Findings – Reflecting the chaotic nature of the web world in which wikis exist, the reality of the situation is that they often do not operate in a positive environment. The author's experience as a wiki participant on several occasions has been far from positive. Originality/value – Provides advice to users of wikis: use with caution.

Journal ArticleDOI
TL;DR: The findings showed that user roles and the level of user involvement did have a significant effect on users' perceptions of requirements, implying that high involvers (core members) are more appropriate for requirement elicitation, even though high involver might not be the only candidates for requirements elicitation.
Abstract: Purpose – The purpose of this study is to investigate the relationship between users' role and users' perception of requirements in a web‐based community of practice.Design/methodology/approach – Online survey and social network analysis were adopted in this study.Findings – The study classified user roles as advisors and two types of regular users (both aware of advisors and not); each type of user demonstrated different sets of requirement perceptions. The findings showed that user roles and the level of user involvement did have a significant effect on users' perceptions of requirements. Advisors paid less attention to the interface, communication and recommendation links than the regular users.Practical implications – It implies that high involvers (core members) are more appropriate for requirement elicitation, even though high involvers might not be the only candidates for requirements elicitation.Originality/value – Few articles discuss the relationship between user roles and requirements; therefor...

Journal ArticleDOI
TL;DR: An investigation into the ways in which Canadian digital library collections have incorporated knowledge organization systems into their search interfaces is reported.
Abstract: Purpose – The purpose of this paper is to report an investigation into the ways in which Canadian digital library collections have incorporated knowledge organization systems into their search interfaces.Design/methodology/approach – A combination of data‐gathering techniques was used. These were as follows: a review of the literature related to the application of knowledge organization systems, deep scanning of Canadian governmental and academic institutions web sites on the web, identify and contact researchers in the area of knowledge organization, and identify and contact people in the governmental organizations who are involved in knowledge organization and information management.Findings – A total of 33 digital collections were identified that have made use of some type of knowledge organization system. Thesauri, subject heading lists and classification schemes were the widely used knowledge organization systems in the surveyed Canadian digital library collections.Research limitations/implications –...

Journal ArticleDOI
TL;DR: A methodological procedure based on keyword analysis and the structuring of data into information systems in the field of functional foods, a newly‐emerging scientific field within the broader scope of food sciences and technology is addressed.
Abstract: Purpose – The article seeks to address a methodological procedure based on keyword analysis and the structuring of data into information systems in the field of functional foods, a newly‐emerging scientific field within the broader scope of food sciences and technology.Design/methodology/approach – An experiment was undertaken by selection of a research field or research subject, selection of search profile, selection and processing of relevant databases, keyword analysis, and the arrangement of data (keywords) according to tree‐structures. Keyword analysis was employed to identify narrower research fields within the broader scientific field. The structuring of data into systems was used to classify the terms within the particular narrow field. Keywords with higher and lower frequency were identified. A classification tree was set up, based on keywords (thesaurus‐based descriptors) extracted from the FSTA (Food Science and Technology Abstracts) database available online. The tree was supplemented and upgr...

Journal ArticleDOI
TL;DR: It is found that overlap is minimal among web‐wide search engines which crawl and index the mostly unstructured open web; and that overlap among Google Scholar and the native search engines is far less than the ideal 100 per cent in the optimal context of crawling and indexing highly‐structured and metadata‐rich collections.
Abstract: Purpose – To compare the performance of different search engines, highlighting the overlap and rank differences.Design/methodology/approach – Presents results of an overlap test search series among traditional CD‐ROM indexing/abstracting databases since the mid‐1980s, web databases, and authors' own polysearch engine, and reviews Google Scholar.Findings – Finds that overlap is minimal among web‐wide search engines which crawl and index the mostly unstructured open web; and that overlap among Google Scholar and the native search engines is far less than the ideal 100 per cent in the optimal context of crawling and indexing highly‐structured and metadata‐rich collections.Originality/value – Reinforces the existing view that for comprehensive searches one must search more than one database. Highlights and recommends several very good search engine sites.

Journal ArticleDOI
TL;DR: There were a certain number of problematic search behavior patterns and obstacles uncovered, which proved to have a critical impact on the search results, such as: difficulties in acquiring new or alternative vocabulary during the search process, perception of the task as the ultimate source of keywords, and perception of commercial sources as unavailable.
Abstract: Purpose – The purpose of the research was to create internet search instructions, to test their effectiveness and to track the search behavior of the first year MLS studentsDesign/methodology/approach – The students received two exercises with similar search tasks. No specific search guidelines were delivered to the students prior to the first exercise. Prior to the second exercise students received a lecture based on the internet search instructions developed by the authors. The results of the two searches were analyzed and compared.Findings – As a result of the lecture students significantly improved the search results. There were a certain number of problematic search behavior patterns and obstacles uncovered, which proved to have a critical impact on the search results, such as: difficulties in acquiring new or alternative vocabulary during the search process, perception of the task as the ultimate source of keywords, and perception of commercial sources as unavailable.Research limitations/implication...


Journal ArticleDOI
TL;DR: An evaluation of the web as a source of scientific bibliographic information with respect to the quantity and quality of retrieved documents and that retrieved through search engines indicates that the web, assuming a professional use of the medium, is a good source of information.
Abstract: Purpose – The aim of this study was an evaluation of the web as a source of scientific bibliographic information.Design/methodology/approach – In order to give this evaluation a quantitative dimension, a comparison with the information obtained from fee‐based bibliographic databases was performed. Based on a concrete search example in the field of plant production biotechnology, a comparative study of selected fee‐based bibliographic databases (CAB Abstracts and Biosis Preview) and a search of the web with selected search engines (Scirus and Google) was carried out. A comparison of the information retrieved through the databases and that retrieved through search engines was conducted with respect to the quantity and quality of retrieved documents, search time, the cost of information, retrieval strategies, the reliability of information and the demands on the skills of the searcher.Findings – The surprising results of this comparison clearly indicate that the web, assuming a professional use of the medium...


Journal ArticleDOI
TL;DR: The design and implementation of the Query Graph Visualizer (QGV), a collaborative querying system which harvests and clusters previously issued queries to form query networks that represent related information needs are described.
Abstract: Purpose – Information overload has led to a situation where users are swamped with too much information, resulting in difficulty sifting through material in search of relevant content. Aims to address this issue from the perspective of collaborative querying, an approach that helps users formulate queries by harnessing the collective knowledge of other searchers.Design/methodology/approach – The design and implementation of the Query Graph Visualizer (QGV), a collaborative querying system which harvests and clusters previously issued queries to form query networks that represent related information needs are described. A preliminary evaluation of the QGV is also described in which a group of participants evaluated the usability and usefulness of the system by completing a set of tasks and a questionnaire based on Nielsen's heuristic evaluation technique.Findings – In the QGV, a submitted query is matched to its closest cluster and a recursive algorithm is applied to find other related clusters, forming a ...

Journal ArticleDOI
TL;DR: The implementation of local online information systems; OPACs, CD‐ROM databases, online searching, networks/databases subscribed to, methods of online services to users, problems and solutions are investigated.
Abstract: Purpose – The study is concerned with gathering factual data on the use of local online information systems, automation, online connections, online public access catalogs (OPACs), CD‐ROM‐based systems in 19 Jordanian university libraries.Design/methodology/approach – Information was sought through a written questionnaire; personal interviews; literature review, and field visits. The implementation of local online information systems; OPACs, CD‐ROM databases, online searching, networks/databases subscribed to, methods of online services to users, problems and solutions are investigated.Findings – Of the 18 (94.7 percent) libraries responded, half (50 percent) are using MINISIS or CDS/ISIS, one‐third (33.3 percent) is using either M2L, or ORACLE based systems, and 3 (16.7 percent) are using locally designed packages. Technical processing, information retrieval, circulation, reference services, and serial control are the prime systems functions. All respondents are linked to the internet. A total of 15 (83.3...

Journal ArticleDOI
TL;DR: The Yahoo! set showed generally no significant difference in inclusion of descriptions and keywords between generator‐identifying and other pages, but the geocities.com set did show a significant difference for both keywords and descriptions.
Abstract: Purpose – To determine the effect of web page editing tools on inclusion and page specificity of meta tagged descriptions and keywords.Design/methodology/approach – Using customized software with Yahoo!'s random page service, data from 2,048 URLs were logged. Generator identification was cross‐tabulated with presence and length of both descriptions and keywords. A second analysis on pages on geocities.com was performed using URLs from Altavista. Local links from a sample of the Yahoo! set were followed and linked‐to pages were examined for presence of description or keywords and whether these differed from those on the linking pages.Findings – The Yahoo! set showed generally no significant difference in inclusion of descriptions and keywords between generator‐identifying and other pages. The geocities.com set did show a significant difference for both keywords and descriptions. Exact repetition of descriptions or keywords between pages on the same site did not generally correlate significantly with identi...


Journal ArticleDOI
TL;DR: It is found that citation counts and impact factors can be easily manipulated.
Abstract: Purpose – This issue aims to look into the subject of citation counts and whether they have an impact on the articles being cited.Design/methodology/approach – Uses Selgens list of citations that have been counted within such areas as: review articles, long articles, short publications and research fields as an example.Findings – Finds that citation counts and impact factors can be easily manipulated.Originality/value – Gives a strong view to a controversial subject.

Journal ArticleDOI
TL;DR: This article is the first in a four‐part series that aims to illustrate the processes involved in planning a portal and creating a portal definition document and will be useful to the information professional who is contemplating portal development.
Abstract: Purpose – This article is the first in a four‐part series that aims to illustrate the processes involved in planning a portal and creating a portal definition document.Design/methodology/approach – Based on a review of the academic and industry literature and using a case study, the authors share their experience in planning the Florida ExpertNet Research portal.Findings – Portal development is a complex and costly endeavor that requires meticulous planning and design. As with any system development initiative, the time and energy spent in the planning process is reflected in the success or failure of the end product. Creating a portal vision is the first step in developing the portal definition document and includes the following areas: mission statement, objectives, needs assessment, creative layout, benefits, and funding and sponsorship.Originality/value – This series will be useful to the information professional who is contemplating portal development and may be used as a model in developing a bluepr...

Journal ArticleDOI
TL;DR: The first year of implementation of the Innovative Interfaces Inc.
Abstract: Purpose – This paper aims to describe the first year of implementation of the Innovative Interfaces Inc. Millennium Access Plus (MAP) portal at the University of Exeter Library, UK.Design/methodology/approach – The paper provides a description of the features of MAP and a case studyFindings – In as little as 25 years, online information provision has been transformed. This has largely been uncontrolled, yet the resources now available offer rich returns for the data‐hungry end user. Increasingly, many users require their information immediately and really do not care where it is stored. Such demands, coupled with librarians’ natural tendency to apply order, have led to the development of library portals which aim to solve the problems arising from trying to navigate this labyrinth of information. The MAP portal is one such product. Using the NISO OpenURL standard, MAP offers three main elements – contextual linking via WebBridge, one step “multi” searching using MetaFind and resource authentication throug...

Journal ArticleDOI
TL;DR: It became evident during its development and integration stages that MyLibrary@Lehigh would greatly enhance the success and usage of the campus portal and has become the one‐stop shop for the library within the campus‐wide, one‐ stop shop of thecampus portal.
Abstract: Purpose – In August 2004, the Library Collections and Systems team at Lehigh University released MyLibrary @ Lehigh within the campus portal to the university community. The purpose of this article is to explain how what began as an integration strategy of the library's electronic resources into one complete stand‐alone application became the library's response to the fast‐growing campus portal.Design/methodology/approach – Explains how MyLibrary@Lehigh was developed and implemented.Findings – It became evident during its development and integration stages that MyLibrary@Lehigh would greatly enhance the success and usage of the campus portal. As a repository of all of the library's electronic resources, MyLibrary@Lehigh has become the one‐stop shop for the library within the campus‐wide, one‐stop shop of the campus portal.Originality/value – MyLibrary has become a clear choice as an open‐source solution.