scispace - formally typeset
Search or ask a question

Showing papers on "Data management published in 2002"


Proceedings ArticleDOI
03 Jun 2002
TL;DR: This paper shows that XML's ordered data model can indeed be efficiently supported by a relational database system, and proposes three order encoding methods that can be used to represent XML order in the relational data model, and also proposes algorithms for translating ordered XPath expressions into SQL using these encoding methods.
Abstract: XML is quickly becoming the de facto standard for data exchange over the Internet. This is creating a new set of data management requirements involving XML, such as the need to store and query XML documents. Researchers have proposed using relational database systems to satisfy these requirements by devising ways to "shred" XML documents into relations, and translate XML queries into SQL queries over these relations. However, a key issue with such an approach, which has largely been ignored in the research literature, is how (and whether) the ordered XML data model can be efficiently supported by the unordered relational data model. This paper shows that XML's ordered data model can indeed be efficiently supported by a relational database system. This is accomplished by encoding order as a data value. We propose three order encoding methods that can be used to represent XML order in the relational data model, and also propose algorithms for translating ordered XPath expressions into SQL using these encoding methods. Finally, we report the results of an experimental study that investigates the performance of the proposed order encoding methods on a workload of ordered XML queries and updates.

2,402 citations


Book ChapterDOI
20 Aug 2002
TL;DR: This paper presents Aurora, a new DBMS that is currently under construction at Brandeis University, Brown University, and M.I.T. and describes the basic system architecture, a stream-oriented set of operators, optimization tactics, and support for real-time operation.
Abstract: This paper introduces monitoring applications, which we will show differ substantially from conventional business data processing. The fact that a software system must process and react to continual inputs from many sources (e.g., sensors) rather than from human operators requires one to rethink the fundamental architecture of a DBMS for this application area. In this paper, we present Aurora, a new DBMS that is currently under construction at Brandeis University, Brown University, and M.I.T. We describe the basic system architecture, a stream-oriented set of operators, optimization tactics, and support for real-time operation.

963 citations


Book ChapterDOI
20 Aug 2002
TL;DR: This work provides a framework to assess the abilities of an XML database to cope with a broad range of different query types typically encountered in real-world scenarios and offers a set of queries where each query is intended to challenge a particular aspect of the query processor.
Abstract: While standardization efforts for XML query languages have been progressing, researchers and users increasingly focus on the database technology that has to deliver on the new challenges that the abundance of XML documents poses to data management: validation, performance evaluation and optimization of XML query processors are the upcoming issues. Following a long tradition in database research, we provide a framework to assess the abilities of an XML database to cope with a broad range of different query types typically encountered in real-world scenarios. The benchmark can help both implementors and users to compare XML databases in a standardized application scenario. To this end, we offer a set of queries where each query is intended to challenge a particular aspect of the query processor. The overall workload we propose consists of a scalable document database and a concise, yet comprehensive set of queries which covers the major aspects of XML query processing ranging from textual features to data analysis queries and ad hoc queries. We complement our research with results we obtained from running the benchmark on several XML database platforms. These results are intended to give a first baseline and illustrate the state of the art.

822 citations


Proceedings ArticleDOI
07 Aug 2002
TL;DR: A novel paradigm for data management in which a third party service provider hosts "database as a service", providing its customers with seamless mechanisms to create, store, and access their databases at the host site is explored.
Abstract: We explore a novel paradigm for data management in which a third party service provider hosts "database as a service", providing its customers with seamless mechanisms to create, store, and access their databases at the host site. Such a model alleviates the need for organizations to purchase expensive hardware and software, deal with software upgrades, and hire professionals for administrative and maintenance tasks which are taken over by the service provider. We have developed and deployed a database service on the Internet, called NetDB2, which is in constant use. In a sense, a data management model supported by NetDB2 provides an effective mechanism for organizations to purchase data management as a service, thereby freeing them to concentrate on their core businesses. Among the primary challenges introduced by "database as a service" are the additional overhead of remote access to data, an infrastructure to guarantee data privacy, and user interface design for such a service. These issues are investigated. We identify data privacy as a particularly vital problem and propose alternative solutions based on data encryption. The paper is meant as a challenge for the database community to explore a rich set of research issues that arise in developing such a service.

707 citations


Patent
30 Jan 2002
TL;DR: In this paper, resource utilization information is collected from one or more subsystems or processing engines of an information management system, and analyzed to gain insights into system performance, such as system capacity management.
Abstract: Logging and analysis of resource utilization information may be implemented, for example, to achieve intelligent capacity management of information management system resources. In one implementation, resource utilization information may be collected from one or more subsystems or processing engines of an information management system, and analyzed to gain insights into system performance. System capacity management may be implemented using collected resource utilization information, for example, to achieve dynamic management of system resources.

661 citations


Journal ArticleDOI
01 May 2002
TL;DR: A high-speed transport service that extends the popular FTP protocol with new features required for Data Grid applications, such as striping and partial file access and a replica management service that integrates a replica catalog with GridFTP transfers to provide for the creation, registration, location, and management of dataset replicas.
Abstract: An emerging class of data-intensive applications involve the geographically dispersed extraction of complex scientific information from very large collections of measured or computed data. Such applications arise, for example, in experimental physics, where the data in question is generated by accelerators, and in simulation science, where the data is generated by supercomputers. So-called Data Grids provide essential infrastructure for such applications, much as the Internet provides essential services for applications such as e-mail and the Web. We describe here two services that we believe are fundamental to any Data Grid: reliable, high-speed transport and replica management. Our high-speed transport service, GridFTP, extends the popular FTP protocol with new features required for Data Grid applications, such as striping and partial file access. Our replica management service integrates a replica catalog with GridFTP transfers to provide for the creation, registration, location, and management of dataset replicas. We present the design of both services and also preliminary performance results. Our implementations exploit security and other services provided by the Globus Toolkit.

633 citations


Proceedings Article
01 Jan 2002
TL;DR: This work motivates special database problems introduced by peer-to-peer computing and proposes the Local Relational Model (LRM) to solve some of them and summarizes a formalization of LRM.
Abstract: We motivate special database problems introduced by peer-to-peer computing and propose the Local Relational Model (LRM) to solve some of them As well, we summarize a formalization of LRM, present an architecture for a prototype implementation, and discuss open research questions

419 citations


Book ChapterDOI
01 Jan 2002
TL;DR: Knowledge management systems, Knowledge management systems , مرکز فناوری اطلاعات و اصاع رسانی, کδاوρزی
Abstract: Knowledge management systems , Knowledge management systems , مرکز فناوری اطلاعات و اطلاع رسانی کشاورزی

416 citations


Journal ArticleDOI
TL;DR: In this paper, the authors explore baselines for moving from performance measurement to performance management and provide a discussion of how the FM performance assessment can be used to manage the FM function effectively.
Abstract: In order for a facilities management (FM) organisation to make effective use of the results of performance measurement it must be able to make the transition from measurement to management. It must also be able to anticipate needed changes in the strategic direction of the organisation and have a methodology in place for effecting strategic change. Successful accomplishment of these two tasks represents the foundation of good performance management. This paper explores baselines for moving from performance measurement to performance management and provides a discussion of how the FM performance assessment can be used to manage the FM function effectively.

346 citations


Journal ArticleDOI
01 Mar 2002
TL;DR: The Zoltan library simplifies the load-balancing, data movement, unstructured-communication, and memory usage difficulties that arise in dynamic applications such as adaptive finite-element methods, particle methods, and crash simulations.
Abstract: The Zoltan library is a collection of data management services for parallel, unstructured, adaptive, and dynamic applications that is available as open-source software. It simplifies the load-balancing, data movement, unstructured-communication, and memory usage difficulties that arise in dynamic applications such as adaptive finite-element methods, particle methods, and crash simulations. Zoltan's data-structure-neutral design also lets a wide range of applications use it without imposing restrictions on application data structures. Its object-based interface provides a simple and inexpensive way for application developers to use the library and researchers to make new capabilities available under a common interface.

307 citations



Journal ArticleDOI
TL;DR: The nature of Knowledge Management is examined - how it differs from Data Management and Information Management, and its relationship to the development of Expert Systems and Decision Support Systems is examined.
Abstract: This article examines the nature of Knowledge Management - how it differs from Data Management and Information Management, and its relationship to the development of Expert Systems and Decision Support Systems. It also examines the importance of Communities of Practice and Tacit Knowledge for Knowledge Management. The discussion is organized around five explicit questions. One: What is "knowledge"? Two: Why are people, especially managers, thinking about Knowledge Management? Three: What are the enabling technologies for Knowledge Management? Four: What are the prerequisites for Knowledge Management? Five: What are the major challenges for Knowledge Management?

Journal ArticleDOI
TL;DR: In this article, the authors present a framework for selecting a knowledge management strategy that is appropriate to the organisational and cultural context of an organization, based on a detailed study of current knowledge management processes in the construction and manufacturing sectors.

Patent
15 May 2002
TL;DR: In this article, an apparatus and method for patient point-of-care data management is presented, where patient point of care data is obtained in real-time or in substantially realtime and is made available to subscribers or authorized users on either a local computer display screen, a remote computer screen, or both.
Abstract: An apparatus and method for patient point-of-care data management is provided. Patient point-of-care data is obtained in realtime or in substantially realtime and is made available to subscribers or authorized users on either a local computer display screen, a remote computer display screen, or both. In some embodiments, data associated with a plurality of patients is made available and is viewable simultaneously on a computer display screen. In some embodiments, patient point-of-care data from a plurality of hospitals is available.

01 Jan 2002
TL;DR: The requirements and innovative technology for an integrated disaster management communication and information system are sketched, addressing in particular network, configuration, scheduling and data management issues during the response and recovery phases.
Abstract: Disaster response and recovery efforts require timely interaction and coordination of public emergency services in order to save lives and property. Today, IT is used in this field only to a limited extent, but there is a tremendous potential for increasing efficiency and effec- tiveness in coping with a disaster. In this paper we sketch requirements and innovative technology for an integrated disaster management communication and information system, addressing in particular network, configuration, scheduling and data management issues during the response and recovery phases.

Book ChapterDOI
20 Aug 2002
TL;DR: ProTDB as mentioned in this paper is a probabilistic tree data base for XML query processing, which is based on the Probabilistic Tree Data Base (PTDB) model for relational databases.
Abstract: Where as traditional databases manage only deterministic information, many applications that use databases involve uncertain data. This paper presents a Probabilistic Tree Data Base (ProTDB) to manage probabilistic data, represented in XML. Our approach differs from previous efforts to develop probabilistic relational systems in that we build a probabilistic XML database. This design is driven by application needs that involve data not readily amenable to a relational representation. XML data poses several modeling challenges: due to its structure, due to the possibility of uncertainty association at multiple granularities, and due to the possibility of missing and repeated sub-elements. We present a probabilistic XML model that addresses all of these challenges. We devise an implementation of XML query operations using our probability model, and demonstrate the efficiency of our implementation experimentally. We have used ProTDB to manage data from two application areas: protein chemistry data from the bioinformatics domain, and information extraction data obtained from the web using a natural language analysis system. We present a brief case study of the latter to demonstrate the value of probabilistic XML data management.


Book
08 Jan 2002
TL;DR: This book reveals how constraint databases bring together techniques from a variety of fields, such as logic and model thoery, algebraic and computational geometry, and symbolic computation, to the design and analysis of data models and query languages.
Abstract: Introduction to Constraint Databases comprehensively covers both constraint-database theory and several sample systems. The book reveals how constraint databases bring together techniques from a variety of fields, such as logic and model thoery, algebraic and computational geometry, and symbolic computation, to the design and analysis of data models and query languages. Constraint databases are shown to be powerful and simple tools for data modeling and querying in application areas---such as environmental modeling, bioinformatics, and computer vision---that are not suitable for relational databases. Specific applications are examined in geographic information systems, spatiotemporal data management, linear programming, genome databases, model checking of automata, and other areas.

Journal ArticleDOI
TL;DR: In this paper, the importance of knowledge management for higher education, in general, and for institutional research, in particular, is discussed and a background chapter defines knowledge management, its components, processes, and outcomes.
Abstract: This background chapter defines knowledge management, its components, processes, and outcomes. It addresses the importance of knowledge management for higher education, in general, and for institutional research, in particular.

Journal ArticleDOI
TL;DR: In this article, the authors proposed a supply network with a global bill of material (BOM) model, which consists of four modules: supply network design optimization module, planning module for production and distribution operations from raw material suppliers to customers, model management module and data management module.

Patent
19 Apr 2002
TL;DR: In this article, a method and system for automating management of document data in a document management system is provided, which is suitably created by selecting from a variety of criteria permit users to automate tasks that are performed on document data which is preferably received by the document management systems.
Abstract: A method and system is provided for automating management of document data in a document management system. The document data management automation system suitably permits document management system users to create and define rules for automated management of document data in the document management system. The rules are suitably created by selecting from a variety of criteria permit users to automate tasks that are performed on document data which is preferably received by the document management system.

Patent
22 Nov 2002
TL;DR: In this article, an enterprise information unification system including an information modeler for modeling business entities and their properties as information model, the information model being comprised of business constructs, a data rationalizer for generating mappings from physical data assets to the information models, and a data thesaurus for automating tasks of data management based on the mappings.
Abstract: An enterprise information unification system including an information modeler for modeling business entities and their properties as an information model, the information model being comprised of business constructs, a data rationalizer for generating mappings from physical data assets to the information model, the physical data assets being comprised of data asset constructs, and a data thesaurus for automating tasks of data management based on the mappings. A method is also described and claimed.

Patent
29 Jan 2002
TL;DR: In this paper, a system and methodology for the acquisition of medical data from multiple, at-home patient measurement devices transfers the data via telephone communications to a remotely accessed data management and display system such as a Web-based data management system.
Abstract: A system and methodology for the acquisition of medical data from multiple, at-home patient measurement devices transfers the data via telephone communications to a remotely accessed data management and display system such as a Web-based data management system. The measurement device, which supports two-way communications between the patient and the database system, is extensible to support multiple plug-in medical devices and devices for patient input or identification. Collection of data is mediated through an interface device that is connected via a digital link (serial, network, or digital RF link) to single or multiple patient measurement devices. This interface device is connected to a network via wide area network means including standard phone lines. In the preferred embodiment, data transfer is mediated through pervasive DTMF phone networks. Although DTMF is the preferred format, alternative analog or digital audio tone formats may be used.

Patent
26 Feb 2002
TL;DR: In this article, a secure database stores risk management information that is accessible by authorized access through a network and a graphics interface generates graphic data of the risk management data in response to the authorized access.
Abstract: A graphical and interactive interface system manages risk management information. A secure database stores risk management information that is accessible by authorized access through a network. A graphics interface generates graphic data of the risk management information in response to the authorized access. One or more workflow process terminals connect in network with the database to provide updates to the risk management information. Summary reporting and statistical processing functionalities facilitate predictive accuracy of the system by permitting a user to compare relevant system inputs when selecting data to provide recommendations to customers for adjustment of insurance policies in accordance with risk management practices.

Journal Article
TL;DR: While the focus in information management is mostly on explicit knowledge, knowledge management brings a new dimension, the need to manage tacit knowledge by focusing on people and enhance their capability by improving communication, information transfer and collaboration.
Abstract: Knowledge management has generated much interest in recent years and has become the latest management buzz in town. Many people start wondering whether knowledge management is here to stay or it is just another consultancy fad. To put things in perspective, it is important to look at the sequence of events that led to the rise of knowledge management. The arrival of the information society and the move toward the knowledge-based economy highlighted the importance of tacit knowledge and the need to manage knowledge resources including skills and competencies. Knowledge management as a concept with people taking the centre stage has prompted us to rethink information management and shift focus from trying to develop intelligent systems to that of developing tools for intelligent people. It is this realization in my opinion that makes knowledge management attractive to many organizations. While the focus in information management is mostly on explicit knowledge, knowledge management brings a new dimension, the need to manage tacit knowledge by focusing on people and enhance their capability by improving communication, information transfer and collaboration.

Journal ArticleDOI
TL;DR: The prerequisites for treering studies are shown, on a conceptual and methodical level, and implications for measurements and data processing for most applications encountered and data analysis tools are discussed.

Journal ArticleDOI
TL;DR: This paper examines, evaluates and organizes a wide variety of knowledge management tools, as it looks at their origins and their opportunities in the knowledge management arena by examining the literature related to the selection and evaluation of theknowledge management tools available on the software market.

Proceedings ArticleDOI
16 Sep 2002
TL;DR: A framework to address the shortcomings in current knowledge management implementation strategies is proposed that consists of three main interlinked components: Knowledge Management of the Organisation, People, and Infrastructure and Processes.
Abstract: This paper outlines ongoing research in the area of knowledge management implementation strategies. An investigation of the literature reveals that when organisations initiate a knowledge management effort, most of them tend to over-emphasise the role of information technology at the expense of the human factor. A preliminary survey of local organisations using Duffy's Knowledge Management Benchmarking Questionnaire confirms these findings. This paper then proposes a framework that will address the shortcomings in current knowledge management implementation strategies. The framework consists of three main interlinked components: Knowledge Management of the Organisation, People, and Infrastructure and Processes. The framework focuses on the importance of aligning the knowledge management strategy of the organisation with the overall business strategy of the organisation. The culture and managing the culture change when implementing knowledge management are also dealt with. Furthermore, the framework recommends a holistic approach to managing knowledge.


Patent
30 Jul 2002
TL;DR: In this paper, a NSP network is disclosed for providing a platform service for providing data communication security in an IP-based communications network, which includes connections to various private networks, a subscriber data management unit for managing information about the mobile terminals using the network, and a platform unit.
Abstract: A NSP network is disclosed for providing a platform service for providing data communication security in an IP-based communications network. The platform service includes connections to various private networks, a subscriber data management unit for managing information about the mobile terminals using the network, and a platform unit. The platform unit determines from the subscriber data management unit the status of the mobile terminals requesting connections to private networks and the status of the mobile terminals for which private networks target for connection. The platform unit includes applications for querying the subscriber data management unit for subscriber status, applications to command the transmission of authorization data, and applications to provide secure virtual private network communication lines for connecting the mobile terminals and the private networks.