scispace - formally typeset
Search or ask a question

Showing papers on "Data management published in 2004"


Journal ArticleDOI
TL;DR: The 'mzXML' format is introduced, an open, generic XML (extensible markup language) representation of MS data that will facilitate data management, interpretation and dissemination in proteomics research.
Abstract: A broad range of mass spectrometers are used in mass spectrometry (MS)-based proteomics research. Each type of instrument possesses a unique design, data system and performance specifications, resulting in strengths and weaknesses for different types of experiments. Unfortunately, the native binary data formats produced by each type of mass spectrometer also differ and are usually proprietary. The diverse, nontransparent nature of the data structure complicates the integration of new instruments into preexisting infrastructure, impedes the analysis, exchange, comparison and publication of results from different experiments and laboratories, and prevents the bioinformatics community from accessing data sets required for software development. Here, we introduce the 'mzXML' format, an open, generic XML (extensible markup language) representation of MS data. We have also developed an accompanying suite of supporting programs. We expect that this format will facilitate data management, interpretation and dissemination in proteomics research.

788 citations


Proceedings Article
01 Aug 2004
TL;DR: This paper provides numerous motivating applications for Trio and lays out preliminary plans for the data model, query language, and prototype system.
Abstract: Trio is a new database system that manages not only data, but also the accuracy and lineage of the data. Approximate (uncertain, probabilistic, incomplete, fuzzy, and imprecise!) databases have been proposed in the past, and the lineage problem also has been studied. The goals of the Trio project are to distill previous work into a simple and usable model, design a query language as an understandable extension to SQL, and most importantly build a working system---a system that augments conventional data management with both accuracy and lineage as an integral part of the data. This paper provides numerous motivating applications for Trio and lays out preliminary plans for the data model, query language, and prototype system.

547 citations



Book ChapterDOI
31 Aug 2004
TL;DR: This paper analyzes the data partitioning (bucketization) technique and algorithmically develops this technique to build privacy-preserving indices on sensitive attributes of a relational table and develops a novel algorithm for achieving the desired balance between privacy and utility of the index.
Abstract: Database outsourcing is an emerging data management paradigm which has the potential to transform the IT operations of corporations. In this paper we address privacy threats in database outsourcing scenarios where trust in the service provider is limited. Specifically, we analyze the data partitioning (bucketization) technique and algorithmically develop this technique to build privacy-preserving indices on sensitive attributes of a relational table. Such indices enable an untrusted server to evaluate obfuscated range queries with minimal information leakage. We analyze the worst-case scenario of inference attacks that can potentially lead to breach of privacy (e.g., estimating the value of a data element within a small error margin) and identify statistical measures of data privacy in the context of these attacks. We also investigate precise privacy guarantees of data partitioning which form the basic building blocks of our index. We then develop a model for the fundamental privacy-utility tradeoff and design a novel algorithm for achieving the desired balance between privacy and utility (accuracy of range query evaluation) of the index.

481 citations


Journal ArticleDOI
TL;DR: This paper describes-several aspects of the Piazza PDMS, including the schema mediation formalism, query answering and optimization algorithms, and the relevance of PDMSs to the semantic Web.
Abstract: Intuitively, data management and data integration tools are well-suited for exchanging information in a semantically meaningful way. Unfortunately, they suffer from two significant problems: They typically require a comprehensive schema design before they can be used to store or share information and they are difficult to extend because schema evolution is heavyweight and may break backward compatibility. As a result, many small-scale data sharing tasks are more easily facilitated by nondatabase-oriented tools that have little support for semantics. The goal of the peer data management system (PDMS) is to address this need: We propose the use of a decentralized, easily extensible data management architecture in which any user can contribute new data, schema information, or even mappings between other peers' schemes. PDMSs represent a natural step beyond data integration systems, replacing their single logical schema with an interlinked collection of semantic mappings between peers' individual schemas. This paper describes-several aspects of the Piazza PDMS, including the schema mediation formalism, query answering and optimization algorithms, and the relevance of PDMSs to the semantic Web.

297 citations


Patent
24 May 2004
TL;DR: In this article, an enterprise resource planning (ERP) system is presented in which information processing and data management are integrated with vehicle diagnostics, including a first device (110) for obtaining the vehicle characteristic information from a vehicle, and a second device (102) for integrating the vehicle information into an information management system (102b).
Abstract: An enterprise resource planning (ERP) system (100) in which information processing and data management are integrated with vehicle diagnostics. The ERP system (100) includes a first device (110) for obtaining the vehicle characteristic information from a vehicle, and a second device (102) for integrating the vehicle characteristic information into an information management system (102b).

284 citations


Journal ArticleDOI
TL;DR: The authors present their research findings, based closely on their report to OECD, on key issues in data access, as well as operating principles and management aspects necessary to successful data access regimes.
Abstract: Access to and sharing of data are essential for the conduct and advancement of science. This article argues that publicly funded research data should be openly available to the maximum extent possible. To seize upon advancements of cyberinfrastructure and the explosion of data in a range of scientific disciplines, this access to and sharing of publicly funded data must be advanced within an international framework, beyond technological solutions. The authors, members of an OECD Follow-up Group, present their research findings, based closely on their report to OECD, on key issues in data access, as well as operating principles and management aspects necessary to successful data access regimes.

274 citations


Journal ArticleDOI
TL;DR: In this paper, an integrated view of quality and knowledge using Nonaka's theory of knowledge creation is proposed to illuminate how quality practices can lead to knowledge creation and retention, and the knowledge perspective also provides insight into what it means to effectively deploy quality management practices.

248 citations


Patent
10 Feb 2004
TL;DR: In this article, a data management system for displaying a data table having a plurality of rows and columns of data cells is presented, each of the columns in the data table has an associated filter cell whereby the user may input filter criteria.
Abstract: System and method for filtering and sorting data. The system includes data management system for displaying a data table having a plurality of rows and columns of data cells. Each of the columns in the data table has an associated filter cell whereby the user may input filter criteria. In response to receiving the filter criteria in the filter cells, a data management system filters and sorts each of the data cells in the data table. In one embodiment of the invention, the data management system records the order in which the user inputs the filter criteria and generates a sorting priority list that defines a sorting priority for the data cells. The data management system sorts the data cells using the sorting priority list. Furthermore, in one embodiment of the invention, the data management system displays the data cells that satisfy the filter criteria contiguously and interleaved between those data cells that do not satisfy the filter criteria.

242 citations


Patent
24 Aug 2004
TL;DR: In this article, the present invention provides systems and methods for efficient data storage, management, and back up, in particular, for efficient replication of data, and provides devices, software, and processes for efficient replicating data.
Abstract: The present invention provides systems and methods for efficient data storage, management, and back up. In particular, the present invention provides devices, software, and processes for efficient replication of data.

234 citations


Book ChapterDOI
01 Jan 2004
TL;DR: This chapter presents the On-To-Knowledge Methodology (OTKM) for introducing and maintaining ontology based knowledge management applications into enterprises with a focus on Knowledge Processes and Knowledge Meta Processes.
Abstract: In this chapter we present the On-To-Knowledge Methodology (OTKM) for introducing and maintaining ontology based knowledge management applications into enterprises with a focus on Knowledge Processes and Knowledge Meta Processes. While the former process circles around the usage of ontologies, the latter process guides their initial set up. We illustrate our methodology by an example from a case study on skills management.

Book ChapterDOI
01 Jan 2004
TL;DR: A brief introduction to RFID technology is presented and a few of the data management challenges of the proliferation of RFID readers and tags are highlighted.
Abstract: Radio-Frequency Identiflcation (RFID) technology enables sensors to e‐ciently and inexpensively track merchandise and other objects. The vast amount of data resulting from the proliferation of RFID readers and tags poses some interesting challenges for data management. We present a brief introduction to RFID technology and highlight a few of the data management challenges.

Proceedings Article
31 Aug 2004
TL;DR: In this paper, a brief introduction to RFID technology and highlight a few of the data management challenges is presented, and the vast amount of data resulting from the proliferation of RFID readers and tags poses some interesting challenges for data management.
Abstract: Radio-Frequency Identification (RFID) technology enables sensors to efficiently and inexpensively track merchandise and other objects. The vast amount of data resulting from the proliferation of RFID readers and tags poses some interesting challenges for data management. We present a brief introduction to RFID technology and highlight a few of the data management challenges.

Patent
27 Dec 2004
TL;DR: In this paper, a software tool for enabling automated tracking of activity related to the status and usage statistics of a plurality of Web sites on a data packet network is provided, which comprises a network communication capability for establishing network communication between the software tool and the tracked Web sites; a plurality-of-data-reporting modules for obtaining and reporting data about tracked web sites; and a data input function for excepting data from the reporting modules and from external sources; a data recording function for recording and logging the data received from the reported data and from the external sources.
Abstract: A software tool for enabling automated tracking of activity related to the status and usage statistics of a plurality of Web sites on a data packet network is provided. The software tool comprises a network communication capability for establishing network communication between the software tool and the tracked Web sites; a plurality of data-reporting modules for obtaining and reporting data about tracked Web sites; a data input function for excepting data from the reporting modules and from external sources; a data recording function for recording and logging the data received from the reporting modules and from the external sources; and a data management function for organizing and storing the received data and rendering the data accessible for use in software development. A software engineer or developer accesses the site-tracking software and connected database through a Web browser from a network-connected workstation in order to utilize data mined from Web sites for the purpose of creating routines enabling automated navigation and site manipulation by proxy for subscribed users.

Patent
15 Oct 2004
TL;DR: In this article, the authors present a ticket entry system and method that includes a program management system, a spatial management system and a data management system for locating a location in a map.
Abstract: In one embodiment, a ticket entry system and method include a program management system, a spatial management system, a data management system, and a user interface. The program management system receives contact information and a search criteria for a locate request. The program management system transmits the search criteria to the spatial management system. The spatial management system geocodes the search criteria and retrieves map data from the data management system corresponding to the search criteria geocode. The spatial management system transmits the map data to the program management system. The program management system transmits the map data to the user system and, in return, receives an indication of a site location where the operation is to be performed. The program management system transmits the site location indication to the spatial management system. The spatial management system geocodes the site location indication and transmits the site location geocode to the program management system. The program management system creates a locate ticket, which comprises the site location geocode and the contact information.

Journal ArticleDOI
TL;DR: This work examines how the quality metrics of source data affect the quality for information outputs produced using the relational algebra operations selection, projection, and Cartesian product.
Abstract: The cost associated with making decisions based on poor-quality data is quite high. Consequently, the management of data quality and the quality of associated data management processes has become critical for organizations. An important first step in managing data quality is the ability to measure the quality of information products (derived data) based on the quality of the source data and associated processes used to produce the information outputs. We present a methodology to determine two data quality characteristics--accuracy and completeness--that are of critical importance to decision makers. We examine how the quality metrics of source data affect the quality for information outputs produced using the relational algebra operations selection, projection, and Cartesian product. Our methodology is general, and can be used to determine how quality characteristics associated with diverse data sources affect the quality of the derived data.

Book
29 Mar 2004
TL;DR: In this paper, the authors discuss the need for performance management as a system, and present a suite of proven methodologies to integrate data and information from Fact-Based Data and Information Technology.
Abstract: About the Author. About the Web Site. Preface. 1. Why the Need for Performance Management as a System? PART ONE: Performance Management Process. 2. Integrating a Suite of Proven Methodologies. 3. Support from Fact-Based Data and Information Technology. PART TWO: Strategy Maps and Balanced Scorecards: The Link between Strategy and Successful Execution by Operations. 4. Measurement Problems and Solutions. 5. Strategy Maps and Scorecards as a Solution. 6. Strategic Objectives' Drive Gears: Cascading Measures. 7. A Recipe for Implementation. 8. The Human Side of Collaboration. 9. Fact-Based Management Accounting Data. 10. Scorecards and Strategy Maps: Enablers for Performance Management. PART THREE: Leveraging Financial Analytical Facts and Truths. 11. If Activity-Based Management Is the Answer, What Is the Question? 12. Activity-Based Management Model Design and Principles: Key to Success. 13. Operational (Local) Activity-Based Management for Continuous Improvement. 14. Strategic Activity-Based Management for Customer and Channel Profitability Analysis. 15. Predictive Costing, Predictive Accounting, and Budgeting. 16. Activity-Based Management Supports Performance Management. PART FOUR: Integrating Performance Management with Core Solutions. 17. Customer Intelligence and Customer Relationship Management. 18. Supplier Intelligence: Managing Economic Profit across the Value Chain. 19. Process Intelligence with Six Sigma Quality and Lean Thinking. 20. Shareholder Intelligence: Return on Whose Investment? 21. Employee Intelligence: Human Capital Management. PART FIVE: Performance Management, Business Intelligence, and Technology. 22. Data Management and Mining with Performance Management. 23. Final Thoughts: Linking Customers to Shareholders. Index.

Book ChapterDOI
01 Jan 2004
TL;DR: This chapter presents an extensive ontology of knowledge management, which identifies and characterizes basic components of knowledge Management episodes, the knowledge resources an organization uses in these episodes, a generic set of elemental knowledge manipulation activities that manifest within knowledge management episodes, and categories of influences on the conduct and outcomes of these episodes.
Abstract: Knowledge-based organizations are hosts for multitudes of knowledge management episodes. Each episode is triggered by a knowledge need (or opportunity) and culminates with the satisfaction of that need (or its abandonment). Within an episode, one or more of the organization’s processors (human and/or computer-based) manipulate knowledge resources in various ways and subject to various influences in an effort to meet the need or take advantage of the opportunity. This chapter presents an extensive ontology of knowledge management. The ontology identifies and characterizes basic components of knowledge management episodes, the knowledge resources an organization uses in these episodes, a generic set of elemental knowledge manipulation activities that manifest within knowledge management episodes, and categories of influences on the conduct and outcomes of these episodes. This ontology was developed using conceptual synthesis and a collaborative methodology involving an international panel of researchers and practitioners in the knowledge management field. The ontology can serve as a common language for discourse about knowledge management. For researchers, it suggests issues that deserve investigation and concepts that must be considered in explorations of knowledge management episodes. For practitioners, the ontology provides a perspective on factors that need to be considered in the implementation of an organization’s knowledge management initiatives.

Journal ArticleDOI
TL;DR: This paper presents a framework to address new data management challenges introduced by data-intensive, pervasive computing environments, including a spatio-temporal variation of data and data source availability, lack of a global catalog and schema, and no guarantee of reconnection among peers due to the serendipitous nature of the environment.
Abstract: This paper presents a framework to address new data management challenges introduced by data-intensive, pervasive computing environments. These challenges include a spatio-temporal variation of data and data source availability, lack of a global catalog and schema, and no guarantee of reconnection among peers due to the serendipitous nature of the environment. An important aspect of our solution is to treat devices as semiautonomous peers guided in their interactions by profiles and context. The profiles are grounded in a semantically rich language and represent information about users, devices, and data described in terms of "beliefs," "desires," and "intentions." We present a prototype implementation of this framework over combined Bluetooth and Ad Hoc 802.11 networks and present experimental and simulation results that validate our approach and measure system performance.

Journal ArticleDOI
TL;DR: ArrayTrack as discussed by the authors is a public microarray data management and analysis software, called ArrayTrack is Minimum Information About a Microarray Experiment (MIAME) supportive for storing both gene expression data and experiment parameters associated with a toxicogenomics study.
Abstract: A robust bioinformatics capability is widely acknowledged as central to realizing the promises of toxicogenomics. Successful application of toxicogenomic approaches, such as DNA microarray, inextricably relies on appropriate data management, the ability to extract knowledge from massive amounts of data and the availability of functional information for data interpretation. At the FDA's National Center for Toxicological Research (NCTR), we are developing a public microarray data management and analysis software, called ArrayTrack. ArrayTrack is Minimum Information About a Microarray Experiment (MIAME) supportive for storing both microarray data and experiment parameters associated with a toxicogenomics study. A quality control mechanism is implemented to assure the fidelity of entered expression data. ArrayTrack also provides a rich collection of functional information about genes, proteins and pathways drawn from various public biological databases for facilitating data interpretation. In addition, several data analysis and visualization tools are available with ArrayTrack, and more tools will be available in the next released version. Importantly, gene expression data, functional information and analysis methods are fully integrated so that the data analysis and interpretation process is simplified and enhanced. ArrayTrack is publicly available online and the prospective user can also request a local installation version by contacting the authors.

Book
01 Jan 2004
TL;DR: In this paper, the authors present a risk management process for Third Party Damage Index Corrosion Index Design Index Incorrect Operations Index Leak Impact Index Data Management and Analyses Additional Risk Modules Service Interruption Distribution Systems Offshore Pipeline Systems Stations and Surface Facilities Absolute Risk Risk Management
Abstract: Acknowledgements... Risk: Theory and Application Risk Management Process Third Party Damage Index Corrosion Index Design Index Incorrect Operations Index Leak Impact Index Data Management and Analyses Additional Risk Modules Service Interruption Distribution Systems Offshore Pipeline Systems Stations and Surface Facilities Absolute Risk Risk Management

Proceedings ArticleDOI
05 Jan 2004
TL;DR: The study found that knowledge management drivers such as learning orientation, knowledge sharing intention, knowledge management system quality, reward, and knowledge management team activity were significantly related to the organizational knowledge management performance - knowledge quality and user knowledge satisfaction.
Abstract: Despite the active interest in managing organizational knowledge as a strategic resource, most organizations do not yet understand the challenges involved in implementing knowledge management initiatives. Much of the knowledge management literature has been either conceptual or based on individual implementation cases. This study aimed at identifying the several key drivers for developing organizational knowledge management capability and examining their relationships with knowledge management performance. Using data collected from the 66 Korean firms, the study found that knowledge management drivers such as learning orientation, knowledge sharing intention, knowledge management system quality, reward, and knowledge management team activity were significantly related to the organizational knowledge management performance - knowledge quality and user knowledge satisfaction. The study also found that the knowledge management stage of an organization moderates the relationship between some of the knowledge management drivers and knowledge management performance variables.

Journal ArticleDOI
TL;DR: This work proposes a comprehensive software architecture for the next-generation grid, which integrates currently available services and components in Semantic Web, Semantic Grid, P2P, and ubiquitous systems.
Abstract: Just as the Internet is shifting its focus from information and communication to a knowledge delivery infrastructure, we see the Grid moving from computation and data management to a pervasive, worldwide knowledge management infrastructure. We have the technology to store and access data, but we seem to lack the ability to transform data tombs into useful data and extract knowledge from them. We review some of the current and future technologies that will impact the architecture, computational model, and applications of future grids. We attempt to forecast the evolution of computational grids into what we call the next-generation grid, with a particular focus on the use of semantics and knowledge discovery techniques and services. We propose a comprehensive software architecture for the next-generation grid, which integrates currently available services and components in Semantic Web, Semantic Grid, P2P, and ubiquitous systems. We'll also discuss a case study that shows how some new technologies can improve grid applications.

Patent
04 Nov 2004
TL;DR: A product data exchange system 300 is used for exchanging technical product data between respective computer systems of a plurality of collaborating companies.
Abstract: A product data exchange system 300 is used for exchanging technical product data between respective computer systems (310, 320, 340, 350, 360) of a plurality of collaborating companies. At least a computer system (310) of a first one of the collaborating companies includes a plurality of distinct data management systems (312, 314, 316), such as CAD, PLM, ERP, each for creating respective technical product data. The system (310) also includes an editing system (318) for importing technical product data relating to a user­selectable project from a plurality of the data management systems, creating an exchange package representing user-selectable parts of the imported technical product data; and providing the exchange package to a computer system located at at least one of the other collaborating companies.

Journal ArticleDOI
TL;DR: In this paper, the authors report the co-production model of knowledge creation and transfer through a novel case of this in practice and outline how academics and managers can work together using a "systematic review" of the science base to synthesize management knowledge and ensure its transfer.
Abstract: The division between academic knowledge and its relevance for practice is an enduring problem across many fields. Nowhere is this division more pronounced, and resolution of its negative features more required, than in academic management research and its relationship to management practice, for the advent of the knowledge revolution requires that organizations capitalize on all available assets including knowledge assets when improving performance either by increasing efficiencies or ensuring mission delivery in the medium term. How companies might achieve this has become a key question. This paper reports the “co‐production” model of knowledge creation and transfer through a novel case of this in practice. It outlines how academics and managers can work together using a “systematic review” of the science base to synthesize management knowledge and ensure its transfer. In doing so it offers management academics and practitioners a new model for the production and application of knowledge.

Journal ArticleDOI
TL;DR: The World Health Organization's Multicentre Growth Reference data management protocol was designed to create and manage a large data bank of information collected from multiple sites over a period of several years, and has accumulated a large body of very high-quality data.
Abstract: The World Health Organization (WHO) Multicentre Growth Reference (MGRS) data management protocol was designed to create and manage a large data bank of information collected from multiple sites over a period of several years Data collection and processing instruments were prepared centrally and used in a standardized fashion across sites The data management system contained internal validation features for timely detection of data errors and its standard operating procedures stipulated a method of master file updating and correction that maintained a clear trail for data auditing purposes Each site was responsible for collecting entering verifying and validating data and for creating site-level master files Data from the sites were sent to the MGRS Coordinating Centre every month for master file consolidation and more extensive quality control checking All errors identified at the Coordinating Centre were communicated to the site for correction at source The protocol imposed transparency on the sites data management activities but also ensured access to technical help with operation and maintenance of the system Through the rigorous implementation of what has been a highly demanding protocol the MGRS has accumulated a large body of very high-quality data (authors)

Journal Article
Yao Min1
TL;DR: According to knowledge life cycle, specialties and applied fields of various knowledge managing technologies are summaried and compared through the researching on knowledge characteristic of different phases of knowledge management.
Abstract: Knowledge management is an important concept and means in contemporary enterprise management, and the aim of knowledge management is to manage knowledge completely, effectively and coherently. According to knowledge life cycle, specialties and applied fields of various knowledge managing technologies are summaried and compared through the researching on knowledge characteristic of different phases of knowledge management. The generalized learning and generalized computing technology are mainly discussed in its action. In the end, the fundamental theories and basic methods of knowledge management are summaried through comparative research on knowledge managing technologies and methods.

Book ChapterDOI
30 Aug 2004
TL;DR: In this paper, the authors address the impact of user-centric context-awareness requirement on data management strategies and solutions by providing a multidimensional view of database access context, taking diverse contextual information into account, using the most fundamental database operation-context-aware query request as a case in point.
Abstract: Ambient Intelligence (AmI) is a vision of future Information Society, where people are surrounded by an electronic environment which is sensitive to their needs, personalized to their requirements, anticipatory of their behavior, and responsive to their presence. It emphasizes on greater user-friendliness, user-empowerment, and more effective service support, with an aim to make peoplersquos daily activities more convenient, thus improving the quality of human life. To make AmI real, effective data management support is indispensable. High-quality information must be available to any user, anytime, anywhere, and on any lightweight device. Beyond that, AmI also raises many new challenges related to context-awareness and natural user interaction, entailing us to re-think current database techniques. The aim of this paper is to address the impact of AmI, particularly its user-centric context-awareness requirement on data management strategies and solutions. We first provide a multidimensional view of database access context. Taking diverse contextual information into account, we then present five context-aware data management strategies, using the most fundamental database operation-context-aware query request as a case in point. We execute the proposed strategies via a two-layered infrastructure, consisting of public data manager(s) and a private data manager. Detailed steps of processing a context-aware query are also described in the paper.

Patent
13 Apr 2004
TL;DR: In this article, the disclosure is directed to a system including a data management system configured to receive sensor data associated with a sensor, a monitoring system coupled to the data management systems, and an alert handling system responsive to the management system and configured to access an alert profile associated with an error object provided by the monitoring system.
Abstract: In a particular embodiment, the disclosure is directed to a system including a data management system configured to receive sensor data associated with a sensor, a monitoring system coupled to the data management system, and an alert handling system responsive to the data management system and configured to access an alert profile associated with an error object provided by the monitoring system. The monitoring system is configured to provide the error object in response to a monitoring algorithm associated with the sensor data received by the data management system.

Patent
07 May 2004
TL;DR: In this article, a data management services framework coupled with a centralized master repository for core enterprise reference data associated with an enterprise is presented, where a set of synonyms representing a mapping between the field in the master schema and a corresponding field in a particular one of the multiple schemas.
Abstract: In one embodiment, a system is provided for managing a centrally managed master repository for core enterprise reference data associated with an enterprise. A centralized master repository contains the reference data, the reference data being associated with multiple schemas, each schema including one or more data models for reference data, each data model including one or more fields. A data management services framework coupled to the repository provides services for managing the reference data in the repository. The services framework supports a master schema including a union of multiple models and associated fields in the multiple schemas. The services framework also supports a thesaurus including, for each field in the master schema, a set of synonyms each representing a mapping between the field in the master schema and a corresponding field in a particular one of the multiple schemas. The master schema and thesaurus facilitate centralized management of the reference data in the repository across multiple heterogeneous external operational systems that have different associated data models and are provided indirect access to the reference data in the repository for operational use of the reference data according to associated business workflows.