scispace - formally typeset
Search or ask a question

Showing papers on "Meta Data Services published in 2018"


Journal ArticleDOI
TL;DR: This study uses data from a large original survey to present how the cataloging and metadata community is approaching new and emerging data standards and technologies, and demonstrates strong professional-development interest in Semantic Web and Linked Data applications.
Abstract: This study uses data from a large original survey (nearly one thousand initial respondents) to present how the cataloging and metadata community is approaching new and emerging data standards and technologies. The data analysis demonstrates strong professional-development interest in Semantic Web and Linked Data applications. With respect to continuing education topics, Linked Data technology, BIBFRAME, and an overview of current and emerging data standards and technologies ranked high. The survey data illustrate that personal continuing education interests often varied from reported institutional needs. These results reflect the fact that library services and projects in these emerging areas have not yet progressed beyond the exploratory stage. They also suggest that cataloging and metadata professionals expect to be able to exercise a mixture of core professional skill sets including teamwork, communication, and subject analysis, and the ability to adapt and accommodate Semantic Web standards and technologies, digital libraries, and other innovations in cataloging and metadata services.

14 citations


Journal ArticleDOI
TL;DR: MetaStore is an adaptive metadata management framework based on a NoSQL database and an RDF triple store that automatically segregates the different categories of metadata in their corresponding data models to maximize the utilization of the data models supported by NoSQL databases.
Abstract: In this paper, we present MetaStore, a metadata management framework for scientific data repositories. Scientific experiments are generating a deluge of data, and the handling of associated metadata is critical, as it enables discovering, analyzing, reusing, and sharing of scientific data. Moreover, metadata produced by scientific experiments are heterogeneous and subject to frequent changes, demanding a flexible data model. Existing metadata management systems provide a broad range of features for handling scientific metadata. However, the principal limitation of these systems is their architecture design that is restricted towards either a single or at the most a few standard metadata models. Support for handling different types of metadata models, i.e., administrative, descriptive, structural, and provenance metadata, and including community-specific metadata models is not possible with these systems. To address this challenge, we present MetaStore, an adaptive metadata management framework based on a NoSQL database and an RDF triple store. MetaStore provides a set of core functionalities to handle heterogeneous metadata models by automatically generating the necessary software code (services) and on-the-fly extends the functionality of the framework. To handle dynamic metadata and to control metadata quality, MetaStore also provides an extended set of functionalities such as enabling annotation of images and text by integrating the Web Annotation Data Model, allowing communities to define discipline-specific vocabularies using Simple Knowledge Organization System, and providing advanced search and analytical capabilities by integrating the ElasticSearch. To maximize the utilization of the data models supported by NoSQL databases, MetaStore automatically segregates the different categories of metadata in their corresponding data models. Complex provenance graphs and dynamic metadata are modeled and stored in an RDF triple store, whereas the static metadata is stored in a NoSQL database. For enabling large-scale harvesting (sharing) of metadata using the METS standard over the OAI-PMH protocol, MetaStore is designed OAI-compliant. Finally, to show the practical usability of the MetaStore framework and that the requirements from the research communities have been realized, we describe our experience in the adoption of MetaStore for three communities.

10 citations


Book ChapterDOI
01 Jan 2018
TL;DR: This chapter introduces the reader to metadata as well as the standards and tools that can be used to generate and manage standardized metadata during a research project.
Abstract: This chapter introduces the reader to metadata as well as the standards and tools that can be used to generate and manage standardized metadata during a research project. First, metadata is defined and many of the benefits that accrue from creating comprehensive metadata are listed. Second, the different types of metadata that may be necessary to understand and use (e.g., analyze, visualize) a data set are described along with some relevant examples. Third, the content, characteristics, similarities and differences among many of the relevant ecological metadata standards are presented. Fourth, the various software tools that enable one to create metadata are described and best practices for creating and managing metadata are recommended.

4 citations



Journal ArticleDOI
TL;DR: This work introduces IntegrityCatalog, a novel software system that can be integrated into any digital repository and introduces a treap‐based persistent authenticated dictionary managing arbitrary length key/value pairs, which it uses to store all integrity metadata.
Abstract: Summary Digital repositories must periodically check the integrity of stored objects to assure users of their correctness. Prior solutions calculate integrity metadata and require the repository to store it alongside the actual data objects. To safeguard and detect damage to this metadata, prior solutions rely on widely visible media (unaffiliated third parties) to store and provide back digests of the metadata to verify it is intact. However, they do not address recovery of the integrity metadata in case of damage or adversarial attack. We introduce IntegrityCatalog, a novel software system that can be integrated into any digital repository. It collects all integrity-related metadata in a single component and treats them as first class objects, managing both their integrity and their preservation. We introduce a treap-based persistent authenticated dictionary managing arbitrary length key/value pairs, which we use to store all integrity metadata, accessible simply by object name. Additionally, IntegrityCatalog is a distributed system that includes a network protocol that manages both corruption detection and preservation of this metadata, using administrator-selected network peers with 2 possible roles. Verifiers store and offer attestations on digests and have minimal storage requirements, while preservers efficiently synchronize a complete copy of the catalog to assist in recovery in case of a detected catalog compromise on the local system. We present our approach in developing the prototype implementation, measure its performance experimentally, and demonstrate its effectiveness in real-world situations. We believe the implementation techniques of our open-source IntegrityCatalog will be useful in the construction of next-generation digital repositories.

1 citations



Book ChapterDOI
01 Jan 2018
TL;DR: The article explains how to optimize metadata and spatial data infrastructure strategy for a successful and sustainable system as well as highlights some emerging trends in the geospatial and general information technology fields that will likely impact future use of these concepts.
Abstract: Geospatial data is most useful when it can be discovered, shared, and used. This article describes the mechanism for describing and organizing geospatial data through the use of metadata as the descriptive element and spatial data infrastructure as the organizational framework. Specific guidance is provided in the text for development of metadata requirements, use of metadata standards, and implementing best practices and automation in creation of metadata. The article then builds on the foundation of good metadata to describe the components of a spatial data infrastructure and how each part is designed and integrated. The focus for the spatial data infrastructure is discoverability and dissemination of geospatial data. Finally, the article explains how to optimize metadata and spatial data infrastructure strategy for a successful and sustainable system as well as highlights some emerging trends in the geospatial and general information technology fields that will likely impact future use of these concepts.

1 citations


Patent
27 Mar 2018
TL;DR: In this article, a file-system metadata management system consisting of a client terminal and a server terminal, wherein the server terminal comprises metadata services and a metadata relationship database, is described.
Abstract: The invention discloses a file-system metadata management system. The file-system metadata management system comprises a client terminal and a server terminal, wherein the server terminal comprises metadata services and a metadata relationship database; the metadata relationship database stores file system metadata, wherein the client terminal queries the metadata relationship database through thedatabase index function and the metadata services, and target metadata is obtained. It is shown that the file system metadata is stored in the relationship database; in this way, the metadata is queried through the database index function, it is avoided that all files are traversed, the required metadata can be rapidly found, and the metadata querying efficiency is improved. A file-system metadata management method has the same advantages.

1 citations