scispace - formally typeset
Search or ask a question
Topic

Metadata repository

About: Metadata repository is a research topic. Over the lifetime, 5841 publications have been published within this topic receiving 121778 citations.


Papers
More filters
Journal ArticleDOI
TL;DR: The relevance of the CRKM project to future archival systems and the deployment of metadata for multiple archival purposes is explored as part of the Smart Metadata and the Archives of the Future session that aims to communicate the progress and findings of several inter-related collaborative research projects and standards initiatives.
Abstract: Metadata is a key component in the creation, management and preservation of electronic records, as well as their innovative use as archives, memory and knowledge. However metadata generation and deployment are currently resource intensive and application specific. Metadata creation is not usually fully automated. Metadata created in one application of potential relevance to other applications is not shared between applications. Although data modeling, mark up language and syntax initiatives are addressing the data representation requirements for metadata translation and exchange, this functionality has not as yet been utilized in the systems that support eGovernment and eBusiness processes, electronic recordkeeping and archival description. Moreover there has been little progress in relation to developing strategies and meta-tools for the translation of metadata attributes and values between schemas in these environments. The Monash Clever Recordkeeping Metadata (CRKM) project addresses the challenge of automating metadata creation and sharing metadata between business systems, current recordkeeping systems and archival systems. This paper explores the relevance of the CRKM project to future archival systems and the deployment of metadata for multiple archival purposes. It is presented as part of the Smart Metadata and the Archives of the Future session that aims to communicate the progress and findings of several inter-related collaborative research projects and standards initiatives. Other papers in the session report on the related work of the InterPARES 2 Description Research Team (Designing a Meta-Registry for the Registration, Analysis and Archival Extension of Pre-Existing Metadata), the San Diego Supercomputing Center’s development of Persistent Archives Technology (Metadata Tools and Sustainable Archives Technologies), and the ISO Metadata for Records Standard (Smart Metadata Research and International Standards).

44 citations

Journal ArticleDOI
01 Jan 2016-Database
TL;DR: The principles and philosophy used to define the ENCODE metadata in order to create a metadata standard that can be applied to diverse assays and multiple genomic projects are outlined.
Abstract: The Encyclopedia of DNA Elements (ENCODE) Data Coordinating Center (DCC) is responsible for organizing, describing and providing access to the diverse data generated by the ENCODE project. The description of these data, known as metadata, includes the biological sample used as input, the protocols and assays performed on these samples, the data files generated from the results and the computational methods used to analyze the data. Here, we outline the principles and philosophy used to define the ENCODE metadata in order to create a metadata standard that can be applied to diverse assays and multiple genomic projects. In addition, we present how the data are validated and used by the ENCODE DCC in creating the ENCODE Portal (https://www.encodeproject.org/). Database URL: www.encodeproject.org.

44 citations

Patent
31 Aug 2007
TL;DR: In this article, the first aspect of the present invention to provide a computer implemented method of performing ad-hoc analysis including the steps of: generating a text index of the textual information items, generating a metadata lookup structure based, at least in part, on the text index, searching the search index using a search query, compiling results of text index search into aggregate information related to characteristics of the search results from the metadata items associated with the text information items in the query results from metadata lookup, and reporting the aggregate information.
Abstract: It is a first aspect of the present invention to provide a computer implemented method of performing ad-hoc analysis including the steps of: generating a text index of the textual information items, generating a metadata lookup structure based, at least in part, on the text index, searching the text index using a search query, compiling results of the text index search into aggregate information related to characteristics of the search results from the metadata items associated with the textual information items in the search results from the metadata lookup structure, and reporting the aggregate information.

44 citations

Proceedings ArticleDOI
Kenneth Haase1
10 Oct 2004
TL;DR: This article argues for the growing importance of quality metadata and the equation of that quality with precision and semantic grounding and looks at a particular richly annotated image collection to show how context dramatically reduces the problem of ambiguity over this particular corpus.
Abstract: This article argues for the growing importance of quality metadata and the equation of that quality with precision and semantic grounding. Such semantic grounding requires metadata that derives from intentional human intervention as well as mechanistic measurement of content media. In both cases, one chief problem in the automatic generation of semantic metadata is ambiguity leading to the overgeneration of inaccurate annotations. We look at a particular richly annotated image collection to show how context dramatically reduces the problem of ambiguity over this particular corpus. In particular, we consider both the abstract measurement of "contextual ambiguity" over the collection and the application of a particular disambiguation algorithm to synthesized keyword searches across the selection.

44 citations

Journal ArticleDOI
TL;DR: The development of OAIster's system to harvest, store, transform the metadata into Digital Library eXtension Service (DLXS) Bibliographic Class format, build indexes and make the metadata searchable through an interface using the XPAT search engine is described.
Abstract: OAIster, at the University of Michigan, University Libraries, Digital Library Production Service (DLPS), is an Andrew W. Mellon Foundation grant‐funded project designed to test the feasibility of using the Open Archives Initiative Protocol for Metadata Harvesting (OAI‐PMH) to harvest digital object metadata from multiple and varied digital object repositories and develop a service to allow end‐users to access that metadata. This article describes in‐depth the development of our system to harvest, store, transform the metadata into Digital Library eXtension Service (DLXS) Bibliographic Class format, build indexes and make the metadata searchable through an interface using the XPAT search engine. Results of the testing of our service and statistics on usage are reported, as well as the issues that we have encountered during our harvesting and transformation operations. The article closes by discussing the future improvements and potential of OAIster and the OAI‐PMH protocol.

44 citations


Network Information
Related Topics (5)
Information system
107.5K papers, 1.8M citations
85% related
User interface
85.4K papers, 1.7M citations
81% related
Software
130.5K papers, 2M citations
80% related
Mobile computing
51.3K papers, 1M citations
80% related
Support vector machine
73.6K papers, 1.7M citations
80% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
202332
202279
202113
202011
201921
201824