Assessing the Impact and Quality of Research Data Using Altmetrics and Other Indicators
Stacy Konkiel
- Vol. 2, Iss: 1
Reads0
Chats0
TLDR
The research into research data metrics, these metrics’ strengths and limitations with regard to formal evaluation practices, and the possible meanings of such indicators are discussed, and heuristics for policymakers and evaluators interested in doing so are suggested.Abstract:
Research data in all its diversity—instrument readouts, observations, images, texts, video and audio files, and so on—is the basis for most advancement in the sciences. Yet the assessment of most research programmes happens at the publication level, and data has yet to be treated like a first-class research object. How can and should the research community use indicators to understand the quality and many potential impacts of research data? In this article, we discuss the research into research data metrics, these metrics’ strengths and limitations with regard to formal evaluation practices, and the possible meanings of such indicators. We acknowledge the dearth of guidance for using altmetrics and other indicators when assessing the impact and quality of research data, and suggest heuristics for policymakers and evaluators interested in doing so, in the absence of formal governmental or disciplinary policies. Policy highlights Research data is an important building block of scientific production, but efforts to develop a framework for assessing data’s impacts have had limited success to date. Indicators like citations, altmetrics, usage statistics, and reuse metrics highlight the influence of research data upon other researchers and the public, to varying degrees. In the absence of a shared definition of “quality”, varying metrics may be used to measure a dataset’s accuracy, currency, completeness, and consistency. Policymakers interested in setting standards for assessing research data using indicators should take into account indicator availability and disciplinary variations in the data when creating guidelines for explaining and interpreting research data’s impact. Quality metrics are context dependent: they may vary based upon discipline, data structure, and repository. For this reason, there is no agreed upon set of indicators that can be used to measure quality. Citations are well-suited to showcase research impact and are the most widely understood indicator. However, efforts to standardize and promote data citation practices have seen limited success, leading to varying rates of citation data availability across disciplines. Altmetrics can help illustrate public interest in research, but availability of altmetrics for research data is very limited. Usage statistics are typically understood to showcase interest in research data, but infrastructure to standardize these measures have only recently been introduced, and not all repositories report their usage metrics to centralized data brokers like DataCite. Reuse metrics vary widely in terms of what kinds of reuse they measure (e.g. educational, scholarly, etc). This category of indicator has the fewest heuristics for collection and use associated with it; think about explaining and interpreting reuse with qualitative data, wherever possible. All research data impact indicators should be interpreted in line with the Leiden Manifesto’s principles, including accounting for disciplinary variation and data availability. Assessing research data impact and quality using numeric indicators is not yet widely practiced, though there is generally support for the practice amongst researchers.read more
Citations
More filters
Journal ArticleDOI
Analysis of shared research data in Spanish scientific papers about COVID‐19: A first approach
Roxana Cerda‐Cosme,Eva Méndez +1 more
TL;DR: In this article , the authors study how many Spanish scientific papers on COVID•19 published during 2020 share their research data and find that only a small percentage of the papers share their data.
Chapter 2 - Information and data ecologies
TL;DR: In this article, the authors focus on issues that have varied interfaces with literacies, but are not literacies in the proper sense of the word, such as openness, reproducibility, credibility, and sharing of digital data.
Journal ArticleDOI
A systematic map of cassava farming practices and their agricultural and environmental impacts using new ontologies: Agri‐ontologies 1.0
Amelia S. C. Hood,Gorm E. Shackelford,Alec P. Christie,Hope O. Usieta,Philip A. Martin,William J. Sutherland +5 more
TL;DR: In this paper , a systematic map of scientific studies about cassava farming practices is presented, with the aim of identifying knowledge gaps and clusters, and developing a classification system for [1] farming interventions and [2] agricultural, economic and environmental outcomes.
References
More filters
Journal ArticleDOI
Data fusion
Jens Bleiholder,Felix Naumann +1 more
TL;DR: This article places data fusion into the greater context of data integration, precisely defines the goals of data fusion, namely, complete, concise, and consistent data, and highlights the challenges of data Fusion.
Journal ArticleDOI
Data quality assessment
TL;DR: Concepts that can help organizations develop usable data quality metrics are described that are suitable for use in practice and not just on a whim.
Journal ArticleDOI
Anchoring data quality dimensions in ontological foundations
Yair Wand,Richard Y. Wang +1 more
TL;DR: A leading computer industry information service firm indicated that it “expects most business process reengineering initiatives to fail through lack of attention to data quality”.
Journal ArticleDOI
Bibliometrics: The Leiden Manifesto for research metrics
TL;DR: Zehn Grundsatze um Forschung zu bewerten, drangen Diana Hicks, Paul Wouters und Kollegen einiges zusammen wirkt.
Journal ArticleDOI
Methodologies for data quality assessment and improvement
TL;DR: Methodologies are compared along several dimensions, including the methodological phases and steps, the strategies and techniques, the data quality dimensions, the types of data, and, finally, thetypes of information systems addressed by each methodology.