Completeness and overlap in open access systems: Search engines, aggregate institutional repositories and physics-related open sources

Authors: Tsay M-y, Wu T-l & Tseng L-l

Comment: This article compares several open access search engines (i.e., Google Scholar (GS), Microsoft Academic (MSA), OAIster, OpenDOAR, arXiv.org and Astrophysics Data System (ADS)) using publications of Nobel Laureates for Physics from 2001 to 2013. A short literature on comparing search engines is given. Both internal and external overlaps are studied. At the time of this work, GS had the highest coverage of this sample, but had a very high percentage of internal overlap (>92%). It actually covers all items in other sources, except for MSA. ADS and MSA both had coverage just below GS, with ADS having the lowest internal overlap of the three (just slightly higher than arXiv.org, which had 0 internal overlap).

Abstract: This study examines the completeness and overlap of coverage in physics of six open access scholarly communication systems, including two search engines (Google Scholar and Microsoft Academic), two aggregate institutional repositories (OAIster and OpenDOAR), and two physics-related open sources (arXiv.org and Astrophysics Data System). The 2001–2013 Nobel Laureates in Physics served as the sample. Bibliographic records of their publications were retrieved and downloaded from each system, and a computer program was developed to perform the analytical tasks of sorting, comparison, elimination, aggregation and statistical calculations. Quantitative analyses and cross-referencing were performed to determine the completeness and overlap of the system coverage of the six open access systems. The results may enable scholars to select an appropriate open access system as an efficient scholarly communication channel, and academic institutions may build institutional repositories or independently create citation index systems in the future. Suggestions on indicators and tools for academic assessment are presented based on the comprehensiveness assessment of each system.

Tsay M-y, Wu T-l, Tseng L-l (2017) Completeness and overlap in open access systems: Search engines, aggregate institutional repositories and physics-related open sources. PLoS ONE 12(12): e0189751. https://doi.org/10.1371/journal.pone.0189751

Source: Completeness and overlap in open access systems: Search engines, aggregate institutional repositories and physics-related open sources

Microsoft Academic is one year old: the Phoenix is ready to leave the nest

Authors: Harzing, AW. & Alakangas, S.

Comment: This is the third of a series of articles, by the first author, investigating the relative citation and publication coverage of Microsoft Academic (MA) within its first year of (re-)launch. Although the studies were of relatively small scale (citation record of 1 and 145 academics), they provided strong evidence for the advantages of MA over other databases. In particular, it possesses high coverage like Google Scholar and, at the same time, structured metadata like in Scopus and Web of Science. These, together with its fast growth, make MA an excellent alternative for bibliometrics and scientometrics studies.

Abstract: We investigate the coverage of Microsoft Academic (MA) just over a year after its re-launch. First, we provide a detailed comparison for the first author’s record across the four major data sources: Google Scholar (GS), MA, Scopus and Web of Science (WoS) and show that for the most important academic publications, journal articles and books, GS and MA display very similar publication and citation coverage, leaving both Scopus and WoS far behind, especially in terms of citation counts. A second, large scale, comparison for 145 academics across the five main disciplinary areas confirms that citation coverage for GS and MA is quite similar for four of the five disciplines. MA citation coverage in the Humanities is still substantially lower than GS coverage, reflecting MA’s lower coverage of non-journal publications. However, we shouldn’t forget that MA coverage for the Humanities still dwarfs coverage for this discipline in Scopus and WoS. It would be desirable for other researchers to verify our findings with different samples before drawing a definitive conclusion about MA coverage. However, based on our current findings we suggest that, only one year after its re-launch, MA is rapidly become the data source of choice; it appears to be combining the comprehensive coverage across disciplines, displayed by GS, with the more structured approach to data presentation, typical of Scopus and WoS. The Phoenix seems to be ready to leave the nest, all set to start its life into an adulthood of research evaluation.

Harzing, AW. & Alakangas, S. (2017) Microsoft Academic is one year old: the Phoenix is ready to leave the nest. Scientometrics 112: 1887. https://doi.org/10.1007/s11192-017-2454-3

Source: Microsoft Academic is one year old: the Phoenix is ready to leave the nest | Springer for Research & Development

Open Access: An Evaluation of its Impact, Obstacles, and Advancements

Author: Rachel A. Miles

Comments: A detailed article providing reviews on OA and Impact Metrics, and discussions on their misconceptions and misunderstandings. A review on OA mandates and policies is also provided. Other interesting discussions include those on Altmetrics, Eigenfactor, SNIP, JOI. An extensive list of potentially useful references are given.

Abstract: Access to research results is imperative in today’s robust digital age, yet access is often prevented by publisher paywalls. Open Access (OA) is the simple idea that all research should be free for all to access, use, and build upon. This paper will focus on three critical areas of the OA landscape: its impact on scholarship and the public, the obstacles to be overcome, and its advancements. The impact of OA actions and initiatives has been difficult to quantify, but the growing number of studies on OA have shown mostly overwhelmingly positive results. Cultural norms within academia, such as the reliance on the journal Impact Factor (IF) to assess the quality of individual research articles, have impeded the progress of OA. Conversely, federal mandates and institutional policies have supported the OA movement by requiring that scholarly publications be deposited into institutional or subject repositories immediately following publication. As information professionals, library and information science (LIS) professionals have a responsibility as practitioners, authors, and editors to support OA and encourage other academics to do the same.

Cite as: Miles, Rachel. (2016). Open Access: An Evaluation of its Impact, Obstacles, and Advancements. Bibliotekar, 58: (1-2).

Source: Open Access: An Evaulation of its Impact, Obstacles, and Advancements

Growth of hybrid open access, 2009–2016

Author: Bo-Christer Bjork

Notes: This 2017 article estimates the growth in hybrid OA journals and articles published within from 2009 to 2016. from 20 publishers Most interesting is the difficulty experienced in obtaining data because the hybridity of a journal is not always indicated. The author used previous studies and more recent data from 15 publishers who agreed to share, plus 5 big publishers. However data are not itemised for each publisher.

Abstract

Hybrid Open Access is an intermediate form of OA, where authors pay scholarly publishers to make articles freely accessible within journals, in which reading the content otherwise requires a subscription or pay-per-view. Major scholarly publishers have in recent years started providing the hybrid option for the vast majority of their journals. Since the uptake usually has been low per journal and scattered over thousands of journals, it has been very difficult to obtain an overview of how common hybrid articles are. This study, using the results of earlier studies as well as a variety of methods, measures the evolution of hybrid OA over time. The number of journals offering the hybrid option has increased from around 2,000 in 2009 to almost 10,000 in 2016. The number of individual articles has in the same period grown from an estimated 8,000 in 2009 to 45,000 in 2016. The growth in article numbers has clearly increased since 2014, after some major research funders in Europe started to introduce new centralized payment schemes for the article processing charges (APCs).

https://peerj.com/articles/3878/

Evaluation of Openness in the Activities of Research Organisations and Research Funding Organisations in 2016

Author: finland Ministry of Education and Culture, Open Science and Research Initiative

Notes: Interesting scoring of Finnish research organisations regarding progress towards openness using data retrieved from public websites. A follow up survey gave organisations the opportunity to correct and supplement data. Also compared with other European research organisations. Because of the rapidly changing landscape the assessment was not repeated in 2017.

Abstract: This evaluation of the openness of Finnish research performing and funding organisations was completed as part of the Open Science and Research Initiative (ATT) by the Ministry of Education and Culture. The target of this evaluation is to assess the openness of operational cultures in research organisations and research funding organisations. The key objectives, against which the assessments are made, are defined in the Open Science and Research Roadmap. More information about the evaluation can be found at openscience.fi/openculture

http://www.doria.fi/handle/10024/127273

Publications | Free Full-Text | Enhancing Institutional Publication Data Using Emergent Open Science Services | HTML

Authors: David Walters and Christopher Daley (Brunel University, London)

Notes: An interesting article looking at integrating data sources to assess OA status and location of OA copies for single UK university. Focusses on data derived from CORE and from Unpaywall and its combination with other information from university systems.

Abstract: The UK open access (OA) policy landscape simultaneously preferences Gold publishing models (Finch Report, RCUK, COAF) and Green OA through repository usage (HEFCE), creating the possibility of confusion and duplication of effort for academics and support staff. Alongside these policy developments, there has been an increase in open science services that aim to provide global data on OA. These services often exist separately to locally managed institutional systems for recording OA engagement and policy compliance. The aim of this study is to enhance Brunel University London’s local publication data using software which retrieves and processes information from the global open science services of Sherpa REF, CORE, and Unpaywall. The study draws on two classification schemes; a ‘best location’ hierarchy, which enables us to measure publishing trends and whether open access dissemination has taken place, and a relational ‘all locations’ dataset to examine whether individual publications appear across multiple OA dissemination models. Sherpa REF data is also used to indicate possible OA locations from serial policies. Our results find that there is an average of 4.767 permissible open access options available to the authors in our sample each time they publish and that Gold OA publications are replicated, on average, in 3 separate locations. A total of 40% of OA works in the sample are available in both Gold and Green locations. The study considers whether this tendency for duplication is a result of localised manual workflows which are necessarily focused on institutional compliance to meet the Research Excellence Framework 2021 requirements, and suggests that greater interoperability between OA systems and services would facilitate a more efficient transformation to open scholarship.

Source: Publications | Free Full-Text | Enhancing Institutional Publication Data Using Emergent Open Science Services | HTML

Over 80% of research outputs meet requirements of REF 2021 open access policy – Research England

Author: Research England (neé HEFCE)

Notes: An important national survey of progress towards Open Access in the context of a strong policy and compliance requirement. Interesting both for the claims it makes about the levels of OA as well as the language and nature of the process by which it is being achieved. Lots of important detail on how metadata is and is not being collected an processed.

Abstract: Sixty one per cent of research outputs known to be in scope for the REF 2021 are meeting open access deposit, discovery and access requirements, with a further twenty per cent reporting a known exception, a report published today shows.The report details the findings of a survey by the former Higher Education Funding Council for England (HEFCE), the Wellcome Trust, the former Research Councils UK (RCUK) and Jisc. The survey sought to assess how the sector is delivering funders’ open access (OA) policies and to understand some of the challenges the sector faces. The four project partners were also interested in understanding the methods and tools being used across the sector to ensure policy compliance.

Source: Over 80% of research outputs meet requirements of REF 2021 open access policy – Research England

Turning FAIR data into reality: interim report from the European Commission Expert Group on FAIR data | Zenodo

Authors: Hodson, Simon; Jones, Sarah; Collins, Sandra; Genova, Françoise; Harrower, Natalie; Laaksonen, Leif; Mietchen, Daniel; Petrauskaité, Rūta; Wittenburg, Peter

Notes: Extensive report that provides a large set of recommendations likely to be taken up by the EC. Focus is on large scale institutions and infrastructures more than universities as well as the broader ecosystem but many of the recommendations have implications for university policy and actions.

Abstract: Interim report of the European Commission Expert Group on Turning FAIR Data into reality. The Group has a remit to provide recommendations, indicators and input on the financing of activities required to turn FAIR data into reality at an EU, Member State and international level. A FAIR Data Action Plan has also been proposed. See https://doi.org/10.5281/zenodo.1285290

The interim report will be formally released at the EOSC Summit on 11 June 2018 in Brussels, where a workshop will be run to consult on the recommendations and Action Plan. The report will be open for comments via a stakeholder consultation in June-August 2018.

The FAIR Data Expert Group was also asked to contribute to the evaluation of the Horizon 2020 Data Management Plan template and future revisions in light of harmonisation with funders across the EU, including the development of additional sector/ discipline specific guidance (if desired). A separate report was published on this in Spring 2018. See https://doi.org/10.5281/zenodo.1120245

Hodson, Simon, Sarah Jones, Sandra Collins, Françoise Genova, Natalie Harrower, Leif Laaksonen, Daniel Mietchen, Rūta Petrauskaité, and Peter Wittenburg. “Turning FAIR Data into Reality: Interim Report from the European Commission Expert Group on FAIR Data,” June 7, 2018. https://doi.org/10.5281/zenodo.1285272.

Source: Turning FAIR data into reality: interim report from the European Commission Expert Group on FAIR data | Zenodo

It’s Time to Make Your Data Count!

Author: Daniella Lowenburg

Notes: The Making Data Count project is a Sloan funded effort to develop standardised metrics for data usage across data repositories. It represents the most general effort to track usage for generic research data to date. Here they report progress within two repositories (California Digital Library and DataONE) and are seeking to get engagement from other repositories to expand the program.

Summary: One year into our Sloan funded Make Data Count project, we are proud to release Version 1 of standardized data usage and citation metrics!

As a community that values research data it is important for us to have a standard and fair way to compare metrics for data sharing. We know of and are involved in a variety of initiatives around data citation infrastructure and best practices; including Scholix, Crossref and DataCite Event Data. But, data usage metrics are tricky and before now there had not been a group focused on processes for evaluating and standardizing data usage. Last June, members from the MDC team and COUNTER began talking through what a recommended standard could look like for research data.

Since the development of our COUNTER Code of Practice for Research Data we have implemented comparable, standardized data usage and citation metrics at Dash (CDL) and DataONE.

Source: It’s Time to Make Your Data Count!

The Landscape of Research Data Repositories in 2015: A re3data Analysis

TiTle: The Landscape of Research Data Repositories in 2015: A re3data Analysis

Authors: Maxi Kindling et al

https://doi.org/10.1045/march2017-kindling

Summary: Analysis of data repositories in re3data shows a range of access, software, APIs, PIDs used as well as content, owners and countries. Limited standard compliance was noted.

re3data now provides much of this info on its metrics page https://www.re3data.org/metrics

D-Lib Magazine March/April 2017
Volume 23, Number 3/4

Abstract

This article provides a comprehensive descriptive and statistical analysis of metadata information on 1,381 research data repositories worldwide and across all research disciplines. The analyzed metadata is derived from the re3data database, enabling search and browse functionalities for the global registry of research data repositories. The analysis focuses mainly on institutions that operate research data repositories, types and subjects of research data repositories (RDR), access conditions as well as services provided by the research data repositories. RDR differ in terms of the service levels they offer, languages they support or standards they comply with. These statements are commonly acknowledged by saying the RDR landscape is heterogeneous. As expected, we found a heterogeneous RDR landscape that is mostly influenced by the repositories’ disciplinary background for which they offer services.

Keywords: Research Data Repositories, RDR, Statistical Analysis, Metadata, re3data, Open Science, Open Access, Research Data, Persistent Identifier, Digital Object Identifier, Licenses

Source: The Landscape of Research Data Repositories in 2015: A re3data Analysis