Research Papers

Sharing gray academic literature with ResearchGate DOIs: Increased discoverability but inaccurate metadata

  • Enrique Orduña-Malea
Expand
  • The iMetrics Lab., Universitat Politècnica de València, 46022 Valencia, Spain
Enrique Orduña-Malea (Email: ; ORCID: 0000-0002-1989-8477).

Received date: 2025-03-31

  Revised date: 2025-05-23

  Accepted date: 2025-06-19

  Online published: 2025-07-07

Abstract

Purpose: To describe the characteristics of research outputs using persistent identifiers generated by ResearchGate to gain insight into what publications are shared and disseminated through this functionality, revealing their academic and non-academic impact.

Design/methodology/approach: A total of 1,092,934 RG-DOIs were collected, using the DataCite API, along with bibliographic metadata for the associated registered output (RG-DOI publications). The subsequent analysis evaluated the publication date, document type, and language. These values were cross-referenced against the full text of a random sample of 666 records to verify accuracy.

Findings: RG-DOIs have served primarily to identify and make accessible scholarly gray literature, including posters, presentations, conference papers, and theses, with notable emphasis on publications in Spanish and Portuguese. Around 41,000 citations from Web of Science indexed publications to RG publications are evidence of their infrequent but perceptible use in scholarly discourse. The declining number of registrations of RG-DOIs observed may indicate a shift in researcher preferences to alternative platforms for DOI generation.

Research limitations: The study uncovered substantial inconsistencies in DataCite metadata, which can be attributed to the automated DOI registration process and internal changes in the available document types on ResearchGate.

Practical implications: The study encountered challenges in conducting a quantitative analysis due to inconsistencies in the metadata. These have potential implications for researchers, practitioners, and librarians relying on RG-DOIs to conduct bibliometric or bibliographic analysis.

Originality/value: This study is the first comprehensive analysis of RG-DOIs and, as such, provides a unique perspective into academic gray literature. It also sheds light on the quality of ResearchGate data transmitted to DataCite when registering DOIs.

Cite this article

Enrique Orduña-Malea . Sharing gray academic literature with ResearchGate DOIs: Increased discoverability but inaccurate metadata[J]. Journal of Data and Information Science, 2025 , 10(4) : 44 -65 . DOI: 10.2478/jdis-2025-0039

1 Introduction

If we consider research work as a sequential process guided by the documentary records thus generated, then the researcher’s work can be seen as constituting five principal stages: learning, elaborating, communicating, depositing, and disseminating.
In the learning phase, the researcher acquires information via specific courses, conferences, work meetings, and by reading scientific publications. In the following phase, that of elaboration, they obtain information using a range of different research methods (e.g. laboratory work, fieldwork, and documentary work). This information (raw or elaborated) is then recorded, and communicated by means of a variety of scholarly outputs.
Some outputs undergo scholarly research quality assurance and assessment (i.e. peer-review) to validate their conclusions, methods, and claims, prior to their communication via formal publication (Tennant & Ross-Hellauer, 2020). Peer-reviewed publications are considered the most authoritative and recognizable research output (Huang & Chang, 2008) and are broadly used in research evaluation tasks as signals of scholarly reputation (Nicholas et al., 2015; Petersen et al., 2014). These publications, moreover, serve an archival function, preserving and making subsequent retrieval possible (Garvey & Griffith, 1972).
Other outputs are shared privately with colleagues or publicly published outside the scholarly publishing circuit. Non-reviewed documents comprise material supporting peer-reviewed publications (e.g. manuscript drafts, methodological designs, raw data, and scientific software), discarded material (e.g. personal notes, negative results, and preliminary designs), early material (e.g. research proposals), and professional publications (e.g. reports, working papers, technical papers, comments, blog posts, and opinion papers).
Scholarly output can subsequently be deposited on online platforms (e.g. publisher websites, personal websites, bibliographic databases, and repositories) for preservation and where their dissemination is ultimately enabled via different media. This last step might involve elaborating additional non-reviewed promotional documents (e.g. press releases, social media posts, informative videos, and infographics).
The advent of the academic web and social networking sites has facilitated the sharing of non-reviewed documents by researchers (Barjak, 2006; Priem, 2013). However, their collection and quantitative analysis via Altmetrics aggregators raises various challenges, as many non-reviewed documents lack persistent identifiers, as they have not gone through a formal publication process managed by publishers, bibliographic databases, or libraries. Thus, both their diffusion and traceability are still limited.
Many repositories allow the inclusion of non-reviewed documents, such as conference presentations, datasets, reports, and images, assigning them with persistent identifiers (e.g. handles). However, this practice does not solve this problem. First, the inclusion of non-reviewed documents depends on policy guidelines, which can differ significantly from one repository to another. Second, repositories are filtered thematically or institutionally, making global analysis unfeasible. And, third, for various reasons - including ignorance or a preference for other external services - many researchers do not use the services offered by repositories (Borrego, 2017).
Among these external services, the role played by academic social networking sites stands out (Jordan, 2014, 2019; Ortega, 2016; Van Noorden, 2014), especially that of ResearchGate (RG). This networking site allows its users to deposit a wide variety of non-reviewed documents and to generate persistent identifiers free of charge using the digital object identifier (DOI) standard (henceforth RG-DOIs), thereby facilitating the dissemination and citation of academic documents that have not been published via formal scientific communication channels, and taking advantage of RG services for their dissemination and monitoring.
Determining the quantity and basic bibliographic properties of the documents deposited in RG and identified with an RG-DOI (i.e. RG-DOI publications) could provide insights into how these informal publications are shared and disseminated, as well as unveiling their scholarly and non-scholarly impact. Nonetheless, using RG for bibliographic analysis poses two significant challenges: First, massive access to data is compromised since there is no API to collect metadata; and, second, the metadata quality might be limited, as users describe their publications without any bibliographic controls (Orduña-Malea & Delgado López-Cózar, 2020).
This study aims to describe the characteristics of this gray literature, while seeking to address obvious data collection challenges and descriptive limitations. To do so, the following research questions are raised:
RQ1. How many RG-DOIs have been assigned over time?
RQ2. What are the bibliographic properties of RG-DOI publications?
RQ3. What is the metadata accuracy of RG-DOI publications?
RQ4. What are the scholarly and non-scholarly impact of RG-DOI publications?

2 Research background

ResearchGate was created in 2008 with the aim of facilitating interconnectivity between academics via different social functionalities, both free and paid (Orduña-Malea & Delgado López-Cózar, 2020).
Among the free functionalities, the most notable are the deposit and dissemination of academic works, the collection of scholarly metrics, asynchronous messaging among researchers, a Q&A service, alerts, and job offers, while among the paid functionalities, we find marketing services (sponsored content, sponsored email, display ads, native ads, and event promotions), branding (publisher solutions), and recruitment (job postings).
According to official information delivered by ResearchGate, as of December 12, 2023, the platform covers 156 million publications and is used by 23 million registered users, mainly from Europe (33%) and Asia (28%), who specialized mainly in engineering (19%), medicine (15%), and biology (13%). The service receives 297 million monthly views on its publication pages, making ResearchGate a relevant platform for carrying out quantitative science studies.
The scientific literature on RG has grown significantly since 2015 (Prieto-Gutierrez, 2019) and has tended to focus on three main aspects: scholarly impact, individual and collective scholarly practices, and social and reputational features (Manca, 2018).
In the case of scholarly impact, ResearchGate has been characterized as a platform with distinctive publication and citation coverage, differing from other bibliographic databases (Singh et al., 2022), indexing a remarkable number of recent publications (Thelwall & Kousha, 2017b) and academic gray literature (Nicholas et al., 2016b), with biases towards experimental science disciplines (Thelwall & Kousha, 2017a), and exhibiting higher visibility for research institutions with a high number of academic staff (Lepori et al., 2018). In this line of research, we find studies on academic publications (Thelwall & Kousha, 2017a), universities (Lepori et al., 2018; Yan et al., 2018; Yan & Zhang, 2018), open data (Raffaghelli & Manca, 2023), and researchers, the latter of which is filtered by research organizations (Mikki et al., 2015; Ortega, 2015) and disciplines (Orduna-Malea & Delgado López-Cózar, 2017; Yan et al., 2021).
The study of scholarly impact has also evidenced the existence of a variety of indicators, where citation-based metrics tend to correlate strongly, not only within ResearchGate but also across platforms, but achieve a lower correlation with alternative metrics (Martín-Martín et al., 2018; Mikki et al., 2015), reflecting the dependency on users’ activity within the platform. Otherwise, RG’s indicators (e.g. RG Score) have been widely criticized for their opacity, lack of reproducibility, and susceptibility to manipulation by fraudulent behaviors aimed at achieving academic status (Copiello, 2019; Copiello & Bonifaci, 2018, 2019; Kirilova & Zoepfl, 2025; Kraker & Lex, 2015; Orduna-Malea et al., 2017; Yu et al., 2016).
Regarding scholarly practices, ResearchGate has been characterized as a platform with a significant representation of early career researchers, a strong bias towards engineering and biomedicine, and an underrepresentation of female authors (Mikki et al., 2015; Ortega, 2017).
While the literature highlights three general ResearchGate user types (information source users, friend users, and information seeker users) (Yan et al., 2018), the platform is mainly used for knowledge sharing, especially using collaborative Q&A strategies (Ebrahimzadeh et al., 2020; Jeng et al., 2017; Li et al., 2018), accessing literature for personal consumption (Mason & Sakurai, 2021), posting English-language outputs, and featuring publications recently published, highly cited, and published by reputable sources (Liu & Fang, 2018), often infringing copyright issues (Borrego, 2017; Jamali, 2017; Lee et al., 2019).
However, despite the advantages of the platform, most users do not use their RG account heavily, spending little time searching and updating their profiles (Meier & Tunger, 2018), and perceiving negative feelings related to productivity and stress (Muscanell & Utz, 2017).
In the case of platform features, the literature has analyzed ResearchGate as a driver of scholarly reputation, through the availability of academic social networking features (e.g. open review, following users or projects, and sending emails), reputational indicators (e.g. RG Score, Reads, and Recommendations), and marketing-based mechanisms (e.g. e-mails informing users on their peers’ contributions, activities, and achievements) (Nicholas et al., 2016a, 2016b; Orduna-Malea et al., 2016; Orduña-Malea & Delgado López-Cózar, 2020) aimed all of them to show academic research as a game to motivate and promote users’ participation (Hammarfelt et al., 2016). Specifically, ResearchGate excels in its analytical capabilities, interactivity, and ease of publication deposit, while its search, browsing, site navigation, and export functions are weaker (Bhardwaj, 2017).
Among its features, ResearchGate launched the free DOI service on August 13, 2014, to allow its users to showcase their research, make it citable, and date their discoveries (ResearchGate, 2014). DOIs created by RG (RG-DOIs) are generated at the user’s request once the publication has been deposited on the platform. This option is available on the “More” tab and is only available for certain document types. The system is automatic and instantaneous, generating a permanent identifier registered by Datacite. The RG-DOIs are embedded in doi.org and rgdoi.net domain names, with “10.13140” as the DOI prefix.
This functionality might influence how researchers share their unpublished work, while impacting the competition between RG and institutional repositories in their role as agents for the preservation of digital scholarly documents. It also raises the question as to whether RG should be considered a “publisher” (Konkiel, 2014).
Given that users directly request DOIs when depositing their work in RG, an analysis of RG-DOIs would facilitate the study of an unknown dimension of DOIs, that is, their use to identify and describe scholarly publications either unpublished or originally published without a DOI.

3 Methods

3.1 RG-DOIs collection

First, the DOI Resolver was used to determine the registration agency that generates RG-DOIs. For this purpose, a random RG-DOI test was used via HTTPS, returning DataCite as the registration agency. A Python script was then designed to recover all findable DOIs generated by RG using the DataCite REST API.
The query returned 1,092,934 RG-DOIs registered as of May 21, 2023. For each RG-DOI, the DataCite REST API response included all available bibliographic metadata of the corresponding RG publication in the JSON format. Python scripts were subsequently built to extract the metadata obtained. The metadata fields eventually analyzed here were the registration date, publication date, language, and Resource Type.

3.2 RG-DOIs analysis

The number of citations from formally published publications determined the citation-based scientific impact of publications identified with RG-DOIs. The Web of Science’s Cited References database was searched through the “cited DOI” search field, using the ResearchGate DOI prefix as a string and the asterisk as Wild Card (i.e. “10.13140*”). This process was carried out on May 22, 2023, obtaining 24,123 records. A manual data cleaning was then carried out to eliminate broken or invalid DOIs, obtaining 21,092 unique RG-DOIs.
Altmetrics mentions were used to determine the non-scholarly impact of the publications with RG-DOIs. A Python script was again designed to collect all RG-DOI mentions using Elsevier’s PlumX Metrics API.
It was decided to carry out two data collections (May 22 and October 22, 2023) to determine the coverage of RG-DOIs in PlumX with greater robustness. All statistics were performed with the most recent data (October). However, the data from May were used to make internal comparisons between the data collections.
PlumX only provides the raw number of Mendeley Readers, so an additional script was created to collect data about Readers’ roles through the Mendeley API. Data were collected on October 29, 2023, yielding 7,603 records with Readers’ data.

3.3 RG-DOIs metadata accuracy

Given that RG users directly deposit and describe their publications, we need to determine how accurate the metadata provided by DataCite is. A random sample of 666 RG-DOIs (99% confidence interval with a margin of error of 5%) was taken to check the metadata accuracy.
The author manually checked the publication date and document type in the DataCite records (resourceType field), RG’s publication page, and the document’s full text for each RG-DOI. This process was carried out between October 12 and October 15, 2023.

4 Results

4.1 Output

A total of 1,092,934 RG-DOIs have been registered as of May 21, 2023 (0.6% of all publications deposited in RG) since the launch of the DOI generation service in 2014. The two years following its implementation recorded the highest annual number of registrations (154,428 and 147,515 RG-DOIs, respectively). However, starting in 2016, the number of registrations decreased sharply, before rising again in 2019 and 2020 due to the increase in platform use during the COVID-19 lockdown (ResearchGate, 2021). Since then, the number of RG-DOIs has once again fallen, with 2022 registering the lowest annual value since the service was launched, with 86,138 new RG-DOIs (Figure 1).
Figure 1. Number of RG-DOIs registered over time. Note: 2023 is incomplete; 184 records appear without the year of registration in the metadata field.

4.2 Bibliographic properties

4.2.1 Year of publication

The years with the most publications correspond to 2015 (123,396), 2016 (105,489), 2019 (103,461), and 2020 (100,929), which coincide with the years recording the most RG-DOI registrations (see Figure 1). However, 332 records have been identified without a publication date, all with a registration date of 2016.
The publication date of 78.53% of the RG-DOI registrations lies between 2014 and 2023. The remaining 21.4% (234,342 RG publications)-a significant percentage-correspond to previous publications, except 2,874 records for which the publication date is later than the RG-DOI registration date.
More specifically, the RG-DOI registration date does not coincide with the publication date in 43.8% of the records. These data were especially at variance in 2014 (the first year of the RG-DOI service), when the discrepancy reached 65.6% (Table 1).
Table 1. Differences between the registration and publication dates.
Year of
Registration
RG-DOIs with other year of publication % Difference
(Avg.)
RG-DOIs with same year of publication % Total RG-DOI
publications
2014 62,114 65.6 5.9 32,636 34.4 94,750
2015 72,265 46.8 6.3 82,164 53.2 154,429
2016 75,617 51.3 5.7 71,900 48.7 147,517
2017 48,624 46.6 5.3 55,822 53.4 104,446
2018 37,188 37.5 5.3 62,043 62.5 99,231
2019 46,372 38.2 5.4 75,137 61.8 121,509
2020 53,984 39.3 5.0 83,401 60.7 137,385
2021 39,705 35.2 5.0 73,049 64.8 112,754
2022 28,579 33.2 4.6 57,559 66.8 86,138
2023 14,121 40.8 4.4 20,473 59.2 34,594
Null N/A N/A N/A N/A N/A 181
All 478,750 43.8 5.2 614,184 56.2 1,092,934

Note: N/A: data not available.

4.2.2 Document types

Nineteen different document types with at least one registered RG-DOI have been identified (Table 2), among which Posters (13%), Presentations (12.6%), and Conference papers (10.8%) are most evident. However, there is no information on the document type for a significant number of records (26.1%).
Table 2. Number of RG-DOI publications according to document type.
Document type RG-DOI publications %
Null 284,985 26.1
Poster 141,725 13.0
Presentation 137,372 12.6
Conference Paper 118,238 10.8
Preprint 101,356 9.3
Thesis 86,767 7.9
Technical Report 78,429 7.2
Other 66,513 6.1
Research Proposal 16,452 1.5
Working Paper 13,526 1.2
Book 13,231 1.2
Experiment Findings 11,926 1.1
Method 9,375 0.9
Publication 6,471 0.6
Code 4,570 0.4
Chapter 1,425 0.1
Negative Results 218 0.0
Patent 157 0.0
Article 143 0.0
Cover Page 55 0.0
Total 1,092,934 100

Note: Categories based on the “ResourceType” metadata field.

On the other hand, a minority presence of formal document types is observed, such as Articles and Chapters, which do not necessarily have a DOI (Khurana et al., 2023).
Considering the last full year available (2022), the most common types are Posters, Presentations, Preprints, and Doctoral Theses (Figure 2). Indeed, these first two types have experienced significant growth over the years.
Figure 2. Number of RG-registered DOIs per document type over time. Note: The year 2023 is incomplete.
Anomalous behavior was detected in 2017, with the number of some document types growing significantly from 2016 to 2017. These included Other (from 2,073 to 8,686 publications), Posters (from 4,469 to 24,176), Presentations (from 3,627 to 21,039), and Working Papers (from 2,067 to 10,797). However, the number of records decreased abruptly in the case of other document types, including Conference Papers (from 14,579 to 558) and, most remarkably, those publications without a defined document type (from 112,319 to 6,856).

4.2.3 Languages

DataCite provides language information for 67.5% of total publications with RG-DOI, identifying up to 26 different languages, with English being the predominant language (74.2%), followed at some distance by Spanish (7.2%) and German (4.6%). If we compare these results with the distribution of languages for publications indexed in Scopus (Table 3), a number of significant differences can be observed. On the one hand, some languages appear overrepresented in RG (e.g. Catalan, Indonesian, Malay Portuguese, Romanian, Slovak, and Spanish). In contrast, other languages are absent (e.g. Chinese, Greek, Japanese, Korean, Russian, and Turkish).
Table 3. Number of publications with RG-DOIs according to language.
Language ResearchGate Scopus
Number of RG-DOI publications (%) Number of publications (%)
English 549,356 74.44 82,061,257 87.45
Spanish 53,403 7.24 946,960 1.01
German 33,689 4.57 2,673,428 2.85
Portuguese 26,097 3.54 354,693 0.38
French 18,012 2.44 1,511,694 1.61
Catalan 11,364 1.54 5,415 0.01
Indonesian 6,695 0.91 2,755 0.00
Malay 6,569 0.89 2,137 0.00
Italian 6,351 0.86 542,782 0.58
Romanian 5,731 0.78 46,341 0.05
Slovak 3,880 0.53 37,182 0.04
Dutch 2,469 0.33 154,781 0.16
Swedish 2,138 0.29 74,894 0.08
Danish 2,092 0.28 68,904 0.07
Hungarian 1,939 0.26 73,064 0.08
Esperanto 1,854 0.25 170 0.00
Other 6,322 0.86 5,286,659 5,63
Total 1,092,934 100 93,843,116 100

Note: Only languages with at least 1,000 records for publications with RG-DOIs.

4.3 Metadata accuracy

For the random sample of 666 documents, the accuracy of the publication date of DataCite records was verified by comparing the date indicated in DataCite’s metadata with the date stated in the document itself. Five RG-DOIs were removed from the analysis as they were no longer registered in DataCite. For the remaining 661 documents, the date was correct in 46.9% of cases and incorrect in 8.6% (most notably the case of Presentations and Theses). However, 44.1% of the documents could not be verified as the date was not indicated in the full text of the deposited document. The absence of a publication date was especially apparent in Posters (58 undated publications), Presentations (35), Conference Papers (34), and Preprints (31), as well as in publications without an assigned category (87). In the case of datasets, even though the date was the same as the deposit date in some cases, there is no guarantee that this dataset had not been created earlier and deposited on some other platform.
In the case of the precision in the identification of document type, for 34.6% of the publications there was no correspondence between the document type indicated in DataCite’s metadata and the type displayed on the publication page. This discrepancy is fundamentally attributable to publications without a type in DataCite (i.e. Null). After accessing the publication page of said RG-DOIs, the author was able to verify that these documents were labeled as Data (45 documents) and Research (43). It should be stressed that the “Research” tag does not appear among the document types that users can employ to categorize publications on RG, nor does it appear as a document type in DataCite.
Likewise, no agreement was found between the label assigned to the publication page and the actual document type, identified after a manual inspection of the full text. In this case, a lack of agreement was detected for 40.5% of the publications in the sample. The errors concentrated in the Research (85 incorrectly assigned documents, of which 26 were preprints, 17 articles, and 13 Reports), Conference Papers (35 incorrectly assigned documents, of which 17 were preprints and 10 posters), and Presentation (20 misassigned documents, 11 of which were preprints) labels.
Agreement between the typology indicated in the metadata and the actual document type was low (43.4%), mainly as a result of documents not being categorized or being categorized as Other. Many discrepancies were also found for the Conference Paper type (40 documents incorrectly categorized, of which 18 were preprints, 10 posters, and 6 presentations).
If we exclude those documents that were not categorized in DataCite, the agreement percentages in the case of document typology increase significantly, especially between DataCite and the Publication page, as can be seen in Table 4.
Table 4. Percentage of agreement between the actual document type, the document type indicated in DataCite, and the document type displayed on the publication page.
Metadata Page Real
All
N= 661
Without null
N= 485
All
N= 661
Without null
N= 485
All
N= 661
Without null
N= 485
Metadata -- -- 65.4% 90% 43.4% 59.8%
Page 65.4% 90% -- -- 59.5% 61.7%
Real 43.4% 59.8% 59.5% 61.7% -- --

4.4 Impact

The number of RG-DOI publications cited by publications indexed in WoS stands at 20,954 (1.92% of the total RG-DOIs identified), accumulating a total of 41,492 citations. The most cited publication is a book cited 1,932 times (Popay et al., 2006). The average number of citations to these publications is low (1.98), exhibiting a non-linear distribution, in which the 10 publications with the most citations concentrate 7.9% of the citations to RG-DOIs, while 73.3% of the RG-DOI publications receive one citation each.
In absolute terms, Books, Conference Papers, Technical Reports, and Theses are the most frequently cited document types, along with preprints (Table 5). Otherwise, despite being the most frequent typologies with RG-DOI, only 0.68% of Posters and 0.60% of Presentations are cited at least once. In addition, the significant average number of citations received by Code stands out, despite the low number of RG-DOIs assigned to this category (74).
Table 5. Number of documents with RG-DOI cited by WoS publications.
Document type Number of records % Citations Average Citations Median Citations Max
Null 5,603 1.97 2.15 1.00 145
Conference Paper 3,357 2.84 1.89 1.00 217
Technical Report 2,936 3.74 1.90 1.00 152
Preprint 1,978 1.95 1.65 1.00 66
Thesis 1,334 1.54 1.41 1.00 19
Other 984 1.48 1.94 1.00 108
Poster 962 0.68 1.19 1.00 17
Presentation 839 0.61 1.28 1.00 26
Book 624 4.72 5.37 1.00 1,932
Working Paper 375 2.77 1.78 1.00 34
Method 190 2.03 2.04 1.00 35
Experiment Findings 134 1.12 1.49 1.00 9
Publication 133 2.06 3.84 1.00 39
Research Proposal 102 0.62 1.55 1.00 30
Code 74 1.62 4.03 1.00 98
Chapter 30 2.11 1.33 1.00 5
Article 4 2.80 1.75 1.00 4
Cover Page 2 3.64 6.50 6.50 10
Patent 2 1.27 1.50 1.50 2
Negative Results 1 0.46 1.00 1.00 1
In any case, the low median values (1 in almost all document types) reflect that most publications accrued just one citation, which a few publications surpassing 100 citations, as the boxplot displayed in Figure 3 shows.
Figure 3. Number of citations received by RG-registered DOIs per document type.
In the case of alternative metrics, only 0.9% (10,335) of the RG-DOIs have received at least one mention. In absolute terms, Conference Papers and, to a lesser extent, Technical Reports, are the most frequently mentioned document types. As can be seen in Table 6, the most important source of mentions is Mendeley, which accrues 81,884 Reads from 7,904 RG-DOIs, while the number of mentions from news (415 mentions to 176 RG-DOIs) and policy documents (1,418 mentions to 921 RG-DOIs) is low.
Table 6. Number of records with Altmetrics performance per record type and metric.
Record types Total % News count Readers count Policy citations
Null 3,725 1.3 84 2,690 428
Conference Paper 3,083 2.6 28 2,683 100
Technical Report 622 0.8 26 326 178
Book 581 4.4 5 436 55
Thesis 492 0.6 2 421 21
Poster 431 0.3 6 332 10
Presentation 386 0.3 1 294 20
Preprint 286 0.3 12 195 17
Other 275 0.4 5 218 29
Working Paper 190 1.4 3 107 48
Publication 85 1.3 1 52 7
Code 69 1.5 2 62 1
Chapter 43 3.0 0 41 1
Method 27 0.3 0 18 3
Research Proposal 19 0.1 0 14 0
Experiment Findings 15 0.1 0 10 1
Article 4 2.8 0 3 0
Cover Page 1 1.8 0 0 1
Patent 1 0.6 0 1 0
TOTAL 10,335 175 7,903 920

Note: One record may have been performed in more than one PlumX metrics. Source: PlumX.

Of the 19,885 Reads captured directly from the Mendeley API, a high percentage are determined as being attributable to students (47.7%), including undergraduates (1,798), postgraduates (3,541) and doctoral students (4,145). The rest of the Reads are divided between researchers (3,524), lecturers and professors (3,877), others (1,410), and librarians (240). The role of the readers of 262 Reads remains unspecified.

5 Discussion

Although a significant number of registered RG-DOIs was identified (around 1 million), the number of registrations has fallen quickly in recent years. This drop in usage could reflect a decrease in the number of users on RG. However, official data actually indicate a continuous increase in users (19 million in 2020 and 22.5 million in 2021). Thus, a plausible explanation might be that users are tending to use the DOI functionality less owing to the existence of other platforms that also permit the generation of DOIs, including OSF, Zenodo, Orvium, and Figshare. In addition, institutional repositories have begun to generate DOIs for certain document types, such as theses and datasets.
While the results indicate that users took advantage of the creation of this functionality to deposit and disseminate gray academic publications, RG-DOI publications concentrated on the years when the DOI registration functionality was created, indicating a preference in RG for recent articles, an aspect previously noted by Thelwall and Kousha (2017a). However, the lack of precision in the publication date data (see Section 4.3) means we need to exercise caution when dealing with the results in Table 4.
In the case of document typology, results show the relevance of publications associated with scientific events (e.g. posters, conference papers, and presentations), doctoral theses, reports (including working papers and technical reports), and preprints. However, the analysis of document types highlights a series of difficulties.
First, more than a quarter of the RG-DOIs (around 285,000 records) are not assigned a document type: that is, no information is included in the “ResourceType” metadata provided by DataCite. The absence of this field for such many records is unexpected, given that users must specify the document type (required field) at the time of deposit, and this category does appear on the publication page. It might be the case that users believe they can change the document type after registration. Yet, RG instructions explicitly state:
“Once a ResearchGate DOI has been generated for a research item, you can no longer edit that item. Instead, you should remove the research completely, re-upload it with the edits and generate a new DOI.”
However, the author was allowed to change a publication assigned as Poster with an RG-DOI to another document type, thus contradicting the official information provided by RG. This could account for the inconsistencies between the DataCite data (which do remain unchanged) and the publication page data, but it does not account for the number of Null fields.
Remarkably, 83.7% of all records without document-type metadata were accumulated between 2014 and 2016. The number of documents without the “ResourceType” metadata fell from 2017 onwards, although it remains significant to this day (4,963 records in 2022). This can probably be attributed to the platform’s internal control process. The author contacted RG to clarify this situation, but has yet to receive a response.
Second, the document typologies identified yielded additional contradictory results. The official information from RG indicates that RG-DOIs cannot be generated for Articles, Books, Chapters, Conference Papers, Cover Pages, Patents, and Posters since they are considered to have been published elsewhere and may already have a DOI. However, RG-DOIs assigned to Articles (143), Patents (157), Chapters (1,425), Books (13,231), and, most significantly, Conference Papers (118,238) and Posters (141,725) have been identified. To verify this situation, the author attempted to create RG-DOIs for each document type manually, finding it possible in the case of Posters but not for Articles, Books, Chapters, Conference Papers, Cover Pages, and Patents. It therefore seems plausible that the generation policy has been changed, with the exception of Posters, which would appear to be an error awaiting correction on RG’s website.
Third, RG’s classification, based on 18 document types, introduces various inconsistencies that jeopardize any quantitative analyzes by document type. For example, the Methods category is inappropriate since it confuses document type (e.g. Journal Article, Book chapter, and Book) with genre (e.g. Method, Bibliographic Review, Book Review, Editorial, Letter to Editor, etc.). In fact, this error is also present in many bibliographic databases. Other categories are ambiguous. For example Presentation, which might correspond to conference keynotes, teaching material, supplementary material, or even a Poster.
Fourth, DataCite’s metadata includes document types not currently available for users when describing their publications. For example, since 2017, there have been no records labeled “Publication.” The case of “Working Paper” is more complex. This document type experienced a peak in records precisely in 2017 (with 10,797 documents) and then practically disappeared and is currently unavailable as a document type. However, DataCite assigned the “Working Paper” label to 20 records in 2022. The type “Other” (with 6,900 records in 2022) is also present in DataCite but is unavailable for users. In addition, the “Research” category is unavailable for users when describing their publications but appears as a document-type label in RG’s publication pages.
Fifth, the changes recorded in 2017 can be attributed to the fact that five publishers (American Chemical Society, Brill, Elsevier, Wiley, and Wolters Kluwer) formed the Coalition for Responsible Sharing following unsuccessful attempts to find ways for RG to run its service without breaching publishers’ copyright (Van Noorden, 2017). Consequently, in 2017, ResearchGate removed many copyrighted articles from public view. This circumstance could explain the data shown in Figure 2. It is not in vain that the Preprints category has been growing since 2018, probably because of this legal issue.
In the case of the document language, the results reported here show the predominance of English, which is consistent with the global distribution of languages by scholarly output. While Scopus was used as a baseline for global language distribution in academic publications, which might differ from that of other bibliographic databases (Vera-Baceta et al., 2019), this comparison provides evidence of a greater than expected number of publications in Portuguese and, especially, in Spanish. These results may reflect the high use of RG in these countries and the activity of users in South America, who make up 8% of all RG members. German, the second language in global scholarly output according to Scopus, and the place of tax residence of RG, also stands out in this regard. Otherwise, the absence of publications with RG-DOIs in languages with non-Latin alphabets (e.g. Russian, Japanese, Korean, and Chinese) is significant, and even more so when there are publications deposited in RG written and described in these languages. It is unlikely that any author from these countries will have submitted a publication to RG without requesting an RG-DOI; thus, the error could be in the detection system. Notwithstanding, Thelwall and Kousha (2015) detected the infrequent use of RG by scholars from China, South Korea, and Russia, a factor that may well be related. The distribution of RG users at the country level should help shed light on this issue. Unfortunately, RG only provides statistics of RG members by continent.
The number of publications with RG-DOIs that do not specify the language in DataCite is also remarkable (354,973). This circumstance can be considered unexpected given that, while RG does not request the language during the deposit process, it should detect the language used automatically. For this reason, it is unclear as to why this information does not appear in DataCite’s metadata.
The high percentage of discrepancies between the actual document type, the document type indicated in DataCite, and the document type displayed on the RG publication page suggests that the metadata provided by DataCite for RG-DOIs are not entirely reliable. These discrepancies are partially associated with human errors when depositing documents (e.g. year of publication and document typology), which are not verified. For example, RG-DOI “10.13140/RG.2.1.3702.8566” is displayed as “Research” on the publication page. However, the full text actually includes a journal article postprint, with its own DOI (“10.1108/IJEM-10-2014-0140”).
The need to establish moderation mechanisms has been discussed in relation to thematic repositories, especially due to the quality of the deposited works (Silagadze, 2023). Indeed, platforms such as OSF have started piloting post-moderation policies as of November 2023 to stop bot-generated, plagiarized, and self-plagiarized content from being submitted to the platform. However, giving users full freedom to describe the publications they deposit means the metadata generated cannot be used to carry out scientific studies owing to problems of consistency.
In some cases, users who deposit publications fail to include the year of publication in the document itself. For example, 42 presentations (51.2%) and 67 posters (70.5%) in the random sample do not include any publication date, but rather the authors assign the date of deposit as the date of publication in the metadata. In other cases, the inconsistency is associated with the document type. For example, authors assign a publication to the category Article but, in fact, deposit a preprint (for which there is insufficient information to determine whether it corresponds to an Article, Chapter, Conference Paper, or Report) or, vice versa, they assign the publication to the preprint category but actually deposit an Article, giving rise to copyright infringements in RG, an occurrence reported by Jamali (2017). Recent cooperation agreements and partnerships entered into between RG and different publishers (e.g. MDPI, Taylor and Francis, Emerald, American Association for the Advancement of Science, Springer, and Wiley) could help improve the quality of metadata in formal publications. However, since RG-DOIs are assigned to material that has not been formally published, it seems unlikely that DataCite will be able to provide useful data for conducting quantitative studies.
In the case of scientific impact, the percentage of publications with RG-DOIs cited is low (1.92%). However, this result is not unexpected considering the informal nature of many of the publications analyzed. Moreover, the introduction of typographical errors when entering the RG-DOI in the references, and parsing errors in the WoS’s Cited references system may well have further reduced the number of RG-DOIs cited, as occurs with all DOIs indexed in bibliographic references (Franceschini et al., 2015; Xu et al., 2019; Zhu et al., 2019).
Interestingly, 1,290 RG-DOIs extracted from WoS’s Cited References are not indexed by DataCite. Even though some of them are indeed no longer valid or have been removed, others do exist and continue to function. For example, DataCite indicates that the RG-DOI “10.13140/RG.2.2.29474.02242/1” does not exist; however, this identifier is correctly linked to a publication accessible on ResearchGate. These results confirm the existence of more RG-DOIs than those reported by DataCite, probably as a consequence of some technical error in the DOI resolver. Since DataCite is the official DOI registrar for RG, this prevents the systematic collection of all RG-DOIs created. However, it is estimated that the percentage of existing RG-DOIs not identified by DataCite is small and should not affect the general outcomes of the present study.
If we assume that those authors who decide to deposit their publications in RG and register an RG-DOI have a certain interest in circulating their work on social networks, the percentage of RG-DOIs indexed in PlumX (0.9%) is smaller than expected. To determine whether this apparently low coverage was due to an error, two data collections were carried out (May and October, 2023). In the first data collection, 10,303 RG-DOIs were found, and in the second, 10,335 (of which 10,054 were included in the first data collection). Thus, the low coverage of RG-DOIs indexed in PlumX is confirmed.
Finally, the low level of online dissemination of RG-DOIs on Twitter (now X) stands out. In PlumX’s May 2023 data collection, only 420 RG-DOIs embedded in the tweets were identified. Twitter data are no longer reported by PlumX, which means more recent data are not available. It might be the case that most users requesting RG-DOIs limit the dissemination of their Works to RG, and do not use other social media platforms; however, this hypothesis would need to be tested in future studies.
The findings demonstrate, on the one hand, the usefulness of analyzing DOIs generated by academic social networking platforms when studying the dissemination of gray academic literature, as the generated DOIs facilitate the indexing, sharing, and citation of this literature. However, on the other hand, errors found in the metadata (e.g. publication years, language, or document type) limit the accuracy of large-scale RG-DOI analyzes. Both meta-researchers and practitioners should exercise extreme caution when drawing conclusions from the analysis of RG-DOI publications unless manual data cleansing has been applied. These errors, in turn, can spread between interoperable bibliographic systems, which should not ingest RG-DOI publications metadata directly. In the same way, ResearchGate users should not rely on the metadata of these publications, especially on the ResearchGate publication pages, as they may not match the actual records deposited, which could confuse them or limit their user experience on the platform.

6 Conclusions

This study has revealed a significant number of registered RG-DOIs (1,092,934), facilitating access to, and the citation of, academic gray literature. Notably, there is a prevalence of congress materials (presentations, posters, and conference papers), reports, books, and theses. However, a decline in this function of ResearchGate has been identified, possibly attributable to the emergence of other platforms. Indeed, future research might usefully examine the registration, dissemination, and citation of informal publications via alternative platforms, most obviously Zenodo, OSF, and Figshare.
A small percentage of publications with RG-DOIs are cited in WoS-indexed publications, underscoring their utility in scholarly discourse. The 41,492 citations identified are deemed relevant, considering that publications with RG-DOIs constitute document types that are infrequently cited. Future research should seek to characterize these citing publications to ascertain whether the citation of these materials is increasing, and to determine the relevance of the formal sources citing them, the languages of the citing documents, and the rate of self-citation.
The societal impact of publications with RG-DOIs lags behind their scientific impact, as evidenced by their low coverage in PlumX. Employing other aggregators (e.g. Altmetric.com) might yield different results; again, something future studies might be usefully considered. Noteworthy is the interest these publications have garnered in Mendeley, where the number of Reads points to their use in bibliographies, especially among students. However, these works have only limited dissemination on Twitter, an aspect that needs to be further analyzed.
The metadata provided by DataCite are inadequate for conducting quantitative studies of RG-DOIs. This is attributable, on the one hand, to ResearchGate’s automated DOI registration system, and, on the other, to the intervention of potentially non-expert researchers in the provision of scientific information. The lack of verification, in the first instance, and of adequate content curation, in the second, results in inconsistencies in the metadata that impact the year of publication and document type. Additionally, discrepancies are evident in ResearchGate’s official documentation, which seems to permit, on occasions, the generation of DOIs for posters and changes to the document type following DOI registration, thus contradicting the site’s official directives. It is recommended, therefore, that the ResearchGate team updates its information.
Finally, the availability of an API would alleviate dependence on DataCite for gathering bibliographic information on publications with RG-DOIs. It would also provide access to usage and impact metrics-aspects not addressed in this study-that might shed more light on the use of these materials within the scientific community.

Acknowledgements

The author thanks Francisco Jesús Martínez Galindo for his valuable comments on the preliminary versions of this manuscript and Iain Robinson for the review of the English version of this manuscript.

Funding information

This study was supported by Grant PID2022-142569NA-I00, funded by MCIN/AEI/10.13039/501100011033 and by the European Union through “ERDF A way of making Europe.”

Declaration of interest

The author declares no competing interests.

Data availability statements

The data that support the findings of this study are openly available at https://riunet.upv.es/handle/10251/222449. Please cite the source when using the data.
[1]
Barjak, F. (2006). The role of the Internet in informal scholarly communication. Journal of the American Society for Information Science and Technology, 57(10), 1350-1367. DOI: 10.1002/asi.20454

[2]
Bhardwaj, R. K. (2017). Academic social networking sites: Comparative analysis of ResearchGate, Academia.edu, Mendeley and Zotero. Information and Learning Science, 118(5/6), 298-316. DOI: 10.1108/ils-03-2017-0012

[3]
Borrego, Á. (2017). Institutional repositories versus ResearchGate: The depositing habits of Spanish researchers. Learned publishing, 30(3), 185-192. DOI: 10.1002/leap.1099

[4]
Copiello, S. (2019). Research Interest: Another undisclosed (and redundant) algorithm by ResearchGate. Scientometrics, 120(1), 351-360. DOI: 10.1007/s11192-019-03124-w

[5]
Copiello, S., & Bonifaci, P. (2018). A few remarks on ResearchGate score and academic reputation. Scientometrics, 114(1), 301-306. DOI: 10.1007/s11192-017-2582-9

[6]
Copiello, S., & Bonifaci, P. (2019). ResearchGate Score, full-text research items, and full-text reads: A follow-up study. Scientometrics, 119(2), 1255-1262. DOI: 10.1007/s11192-019-03063-6

[7]
Ebrahimzadeh, S., Rezaei Sharifabadi, S., Karbala Aghaie Kamran, M., & Dalkir, K. (2020). Triggers and strategies related to the collaborative information-seeking behaviour of researchers in ResearchGate. Online Information Review, 44 (5), 1077-1096. DOI: 10.1108/oir-12-2019-0380

[8]
Franceschini, F., Maisano, D., & Mastrogiacomo, L. (2015). Errors in DOI indexing by bibliometric databases. Scientometrics, 102(3), 2181-2186. DOI: 10.1007/s11192-014-1503-4

[9]
Garvey, W. D., & Griffith, B. C. (1972). Communication and information processing within scientific disciplines: Empirical findings for psychology. Information Storage and Retrieval, 8, 123-136. DOI: 10.1016/b978-0-08-023344-4.50011-8

[10]
Hammarfelt, B., de Rijcke, S., & Rushforth, A. D. (2016). Quantified academic selves: The gamification of research through social networking services. Information Research, 21(2). Retrieved from https://informationr.net/ir/21-2/SM1.html

[11]
Huang, M. H., & Chang, Y. W. (2008). Characteristics of research output in social sciences and humanities: From a research evaluation perspective. Journal of the American Society for Information Science and Technology, 59(11), 1819-1828. DOI: 10.1002/asi.20885

[12]
Jamali, H. R. (2017). Copyright compliance and infringement in ResearchGate full-text journal articles. Scientometrics, 112(1), 241-254. DOI: 10.1007/s11192-017-2291-4

[13]
Jeng, W., DesAutels, S., He, D., & Li, L. (2017). Information exchange on an academic social networking site: A multidiscipline comparison on ResearchGate Q&A. Journal of the Association for Information Science and Technology, 68(3), 638-652. DOI: 10.1002/asi.23692

[14]
Jordan, K. (2014). Academics and their social networks: Exploring the role of academic social networking sites. First Monday, 19(11). DOI: 10.5210/fm.v19i11.4937

[15]
Jordan, K. (2019). From social networks to publishing platforms: A review of the history and scholarship of academic social network site. Frontiers in Digital Humanities, 6. DOI: 10.3389/fdigh.2019.00005

[16]
Khurana, P., Ganesan, G., Kumar, G., & Sharma, K. (2023). A bibliometric analysis to unveil the impact of digital object identifiers (DOI) on bibliometric indicators. In Singh, P. K., Wierzchoń, S. T., Tanwar, S., Rodrigues, J. J. P. C., Ganzha, M. (Eds). Lecture Notes in Networks and Systems (pp. 859-869). Springer. DOI: 10.1007/978-981-19-1142-2_67

[17]
Kirilova, S., & Zoepfl, F. (2025). Metrics fraud on ResearchGate. Journal of Informetrics, 19(1). DOI: 10.1016/j.joi.2024.101604

[18]
Konkiel, S. (2014). Is ResearchGate’s new DOI feature a game-changer?. OurResearch blog. Retrieved from https://blog.ourresearch.org/researchgate-doi

[19]
Kraker, P., & Lex, E. (2015). A critical look at the ResearchGate score as a measure of scientific reputation. Proceedings of the quantifying and analysing scholarly communication on the web workshop (ASCW’15). Retrieved from https://zenodo.org/records/35401

[20]
Lee, J., Oh, S., Dong, H., Wang, F., & Burnett, G. (2019). Motivations for self-archiving on an academic social networking site: A study on ResearchGate. Journal of the Association for Information Science and Technology, 70(6), 563-574. DOI: 10.1002/asi.24138

[21]
Lepori, B., Thelwall, M., & Hoorani, B. H. (2018). Which US and European Higher Education Institutions are visible in ResearchGate and what affects their RG score?. Journal of Informetrics, 12(3), 806-818. DOI: 10.1016/j.joi.2018.07.001

[22]
Li, L., He, D., Zhang, C., Geng, L., & Zhang, K. (2018). Characterizing peer-judged answer quality on academic Q&A sites: A cross-disciplinary case study on ResearchGate. Aslib Journal of Information Management, 70(3), 269-287. DOI: 10.1108/ajim-11-2017-0246

[23]
Liu, X. Z., & Fang, H. (2018). Which academic papers do researchers tend to feature on ResearchGate?. Information Research, 23(1). Retrieved from http://www.informationr.net/ir/23-1/paper785.html

[24]
Manca, S. (2018). ResearchGate and Academia. edu as networked socio-technical systems for scholarly communication: A literature review. Research in Learning Technology, 26. DOI: 10.25304/rlt.v26.2008

[25]
Martín-Martín, A., Orduña-Malea, E., & Delgado López-Cózar, E. (2018). Author-level metrics in the new academic profile platforms: The online behaviour of the Bibliometrics community. Journal of informetrics, 12(2), 494-509.

DOI

[26]
Mason, S., & Sakurai, Y. (2021). A ResearchGate-way to an international academic community?. Scientometrics, 126(2), 1149-1171. DOI: 10.1007/s11192-020-03772-3

[27]
Meier, A., & Tunger, D. (2018). Survey on opinions and usage patterns for the ResearchGate platform. Plos One, 13(10). DOI: 10.1371/journal.pone.0204945

[28]
Mikki, S., Zygmuntowska, M., Gjesdal, Ø. L., & Al Ruwehy, H. A. (2015). Digital presence of Norwegian scholars on academic network sites - where and who are they?. PloS One, 10(11). DOI: 10.1371/journal.pone.0142709

[29]
Muscanell, N., & Utz, S. (2017). Social networking for scientists: an analysis on how and why academics use ResearchGate. Online information review, 41(5), 744-759. DOI: 10.1108/oir-07-2016-0185

[30]
Nicholas, D., Clark, D., & Herman, E. (2016a). ResearchGate: reputation uncovered. Learned Publishing, 29(3), 173-182. DOI: 10.1002/leap.1035

[31]
Nicholas, D., Herman, E., & Clark, D. (2016b). Scholarly Reputation Building: How does ResearchGate Fare?. International Journal of Knowledge Content Development & Technology, 6(2). Retrieved from https://journals.sfu.ca/ijkcdt/index.php/ijkcdt/article/view/85

[32]
Nicholas, D., Herman, E., Jamali, H., Rodríguez-Bravo, B., Boukacem-Zeghmouri, C., Dobrowolski, T., & Pouchot, S. (2015). New ways of building, showcasing, and measuring scholarly reputation. Learned Publishing, 28(3), 169-183. DOI: 10.1087/20150303

[33]
Orduna-Malea, E., & Delgado López-Cózar, E. (2017). Performance behavior patterns in author-level metrics: A disciplinary comparison of Google Scholar Citations, ResearchGate, and ImpactStory. Frontiers in Research Metrics and Analytics, 2. DOI: 10.3389/frma.2017.00014

[34]
Orduña-Malea, E., & Delgado López-Cózar, E. (2020). ResearchGate and the Academic Social Network Sites: New Environments for New Bibliometrics?. In Ball, R. (ed.). Handbook Bibliometrics (pp. 265-280). De Gruyter. DOI: 10.1515/9783110646610-027

[35]
Orduna-Malea, E., Martin-Martin, A., & Delgado López-Cózar, E. (2016). ResearchGate as a source for scientific evaluation: Revealing its bibliometric applications. Profesional de la información, 25(2), 303-310. DOI: 10.3145/epi.2016.mar.18

[36]
Orduna-Malea, E., Martín-Martín, A., Thelwall, M., & Delgado López-Cózar, E. (2017). Do ResearchGate Scores create ghost academic reputations?. Scientometrics, 112(1), 443-460. DOI: 10.1007/s11192-017-2396-9

[37]
Ortega, J. L. (2015). Relationship between altmetric and bibliometric indicators across academic social sites: The case of CSIC’s members. Journal of informetrics, 9(1), 39-49. DOI: 10.1016/j.joi.2014.11.004

[38]
Ortega, J. L. (2016). Social network sites for scientists: A quantitative survey. Chandos. DOI: 10.1016/c2015-0-00021-1

[39]
Ortega, J. L. (2017). Towards a homogenization of academic social sites: A longitudinal study of profiles in Academia.edu, Google Scholar Citations and ResearchGate. Online Information Review, 41(6), 812-825. DOI: 10.1108/oir-01-2016-0012

[40]
Petersen, A. M., Fortunato, S., Pan, R. K., Kaski, K., Penner, O., Rungi, A., Riccaboni, R., Stanley, H. E., & Pammolli, F. (2014). Reputation and impact in academic careers. Proceedings of the National Academy of Sciences, 111(43), 15316-15321. DOI: 10.1073/pnas.1323111111

[41]
Popay, J., Roberts, H., Sowden, A., Petticrew, M., Arai, L., Rodgers, M., Britten, N., Roen, K., & Duffy, S. (2006). Guidance on the conduct of narrative synthesis in systematic reviews: A product from the ESRC methods programme. DOI: 10.13140/2.1.1018.4643

[42]
Priem, J. (2013). Beyond the paper. Nature, 495(7442), 437-440. DOI: 10.1038/495437a

[43]
Prieto-Gutierrez, J. J. (2019). Ten years of research on ResearchGate: A scoping review using Google Scholar (2008-2017). European Science Editing, 45(3), 60-64. DOI: 10.20316/ese.2019.45.18023

[44]
Raffaghelli, J. E., & Manca, S. (2023). Exploring the social activity of open research data on ResearchGate: Implications for the data literacy of researchers. Online Information Review, 47(1), 197-217. DOI: 10.1108/oir-05-2021-0255

[45]
ResearchGate (2014, August 13). Celebrating five million members with free DOIs [Web blog]. Retrieved from https://www.researchgate.net/blog/post/celebrating-five-million-members-with-free-dois

[46]
ResearchGate (2021, April 12). Science hasn’t slowed, it’s just moved online:How the COVID-19 pandemic continues to change how research is happening in 2021[Video file]. Retrieved from https://www.youtube.com/watch?v=klTF5yK7rxU

[47]
Silagadze, Z. K. (2023). On arXiv moderation system. Journal of Informetrics, 17(3). DOI: 10.1016/j.joi.2023.101433

[48]
Singh, V. K., Srichandan, S. S., & Lathabai, H. H. (2022). ResearchGate and Google Scholar: How much do they differ in publications, citations and different metrics and why?. Scientometrics, 127(3), 1515-1542. DOI: 10.1007/s11192-022-04264-2

[49]
Tennant, J. P., & Ross-Hellauer, T. (2020). The limitations to our understanding of peer review. Research Integrity and Peer Review, 5. DOI: 10.1186/s41073-020-00092-1

[50]
Thelwall, M., & Kousha, K. (2015). ResearchGate: Disseminating, Communicating, and Measuring Scholarship?. Journal of the American Society for Information Science and Technology, 66: 876-889. DOI: https://doi.org/10.1002/asi.23236

[51]
Thelwall, M., & Kousha, K. (2017a). ResearchGate articles: Age, discipline, audience size, and impact. Journal of the Association for Information Science and Technology, 68(2), 468-479. DOI: 10.1002/asi.23675

[52]
Thelwall, M., & Kousha, K. (2017b). ResearchGate versus Google Scholar: Which finds more early citations?. Scientometrics, 112(2), 1125-1131. DOI: 10.1007/s11192-017-2400-4

[53]
Van Noorden, R. (2014). Online collaboration: scientists and the social network. Nature, 512(7513), 126-129. DOI: 10.1038/512126a

[54]
Van Noorden, R. (2017). Publishers threaten to remove millions of papers from ResearchGate. Nature News. DOI: 10.1038/nature.2017.22793

[55]
Vera-Baceta, M. A., Thelwall, M., & Kousha, K. (2019). Web of Science and Scopus language coverage. Scientometrics, 121(3), 1803-1813. DOI: 10.1007/s11192-019-03264-z

[56]
Xu, S., Hao, L., An, X., Zhai, D., & Pang, H. (2019). Types of DOI errors of cited references in Web of Science with a cleaning method. Scientometrics, 120(3), 1427-1437. DOI: 10.1007/s11192-019-03162-4

[57]
Yan, W., & Zhang, Y. (2018). Research universities on the ResearchGate social networking site: An examination of institutional differences, research activity level, and social networks formed. Journal of Informetrics, 12(1), 385-400. DOI: 10.1016/j.joi.2017.08.002

[58]
Yan, W., Zhang, Y., & Bromfield, W. (2018). Analyzing the follower-followee ratio to determine user characteristics and institutional participation differences among research universities on ResearchGate. Scientometrics, 115(1), 299-316. DOI: 10.1007/s11192-018-2637-6

[59]
Yan, W., Zhang, Y., Hu, T., & Kudva, S. (2021). How does scholarly use of academic social networking sites differ by academic discipline? A case study using ResearchGate. Information Processing & Management, 58(1). DOI: 10.1016/j.ipm.2020.102430

[60]
Yu, M. C., Wu, Y. C. J., Alhalabi, W., Kao, H. Y., & Wu, W. H. (2016). ResearchGate: An effective altmetric indicator for active researchers?. Computers in Human Behavior, 55, 1001-1006. DOI: 10.1016/j.chb.2015.11.007

[61]
Zhu, J., Hu, G., & Liu, W. (2019). DOI errors and possible solutions for Web of Science. Scientometrics, 118(2), 709-718. DOI: 10.1007/s11192-018-2980-7

Outlines

/

京ICP备05002861号-43

Copyright © 2023 All rights reserved Journal of Data and Information Science

E-mail: jdis@mail.las.ac.cn Add:No.33, Beisihuan Xilu, Haidian District, Beijing 100190, China

Support by Beijing Magtech Co.ltd E-mail: support@magtech.com.cn