âčïž Skipped - page is already crawled
| Filter | Status | Condition | Details |
|---|---|---|---|
| HTTP status | PASS | download_http_code = 200 | HTTP 200 |
| Age cutoff | PASS | download_stamp > now() - 6 MONTH | 0 months ago |
| History drop | PASS | isNull(history_drop_reason) | No drop reason |
| Spam/ban | PASS | fh_dont_index != 1 AND ml_spam_score = 0 | ml_spam_score=0 |
| Canonical | PASS | meta_canonical IS NULL OR = '' OR = src_unparsed | Not set |
| Property | Value | ||||||
|---|---|---|---|---|---|---|---|
| URL | https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/ | ||||||
| Last Crawled | 2026-04-22 10:07:14 (1 day ago) | ||||||
| First Indexed | 2024-10-21 13:49:01 (1 year ago) | ||||||
| HTTP Status Code | 200 | ||||||
| Content | |||||||
| Meta Title | Ouvrir la Science - Excessive growth in the number of scientific publications | ||||||
| Meta Description | Un site internet du comité pour la science ouverte - coso, Benoßt Pier et Laurent Romary nous proposent un billet sur l'augmentation récente du nombre de publications scientifiques et son effet sur la recherche scientifique | ||||||
| Meta Canonical | null | ||||||
| Boilerpipe Text | In this post, BenoĂźt Pier (CNRS Research Director at the Laboratory of Fluid Mechanics and Acoustics â LMFA) and Laurent Romary (Research Director, Inria) write about the recent increase in the number of scientific publications and its effect on scientific research, based on the analysis made in the âpreprintâ published on arXiv :Â
âHanson, M. A., Barreiro, P. G., Crosetto, P., & Brockington, D. (2023). The strain on scientific publishing. arXiv preprint arXiv:2309.15884.
https://doi.org/10.48550/arXiv.2309.15884
â
.
A study published in arXiv, entitled â
The strain on scientific publishing
â, addresses the explosion in the number of scientific publications and tries to document this phenomenon precisely, by resorting to quantitative data. The aim is to identify the ingredients of this growth and to understand the consequences for research activity, in particular peer review and bibliographic monitoring.
The authors are specialists in bibliometrics and exploit a range of sources for their quantitative analyses: mainly Scopus and Web of Science, but also directly harvesting the web pages of various journals and publishers. The period analysed covers the last ten years, with a particular focus on 2016-2022.
The first result concerns the total number of articles published, which follows very closely an exponential growth (+5.6% per year). Even taking into account the increase in the number of researchers over this period, we can deduce that the time spent on obtaining the results, validating and peer reviewing them has decreased significantly.
The authors point to a number of factors that could explain this growth, including greater accessibility to scientific publications in the countries of the âGlobal Southâ. However, certain factors seem to predominate, and among them, primarily the new editorial practices. By breaking down the annual number of articles by publisher, the authors show that this strong increase is essentially due to a few publishers who have increased their publication volume enormously: MDPI and Frontiers are clearly on very different trajectories compared to most other publishers. By computing the average annual number of articles published per journal, the authors observe that the growth (significant, but not overwhelming) of the traditional publishers is mainly due to the expansion of the number of journals in their catalogues, whereas the very strong growth of Frontiers and MDPI is the result of an explosive increase in the number of articles per journal. It should be noted that these two publishers, which appeared more recently, are thriving through publication fees paid by authors.
To explore this question further, the authors also compared the number of articles published in special issues with those in regular issues. They found that for MDPI, Hindawi and Frontiers, the very strong growth was almost exclusively due to the increase in articles published in special issues â in 2022, the vast majority of these publishersâ articles were published in special issues.
To provide a more detailed characterisation of publication practices, the authors carried out a statistical analysis of the duration of the peer review cycle, i.e. the time from submission to acceptance. These data are not readily available, but the authors were able to harvest them for a sufficient number of articles to obtain statistical distributions by major publisher, and compare their evolution over the years 2016, 2019 and 2022. This statistical analysis shows that, in general, these distributions have hardly changed between 2016 and 2022, and are asymmetrical, with a long tail at long times, corresponding to the ârare eventsâ consisting of articles that are particularly difficult to review. But here again, Frontiers, MDPI and Hindawi display a quite unusual trend r: in 2022, the article turnaround times is much shorter than 2016 and is accompanied by the loss of the long tail. This triad of publishing houses has therefore moved to a much faster and more homogeneous process, while the disciplines covered are quite diverse and the quality of the manuscripts probably just as variable as in the past.
This study highlights two main factors responsible for the explosion in the number of published articles:
Well established publishers are mainly expanding their catalogues, although there has also been an increase in the number of articles per journal.
The triad (MDPI, Frontiers, Hindawi) increasingly use special issues to publish more and more articles, and this phenomenon is accompanied by a significant shortening of the time allotted to the peer-review process.
This study has not failed to trigger reactions, particularly from Frontiers in
a blog post
which disputes the figures, accuses the authors of bias and continues to promote exponential growth
. The authors of this study
have pointed out
that, unlike other publishers, Frontiers did not provide the requested data, and they methodically dismantle Frontiersâ analytical errors in their blog post
.
This battle over figures underlines the importance of having accessible and open data in order to carry out this kind of study, according to criteria that guarantee the scientific validity of the results obtained. It is therefore regrettable that this work is largely based on the closed databases of Scopus and the Web of Science. Nevertheless, we have attempted to reproduce certain aspects by querying the open data of the
OpenAlex
database and we have found exactly the same trends in the number of articles published per year, per journal and per publisher (see the figure below). However, it is not currently possible to carry out exactly the same detailed study using open data.
Letâs hope that the
Barcelona Declaration on Open Research Information
(of which the French Committee for Open Science is a signatory) will facilitate this in the future.
Figure 1: Evolution of the total number of articles published per year and per publisher, data from OpenAlex.
In conclusion, based on the results of this study, we can ask ourselves whether the current system is sustainable. The answer is clearly no. Exponential growth in the number of publications is not compatible with maintaining scientific quality and ensuring confidence in the results, guaranteed by a careful and thorough peer review process. This is all the more true given that the number of members of the scientific community is increasing little, if at all, if we look at the trend in the number of PhD graduates in OECD countries, which, as the authors of the study presented here observe, has been in slight recession since 2018. However, to maintain a high standard, each published article needs to be validated by at least two independent reviewers who can devote sufficient time to this task; and the editorial boards of most scientific journals are having more and more difficulties to find reviewers. If we take the rejection rate into account, we suggest the following golden rule: to maintain a high-quality publication system, each researcher needs to review around ten times as many articles by his or her peers as he or she writes.
Faced with the current inflation in scientific publication, some people are advocating the development of new tools (using âartificial intelligenceâ, for example) to identify what is relevant, evaluate only worthwhile manuscripts, track down texts that have themselves been generated by machines⊠But just as with climate change, the deterioration of the scientific publishing system is unlikely to be solved by purely technical means, unless we first rely on human intelligence, aided by a little sobriety.
At the same time, institutions, universities and research organisations, as well as funding agencies, must unambiguously support a reform of evaluation based on the principles of
DORA
and
COARA
(see the French chapter of
COARA
), by banning indicators such as the âh-indexâ and the âjournal impact factorâ and encouraging narrative CVs. Publication practices can evolve towards a more virtuous system only if evaluation practices follow the same path.
Rather than to aggravate the current race to publish, it would be better to devote our time to doing better science and therefore to publish fewer and better papers.
Update 13th of November 2024
Since this post was published, the preprint âThe strain on scientific publishingâ has become a publication and
Le Monde
has covered the subject. We provide the references below.
Hanson, Mark A., Pablo GĂłmez Barreiro, Paolo Crosetto, et Dan Brockington. « The strain on scientific publishing ».Â
Quantitative Science Studies
, 8 novembre 2024, 1â21.Â
https://doi.org/10.48550/arXiv.2309.15884
« Lâinflation du nombre de publications scientifiques interroge ». 12 novembre 2024 â access reserved to subscribers.
https://www.lemonde.fr/sciences/article/2024/11/12/l-inflation-du-nombre-de-publications-scientifiques-interroge_6389778_1650684.html | ||||||
| Markdown | - [Aller au menu](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/#menuPrincipal)
- [Aller au menu](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/#menuHamburger)
- [Aller au contenu](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/#main)
- [Aller Ă la recherche](https://www.ouvrirlascience.fr/?s)
 [](https://www.ouvrirlascience.fr/home) 
- Open Science
- Our actions
- [Resources](https://www.ouvrirlascience.fr/category/resources/)
- News
- [Who are we ?](https://www.ouvrirlascience.fr/the-committee-for-open-science/)
- [FR](https://www.ouvrirlascience.fr/mesures-et-demesure-de-la-publication-scientifique/)
- [EN](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/)
[Contact](https://www.ouvrirlascience.fr/home/contact-en)
Back
Open Science
[Get started with Open Science](https://www.ouvrirlascience.fr/get-started-with-open-science/)
[National Plan for Open Science](https://www.ouvrirlascience.fr/second-national-plan-for-open-science-npos/)
[Glossary](https://www.ouvrirlascience.fr/category/open-science/glossary/)
[Small Encyclopedia of Open Science](https://encyclo.ouvrirlascience.fr/)
[Stories](https://www.ouvrirlascience.fr/category/open-science/stories/)
Back
Our actions
The Committee at work
[Colleges, groups & projects](https://www.ouvrirlascience.fr/category/projects/)
[Open Science prizes](https://www.ouvrirlascience.fr/category/openscience-prices/)
National fund for open science
[Presentation of the FNSO](https://www.ouvrirlascience.fr/national-fund-for-open-science/)
[Calls for Projects of the FNSO](https://www.ouvrirlascience.fr/category/nfos/calls-for-projects-nfos/)
[FNSO supported actions](https://www.ouvrirlascience.fr/category/nfos/projects-initiatives-nfos/)
Back
News
[Blog](https://www.ouvrirlascience.fr/category/news/blog-en/)
[Schedule](https://www.ouvrirlascience.fr/category/news/schedule/)
[](http://www.enseignementsup-recherche.gouv.fr/)
- [Open Science](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/)
- [Get started with Open Science](https://www.ouvrirlascience.fr/get-started-with-open-science/)
- [National Plan for Open Science](https://www.ouvrirlascience.fr/second-national-plan-for-open-science-npos/)
- [Glossary](https://www.ouvrirlascience.fr/category/open-science/glossary/)
- [Small Encyclopedia of Open Science](https://encyclo.ouvrirlascience.fr/)
- [Stories](https://www.ouvrirlascience.fr/category/open-science/stories/)
- [Our actions](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/)
- The Committee at work
- [Colleges, groups & projects](https://www.ouvrirlascience.fr/category/projects/)
- [Open Science prizes](https://www.ouvrirlascience.fr/category/openscience-prices/)
- National fund for open science
- [Presentation of the FNSO](https://www.ouvrirlascience.fr/national-fund-for-open-science/)
- [Calls for Projects of the FNSO](https://www.ouvrirlascience.fr/category/nfos/calls-for-projects-nfos/)
- [FNSO supported actions](https://www.ouvrirlascience.fr/category/nfos/projects-initiatives-nfos/)
- [Resources](https://www.ouvrirlascience.fr/category/resources/)
- [News](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/)
- [Blog](https://www.ouvrirlascience.fr/category/news/blog-en/)
- [Schedule](https://www.ouvrirlascience.fr/category/news/schedule/)
- [Who are we ?](https://www.ouvrirlascience.fr/the-committee-for-open-science/)
- [FR](https://www.ouvrirlascience.fr/mesures-et-demesure-de-la-publication-scientifique/)
- [EN](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/)
- [](https://www.ouvrirlascience.fr/search)
[Home](https://www.ouvrirlascience.fr/home/) \> [Blog](https://www.ouvrirlascience.fr/category/blog-en) \> Excessive growth in the number of scientific publications
# Excessive growth in the number of scientific publications
[](https://www.ouvrirlascience.fr/#facebook) [](https://www.ouvrirlascience.fr/#bluesky) [](https://www.ouvrirlascience.fr/#linkedin)
News from the Committee
21/10/2024
*In this post, BenoĂźt Pier (CNRS Research Director at the Laboratory of Fluid Mechanics and Acoustics â LMFA) and Laurent Romary (Research Director, Inria) write about the recent increase in the number of scientific publications and its effect on scientific research, based on the analysis made in the âpreprintâ published on arXiv :*
*âHanson, M. A., Barreiro, P. G., Crosetto, P., & Brockington, D. (2023). The strain on scientific publishing. arXiv preprint arXiv:2309.15884.* <https://doi.org/10.48550/arXiv.2309.15884>*â*.
A study published in arXiv, entitled â[The strain on scientific publishing](https://doi.org/10.48550/arXiv.2309.15884)â, addresses the explosion in the number of scientific publications and tries to document this phenomenon precisely, by resorting to quantitative data. The aim is to identify the ingredients of this growth and to understand the consequences for research activity, in particular peer review and bibliographic monitoring.
The authors are specialists in bibliometrics and exploit a range of sources for their quantitative analyses: mainly Scopus and Web of Science, but also directly harvesting the web pages of various journals and publishers. The period analysed covers the last ten years, with a particular focus on 2016-2022.
The first result concerns the total number of articles published, which follows very closely an exponential growth (+5.6% per year). Even taking into account the increase in the number of researchers over this period, we can deduce that the time spent on obtaining the results, validating and peer reviewing them has decreased significantly.
The authors point to a number of factors that could explain this growth, including greater accessibility to scientific publications in the countries of the âGlobal Southâ. However, certain factors seem to predominate, and among them, primarily the new editorial practices. By breaking down the annual number of articles by publisher, the authors show that this strong increase is essentially due to a few publishers who have increased their publication volume enormously: MDPI and Frontiers are clearly on very different trajectories compared to most other publishers. By computing the average annual number of articles published per journal, the authors observe that the growth (significant, but not overwhelming) of the traditional publishers is mainly due to the expansion of the number of journals in their catalogues, whereas the very strong growth of Frontiers and MDPI is the result of an explosive increase in the number of articles per journal. It should be noted that these two publishers, which appeared more recently, are thriving through publication fees paid by authors.
To explore this question further, the authors also compared the number of articles published in special issues with those in regular issues. They found that for MDPI, Hindawi and Frontiers, the very strong growth was almost exclusively due to the increase in articles published in special issues â in 2022, the vast majority of these publishersâ articles were published in special issues.
To provide a more detailed characterisation of publication practices, the authors carried out a statistical analysis of the duration of the peer review cycle, i.e. the time from submission to acceptance. These data are not readily available, but the authors were able to harvest them for a sufficient number of articles to obtain statistical distributions by major publisher, and compare their evolution over the years 2016, 2019 and 2022. This statistical analysis shows that, in general, these distributions have hardly changed between 2016 and 2022, and are asymmetrical, with a long tail at long times, corresponding to the ârare eventsâ consisting of articles that are particularly difficult to review. But here again, Frontiers, MDPI and Hindawi display a quite unusual trend r: in 2022, the article turnaround times is much shorter than 2016 and is accompanied by the loss of the long tail. This triad of publishing houses has therefore moved to a much faster and more homogeneous process, while the disciplines covered are quite diverse and the quality of the manuscripts probably just as variable as in the past.
This study highlights two main factors responsible for the explosion in the number of published articles:
- Well established publishers are mainly expanding their catalogues, although there has also been an increase in the number of articles per journal.
- The triad (MDPI, Frontiers, Hindawi) increasingly use special issues to publish more and more articles, and this phenomenon is accompanied by a significant shortening of the time allotted to the peer-review process.
This study has not failed to trigger reactions, particularly from Frontiers in [a blog post](https://www.frontiersin.org/news/2024/02/21/bad-bibliometrics-dont-add-up-for-research-or-why-research-publishing-policy) which disputes the figures, accuses the authors of bias and continues to promote exponential growth. The authors of this study [have pointed out](https://the-strain-on-scientific-publishing.github.io/website/posts/response_to_frontiers/) that, unlike other publishers, Frontiers did not provide the requested data, and they methodically dismantle Frontiersâ analytical errors in their blog post.
This battle over figures underlines the importance of having accessible and open data in order to carry out this kind of study, according to criteria that guarantee the scientific validity of the results obtained. It is therefore regrettable that this work is largely based on the closed databases of Scopus and the Web of Science. Nevertheless, we have attempted to reproduce certain aspects by querying the open data of the [OpenAlex](https://openalex.org/) database and we have found exactly the same trends in the number of articles published per year, per journal and per publisher (see the figure below). However, it is not currently possible to carry out exactly the same detailed study using open data.
Letâs hope that the [Barcelona Declaration on Open Research Information](https://www.ouvrirlascience.fr/the-french-committee-for-open-science-is-a-signatory-to-the-barcelona-declaration-on-open-research-information/) (of which the French Committee for Open Science is a signatory) will facilitate this in the future.

*Figure 1: Evolution of the total number of articles published per year and per publisher, data from OpenAlex.*
In conclusion, based on the results of this study, we can ask ourselves whether the current system is sustainable. The answer is clearly no. Exponential growth in the number of publications is not compatible with maintaining scientific quality and ensuring confidence in the results, guaranteed by a careful and thorough peer review process. This is all the more true given that the number of members of the scientific community is increasing little, if at all, if we look at the trend in the number of PhD graduates in OECD countries, which, as the authors of the study presented here observe, has been in slight recession since 2018. However, to maintain a high standard, each published article needs to be validated by at least two independent reviewers who can devote sufficient time to this task; and the editorial boards of most scientific journals are having more and more difficulties to find reviewers. If we take the rejection rate into account, we suggest the following golden rule: to maintain a high-quality publication system, each researcher needs to review around ten times as many articles by his or her peers as he or she writes.
Faced with the current inflation in scientific publication, some people are advocating the development of new tools (using âartificial intelligenceâ, for example) to identify what is relevant, evaluate only worthwhile manuscripts, track down texts that have themselves been generated by machines⊠But just as with climate change, the deterioration of the scientific publishing system is unlikely to be solved by purely technical means, unless we first rely on human intelligence, aided by a little sobriety.
At the same time, institutions, universities and research organisations, as well as funding agencies, must unambiguously support a reform of evaluation based on the principles of [DORA](https://www.ouvrirlascience.fr/19th-munin-conference-on-scholarly-publishing-2/) and [COARA](https://coara.eu/) (see the French chapter of [COARA](https://coara.fr/)), by banning indicators such as the âh-indexâ and the âjournal impact factorâ and encouraging narrative CVs. Publication practices can evolve towards a more virtuous system only if evaluation practices follow the same path.
Rather than to aggravate the current race to publish, it would be better to devote our time to doing better science and therefore to publish fewer and better papers.
**Update 13th of November 2024**
Since this post was published, the preprint âThe strain on scientific publishingâ has become a publication and *Le Monde* has covered the subject. We provide the references below.
- Hanson, Mark A., Pablo GĂłmez Barreiro, Paolo Crosetto, et Dan Brockington. « The strain on scientific publishing ». *Quantitative Science Studies*, 8 novembre 2024, 1â21. <https://doi.org/10.48550/arXiv.2309.15884>
**
*Next post  News from the Committee 07/10/2024 [Wikimedia residencies in higher education and research institutions](https://www.ouvrirlascience.fr/wikimedia-residencies-in-higher-education-and-research-institutions/) In 2023-2024, the French Ministry of Higher Education and Research (MESR) has supported three Wikimedia residencies\[1\]A Wikimedia residency is a partnership between Wikimedia and a cultural, educational or scientific institution, during which a Wikimedia contributor works on site to improve⊠[Agenda Agenda Upcoming events](https://www.ouvrirlascience.fr/category/schedule) [Resources Resources The founding texts and resources](https://www.ouvrirlascience.fr/category/resources)*
* [](http://www.enseignementsup-recherche.gouv.fr/) [Projects](https://www.ouvrirlascience.fr/category/projects/) [Blog](https://www.ouvrirlascience.fr/category/news/blog-en/) [Schedule](https://www.ouvrirlascience.fr/category/news/schedule/) [Resources](https://www.ouvrirlascience.fr/category/resources/) [About us](https://www.ouvrirlascience.fr/about-the-committee-for-open-science/) [Credits and notices](https://www.ouvrirlascience.fr/credits-and-notices/) [Sitemap](https://www.ouvrirlascience.fr/sitemap/) [Contact](https://www.ouvrirlascience.fr/home/contact-en) Follow us: [](https://www.linkedin.com/company/ouvrir-la-science) [](https://www.canal-u.tv/chaines/ouvrirlascience) [](https://social.numerique.gouv.fr/@ouvrirlascience) [](https://bsky.app/profile/ouvrirlascience.bsky.social)  En poursuivant votre navigation, sans modifier vos paramÚtres, vous acceptez l'utilisation et le dépÎt de cookies destinés à mesurer la fréquentation du site grùce au logiciel Matomo. OK Modifier les paramÚtres*
â
Thanks for sharing\!
[AddToAny](https://www.addtoany.com/ "Share Buttons")
[MoreâŠ](https://www.ouvrirlascience.fr/excessive-growth-in-the-number-of-scientific-publications/#addtoany "Show All") | ||||||
| Readable Markdown | *In this post, BenoĂźt Pier (CNRS Research Director at the Laboratory of Fluid Mechanics and Acoustics â LMFA) and Laurent Romary (Research Director, Inria) write about the recent increase in the number of scientific publications and its effect on scientific research, based on the analysis made in the âpreprintâ published on arXiv :*
*âHanson, M. A., Barreiro, P. G., Crosetto, P., & Brockington, D. (2023). The strain on scientific publishing. arXiv preprint arXiv:2309.15884.* <https://doi.org/10.48550/arXiv.2309.15884>*â*.
A study published in arXiv, entitled â[The strain on scientific publishing](https://doi.org/10.48550/arXiv.2309.15884)â, addresses the explosion in the number of scientific publications and tries to document this phenomenon precisely, by resorting to quantitative data. The aim is to identify the ingredients of this growth and to understand the consequences for research activity, in particular peer review and bibliographic monitoring.
The authors are specialists in bibliometrics and exploit a range of sources for their quantitative analyses: mainly Scopus and Web of Science, but also directly harvesting the web pages of various journals and publishers. The period analysed covers the last ten years, with a particular focus on 2016-2022.
The first result concerns the total number of articles published, which follows very closely an exponential growth (+5.6% per year). Even taking into account the increase in the number of researchers over this period, we can deduce that the time spent on obtaining the results, validating and peer reviewing them has decreased significantly.
The authors point to a number of factors that could explain this growth, including greater accessibility to scientific publications in the countries of the âGlobal Southâ. However, certain factors seem to predominate, and among them, primarily the new editorial practices. By breaking down the annual number of articles by publisher, the authors show that this strong increase is essentially due to a few publishers who have increased their publication volume enormously: MDPI and Frontiers are clearly on very different trajectories compared to most other publishers. By computing the average annual number of articles published per journal, the authors observe that the growth (significant, but not overwhelming) of the traditional publishers is mainly due to the expansion of the number of journals in their catalogues, whereas the very strong growth of Frontiers and MDPI is the result of an explosive increase in the number of articles per journal. It should be noted that these two publishers, which appeared more recently, are thriving through publication fees paid by authors.
To explore this question further, the authors also compared the number of articles published in special issues with those in regular issues. They found that for MDPI, Hindawi and Frontiers, the very strong growth was almost exclusively due to the increase in articles published in special issues â in 2022, the vast majority of these publishersâ articles were published in special issues.
To provide a more detailed characterisation of publication practices, the authors carried out a statistical analysis of the duration of the peer review cycle, i.e. the time from submission to acceptance. These data are not readily available, but the authors were able to harvest them for a sufficient number of articles to obtain statistical distributions by major publisher, and compare their evolution over the years 2016, 2019 and 2022. This statistical analysis shows that, in general, these distributions have hardly changed between 2016 and 2022, and are asymmetrical, with a long tail at long times, corresponding to the ârare eventsâ consisting of articles that are particularly difficult to review. But here again, Frontiers, MDPI and Hindawi display a quite unusual trend r: in 2022, the article turnaround times is much shorter than 2016 and is accompanied by the loss of the long tail. This triad of publishing houses has therefore moved to a much faster and more homogeneous process, while the disciplines covered are quite diverse and the quality of the manuscripts probably just as variable as in the past.
This study highlights two main factors responsible for the explosion in the number of published articles:
- Well established publishers are mainly expanding their catalogues, although there has also been an increase in the number of articles per journal.
- The triad (MDPI, Frontiers, Hindawi) increasingly use special issues to publish more and more articles, and this phenomenon is accompanied by a significant shortening of the time allotted to the peer-review process.
This study has not failed to trigger reactions, particularly from Frontiers in [a blog post](https://www.frontiersin.org/news/2024/02/21/bad-bibliometrics-dont-add-up-for-research-or-why-research-publishing-policy) which disputes the figures, accuses the authors of bias and continues to promote exponential growth. The authors of this study [have pointed out](https://the-strain-on-scientific-publishing.github.io/website/posts/response_to_frontiers/) that, unlike other publishers, Frontiers did not provide the requested data, and they methodically dismantle Frontiersâ analytical errors in their blog post.
This battle over figures underlines the importance of having accessible and open data in order to carry out this kind of study, according to criteria that guarantee the scientific validity of the results obtained. It is therefore regrettable that this work is largely based on the closed databases of Scopus and the Web of Science. Nevertheless, we have attempted to reproduce certain aspects by querying the open data of the [OpenAlex](https://openalex.org/) database and we have found exactly the same trends in the number of articles published per year, per journal and per publisher (see the figure below). However, it is not currently possible to carry out exactly the same detailed study using open data.
Letâs hope that the [Barcelona Declaration on Open Research Information](https://www.ouvrirlascience.fr/the-french-committee-for-open-science-is-a-signatory-to-the-barcelona-declaration-on-open-research-information/) (of which the French Committee for Open Science is a signatory) will facilitate this in the future.

*Figure 1: Evolution of the total number of articles published per year and per publisher, data from OpenAlex.*
In conclusion, based on the results of this study, we can ask ourselves whether the current system is sustainable. The answer is clearly no. Exponential growth in the number of publications is not compatible with maintaining scientific quality and ensuring confidence in the results, guaranteed by a careful and thorough peer review process. This is all the more true given that the number of members of the scientific community is increasing little, if at all, if we look at the trend in the number of PhD graduates in OECD countries, which, as the authors of the study presented here observe, has been in slight recession since 2018. However, to maintain a high standard, each published article needs to be validated by at least two independent reviewers who can devote sufficient time to this task; and the editorial boards of most scientific journals are having more and more difficulties to find reviewers. If we take the rejection rate into account, we suggest the following golden rule: to maintain a high-quality publication system, each researcher needs to review around ten times as many articles by his or her peers as he or she writes.
Faced with the current inflation in scientific publication, some people are advocating the development of new tools (using âartificial intelligenceâ, for example) to identify what is relevant, evaluate only worthwhile manuscripts, track down texts that have themselves been generated by machines⊠But just as with climate change, the deterioration of the scientific publishing system is unlikely to be solved by purely technical means, unless we first rely on human intelligence, aided by a little sobriety.
At the same time, institutions, universities and research organisations, as well as funding agencies, must unambiguously support a reform of evaluation based on the principles of [DORA](https://www.ouvrirlascience.fr/19th-munin-conference-on-scholarly-publishing-2/) and [COARA](https://coara.eu/) (see the French chapter of [COARA](https://coara.fr/)), by banning indicators such as the âh-indexâ and the âjournal impact factorâ and encouraging narrative CVs. Publication practices can evolve towards a more virtuous system only if evaluation practices follow the same path.
Rather than to aggravate the current race to publish, it would be better to devote our time to doing better science and therefore to publish fewer and better papers.
**Update 13th of November 2024**
Since this post was published, the preprint âThe strain on scientific publishingâ has become a publication and *Le Monde* has covered the subject. We provide the references below.
- Hanson, Mark A., Pablo GĂłmez Barreiro, Paolo Crosetto, et Dan Brockington. « The strain on scientific publishing ». *Quantitative Science Studies*, 8 novembre 2024, 1â21. <https://doi.org/10.48550/arXiv.2309.15884> | ||||||
| ML Classification | |||||||
| ML Categories |
Raw JSON{
"/Science": 980,
"/Science/Scientific_Institutions": 866
} | ||||||
| ML Page Types |
Raw JSON{
"/Article": 988,
"/Article/Study_or_Research_Findings": 663
} | ||||||
| ML Intent Types |
Raw JSON{
"Informational": 999
} | ||||||
| Content Metadata | |||||||
| Language | en | ||||||
| Author | Anne-Marie Badolato | ||||||
| Publish Time | 2024-10-21 10:15:15 (1 year ago) | ||||||
| Original Publish Time | 2024-10-21 10:15:15 (1 year ago) | ||||||
| Republished | No | ||||||
| Word Count (Total) | 1,686 | ||||||
| Word Count (Content) | 1,379 | ||||||
| Links | |||||||
| External Links | 19 | ||||||
| Internal Links | 37 | ||||||
| Technical SEO | |||||||
| Meta Nofollow | No | ||||||
| Meta Noarchive | No | ||||||
| JS Rendered | Yes | ||||||
| Redirect Target | null | ||||||
| Performance | |||||||
| Download Time (ms) | 291 | ||||||
| TTFB (ms) | 275 | ||||||
| Download Size (bytes) | 18,471 | ||||||
| Shard | 115 (laksa) | ||||||
| Root Hash | 14061789285859253515 | ||||||
| Unparsed URL | fr,ouvrirlascience!www,/excessive-growth-in-the-number-of-scientific-publications/ s443 | ||||||