Rethinking the Evaluation of Scientific Activity

from the costs of hegemonic indicators to the urgency of inclusive and multidimensional models

Authors

DOI:

https://doi.org/10.5195/biblios.2025.1438

Keywords:

bibliometrics, global south, research evaluation, research metrics

Abstract

This editorial analyzes the necessary transition from a purely quantitative evaluation culture to inclusive and multidimensional models, structured around four main themes. First, it discusses the historical role of Metrics Studies in legitimizing evaluation systems, advocating a critical stance toward instruments that have assumed normative functions regarding scientific behavior. The evidence discussed confirms the exhaustion of hegemonic models and the crisis of representativeness in databases, emphasizing the costs of relying on commercial metrics that neglect research output from the Global South and the Humanities. The text denounces how the transformation of indicators into performance targets compromises the integrity of research, proposing approaches that value social impact and institutional diversity. Finally, this special issue of the 9th EBBC is part of a historical debate on metrics, calling on the community to embrace ethical and epistemological co-responsibility. Rethinking metrics means reflecting on what kind of science we wish to produce and for whom. Bibliometrics, with rigor and critical sensitivity, proves essential for guiding scientific production not merely by numbers, but through ethical relevance and the strengthening of the academic ecosystem.

Author Biography

João de Melo Maricato, Universidade de Brasília

   

References

Arocena, R, & Sutz, J. (2005). Latin American universities: from an original revolution to an uncertain transition. Higher Education, 50(4), 573–592. https://doi.org/10.1007/s10734-004-6367-8

Beigel, F., & Gomez, S. (2025). Uruguay: a case of multi-level research assessment, evaluation burn-out and an autonomist vocation. Zenodo. https://doi.org/10.5281/zenodo.16802953

Brasil Varandas Pinto, A. L. (2023). Advancing the evaluation of graduate education: towards a multidimensional model in Brazil. [Tese de Doutorado, Leiden University]. Scholarly Publications Leiden University. https://hdl.handle.net/1887/3645840

Bradford, S. C. (1985). Sources of information on specific subjects. Journal of Information Science, 10(4), 173-175. https://doi.org/10.1177/016555158501000406

CAPES. (2025). Diretrizes comuns da Avaliação de Permanência dos Programas de Pós Graduação stricto sensu. CAPES. https://www.gov.br/capes/pt-br/centrais-de-conteudo/documentos/avaliacao/19052025_20250502_DocumentoReferencial_FICHA.pdf

CLACSO – Consejo Latinoamericano de Ciencias Sociales. (2024). Manifiesto por las métricas socioterritoriales de Ciencia, Tecnología e Innovación. CLACSO. https://www.clacso.org/manifiesto-por-las-metricas-socioterritoriales-de-ciencia-tecnologia-e-innovacion/

CNRS. (2025, Dezembro 1). CNRS breaking free from Web of Science. CNRS. https://www.cnrs.fr/en/update/cnrs-breaking-free-web-science

CoARA – Coalition for Advancing Research Assessment.(2022). Agreement on Reforming Research Assessment. CoARA. https://coara.eu/agreement/the-agreement-full-text/

Curry, S., Gadd, E, & Wilsdon, J. (2022). Harnessing the Metric Tide: indicators, infrastructures & priorities for UK responsible research assessment. Research on Research Institute. https://doi.org/10.6084/m9.figshare.21701624

DORA. (2012). San Francisco Declaration on Research Assessment. DORA. https://sfdora.org/read/

Araújo, K. M., Araújo, P. C., & Vogel, M. J. M. (2024). Manifesto por uma Política de Acesso Aberto e Melhores Práticas de Avaliação da Ciência. Change.org. https://www.change.org/p/manifesto-por-uma-pol%C3%ADtica-de-acesso-aberto-e-melhores-pr%C3%A1ticas-de-avalia%C3%A7%C3%A3o-da-ci%C3%AAncia

Fávero, A. A., Consaltér, E., & Tonieto, C. (2019). Revistas predatórias: uma ameaça à integridade da ciência. Revista Internacional de Educação Superior, 6.

Fire, M., & Guestrin, C. Over-optimization of academic publishing metrics. GigaScience, 8(6). https://doi.org/10.1093/gigascience/giz053

Gärtner, A., Leising, D., Freyer, N., Musfeld, P., Lange, J., & Schönbrodt, F. (2022). Responsible Research Assessment II: A specific proposal for hiring and promotion in psychology. OSF. https://osf.io/5yexm

Haustein, S., & Larivière, V. (2015). The Use of Bibliometrics for Assessing Research: Possibilities, Limitations and Adverse Effects. In I. M. Welpe, J. Wollersheim, S. Ringelhan, & M. Osterloh. Incentives and Performance: governance and research organizations (pp. 121-139). Springer. https://doi.org/10.1007/978-3-319-09785-5_8

Hicks, D., Wouters, P., Waltman, L., de Rijcke, S. & Rafols, I. (2015). Bibliometrics: the Leiden Manifesto. Nature, 520, 429–431. https://doi.org/10.1038/520429a

INORMS Research Evaluation Group (2022). The SCOPE Framework: A five-stage process for evaluating research responsibly. DORA. https://sfdora.org/resource/the-scope-framework-a-five-stage-process-for-evaluating-research-responsibly/

JYU – University of Jyväskylä. (2025). The subscription to the Web of Science database will end on January 1, 2026. JYU. https://www.jyu.fi/en/news/the-subscription-to-web-of-science-database-will-end-on-january-1-2026

KNAW, & VSNU (2019). Room for everyone’s talent: towards a new balance in the recognition and rewards of academics. The Hague. https://www.nwo.nl/sites/nwo/files/media-files/2019-Recognition-Rewards-Position-Paper_EN.pdf

Kuhn, T. S. (2011). A estrutura das revoluções científicas (11th ed.). Perspectiva.

Latour, B. (1987). Science in Action. Harvard University Press.

Lotka, A. J. (1926). The frequency distribution of scientific productivity. Journal of the Washington Academy of Sciences, 16(12), 317-323. https://www.jstor.org/stable/24522522

Maricato, J. M., Mazoni, A., Mugnaini, R., Packer, A. L., & Costas, R. SciELO as an open scientometric research infrastructure: General discussion of coverage in OpenAlex, WoS, scopus and dimensions. 27th International Conference on Science, Technology and Innovation Indicators (STI 2023), Leiden, 2023. https://dapp.orvium.io/deposits/6442c231903ef57acd6dc640/view

Melo, J. H. N. D., Trinca, T. P., & Maricato, J. D. M. (2021). Limites dos indicadores bibliométricos de bases de dados internacionais para avaliação da Pós-Graduação brasileira: a cobertura da Web of Science nas diferentes áreas do conhecimento. Transinformação, 33. https://doi.org/10.1590/2318-0889202133e200071

MINISTÉRIO DA EDUCAÇÃO DA CHINA. (2020, Fevereiro 23) 关于规范高等学校 SCI 论文相关指标使用的若干意见. [Opiniões sobre a padronização dos indicadores relacionados a artigos SCI de papel em instituições de ensino superior]. http://www.moe.gov.cn/srcsite/A16/moe_784/202002/t20200223_423334.html

Moed, H. F. (2017). Applied Evaluative Informetrics. Springer. https://arxiv.org/pdf/1705.06110

Mongeon, P., & Paul Hus, A. (2016). The journal coverage of Web of Science and Scopus: a comparative analysis. Scientometrics, 106(1), 213-228. https://doi.org/10.1007/s11192-015-1765-5

OECD (1995). Canberra Manual: The Measurement of Human Resources in Science and Technology. OECD. https://www.oecd.org/en/publications/measurement-of-scientific-and-technological-activities_9789264065581-en.html

OECD (2015). Frascati Manual: Proposed Standard Practice for Surveys on Research and Experimental Development. OECD. https://www.oecd.org/content/dam/oecd/en/publications/reports/2015/10/frascati-manual-2015_g1g57dcb/9789264239012-en.pdf

OECD (2019). Measuring Innovation in Education 2019: Indicators and Implications. OECD Publishing. https://www.oecd.org/content/dam/oecd/en/publications/reports/2019/03/measuring-innovation-in-education-2019_g1g9e098/9789264311671-en.pdf

OECD/Eurostat. (2018). Oslo Manual: Guidelines for Collecting, Reporting and Using Data on Innovation (4th ed.). OECD Publishing. https://www.oecd.org/content/dam/oecd/en/publications/reports/2018/10/oslo-manual-2018_g1g9373b/9789264304604-en.pdf

OVERTON (2023). Overton Database. https://www.overton.io

Priem, J., Piwowar, H., & Orr, R. (2023). OpenAlex: A fully open index of scholarly works. Cornell University. https://doi.org/10.48550/arXiv.2205.01833

REF (2014). Research Excellence Framework 2014: The Results. Higher Education Funding Council for England. https://2014.ref.ac.uk/media/ref/content/pub/REF%2001%202014%20-%20full%20document.pdf

Schmidt, R., Curry, S., & Hatch, A. (2021). Research Culture: Creating SPACE to evolve academic assessment. eLife. https://doi.org/10.7554/eLife.70929.

Sorbonne University. (2024). Sorbonne University unsubscribes from Web of Science. Sorbonne University. https://www.sorbonne-universite.fr/en/news/sorbonne-university-unsubscribes-web-science

Sugimoto, C. R., & Larivière, V. (2018). Measuring Research: What Everyone Needs to Know. Oxford University Press.

UNESCO (2021). UNESCO Recommendation on Open Science. UNESCO.

Times Higher Education. (2019). Impact Rankings 2020: Methodology. London. https://www.timeshighereducation.com/rankings/impact/2020

TU DELFT. (2025). Access to Web of Science and Journal Citation Reports will end on 1 January 2026. https://www.tudelft.nl/en/2025/library/access-to-web-of-science-and-journal-citation-reports-will-end-on-1-january-2026

UTWENTE – University of Twente. (2025). Web of Science licence will end on 31 December 2025. UTWENTE. https://www.utwente.nl/en/education/student-services/news-events/news/2025/11/622976/web-of-science-licence-will-end-on-31-december-2025

UU – Utrecht University. (2024). Reminder: access to Web of Science will end on 1 January 2026. UU. https://www.uu.nl/en/news/reminder-access-to-web-of-science-will-end-on-1-january-2026

Vessuri, H., Guédon, J-C., & Cetto, A. M.(2014). Excellence or quality? Impact of the current competition regime on science and scientific publishing in Latin America and its implications for development. Current Sociology, 62(5), 647-665. https://doi.org/10.1177/0011392113512839

VUA – Vrije Universiteit Amsterdam. (2025). Termination of access to Web of Science as of January 1, 2026. VUA. https://vu.nl/en/employee/university-library/termination-of-access-to-web-of-science-as-of-january-1-2026

Zipf, G. K. (1949). Human Behavior and the Principle of Least Effort. Addison-Wesley. https://web.stanford.edu/class/psych227/Zipf_Words.pdf

Published

2026-05-13

How to Cite

Maricato, J. de M. (2026). Rethinking the Evaluation of Scientific Activity: from the costs of hegemonic indicators to the urgency of inclusive and multidimensional models. Biblios Journal of Librarianship and Information Science, (esp.), e025. https://doi.org/10.5195/biblios.2025.1438

Issue

Section

Editorial