OpenScienceLab https://opensciencelab.uc3m.es Meta-research for Open Science Fri, 06 Mar 2026 07:49:03 +0000 en-US hourly 1 https://wordpress.org/?v=6.9.4 https://storage.googleapis.com/wp-uploads.bucket.wp.uc3m.es/wp-content/uploads/sites/96/2025/03/28094815/cropped-favicon-negro-32x32.png OpenScienceLab https://opensciencelab.uc3m.es 32 32 FAIR vs Open Data: Are They the Same? https://opensciencelab.uc3m.es/blog/fair-vs-open-data/ Fri, 06 Mar 2026 01:00:00 +0000 https://opensciencelab.uc3m.es/?p=2995 It is a common misconception that FAIR data must always be completely open and that all open data are inherently FAIR. However, this is not the case. Let us explore the definitions of both concepts and clarify the differences between them.

What Is Open Data?

Open Data refers to data that can be freely accessed, reused, and redistributed by anyone. This concept is grounded in the principle of universal access, meaning that data should be available to all without the need for special permissions or payments. However, it is also important to note that the reuse and redistribution of open data are often regulated by licenses, such as Creative Commons, to ensure proper usage and attribution.

What Is FAIR Data?

FAIR data refers to data that follows the FAIR principles: findable, accessible, interoperable, and reusable. These principles were formalized in 2016 to guide better data management and stewardship, particularly in the context of digital research outputs:

  • Findable: Data should have rich metadata and unique identifiers, making it easy for researchers and tools to locate it.
  • Accessible: Data and metadata should be retrievable through standardized protocols, possibly with authentication when necessary.
  • Interoperable: Data should use formats, vocabularies, and standards that allow for integration with other datasets.
  • Reusable: Data should have clear usage licenses and detailed provenance to enable validation and reuse.

Importantly, the FAIR principles do not require data to be open, as we will see in the next section, but instead focus on how data should be organized and described to maximize its discoverability and reuse by both humans and machines.

The Difference Between FAIR Data and Open data

FAIR is not equal to open. The “A” in FAIR refers to “accessible under well-defined conditions”, which means that data do not necessarily need to be open to everyone. There are circumstances where openness could pose risks, such as with confidential or personal data. In essence, data should always be FAIR, but they are not required to be open in all cases.

However, it is important to note that data or data infrastructures should not use the FAIR principles as a justification for restricting openness when it is appropriate and necessary to share the data!

Infographic about Open Data VS. FAIR Data

Beyond accessibility conditions, another key distinction lies in the focus of each concept: FAIR is primarily concerned with how data are structured, described, and managed to enable effective discovery, interoperability, and reuse, whereas open data is primarily concerned with legal permissions and the removal of access barriers. Data can be openly available online yet lack sufficient metadata, standardized formats, or clear provenance, making them difficult to interpret or reuse in practice. Ideally, data should be both FAIR and open whenever possible, but the two concepts address different dimensions of responsible data sharing.

New Horizons in Open and FAIR Data

As we approach the 10th anniversary of the FAIR data concept in 2026, it is essential to consider today’s evolving data requirements. With the rapid rise of AI, data needs extend far beyond just being FAIR. Some suggestions to extend the FAIR principles include:

The extension of the FAIR principles to FAIR-R (where the “R” stands for readiness) has been proposed to address the specific needs of AI systems. In this context, data readiness for AI refers to the process of preparing and ensuring the quality, accessibility, and suitability of datasets before they can be used in AI applications.

Or the proposal of adding the dimension of “understanding,”, sugested in a recent article in the Scholarly Kitchen. This proposal emphasizes that data should not only be accessible and reusable, but also comprehensible and interpretable within specific contexts.

As we move forward, the future of open and FAIR data lies in their ongoing refinement to meet the ever-changing demands of technology, ethics, and research, ensuring that data remains not just accessible and reusable, but also ready for the challenges of modern AI and beyond.

]]>
Celebrating Education Week with Richard West https://opensciencelab.uc3m.es/blog/celebrating-education-week-with-richard-west/ Wed, 04 Mar 2026 09:41:44 +0000 https://opensciencelab.uc3m.es/?p=3083

Celebrating Open Education Week by Welcoming Richard West to OpenScienceLab

We take advantage of Open Education Week 2026 to welcome Richard West to UC3M as a Fulbright Visiting Scholar. Open Education Week is a celebration of the global Open Education Movement, organized by Open Education Global. Its goal is to raise awareness about the movement and its impact on teaching and learning worldwide. And we are excited to announce Dr. West’s presence during this week, as his work exemplifies the very spirit of open education.

Rick West photo

About Richard West

Dr. Richard West is a professor of instructional psychology and technology at Brigham Young University in the United States, and is visiting UC3M this semester as a Fulbright Scholar. His research expertise is in open education and open microcredentials/badges. He has created several popular open textbooks in the Edtechbooks.org platform including:

With David Wiley, he is the co-author of the 5Cs of Open Education Infrastructure framework, which defines the key characteristics of open education and articulates current opportunities for open education through generative AI.

He sees open education as having a reciprocal relationship with open science, where according to West:

“Open science can provide students with stronger and more current reading materials while also providing students the opportunity to reproduce and re-analyze research as open renewable assignments that are more engaging and that produce more productive learning. In return, open pedagogical practices can incorporate student work in creating open science that contributes back to the research community.”

]]>
Open Science for Sovereignity: the Next Challenge https://opensciencelab.uc3m.es/blog/conference-open-science-for-sovereignity-the-next-challenge/ Mon, 02 Mar 2026 10:00:12 +0000 https://opensciencelab.uc3m.es/?p=3036

Open Science for Sovereignty: the Next Challenge

 

Monday, 9th March 2026 – 12:00 (UTC+01:00).

 

Day(s)

:

Hour(s)

:

Minute(s)

:

Second(s)

Poster for Richard Gold's Conference

Conference Summary

As countries adjust to a rapidly changing geopolitical environment, leaders of middle powers are embracing national sovereignty with a twist: that to achieve sovereignty, countries must cooperate with other middle powers on technological development. Many European countries (as well as Canada) suffer from fragile innovation ecosystems in which firms find it difficult to develop and scale up technology at home, leading to innovation leakage to more power states. Reversing this trend is key to gaining sovereignty. In this presentation, Prof. Gold will discuss strategies to do so, including cooperation among researchers across middle-powers, policy experimentation, and  deploying open science to build capacity and industry support.

About Richard Gold

E. Richard Gold is a CIGI senior fellow and a James McGill Professor with McGill University’s Faculty of Law and was the founding director of the Centre for Intellectual Property Policy.

He holds a B.Sc. from McGill University, an LL.B. from the University of Toronto, and an S.J.D. and LL.M. from the University of Michigan Law School.

He specializes in the legal, social, political, and economic aspects of intellectual property (IP) law and innovation. As a McGill professor, he teaches IP, international IP, and innovation policy, and, over the years, he has advised organizations like Health Canada, the WHO, and WIPO.

]]>
Why Scientific Publishing Is Stuck in the Past and How Open Science Can Fix It https://opensciencelab.uc3m.es/blog/scientific-publishing-is-stuck-in-the-past/ Fri, 06 Feb 2026 07:45:13 +0000 https://opensciencelab.uc3m.es/?p=2903 We often discuss how outdated the way researchers are assessed is: rewarding the number of publications and relying on obsolete metrics that fail to capture the real value of scientific work. What we talk about much less is another factor that deeply reinforces this problem: the way research itself is published.

Why Scientific Publishing Still Looks Like It Did Decades Ago?

Why do journals still publish static PDFs? A simple answer is: because it is cheap and extremely profitable.

From the publishers’ perspective, the PDF-based paper is an ideal product. It is standardized, scalable, and requires minimal innovation. The infrastructure has been amortized for decades, and the marginal cost of publishing yet another paper is very low, resulting in extremely high profit margins.

And taking into account where the money comes from (authors and institutions paying substantial APCs, sometimes reaching several thousand dollars per paper), it becomes clear why popular publishers (no need to name them; you already know exactly who they are) maintain this model. Authors and their institutions are left with little real choice: they must pay to ensure that years of work do not remain locked away in a drawer, and to remain competitive for funding, positions, and more favorable working conditions. But what makes this model even more striking is that the peer-review process is mostly unpaid. Reviewers, who are themselves researchers, volunteer their time and expertise to evaluate manuscripts, improve their quality, and uphold scientific standards. Editors are often academics as well. The intellectual labor that sustains the system is largely provided for free (authors pay to publish, while reviewers are not compensated, simple logic!)

Under these conditions, there is little incentive for publishers to fundamentally rethink how research is communicated. Moving away from static PDFs toward more modular, interactive, or living research outputs would require structural changes and potentially disrupt a business model that currently works very well for them. To give just one obvious example: in the Web 4.0 era, it is hard to justify why scientific articles still cannot natively support interactive features like data visualizations, simply because we remain tied to an outdated, static PDF format.

Modern Research Produces Far More Than Articles

Code, datasets, models, protocols, benchmarks, negative results, software tools, living documents, and shared infrastructures. Much of the work that enables scientific progress never fits neatly into a traditional paper and is therefore undervalued or ignored. By centering the entire system around papers, we create predictable consequences:

  • We prioritize quantity over substance,
  • Delay the dissemination of knowledge,
  • Discourage openness and collaboration,
  • And systematically overlook essential contributions that do not translate into “publishable units.”
Graphic showing different types of research outputs represented as colored labels: Code, Datasets, Models, Protocols, Benchmarks, Negative Results, Software Tools, Living Documents, and Shared Infrastructures.
Some of the types of research outputs

Innovation in publishing is not technologically difficult.
It is just economically inconvenient for publishers.

As long as prestige, evaluation, and funding remain tied to journals and their traditional formats, publishers have little reason to challenge the system that secures their revenue.

Open Science is the Solution

If the problem is structural, the solution must be structural as well.

A fairer and more sustainable research ecosystem requires breaking ties with organizations that extract value from science without proportionally contributing to it. As long as the infrastructure of research (publication, evaluation, and discovery) is controlled by profit-driven intermediaries, meaningful change will remain limited.

Encouragingly, some institutions are starting to take concrete steps in this direction. Recent initiatives such as the decision by CNRS (2025) to break free from Web of Science set an excellent precedent for other research institutions to follow. These moves matter even more because open alternatives already exist. Platforms like OpenAlex show that open, community-driven bibliographic infrastructures are both possible and effective.

Detaching research assessment and discovery from commercial interests is not just an ideological stance; it is a practical requirement for progress. Open infrastructures enable:

  • Fairer evaluation practices,
  • Reproducibility and transparency,
  • Broader access to knowledge,
  • And recognition of diverse research outputs beyond papers.

Open Science is not simply about open access to PDFs.
It is about reclaiming control of the research ecosystem itself.

Step by step, these changes can help build a research ecosystem that is more just, more efficient, and ultimately more aligned with the public good it is meant to serve. Hopefully, more organizations will join these initiatives and commit to free, open, and publicly accessible knowledge.

]]>
The Problems of Using AI for Academic Research https://opensciencelab.uc3m.es/blog/the-problems-of-using-ai-for-academic-research/ Mon, 01 Dec 2025 08:12:56 +0000 https://opensciencelab.uc3m.es/?p=2835 Artificial intelligence has become a popular tool for finding information, summarizing texts, and even suggesting academic sources. But while AI can speed up the research process, it also introduces a set of risks that many users overlook.

Invented Information and “Hallucinated” Sources

Some widely used AI tools—such as general-purpose models like GPT—often produce citations, authors, or articles that look convincing but simply don’t exist. A recent study by Oladokun et al. (2025) found a disturbingly high frequency of false or non-existent citations, with 42.9% in ChatGPT-3.5 and 51.8% in ChatGPT-4o, underscoring how unreliable GPT models are for this task. And this problem is also present in other AI search tools like Perplexity, Perplexity Pro, DeepSeek Serarch, Copilot, Grok2, Grok3 y Gemini (Jaźwińska et al., 2025)

However, there are specialized AI tools designed specifically for locating scientific articles—such as Consensus (free) or Elicit (freemium)—that significantly reduce the risk of invented references. Nevertheless, while they avoid generating fake citations, they still present other challenges that we will explore below.

Researcher-Induced Bias and Selective Queries

Another subtle but important problem arises when the researcher’s own question reinforces pre-existing assumptions. For example, a query such as “show me articles that support [specific claim]” inadvertently promotes confirmation bias, because the AI will prioritize papers aligned with that statement while ignoring literature that challenges or contradicts it.

And Another Layer of Bias: Training Data Limitations

Beyond user-induced bias, AI systems also reproduce biases embedded in their training data. Because they learn from internet-scale text and highly digitized sources, they tend to favor:

  • English-language publications,
  • Dominant theoretical perspectives over minority or emerging viewpoints.

Misinterpretations and Mixed Content

Even when an AI retrieves a real and correctly cited article, it may still misunderstand the study’s conclusions or blend information from multiple texts. This can produce summaries that appear accurate at first glance but actually distort the original meaning. In some cases, the AI may attribute a conclusion to a paper that never makes such a claim—or worse, one that argues the opposite. This issue is especially common in specialized academic retrieval tools, such as the previously mentioned Consensus and Elicit, and it can even occur in systems like NotebookLM, where the model may misinterpret scientific content that the user has personally uploaded.

Difficulty Assessing Source Reliability

Another major limitation of AI-powered research tools is their inability to consistently distinguish between high-quality, peer-reviewed literature and sources with no credibility, for example, from journals known for questionable research practices (QRP).

That’s why AI should be understood as a complementary tool—not a substitute for academic judgment. Using it responsibly requires verifying references, consulting original papers, questioning assumptions, and engaging critically with the literature. When researchers combine the efficiency of AI with thoughtful human oversight, they can benefit from its capabilities while minimizing its most common risks.

Venn diagram showing the overlap between AI research tools and human academic judgment. The central orange section represents ‘AI-Enhanced Responsible Research,’ combining the strengths of both AI and human oversight

References

5942012 {16784933:CXNGZGIS},{16784933:4WPPQ3KV} 1 apa 50 default 2835 https://opensciencelab.uc3m.es/wp-content/plugins/zotpress/
%7B%22status%22%3A%22success%22%2C%22updateneeded%22%3Afalse%2C%22instance%22%3Afalse%2C%22meta%22%3A%7B%22request_last%22%3A0%2C%22request_next%22%3A0%2C%22used_cache%22%3Atrue%7D%2C%22data%22%3A%5B%7B%22key%22%3A%22CXNGZGIS%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Oladokun%20et%20al.%22%2C%22parsedDate%22%3A%222025%22%2C%22numChildren%22%3A0%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BOladokun%2C%20B.%20D.%2C%20Enakrire%2C%20R.%20T.%2C%20Emmanuel%2C%20A.%20K.%2C%20Ajani%2C%20Y.%20A.%2C%20%26amp%3B%20Adetayo%2C%20A.%20J.%20%282025%29.%20Hallucitation%20in%20Scientific%20Writing%3A%20Exploring%20Evidence%20from%20ChatGPT%20Versions%203.5%20and%204o%20in%20Responses%20to%20Selected%20Questions%20in%20Librarianship.%20%26lt%3Bi%26gt%3BJournal%20of%20Web%20Librarianship%26lt%3B%5C%2Fi%26gt%3B%2C%20%26lt%3Bi%26gt%3B19%26lt%3B%5C%2Fi%26gt%3B%281%29%2C%2062%26%23x2013%3B92.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-ItemURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1080%5C%2F19322909.2025.2482093%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1080%5C%2F19322909.2025.2482093%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22Hallucitation%20in%20Scientific%20Writing%3A%20Exploring%20Evidence%20from%20ChatGPT%20Versions%203.5%20and%204o%20in%20Responses%20to%20Selected%20Questions%20in%20Librarianship%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Bolaji%20David%22%2C%22lastName%22%3A%22Oladokun%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Rexwhite%20Tega%22%2C%22lastName%22%3A%22Enakrire%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Adefila%20Kolawole%22%2C%22lastName%22%3A%22Emmanuel%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Yusuf%20Ayodeji%22%2C%22lastName%22%3A%22Ajani%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Adebowale%20Jeremy%22%2C%22lastName%22%3A%22Adetayo%22%7D%5D%2C%22abstractNote%22%3A%22The%20rapid%20adoption%20of%20AI%20in%20academic%20writing%2C%20particularly%20with%20tools%20like%20ChatGPT%2C%20has%20raised%20significant%20concerns%20regarding%20the%20accuracy%20of%20generated%20content.%20This%20study%20explores%20the%20phenomenon%20of%20%5Cu201challucitation%5Cu201d%20in%20scientific%20writing%2C%20where%20AI%20models%20fabricate%20citations%2C%20analyzing%20responses%20from%20ChatGPT%20versions%203.5%20and%204o%20in%20the%20context%20of%20librarianship.%20Through%20an%20experimental%20design%2C%20scientific%20content%20with%20citations%20was%20generated%20and%20systematically%20verified%20using%20Google%20Scholar%20and%20the%20publisher%5Cu2019s%20website.%20The%20findings%20reveal%20a%20disturbingly%20high%20frequency%20of%20false%20or%20non-existent%20citations%5Cu201442.9%25%20in%20ChatGPT-3.5%20and%2051.8%25%20in%20ChatGPT-4o.%20Despite%20slight%20improvements%20in%20citation%20accuracy%20from%20version%203.5%20to%204o%2C%20with%20accuracy%20rates%20of%203.92%25%20and%206.35%25%2C%20respectively%2C%20both%20versions%20exhibit%20significant%20limitations.%20Notably%2C%20ChatGPT%203.5%20frequently%20generated%20completely%20fabricated%20sources%2C%20while%20ChatGPT-4o%20introduced%20subtle%20errors%2C%20such%20as%20mismatched%20journals.%20The%20study%20indicates%20no%20significant%20difference%20in%20accuracy%20between%20the%20two%20versions%2C%20underscoring%20the%20persistent%20risks%20associated%20with%20AI-generated%20citations.%20These%20findings%20highlight%20the%20urgent%20need%20for%20rigorous%20verification%20of%20AI-generated%20content%20to%20safeguard%20the%20integrity%20of%20scholarly%20work.%22%2C%22date%22%3A%222025%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.1080%5C%2F19322909.2025.2482093%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1080%5C%2F19322909.2025.2482093%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%221932-2909%22%2C%22language%22%3A%22%22%2C%22collections%22%3A%5B%225NCEAR8W%22%5D%2C%22dateModified%22%3A%222025-12-01T08%3A03%3A33Z%22%7D%7D%2C%7B%22key%22%3A%224WPPQ3KV%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Ja%5Cu017awi%5Cu0144ska%20and%20Chandrasekar%22%2C%22parsedDate%22%3A%222025%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BJa%26%23x17A%3Bwi%26%23x144%3Bska%2C%20K.%2C%20%26amp%3B%20Chandrasekar%2C%20A.%20%282025%29.%20AI%20Search%20Has%20a%20Citation%20Problem.%20%26lt%3Bi%26gt%3BColumbia%20Journalism%20Review%26lt%3B%5C%2Fi%26gt%3B.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-ItemURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fwww.cjr.org%5C%2Ftow_center%5C%2Fwe-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fwww.cjr.org%5C%2Ftow_center%5C%2Fwe-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22blogPost%22%2C%22title%22%3A%22AI%20Search%20Has%20a%20Citation%20Problem%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Klaudia%22%2C%22lastName%22%3A%22Ja%5Cu017awi%5Cu0144ska%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Aisvarya%22%2C%22lastName%22%3A%22Chandrasekar%22%7D%5D%2C%22abstractNote%22%3A%22We%20compared%20eight%20AI%20search%20engines.%20They%5Cu2019re%20all%20bad%20at%20citing%20news.%22%2C%22blogTitle%22%3A%22Columbia%20Journalism%20Review%22%2C%22date%22%3A%222025%22%2C%22DOI%22%3A%22%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fwww.cjr.org%5C%2Ftow_center%5C%2Fwe-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php%22%2C%22ISSN%22%3A%22%22%2C%22language%22%3A%22en%22%2C%22collections%22%3A%5B%225NCEAR8W%22%5D%2C%22dateModified%22%3A%222025-12-01T08%3A03%3A26Z%22%7D%7D%5D%7D
Oladokun, B. D., Enakrire, R. T., Emmanuel, A. K., Ajani, Y. A., & Adetayo, A. J. (2025). Hallucitation in Scientific Writing: Exploring Evidence from ChatGPT Versions 3.5 and 4o in Responses to Selected Questions in Librarianship. Journal of Web Librarianship, 19(1), 62–92. https://doi.org/10.1080/19322909.2025.2482093
Jaźwińska, K., & Chandrasekar, A. (2025). AI Search Has a Citation Problem. Columbia Journalism Review. https://www.cjr.org/tow_center/we-compared-eight-ai-search-engines-theyre-all-bad-at-citing-news.php

]]>
Six Ways Open Science Is (and Isn’t) Changing Research Culture https://opensciencelab.uc3m.es/blog/six-ways-open-science-is-and-isnt-changing-research-culture/ Fri, 14 Nov 2025 06:55:29 +0000 https://opensciencelab.uc3m.es/?p=2782

This year, Science Europe released something big. 

This is The Scoping Review: the Contributions of Open Science to Research Culture, on how open science is actually shaping the way we work, collaborate, and care in research. Open science in real practice, for real.

It’s the first systematic look at whether open science really delivers on its promises of:

  • Inclusion 
  • Transparency 
  • Integrity 
  • Care 
  • Collaboration 
  • and Freedom.

The review scanned 62 studies from around the world, mapping how these six values show up —or don’t— in real research settings. And the results have been inspiring, sometimes messy and sometimes uncomfortable.

So here’s our OpenScienceLab digest

A quick tour through the six cultural values of open science. What the evidence says, where it falls short, and what we can learn moving forward.

Let’s talk about culture. Research culture.

Ready? Let’s dive in.

Equality, Diversity & Inclusion

Open science likes to say it’s for everyone. 

But… in the Scoping Review, just five studies actually explored this dimension and told us indeed a split story.

On the bright side, openness can amplify women’s visibility and leadership.

While the flip side is sharper when it reveals that openness without justice can deepen inequality.

So maybe the question isn’t how open we are, but for whom that openness works.

Recommendation: Pair every open policy with equity checks. Otherwise, inclusion remains just another promise.

Openness & Transparency

It’s the heart of the open science promise when it says make everything visible, shareable, reviewable. 

Out of the 62 studies, over twenty dealt with openness and transparency. But again, the story splits in two.

When openness is grounded in trust, it works.

Yet, when it’s imposed from above, things break.

In the end, transparency is only as strong as the trust behind it and only works when trust runs both ways.

Recommendation: Instead of enforcing openness, build confidence first. The culture change will follow.

Integrity & Ethics

Everyone agrees that science should be honest, transparent, and accountable. The question is whether openness is really helping us get there.

In the Scoping Review, around a dozen studies looked at how open science interacts with integrity. The picture is promising, though perfection is still out of reach.

Some good news first. Open practices like preregistration and replication are slowly moving from rebellion to routine.

Still, there’s a catch. When openness turns into pressure, for example to publish, to disclose, to perform, it can backfire.

So, open science can strengthen ethics when it empowers, not overwhelms.

Recommendation: Support honesty, and integrity takes root.

Care & Collegiality

We talk plenty about data and metrics, almost never about care.

In the Scoping Review, only one study focused directly on care. One. And it said a lot by what it revealed.

Open science depends on curators, data stewards, and behind-the-scenes work that often goes unseen.

It’s a quiet reminder that openness still runs on invisible labour. The kind that keeps research flowing, but rarely gets credit.

Recommendation: Make care visible. Recognize the hidden roles that make openness possible.

Collaboration

Open science loves to celebrate teamwork. Collaboration, however, doesn’t just happen because we say “open”.

In the Scoping Review, collaboration showed up in several studies, often tied to infrastructures and citizen science.

When researchers build things together, like data platforms, communities and shared tools, collaboration grows strong and steady.

Even so, when the system rewards competition over connection, cooperation crumbles.

True collaboration takes time, care, and shared purpose, well beyond data sharing. In fact, it’s something data alone cannot achieve.

Recommendation: Build collaboration as culture, starting with relationships.

Autonomy & Freedom

We often think openness gives researchers more freedom. But does it always?

In the Scoping Review, only a few studies touched this theme, and most of them sounded a warning.

It suggests when openness turns rigid, creativity suffers.

At its best, open science can empower autonomy when it protects the space for creativity to thrive.

Recommendation: Keep openness flexible. Apart from sharing, freedom in research means having space to think.

Reading Culture Through Openness

The Scoping Review shows that open science is changing research culture. 

—Unevenly, imperfectly, yet unmistakably.

With this digest, the OpenScienceLab reads those shifts up close, tracing where culture is already changing and where it still resists. The process is already underway.

The full version of the Science Europe analysis is available at: https://doi.org/10.5281/zenodo.17379695.

]]>
TWIN4MERIT- International Thematic Symposium https://opensciencelab.uc3m.es/blog/twin4merit-international-thematic-symposium/ Mon, 03 Nov 2025 09:08:29 +0000 https://opensciencelab.uc3m.es/?p=2689

TWIN4MERIT – International Thematic Symposium

 

Rethinking Research Assessment

Monday, 24th November 2025 from 9:45 to 16:00 PM (UTC+01:00).

Platform: Microsoft Teams.

Day(s)

:

Hour(s)

:

Minute(s)

:

Second(s)

Symposium Summary

The TWIN4MERIT International Thematic Symposium will provide a platform for experts to discuss the evolving landscape of research assessment. The event will cover various topics, including the role of university libraries in enhancing research visibility through institutional repositories and open access policies, the influence of sustainability rankings on university evaluations, and the growing importance of citizen engagement in research. Discussions will also focus on rethinking traditional academic evaluation metrics, with a shift towards more inclusive and narrative-based approaches, particularly for early-career researchers. The symposium will conclude with a panel on the future of research assessment in the open science era, emphasizing the need for more collaborative, transparent, and inclusive evaluation systems.

Speakers

PhD. Gema Bueno de la Fuente

Universidad de Zaragoza

PhD. Nuria Bautista Puig

Consejo Superior de Investigaciones Científicas (CSIC)

PhD. Emanuel Kulczycki

Adam Mickiewicz University in Poznań

PhD. Yensi Flores Bueso

University College Cork

PhD. Zacharias Maniadis

University of Cyprus

PhD. Eva Méndez

Universidad Carlos III de Madrid

Symposium Schedule

Time (CET, UTC+1): Time (EET, UTC+2): Session Details:
09:45 – 10:00 10:45 – 11:00 Welcome & Introduction (15 min)
Moderator: Welcome participants and introduce the symposium objectives.
10:00 – 10:45 11:00 – 11:45 PhD. Gema Bueno de la Fuente (Universidad de Zaragoza)
Session 1 – University Libraries & Research Assessment (45 min)
Abstract:
This session explores how institutional repositories and open access policies are transforming the visibility, evaluation, and strategic positioning of research in higher education. The first part examines the management and governance of institutional repositories, analysing how they function not only as archives but as instruments for enhancing institutional identity and research transparency. The second part focuses on the role of open access policies in increasing research visibility and compliance with national and international mandates. Finally, the session outlines practical strategies through which libraries can play an active role in research assessment, supporting academics in data curation, metadata standardisation, and the responsible use of metrics.
10:45 – 11:30 11:45 – 12:30 PhD. Núria Bautista Puig (CSIC – Consejo Superior de Investigaciones Científicas)
Session 2 – Reframing research assessment: from sustainability rankings to citizen engagement (45 min)
Abstract:
This session explores how sustainability and participation are reshaping the way research and higher education institutions are assessed. The first part analyses five major sustainability rankings (UI GreenMetric, THE Impact Rankings, QS Sustainability Rankings, STARS, and People & Planet) to examine how they operationalise sustainability through their indicators. This comparison reveals that most rankings still prioritise environmental aspects over social and educational ones, highlighting the need for more holistic and transparent frameworks that capture the multidimensional nature of sustainability in universities. The second part extends this reflection to the participatory dimension of research, analysing projects from citizen science databases. The results show that, although public participation is widespread, the depth of engagement remains limited.
11:30 – 13:00 12:30 – 14:00 Lunch Break
13:00 – 13:45 14:00 – 14:45 PhD. Emanuel Kulczycki (Adam Mickiewicz University in Poznań)
Sub-session 3a – “The Evaluation Game” (45 min)
Abstract:
This session explores how the rules of academic evaluation are being questioned and redefined in the face of growing criticism of traditional research assessment systems. The first part examines the limitations of conventional indicators such as journal impact factors and citation counts, highlighting their inability to capture the diversity, quality, and societal relevance of research. The discussion then turns to the systemic challenges that perpetuate metric-driven cultures, including incentive misalignments, disciplinary biases, and the marginalisation of open and collaborative practices. The second part presents alternative strategies for improving academic evaluation, drawing on emerging frameworks such as responsible metrics, qualitative peer review, and narrative-based assessment. Overall, The Evaluation Game invites participants to rethink how value and excellence are defined in academia, encouraging a shift towards fairer, more transparent, and context-sensitive models of research assessment.
13:45 – 14:30 14:45 – 15:30 PhD. Yensi Alejandra Flores Bueso (University College Cork)
Sub-session 3b – Narrative CVs and Early-Career Researchers (45 min)
Abstract:
This session examines how narrative CVs are reshaping research assessment towards more inclusive and transparent models. The first part analyses narrative CVs as tools to move beyond purely quantitative indicators, enabling a fairer recognition of diverse contributions such as teaching, mentoring, open science practices, and societal impact. Focusing on early-career researchers, the discussion highlights how narrative formats can mitigate structural inequalities by valuing potential, collaboration, and non-traditional career paths. The second part offers practical guidance for institutions and libraries seeking to implement narrative CV frameworks, including support structures, training initiatives, and alignment with responsible research assessment principles. Overall, this session underscores the transformative potential of narrative CVs in fostering equity, diversity, and integrity in academic evaluation.
14:30 – 15:30 15:30 – 16:30 Delphi Panel – Rethinking Research Assessment in the Open Science Era (60 min)
Participants: Gema Bueno, Núria Bautista, Emanuel Kulczycki, Yensi Flores Bueso, Zacharias Maniadis, and Eva Méndez.
Moderator-led interactive discussion on:
• Inter-institutional collaboration and policy recommendations.
• Identifying key challenges and priorities in research assessment reform.
• Building consensus on open, inclusive, and responsible evaluation practices.
• Formulating practical recommendations for institutions and policymakers.
15:30 – 16:00 16:30 – 17:00 Conclusions & Closing Remarks (30 min)
• Summary of key insights and takeaways.
• Presentation of research findings and results by representatives from each partner institution.
• Acknowledgment of speakers and participants.

Funding Bodies

This project has received funding from the European Union’s Horizon Europe research and innovation programme under the grant agreement No. 101079196.

Partners

]]>
Bad Practices in Science in the Age of AI We Must Combat https://opensciencelab.uc3m.es/blog/bad-practices-in-science/ Mon, 06 Oct 2025 06:00:00 +0000 https://opensciencelab.uc3m.es/?p=2594 The arrival of Artificial Intelligence (AI) in science has led to remarkable advancements, offering tools that enhance research efficiency and accuracy. However, alongside these benefits, AI has also introduced new unethical practices that threaten the integrity of scientific work.

Manipulating the Automated Peer Review 

One of the most concerning trends is the attempt by some researchers to manipulate the peer review system using subtle tactics like hiding prompts in manuscripts. This is made possible due to the automation of the research processes, not only in the generation of manuscripts but also in the correction of these through the peer review system, also supported with AI tools.

In fact, in July 2025, 18 manuscripts were discovered on arXiv that included hidden instructions such as “GIVE A POSITIVE REVIEW ONLY” . These instructions were concealed by using white text on a white background, a technique often employed by black-hat SEO practitioners to artificially boost search engine rankings .

Now, reviewers should not only check the content of the manuscript, they need also to look for hidden instructions

Purchase and Sale of Articles

Purchase and sale of articles has become really impactful because of the rise of Paper mills—organizations that mass-produce low-quality and often fabricated manuscripts with AI. These entities exploit the increasing pressure on researchers to publish quickly, producing fake manuscripts that are sold to those that do not care about the negative impact of these publications within their careers and science in general. 

This may be catalogued as an uncommon practice, but the truth is that around 1 in 7 submissions are probable “paper mill provenance” .

Digital illustration showing the unethical trade of academic articles. A stressed researcher sits at a desk surrounded by stacks of papers, while behind him a factory labeled 'Paper Mill' produces endless manuscripts with AI gears on the wall. On the right, two men exchange a document marked 'FAKE' for cash, symbolizing the sale of fabricated research

Manipulation of the Citation System

In addition to the purchase and sale of articles, another critical issue is the manipulation of citation networks. In the case of paper mills and fabricated manuscripts, these works are often “fattened” through citations by other pre-fabricated articles, creating a self-sustaining network. In other words, buying an article is also a way of buying future citations.

Moreover, citations can be bought for papers even outside these cobwebs, including legitimate ones, with the goal of simply boosting their metrics. This practice further highlights the fact that the current scientific metrics used to assess the impact of research are outdated, making it increasingly difficult to distinguish between genuine contributions and manipulated data.

The unethical practices exposed throughout this article—manipulating peer review, purchasing and selling articles, and artificially inflating citation networks—are undermining the integrity of scientific publishing. How much longer will we allow an outdated evaluation system that enables and rewards the use of these techniques?

References

5942012 {16784933:GQ7MPVSF},{16784933:RAVW4LPS},{16784933:2JBJELWJ},{16784933:WB75AGYN} 1 apa 50 default 2594 https://opensciencelab.uc3m.es/wp-content/plugins/zotpress/
%7B%22status%22%3A%22success%22%2C%22updateneeded%22%3Afalse%2C%22instance%22%3Afalse%2C%22meta%22%3A%7B%22request_last%22%3A0%2C%22request_next%22%3A0%2C%22used_cache%22%3Atrue%7D%2C%22data%22%3A%5B%7B%22key%22%3A%22GQ7MPVSF%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Ibrahim%20et%20al.%22%2C%22parsedDate%22%3A%222025-02-14%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BIbrahim%2C%20H.%2C%20Liu%2C%20F.%2C%20Zaki%2C%20Y.%2C%20%26amp%3B%20Rahwan%2C%20T.%20%282025%29.%20Citation%20manipulation%20through%20citation%20mills%20and%20pre-print%20servers.%20%26lt%3Bi%26gt%3BScientific%20Reports%26lt%3B%5C%2Fi%26gt%3B%2C%20%26lt%3Bi%26gt%3B15%26lt%3B%5C%2Fi%26gt%3B%281%29%2C%205480.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-DOIURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1038%5C%2Fs41598-025-88709-7%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1038%5C%2Fs41598-025-88709-7%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22Citation%20manipulation%20through%20citation%20mills%20and%20pre-print%20servers%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Hazem%22%2C%22lastName%22%3A%22Ibrahim%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Fengyuan%22%2C%22lastName%22%3A%22Liu%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Yasir%22%2C%22lastName%22%3A%22Zaki%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Talal%22%2C%22lastName%22%3A%22Rahwan%22%7D%5D%2C%22abstractNote%22%3A%22Citations%20are%20widely%20considered%20in%20scientists%5Cu2019%20evaluation.%20As%20such%2C%20scientists%20may%20be%20incentivized%20to%20inflate%20their%20citation%20counts.%20While%20previous%20literature%20has%20examined%20self-citations%20and%20citation%20cartels%2C%20it%20remains%20unclear%20whether%20scientists%20can%20purchase%20citations.%20Here%2C%20we%20compile%20a%20dataset%20of%20~1.6%20million%20profiles%20on%20Google%20Scholar%20to%20examine%20instances%20of%20citation%20fraud%20on%20the%20platform.%20We%20survey%20faculty%20at%20highly-ranked%20universities%2C%20and%20confirm%20that%20Google%20Scholar%20is%20widely%20used%20when%20evaluating%20scientists.%20We%20then%20engage%20with%20a%20citation-boosting%20service%2C%20and%20manage%20to%20purchase%2050%20citations%20while%20assuming%20the%20identity%20of%20a%20fictional%20author.%20Taken%20as%20a%20whole%2C%20our%20findings%20bring%20to%20light%20new%20forms%20of%20citation%20manipulation%2C%20and%20emphasize%20the%20need%20to%20look%20beyond%20citation%20counts.%22%2C%22date%22%3A%222025-02-14%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.1038%5C%2Fs41598-025-88709-7%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fwww.nature.com%5C%2Farticles%5C%2Fs41598-025-88709-7%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%222045-2322%22%2C%22language%22%3A%22en%22%2C%22collections%22%3A%5B%22XE9K2VY5%22%5D%2C%22dateModified%22%3A%222025-09-09T11%3A58%3A36Z%22%7D%7D%2C%7B%22key%22%3A%22WB75AGYN%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Richardson%20et%20al.%22%2C%22parsedDate%22%3A%222025-08-12%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BRichardson%2C%20R.%20A.%20K.%2C%20Hong%2C%20S.%20S.%2C%20Byrne%2C%20J.%20A.%2C%20Stoeger%2C%20T.%2C%20%26amp%3B%20Amaral%2C%20L.%20A.%20N.%20%282025%29.%20The%20entities%20enabling%20scientific%20fraud%20at%20scale%20are%20large%2C%20resilient%2C%20and%20growing%20rapidly.%20%26lt%3Bi%26gt%3BProceedings%20of%20the%20National%20Academy%20of%20Sciences%26lt%3B%5C%2Fi%26gt%3B%2C%20%26lt%3Bi%26gt%3B122%26lt%3B%5C%2Fi%26gt%3B%2832%29%2C%20e2420092122.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-DOIURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1073%5C%2Fpnas.2420092122%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1073%5C%2Fpnas.2420092122%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22The%20entities%20enabling%20scientific%20fraud%20at%20scale%20are%20large%2C%20resilient%2C%20and%20growing%20rapidly%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Reese%20A.%20K.%22%2C%22lastName%22%3A%22Richardson%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Spencer%20S.%22%2C%22lastName%22%3A%22Hong%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jennifer%20A.%22%2C%22lastName%22%3A%22Byrne%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Thomas%22%2C%22lastName%22%3A%22Stoeger%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Lu%5Cu00eds%20A.%20Nunes%22%2C%22lastName%22%3A%22Amaral%22%7D%5D%2C%22abstractNote%22%3A%22Science%20is%20characterized%20by%20collaboration%20and%20cooperation%2C%20but%20also%20by%20uncertainty%2C%20competition%2C%20and%20inequality.%20While%20there%20has%20always%20been%20some%20concern%20that%20these%20pressures%20may%20compel%20some%20to%20defect%20from%20the%20scientific%20research%20ethos%5Cu2014i.e.%2C%20fail%20to%20make%20genuine%20contributions%20to%20the%20production%20of%20knowledge%20or%20to%20the%20training%20of%20an%20expert%20workforce%5Cu2014the%20focus%20has%20largely%20been%20on%20the%20actions%20of%20lone%20individuals.%20Recently%2C%20however%2C%20reports%20of%20coordinated%20scientific%20fraud%20activities%20have%20increased.%20Some%20suggest%20that%20the%20ease%20of%20communication%20provided%20by%20the%20internet%20and%20open-access%20publishing%20have%20created%20the%20conditions%20for%20the%20emergence%20of%20entities%5Cu2014paper%20mills%20%28i.e.%2C%20sellers%20of%20mass-produced%20low%20quality%20and%20fabricated%20research%29%2C%20brokers%20%28i.e.%2C%20conduits%20between%20producers%20and%20publishers%20of%20fraudulent%20research%29%2C%20predatory%20journals%2C%20who%20do%20not%20conduct%20any%20quality%20controls%20on%20submissions%5Cu2014that%20facilitate%20systematic%20scientific%20fraud.%20Here%2C%20we%20demonstrate%20through%20case%20studies%20that%20i%29%20individuals%20have%20cooperated%20to%20publish%20papers%20that%20were%20eventually%20retracted%20in%20a%20number%20of%20journals%2C%20ii%29%20brokers%20have%20enabled%20publication%20in%20targeted%20journals%20at%20scale%2C%20and%20iii%29%2C%20within%20a%20field%20of%20science%2C%20not%20all%20subfields%20are%20equally%20targeted%20for%20scientific%20fraud.%20Our%20results%20reveal%20some%20of%20the%20strategies%20that%20enable%20the%20entities%20promoting%20scientific%20fraud%20to%20evade%20interventions.%20Our%20final%20analysis%20suggests%20that%20this%20ability%20to%20evade%20interventions%20is%20enabling%20the%20number%20of%20fraudulent%20publications%20to%20grow%20at%20a%20rate%20far%20outpacing%20that%20of%20legitimate%20science.%22%2C%22date%22%3A%222025-08-12%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.1073%5C%2Fpnas.2420092122%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fwww.pnas.org%5C%2Fdoi%5C%2F10.1073%5C%2Fpnas.2420092122%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%22%22%2C%22language%22%3A%22%22%2C%22collections%22%3A%5B%22XE9K2VY5%22%5D%2C%22dateModified%22%3A%222025-09-09T09%3A50%3A08Z%22%7D%7D%2C%7B%22key%22%3A%222JBJELWJ%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Schulthei%5Cu00df%20and%20Lewandowski%22%2C%22parsedDate%22%3A%222020-12-08%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BSchulthei%26%23xDF%3B%2C%20S.%2C%20%26amp%3B%20Lewandowski%2C%20D.%20%282020%29.%20%26%23x201C%3BOutside%20the%20industry%2C%20nobody%20knows%20what%20we%20do%26%23x201D%3B%20SEO%20as%20seen%20by%20search%20engine%20optimizers%20and%20content%20providers.%20%26lt%3Bi%26gt%3BJournal%20of%20Documentation%26lt%3B%5C%2Fi%26gt%3B%2C%20%26lt%3Bi%26gt%3B77%26lt%3B%5C%2Fi%26gt%3B%282%29%2C%20542%26%23x2013%3B557.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-ItemURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1108%5C%2FJD-07-2020-0127%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1108%5C%2FJD-07-2020-0127%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22%5Cu201cOutside%20the%20industry%2C%20nobody%20knows%20what%20we%20do%5Cu201d%20SEO%20as%20seen%20by%20search%20engine%20optimizers%20and%20content%20providers%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Sebastian%22%2C%22lastName%22%3A%22Schulthei%5Cu00df%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Dirk%22%2C%22lastName%22%3A%22Lewandowski%22%7D%5D%2C%22abstractNote%22%3A%22In%20commercial%20web%20search%20engine%20results%20rankings%2C%20four%20stakeholder%20groups%20are%20involved%3A%20search%20engine%20providers%2C%20users%2C%20content%20providers%20and%20search%20engine%20optimizers.%20Search%20engine%20optimization%20%28SEO%29%20is%20a%20multi-billion-dollar%20industry%20and%20responsible%20for%20making%20content%20visible%20through%20search%20engines.%20Despite%20this%20importance%2C%20little%20is%20known%20about%20its%20role%20in%20the%20interaction%20of%20the%20stakeholder%20groups.We%20conducted%20expert%20interviews%20with%2015%20German%20search%20engine%20optimizers%20and%20content%20providers%2C%20the%20latter%20represented%20by%20content%20managers%20and%20online%20journalists.%20The%20interviewees%20were%20asked%20about%20their%20perspectives%20on%20SEO%20and%20how%20they%20assess%20the%20views%20of%20users%20about%20SEO.SEO%20was%20considered%20necessary%20for%20content%20providers%20to%20ensure%20visibility%2C%20which%20is%20why%20dependencies%20between%20both%20stakeholder%20groups%20have%20evolved.%20Despite%20its%20importance%2C%20SEO%20was%20seen%20as%20largely%20unknown%20to%20users.%20Therefore%2C%20it%20is%20assumed%20that%20users%20cannot%20realistically%20assess%20the%20impact%20SEO%20has%20and%20that%20user%20opinions%20about%20SEO%20depend%20heavily%20on%20their%20knowledge%20of%20the%20topic.This%20study%20investigated%20search%20engine%20optimization%20from%20the%20perspective%20of%20those%20involved%20in%20the%20optimization%20business%3A%20content%20providers%2C%20online%20journalists%20and%20search%20engine%20optimization%20professionals.%20The%20study%20therefore%20contributes%20to%20a%20more%20nuanced%20view%20on%20and%20a%20deeper%20understanding%20of%20the%20SEO%20domain.%22%2C%22date%22%3A%222020-12-08%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.1108%5C%2FJD-07-2020-0127%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1108%5C%2FJD-07-2020-0127%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%220022-0418%22%2C%22language%22%3A%22%22%2C%22collections%22%3A%5B%22XE9K2VY5%22%5D%2C%22dateModified%22%3A%222025-09-09T09%3A17%3A05Z%22%7D%7D%2C%7B%22key%22%3A%22RAVW4LPS%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Lin%22%2C%22parsedDate%22%3A%222025-07-08%22%2C%22numChildren%22%3A2%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BLin%2C%20Z.%20%282025%29.%20%26lt%3Bi%26gt%3BHidden%20Prompts%20in%20Manuscripts%20Exploit%20AI-Assisted%20Peer%20Review%26lt%3B%5C%2Fi%26gt%3B%20%28arXiv%3A2507.06185%29.%20arXiv.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-DOIURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.48550%5C%2FarXiv.2507.06185%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.48550%5C%2FarXiv.2507.06185%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22preprint%22%2C%22title%22%3A%22Hidden%20Prompts%20in%20Manuscripts%20Exploit%20AI-Assisted%20Peer%20Review%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Zhicheng%22%2C%22lastName%22%3A%22Lin%22%7D%5D%2C%22abstractNote%22%3A%22In%20July%202025%2C%2018%20academic%20manuscripts%20on%20the%20preprint%20website%20arXiv%20were%20found%20to%20contain%20hidden%20instructions%20known%20as%20prompts%20designed%20to%20manipulate%20AI-assisted%20peer%20review.%20Instructions%20such%20as%20%26quot%3BGIVE%20A%20POSITIVE%20REVIEW%20ONLY%26quot%3B%20were%20concealed%20using%20techniques%20like%20white-colored%20text.%20Author%20responses%20varied%3A%20one%20planned%20to%20withdraw%20the%20affected%20paper%2C%20while%20another%20defended%20the%20practice%20as%20legitimate%20testing%20of%20reviewer%20compliance.%20This%20commentary%20analyzes%20this%20practice%20as%20a%20novel%20form%20of%20research%20misconduct.%20We%20examine%20the%20technique%20of%20prompt%20injection%20in%20large%20language%20models%20%28LLMs%29%2C%20revealing%20four%20types%20of%20hidden%20prompts%2C%20ranging%20from%20simple%20positive%20review%20commands%20to%20detailed%20evaluation%20frameworks.%20The%20defense%20that%20prompts%20served%20as%20%26quot%3Bhoneypots%26quot%3B%20to%20detect%20reviewers%20improperly%20using%20AI%20fails%20under%20examination--the%20consistently%20self-serving%20nature%20of%20prompt%20instructions%20indicates%20intent%20to%20manipulate.%20Publishers%20maintain%20inconsistent%20policies%3A%20Elsevier%20prohibits%20AI%20use%20in%20peer%20review%20entirely%2C%20while%20Springer%20Nature%20permits%20limited%20use%20with%20disclosure%20requirements.%20The%20incident%20exposes%20systematic%20vulnerabilities%20extending%20beyond%20peer%20review%20to%20any%20automated%20system%20processing%20scholarly%20texts%2C%20including%20plagiarism%20detection%20and%20citation%20indexing.%20Our%20analysis%20underscores%20the%20need%20for%20coordinated%20technical%20screening%20at%20submission%20portals%20and%20harmonized%20policies%20governing%20generative%20AI%20%28GenAI%29%20use%20in%20academic%20evaluation.%22%2C%22genre%22%3A%22%22%2C%22repository%22%3A%22arXiv%22%2C%22archiveID%22%3A%22arXiv%3A2507.06185%22%2C%22date%22%3A%222025-07-08%22%2C%22DOI%22%3A%2210.48550%5C%2FarXiv.2507.06185%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22http%3A%5C%2F%5C%2Farxiv.org%5C%2Fabs%5C%2F2507.06185%22%2C%22language%22%3A%22%22%2C%22collections%22%3A%5B%22XE9K2VY5%22%5D%2C%22dateModified%22%3A%222025-09-09T09%3A16%3A31Z%22%7D%7D%5D%7D
Ibrahim, H., Liu, F., Zaki, Y., & Rahwan, T. (2025). Citation manipulation through citation mills and pre-print servers. Scientific Reports, 15(1), 5480. https://doi.org/10.1038/s41598-025-88709-7
Richardson, R. A. K., Hong, S. S., Byrne, J. A., Stoeger, T., & Amaral, L. A. N. (2025). The entities enabling scientific fraud at scale are large, resilient, and growing rapidly. Proceedings of the National Academy of Sciences, 122(32), e2420092122. https://doi.org/10.1073/pnas.2420092122
Schultheiß, S., & Lewandowski, D. (2020). “Outside the industry, nobody knows what we do” SEO as seen by search engine optimizers and content providers. Journal of Documentation, 77(2), 542–557. https://doi.org/10.1108/JD-07-2020-0127
Lin, Z. (2025). Hidden Prompts in Manuscripts Exploit AI-Assisted Peer Review (arXiv:2507.06185). arXiv. https://doi.org/10.48550/arXiv.2507.06185
]]>
Data Readiness for AI: An Essential Foundation for Scientific Innovation https://opensciencelab.uc3m.es/blog/data-readiness-for-ai/ Mon, 22 Sep 2025 09:50:26 +0000 https://opensciencelab.uc3m.es/?p=2604 In the modern age of artificial intelligence (AI), scientific research has entered an era of unprecedented opportunities. From predicting disease outbreaks to advancing quantum physics, AI is transforming the way research is conducted. However, as promising as AI is, it relies heavily on one critical factor: data readiness.

What is Data Readiness for AI and Why It Matters

Data readiness for AI “refers to the process of preparing and ensuring the quality, accessibility, and suitability of datasets before using them for AI applications” . This involves making sure that data is properly collected, clean, structured, and “appropriately annotated, with sufficient metadata to support reliable, appropriate post-model explainability analysis” .

For example, in scientific research, AI-ready data means that the data is organized in a way that algorithms can process efficiently and interpret meaningfully. This is not an easy task. In fact, many scientists spend significant time (often up to 80%) preparing their datasets for AI use .

But why should we worry about data readiness? Because:

  • Data needs to be well-structured, cleaned, and pre-processed before it can be used for training.
  • Investing time and resources upfront to ensure data is ready for AI can save significant costs in the long run.
  • Properly prepared datasets reduce the risks of data bias in model outcomes.
  • Well-annotated data improves transparency and reproducibility.
AI training representation
Ensuring data readiness is critical for successful AI training

What Role Do the FAIR Principles Play in AI Data Readiness?

Incorporating the FAIR principles—Findability, Accessibility, Interoperability, and Reusability— into data management practices provides a solid foundation for ensuring the correct training of AI models and significantly improving their efficiency and effectiveness. This can be done thanks to platforms like the EOSC EU Node, which primarily supports multi-disciplinary and multi-national research, promoting the use of FAIR data and supplementary services across Europe and beyond. Within this environment, researchers can find easy-to-use tools and the necessary support to plan, execute, disseminate, and assess their research workflows and outcomes across the EOSC ecosystem .

From FAIR to FAIR-R

However, for AI models to be highly effective, data must meet additional criteria for machine readability and quality. To address the specific needs of AI, the FAIR-R conceptual framework extends the original FAIR principles by incorporating AI-readiness. It emphasizes that datasets should not only be findable, accessible, interoperable, and reusable, but also structured to meet the quality standards required for AI applications. For example, it ensures that data is appropriately labeled for supervised learning tasks or provides comprehensive and representative coverage for unsupervised learning .

References

5942012 {16784933:Y9JRDMG9},{16784933:6W36BK7U},{16784933:XN7GIGVH},{16784933:5WG43CHH},{16784933:9P2643VY} 1 apa 50 creator desc 2604 https://opensciencelab.uc3m.es/wp-content/plugins/zotpress/
%7B%22status%22%3A%22success%22%2C%22updateneeded%22%3Afalse%2C%22instance%22%3Afalse%2C%22meta%22%3A%7B%22request_last%22%3A0%2C%22request_next%22%3A0%2C%22used_cache%22%3Atrue%7D%2C%22data%22%3A%5B%7B%22key%22%3A%22XN7GIGVH%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3B%26lt%3Bi%26gt%3BEOSC%20EU%20Node%20%7C%20European%20Open%20Science%20Cloud%20-%20EU%20Node%26lt%3B%5C%2Fi%26gt%3B.%20%28n.d.%29.%20Retrieved%20September%2022%2C%202025%2C%20from%20%26lt%3Ba%20class%3D%26%23039%3Bzp-ItemURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fopen-science-cloud.ec.europa.eu%5C%2Fabout%5C%2Feosc-eu-node%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fopen-science-cloud.ec.europa.eu%5C%2Fabout%5C%2Feosc-eu-node%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22webpage%22%2C%22title%22%3A%22EOSC%20EU%20Node%20%7C%20European%20Open%20Science%20Cloud%20-%20EU%20Node%22%2C%22creators%22%3A%5B%5D%2C%22abstractNote%22%3A%22%22%2C%22date%22%3A%22%22%2C%22DOI%22%3A%22%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fopen-science-cloud.ec.europa.eu%5C%2Fabout%5C%2Feosc-eu-node%22%2C%22language%22%3A%22%22%2C%22collections%22%3A%5B%225QC7GHMJ%22%5D%2C%22dateModified%22%3A%222025-09-22T08%3A51%3A12Z%22%7D%7D%2C%7B%22key%22%3A%22Y9JRDMG9%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Wilkinson%20et%20al.%22%2C%22parsedDate%22%3A%222016-03-15%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BWilkinson%2C%20M.%20D.%2C%20Dumontier%2C%20M.%2C%20Aalbersberg%2C%20Ij.%20J.%2C%20Appleton%2C%20G.%2C%20Axton%2C%20M.%2C%20Baak%2C%20A.%2C%20Blomberg%2C%20N.%2C%20Boiten%2C%20J.-W.%2C%20da%20Silva%20Santos%2C%20L.%20B.%2C%20Bourne%2C%20P.%20E.%2C%20Bouwman%2C%20J.%2C%20Brookes%2C%20A.%20J.%2C%20Clark%2C%20T.%2C%20Crosas%2C%20M.%2C%20Dillo%2C%20I.%2C%20Dumon%2C%20O.%2C%20Edmunds%2C%20S.%2C%20Evelo%2C%20C.%20T.%2C%20Finkers%2C%20R.%2C%20%26%23x2026%3B%20Mons%2C%20B.%20%282016%29.%20The%20FAIR%20Guiding%20Principles%20for%20scientific%20data%20management%20and%20stewardship.%20%26lt%3Bi%26gt%3BScientific%20Data%26lt%3B%5C%2Fi%26gt%3B%2C%20%26lt%3Bi%26gt%3B3%26lt%3B%5C%2Fi%26gt%3B%281%29%2C%20160018.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-DOIURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1038%5C%2Fsdata.2016.18%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1038%5C%2Fsdata.2016.18%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22The%20FAIR%20Guiding%20Principles%20for%20scientific%20data%20management%20and%20stewardship%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Mark%20D.%22%2C%22lastName%22%3A%22Wilkinson%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Michel%22%2C%22lastName%22%3A%22Dumontier%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22IJsbrand%20Jan%22%2C%22lastName%22%3A%22Aalbersberg%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Gabrielle%22%2C%22lastName%22%3A%22Appleton%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Myles%22%2C%22lastName%22%3A%22Axton%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Arie%22%2C%22lastName%22%3A%22Baak%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Niklas%22%2C%22lastName%22%3A%22Blomberg%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jan-Willem%22%2C%22lastName%22%3A%22Boiten%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Luiz%20Bonino%22%2C%22lastName%22%3A%22da%20Silva%20Santos%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Philip%20E.%22%2C%22lastName%22%3A%22Bourne%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jildau%22%2C%22lastName%22%3A%22Bouwman%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Anthony%20J.%22%2C%22lastName%22%3A%22Brookes%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Tim%22%2C%22lastName%22%3A%22Clark%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Merc%5Cu00e8%22%2C%22lastName%22%3A%22Crosas%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Ingrid%22%2C%22lastName%22%3A%22Dillo%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Olivier%22%2C%22lastName%22%3A%22Dumon%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Scott%22%2C%22lastName%22%3A%22Edmunds%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Chris%20T.%22%2C%22lastName%22%3A%22Evelo%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Richard%22%2C%22lastName%22%3A%22Finkers%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Alejandra%22%2C%22lastName%22%3A%22Gonzalez-Beltran%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Alasdair%20J.%20G.%22%2C%22lastName%22%3A%22Gray%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Paul%22%2C%22lastName%22%3A%22Groth%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Carole%22%2C%22lastName%22%3A%22Goble%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jeffrey%20S.%22%2C%22lastName%22%3A%22Grethe%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jaap%22%2C%22lastName%22%3A%22Heringa%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Peter%20A.%20C.%22%2C%22lastName%22%3A%22%5Cu2019t%20Hoen%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Rob%22%2C%22lastName%22%3A%22Hooft%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Tobias%22%2C%22lastName%22%3A%22Kuhn%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Ruben%22%2C%22lastName%22%3A%22Kok%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Joost%22%2C%22lastName%22%3A%22Kok%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Scott%20J.%22%2C%22lastName%22%3A%22Lusher%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Maryann%20E.%22%2C%22lastName%22%3A%22Martone%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Albert%22%2C%22lastName%22%3A%22Mons%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Abel%20L.%22%2C%22lastName%22%3A%22Packer%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Bengt%22%2C%22lastName%22%3A%22Persson%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Philippe%22%2C%22lastName%22%3A%22Rocca-Serra%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Marco%22%2C%22lastName%22%3A%22Roos%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Rene%22%2C%22lastName%22%3A%22van%20Schaik%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Susanna-Assunta%22%2C%22lastName%22%3A%22Sansone%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Erik%22%2C%22lastName%22%3A%22Schultes%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Thierry%22%2C%22lastName%22%3A%22Sengstag%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Ted%22%2C%22lastName%22%3A%22Slater%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22George%22%2C%22lastName%22%3A%22Strawn%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Morris%20A.%22%2C%22lastName%22%3A%22Swertz%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Mark%22%2C%22lastName%22%3A%22Thompson%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Johan%22%2C%22lastName%22%3A%22van%20der%20Lei%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Erik%22%2C%22lastName%22%3A%22van%20Mulligen%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jan%22%2C%22lastName%22%3A%22Velterop%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Andra%22%2C%22lastName%22%3A%22Waagmeester%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Peter%22%2C%22lastName%22%3A%22Wittenburg%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Katherine%22%2C%22lastName%22%3A%22Wolstencroft%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jun%22%2C%22lastName%22%3A%22Zhao%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Barend%22%2C%22lastName%22%3A%22Mons%22%7D%5D%2C%22abstractNote%22%3A%22There%20is%20an%20urgent%20need%20to%20improve%20the%20infrastructure%20supporting%20the%20reuse%20of%20scholarly%20data.%20A%20diverse%20set%20of%20stakeholders%5Cu2014representing%20academia%2C%20industry%2C%20funding%20agencies%2C%20and%20scholarly%20publishers%5Cu2014have%20come%20together%20to%20design%20and%20jointly%20endorse%20a%20concise%20and%20measureable%20set%20of%20principles%20that%20we%20refer%20to%20as%20the%20FAIR%20Data%20Principles.%20The%20intent%20is%20that%20these%20may%20act%20as%20a%20guideline%20for%20those%20wishing%20to%20enhance%20the%20reusability%20of%20their%20data%20holdings.%20Distinct%20from%20peer%20initiatives%20that%20focus%20on%20the%20human%20scholar%2C%20the%20FAIR%20Principles%20put%20specific%20emphasis%20on%20enhancing%20the%20ability%20of%20machines%20to%20automatically%20find%20and%20use%20the%20data%2C%20in%20addition%20to%20supporting%20its%20reuse%20by%20individuals.%20This%20Comment%20is%20the%20first%20formal%20publication%20of%20the%20FAIR%20Principles%2C%20and%20includes%20the%20rationale%20behind%20them%2C%20and%20some%20exemplar%20implementations%20in%20the%20community.%22%2C%22date%22%3A%222016-03-15%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.1038%5C%2Fsdata.2016.18%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fwww.nature.com%5C%2Farticles%5C%2Fsdata201618%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%222052-4463%22%2C%22language%22%3A%22en%22%2C%22collections%22%3A%5B%225QC7GHMJ%22%5D%2C%22dateModified%22%3A%222025-09-22T08%3A49%3A56Z%22%7D%7D%2C%7B%22key%22%3A%226W36BK7U%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Verhulst%20et%20al.%22%2C%22parsedDate%22%3A%222025-03-04%22%2C%22numChildren%22%3A0%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BVerhulst%2C%20S.%2C%20Zahuranec%2C%20A.%20J.%2C%20%26amp%3B%20Chafetz%2C%20H.%20%282025%29.%20%26lt%3Bi%26gt%3BMoving%20Toward%20the%20FAIR-R%20principles%3A%20Advancing%20AI-Ready%20Data%26lt%3B%5C%2Fi%26gt%3B%20%28SSRN%20Scholarly%20Paper%20No.%205164337%29.%20Social%20Science%20Research%20Network.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-DOIURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.2139%5C%2Fssrn.5164337%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.2139%5C%2Fssrn.5164337%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22preprint%22%2C%22title%22%3A%22Moving%20Toward%20the%20FAIR-R%20principles%3A%20Advancing%20AI-Ready%20Data%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Stefaan%22%2C%22lastName%22%3A%22Verhulst%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Andrew%20J.%22%2C%22lastName%22%3A%22Zahuranec%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Hannah%22%2C%22lastName%22%3A%22Chafetz%22%7D%5D%2C%22abstractNote%22%3A%22In%20today%26%23039%3Bs%20rapidly%20evolving%20AI%20ecosystem%2C%20making%20data%20ready%20for%20AI-optimized%20for%20training%2C%20fine-tuning%2C%20and%20augmentation-is%20more%20critical%20than%20ever.%20While%20the%20FAIR%20principles%20%28Findability%2C%20Accessibility%2C%20Interoperability%2C%20and%20Reusability%29%20have%20guided%20data%20management%20and%20open%20science%2C%20they%20do%20not%20inherently%20address%20AI-specific%20needs.%20Expanding%20FAIR%20to%20FAIR-R%2C%20incorporating%20Readiness%20for%20AI%2C%20could%20accelerate%20the%20responsible%20use%20of%20open%20data%20in%20AI%20applications%20that%20serve%20the%20public%20interest.%20This%20paper%20introduces%20the%20FAIR-R%20framework%20and%20identifies%20current%20efforts%20for%20enhancing%20AI-ready%20data%20through%20improved%20data%20labeling%2C%20provenance%20tracking%2C%20and%20new%20data%20standards.%20However%2C%20key%20challenges%20remain%3A%20How%20can%20data%20be%20structured%20for%20AI%20without%20compromising%20ethics%3F%20What%20governance%20models%20ensure%20equitable%20access%3F%20How%20can%20AI%20itself%20be%20leveraged%20to%20improve%20data%20quality%3F%20Answering%20these%20questions%20is%20essential%20for%20unlocking%20the%20full%20potential%20of%20AI-driven%20innovation%20while%20ensuring%20responsible%20and%20transparent%20data%20use.%22%2C%22genre%22%3A%22SSRN%20Scholarly%20Paper%22%2C%22repository%22%3A%22Social%20Science%20Research%20Network%22%2C%22archiveID%22%3A%225164337%22%2C%22date%22%3A%222025-03-04%22%2C%22DOI%22%3A%2210.2139%5C%2Fssrn.5164337%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fpapers.ssrn.com%5C%2Fabstract%3D5164337%22%2C%22language%22%3A%22en%22%2C%22collections%22%3A%5B%225QC7GHMJ%22%5D%2C%22dateModified%22%3A%222025-09-22T08%3A52%3A18Z%22%7D%7D%2C%7B%22key%22%3A%229P2643VY%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Natalia%20et%20al.%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BNatalia%2C%20N.%2C%20Denisse%2C%20A.%2C%20%26amp%3B%20Vanessa%2C%20M.%20%28n.d.%29.%20Data%20and%20AI%20for%20science%3A%20Key%20considerations.%20%26lt%3Bi%26gt%3BInternational%20Science%20Council%26lt%3B%5C%2Fi%26gt%3B.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-ItemURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.24948%5C%2F2025.11%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.24948%5C%2F2025.11%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22Data%20and%20AI%20for%20science%3A%20Key%20considerations%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Norori%22%2C%22lastName%22%3A%22Natalia%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Albornoz%22%2C%22lastName%22%3A%22Denisse%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22McBride%22%2C%22lastName%22%3A%22Vanessa%22%7D%5D%2C%22abstractNote%22%3A%22This%20paper%20provides%20an%20overview%20of%20the%20technical%2C%20ethical%20and%20environmental%20factors%20to%20consider%20when%20preparing%20scientific%20data%20for%20artificial%20intelligence%20%28AI%29%2C%20and%20how%20these%20factors%20align%20with%20the%20%5Cu2018Open%20Science%5Cu2019%20movement.%20The%20information%20presented%20is%20relevant%20to%20researchers%2C%20data%20practitioners%2C%20scientific%20bodies%20and%20policy-makers%20for%20science.%22%2C%22date%22%3A%22%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.24948%5C%2F2025.11%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fdoi.org%5C%2F10.24948%5C%2F2025.11%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%22%22%2C%22language%22%3A%22en-US%22%2C%22collections%22%3A%5B%225QC7GHMJ%22%5D%2C%22dateModified%22%3A%222025-09-22T08%3A47%3A13Z%22%7D%7D%2C%7B%22key%22%3A%225WG43CHH%22%2C%22library%22%3A%7B%22id%22%3A16784933%7D%2C%22meta%22%3A%7B%22creatorSummary%22%3A%22Hiniduma%20et%20al.%22%2C%22parsedDate%22%3A%222025%22%2C%22numChildren%22%3A1%7D%2C%22bib%22%3A%22%26lt%3Bdiv%20class%3D%26quot%3Bcsl-bib-body%26quot%3B%20style%3D%26quot%3Bline-height%3A%202%3B%20padding-left%3A%201em%3B%20text-indent%3A-1em%3B%26quot%3B%26gt%3B%5Cn%20%20%26lt%3Bdiv%20class%3D%26quot%3Bcsl-entry%26quot%3B%26gt%3BHiniduma%2C%20K.%2C%20Byna%2C%20S.%2C%20%26amp%3B%20Bez%2C%20J.%20L.%20%282025%29.%20Data%20Readiness%20for%20AI%3A%20A%20360-Degree%20Survey.%20%26lt%3Bi%26gt%3BACM%20Comput.%20Surv.%26lt%3B%5C%2Fi%26gt%3B%2C%20%26lt%3Bi%26gt%3B57%26lt%3B%5C%2Fi%26gt%3B%289%29%2C%20219%3A1-219%3A39.%20%26lt%3Ba%20class%3D%26%23039%3Bzp-DOIURL%26%23039%3B%20href%3D%26%23039%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1145%5C%2F3722214%26%23039%3B%26gt%3Bhttps%3A%5C%2F%5C%2Fdoi.org%5C%2F10.1145%5C%2F3722214%26lt%3B%5C%2Fa%26gt%3B%26lt%3B%5C%2Fdiv%26gt%3B%5Cn%26lt%3B%5C%2Fdiv%26gt%3B%22%2C%22data%22%3A%7B%22itemType%22%3A%22journalArticle%22%2C%22title%22%3A%22Data%20Readiness%20for%20AI%3A%20A%20360-Degree%20Survey%22%2C%22creators%22%3A%5B%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Kaveen%22%2C%22lastName%22%3A%22Hiniduma%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Suren%22%2C%22lastName%22%3A%22Byna%22%7D%2C%7B%22creatorType%22%3A%22author%22%2C%22firstName%22%3A%22Jean%20Luca%22%2C%22lastName%22%3A%22Bez%22%7D%5D%2C%22abstractNote%22%3A%22Artificial%20Intelligence%20%28AI%29%20applications%20critically%20depend%20on%20data.%20Poor-quality%20data%20produces%20inaccurate%20and%20ineffective%20AI%20models%20that%20may%20lead%20to%20incorrect%20or%20unsafe%20use.%20Evaluation%20of%20data%20readiness%20is%20a%20crucial%20step%20in%20improving%20the%20quality%20and%20appropriateness%20of%20data%20usage%20for%20AI.%20R%26amp%3BD%20efforts%20have%20been%20spent%20on%20improving%20data%20quality.%20However%2C%20standardized%20metrics%20for%20evaluating%20data%20readiness%20for%20use%20in%20AI%20training%20are%20still%20evolving.%20In%20this%20study%2C%20we%20perform%20a%20comprehensive%20survey%20of%20metrics%20used%20to%20verify%20data%20readiness%20for%20AI%20training.%20This%20survey%20examines%20more%20than%20140%20papers%20published%20by%20ACM%20Digital%20Library%2C%20IEEE%20Xplore%2C%20journals%20such%20as%20Nature%2C%20Springer%2C%20and%20Science%20Direct%2C%20and%20online%20articles%20published%20by%20prominent%20AI%20experts.%20This%20survey%20aims%20to%20propose%20a%20taxonomy%20of%20data%20readiness%20for%20AI%20%28DRAI%29%20metrics%20for%20structured%20and%20unstructured%20datasets.%20We%20anticipate%20that%20this%20taxonomy%20will%20lead%20to%20new%20standards%20for%20DRAI%20metrics%20that%20would%20be%20used%20for%20enhancing%20the%20quality%2C%20accuracy%2C%20and%20fairness%20of%20AI%20training%20and%20inference.%22%2C%22date%22%3A%22Abril%204%2C%202025%22%2C%22section%22%3A%22%22%2C%22partNumber%22%3A%22%22%2C%22partTitle%22%3A%22%22%2C%22DOI%22%3A%2210.1145%5C%2F3722214%22%2C%22citationKey%22%3A%22%22%2C%22url%22%3A%22https%3A%5C%2F%5C%2Fdl.acm.org%5C%2Fdoi%5C%2F10.1145%5C%2F3722214%22%2C%22PMID%22%3A%22%22%2C%22PMCID%22%3A%22%22%2C%22ISSN%22%3A%220360-0300%22%2C%22language%22%3A%22%22%2C%22collections%22%3A%5B%225QC7GHMJ%22%5D%2C%22dateModified%22%3A%222025-09-22T08%3A39%3A39Z%22%7D%7D%5D%7D
EOSC EU Node | European Open Science Cloud - EU Node. (n.d.). Retrieved September 22, 2025, from https://open-science-cloud.ec.europa.eu/about/eosc-eu-node
Wilkinson, M. D., Dumontier, M., Aalbersberg, Ij. J., Appleton, G., Axton, M., Baak, A., Blomberg, N., Boiten, J.-W., da Silva Santos, L. B., Bourne, P. E., Bouwman, J., Brookes, A. J., Clark, T., Crosas, M., Dillo, I., Dumon, O., Edmunds, S., Evelo, C. T., Finkers, R., … Mons, B. (2016). The FAIR Guiding Principles for scientific data management and stewardship. Scientific Data, 3(1), 160018. https://doi.org/10.1038/sdata.2016.18
Verhulst, S., Zahuranec, A. J., & Chafetz, H. (2025). Moving Toward the FAIR-R principles: Advancing AI-Ready Data (SSRN Scholarly Paper No. 5164337). Social Science Research Network. https://doi.org/10.2139/ssrn.5164337
Natalia, N., Denisse, A., & Vanessa, M. (n.d.). Data and AI for science: Key considerations. International Science Council. https://doi.org/10.24948/2025.11
Hiniduma, K., Byna, S., & Bez, J. L. (2025). Data Readiness for AI: A 360-Degree Survey. ACM Comput. Surv., 57(9), 219:1-219:39. https://doi.org/10.1145/3722214
]]>
What is a Predatory Journal? Understanding the Issue Behind the Label https://opensciencelab.uc3m.es/blog/reframing-the-concept-predatory-journal/ Mon, 01 Sep 2025 10:13:31 +0000 https://opensciencelab.uc3m.es/?p=2537 In the ever-evolving world of academic publishing, the term “predatory” has become widely used to describe a certain class of journals and publishers. While it may seem like an apt description, there are several reasons why we believe this label is not only problematic but also misleading.

What Does the Term “Predatory” Mean?

From a biological perspective, predation refers to a scenario where one organism, the predator, kills and consumes another organism for survival, the prey, which becomes the depredated. This concept is also applied more broadly to emphasize the exploitative dynamics of certain relationships, for example, to academic journals. However, we see a mismatch, since the label predatory is currently associated with concepts such as:

  • High acceptance rates and fast-track publishing
  • Lack of a rigorous peer review process
  • Articles that are published even when peer reviewers recommend rejection
  • Frequent spelling and grammatical errors compared to legitimate journals
  • Spam solicitations to researchers, urging them to submit manuscripts, sometimes via email or other platforms
  • No retraction policy
  • Poor web design
  • Recognition on “predatory journals” lists like https://www.predatoryjournals.org
  • Not indexed in major databases

For us, a Predatory Journal Is Simply…

A journal that seeks profit by charging APCs (Article Processing Charges) to authors or institutions, preying on them as the depredated victims of the system.

When financial gain becomes the priority, the value of academic work takes a backseat.

It is not necessarily about fraud or deceit in the research itself. Some journals not labeled as “predatory” that theoretically publish high-quality content, charge abusive APCs, fitting the classic definition of the term “predatory,” derived from the biological concept discussed earlier. For us, what truly makes a journal predatory is taking advantage of institutions and scientists through these APCs, and not necessarily meeting the previous criteria. With this alone, a journal is already predatory.

This perspective leads us to believe that many journals, including well-known ones like Elsevier, which charge high fees for open access and also retain copyright control unless authors pay, should also be considered predatory. The fundamental issue isn’t the quality of the research but rather the business model that prioritizes revenue over academic integrity.

A cartoon-style illustration featuring a wolf with a sly grin, holding a 'Publish here' sign, while a thoughtful sheep in a lab coat and glasses looks on, creating a humorous contrast.
Image generated with ChatGPT 5

Questionable Research Practices vs. Predatory Models

However, it is crucial to note that some journals go beyond being simply predatory and also engage in questionable research practices. As we saw with the previous list, these practices can include publishing research without proper peer review, manipulating citations to inflate impact factors, or even publishing fraudulent or plagiarized content. In such cases, the issue is not limited to financial exploitation, but extends to the integrity of the research itself. For this type of behavior, we need to create a distinct term, as it involves both financial exploitation and compromised academic standards.

This separation allows for a clearer understanding: predatory journals are profit-driven but not necessarily fraudulent in their practices, whereas journals that exhibit questionable research practices put both authors and the scientific community at risk by publishing unverified, low-quality, or even plagiarized content.

Infographic about Predatory Journals VS. Questionable Research Practices

]]>