The Rise of AI Scientists: A New Era for Science?

The rise of AI scientists has the potential to revolutionize the field of science, but it also raises important questions about the role of human scientists and the integrity of the scientific process.
The Rise of AI Scientists: A New Era for Science?
Photo by Luca Bravo on Unsplash

Can AI Scientists Revolutionize the Field of Science?

The concept of AI scientists has been gaining traction in recent years, with many experts believing that artificial intelligence can revolutionize the field of science. One company, Sakana AI Labs, has taken this idea to the next level by creating an AI system that can make scientific discoveries in the area of machine learning in a fully automated way.

How a Computer Can ‘Do Science’

A lot of science is done in the open, and almost all scientific knowledge has been written down somewhere. Millions of scientific papers are freely available online in repositories such as arXiv and PubMed. Generative large language models (LLMs) trained with this data capture the language of science and its patterns. It is therefore perhaps not at all surprising that a generative LLM can produce something that looks like a good scientific paper – it has ingested many examples that it can copy.

Artificial intelligence

However, what is less clear is whether an AI system can produce an interesting scientific paper. Crucially, good science requires novelty. Scientists don’t want to be told about things that are already known. Rather, they want to learn new things, especially new things that are significantly different from what is already known. This requires judgement about the scope and value of a contribution.

The Sakana System

The Sakana system tries to address interestingness in two ways. First, it “scores” new paper ideas for similarity to existing research (indexed in the Semantic Scholar repository). Anything too similar is discarded.

Second, Sakana’s system introduces a “peer review” step – using another LLM to judge the quality and novelty of the generated paper. Here again, there are plenty of examples of peer review online on sites such as openreview.net that can guide how to critique a paper. LLMs have ingested these, too.

“The ability of LLMs to judge the quality of research is also an open question. My own work (soon to be published in Research Synthesis Methods) shows LLMs are not great at judging the risk of bias in medical research studies, though this too may improve over time.”

AI May Be a Poor Judge of AI Output

Feedback is mixed on Sakana AI’s output. Some have described it as producing “endless scientific slop”. Even the system’s own review of its outputs judges the papers weak at best. This is likely to improve as the technology evolves, but the question of whether automated scientific papers are valuable remains.

AI Tools to Support Scientists, Not Replace Them

AI researchers have been developing systems to support science for decades. Given the huge volumes of published research, even finding publications relevant to a specific scientific question can be challenging. Specialised search tools make use of AI to help scientists find and synthesise existing work. These include the above-mentioned Semantic Scholar, but also newer systems such as Elicit, Research Rabbit, scite and Consensus.

Scientific research

Text mining tools such as PubTator dig deeper into papers to identify key points of focus, such as specific genetic mutations and diseases, and their established relationships. This is especially useful for curating and organising scientific information.

Machine learning has also been used to support the synthesis and analysis of medical evidence, in tools such as Robot Reviewer. Summaries that compare and contrast claims in papers from Scholarcy help to perform literature reviews. All these tools aim to help scientists do their jobs more effectively, not to replace them.

AI Research May Exacerbate Existing Problems

While Sakana AI states it doesn’t see the role of human scientists diminishing, the company’s vision of “a fully AI-driven scientific ecosystem” would have major implications for science.

One concern is that, if AI-generated papers flood the scientific literature, future AI systems may be trained on AI output and undergo model collapse. This means they may become increasingly ineffectual at innovating.

However, the implications for science go well beyond impacts on AI science systems themselves. There are already bad actors in science, including “paper mills” churning out fake papers. This problem will only get worse when a scientific paper can be produced with US$15 and a vague initial prompt.

Artificial intelligence in science

The need to check for errors in a mountain of automatically generated research could rapidly overwhelm the capacity of actual scientists. The peer review system is arguably already broken, and dumping more research of questionable quality into the system won’t fix it.

Science is fundamentally based on trust. Scientists emphasise the integrity of the scientific process so we can be confident our understanding of the world (and now, the world’s machines) is valid and improving. A scientific ecosystem where AI systems are key players raises fundamental questions about the meaning and value of this process, and what level of trust we should have in AI scientists. Is this the kind of scientific ecosystem we want?