Transforming Science with Large Language Models: A Survey on AI-assisted Scientific Discovery, Experimentation, Content Generation, and Evaluation

This survey provides a comprehensive overview of the emerging ecosystem of large language models and tools that support researchers across the scientific lifecycle, covering key tasks from literature search and idea generation to content creation, experimentation, and evaluation, while addressing associated datasets, methods, limitations, and ethical concerns.

Steffen Eger, Yong Cao, Jennifer D'Souza, Andreas Geiger, Christian Greisinger, Stephanie Gross, Yufang Hou, Brigitte Krenn, Anne Lauscher, Yizhi Li, Chenghua Lin, Nafise Sadat Moosavi, Wei Zhao, Tristan Miller

Published 2026-03-09
📖 5 min read🧠 Deep dive

Imagine the world of science as a massive, bustling library that is growing so fast it's about to burst its shelves. For centuries, scientists have been the librarians and explorers of this library, doing all the heavy lifting: finding books, reading them, figuring out what's missing, building experiments, and writing reports.

Now, a new kind of super-intelligent assistant has arrived: Large Language Models (LLMs). Think of these as a team of incredibly fast, well-read, but sometimes overly confident robots. This paper is a giant "user manual" and "safety guide" for how we can use these robots to help scientists, while being very careful not to let them run the show.

Here is a breakdown of how these AI assistants are changing the scientific game, using simple analogies:

1. The Super-Librarian (Literature Search)

The Old Way: A scientist spends weeks digging through piles of papers, trying to find the one sentence that answers their question. It's like looking for a needle in a haystack made of other needles.
The AI Way: The AI is a Super-Librarian. You can ask it, "What do we know about curing this specific type of cancer?" and it instantly scans millions of books, summarizes the key points, and hands you a neat report.

  • The Catch: Sometimes the librarian gets a bit confused and invents a book that doesn't exist (called a "hallucination"). So, you still have to double-check the shelf.

2. The Creative Brainstormer (Ideas & Hypotheses)

The Old Way: A scientist sits in a quiet room, staring at a wall, trying to come up with a new idea. It's like trying to invent a new flavor of ice cream from scratch.
The AI Way: The AI is a Creative Brainstormer. It reads everything ever written about ice cream and suggests, "What if we mix mint with spicy chili?" It can generate hundreds of wild ideas in seconds.

  • The Catch: While it's great at being creative, it's not always practical. It might suggest an idea that sounds cool but is impossible to build or test. It needs a human to say, "Okay, that's a fun idea, but let's try this one instead."

3. The Speed-Writer (Writing Papers)

The Old Way: Writing a scientific paper is like building a house brick by brick. You have to write the title, the abstract, the introduction, and the conclusion, making sure every fact is perfect.
The AI Way: The AI is a Speed-Writer. It can draft the title, summarize the abstract, and even write the "Related Work" section (which is like listing who built houses before you).

  • The Catch: The AI is great at grammar and style, but it's terrible at facts. It might write a beautiful paragraph about a bridge that collapses because it forgot to mention the steel beams. Also, it can't be the "author" of the house; the human scientist must take responsibility for the blueprint.

4. The Artist (Figures, Tables, and Slides)

The Old Way: Turning data into a chart or a slide deck is like painting a picture by hand. It takes time and skill to make sure the colors match the data.
The AI Way: The AI is a Digital Artist. You give it a description like "Show me a graph of rising temperatures," and it instantly paints the picture, draws the chart, or even designs a presentation slide.

  • The Catch: Sometimes the artist gets the colors wrong. The graph might show a line going up when the data says it's going down. You need a human to look at the painting and say, "Wait, that's not right."

5. The Critic (Peer Review)

The Old Way: Before a paper is published, other scientists (peers) read it to check for mistakes. It's like a food critic tasting a dish before it goes to the customers.
The AI Way: The AI is a Junior Critic. It can read the paper quickly, check for grammar, and even spot if a claim sounds suspicious. It helps the human critics by doing the boring first pass.

  • The Catch: The AI doesn't have "common sense" or deep intuition. It might miss a subtle lie or a brilliant new idea that breaks the rules. We can't let the robot be the only judge; a human must have the final say.

The Big Warning: The "Fake News" Problem

The paper sounds a lot of alarms about Ethics.

  • The "Hallucination" Risk: The AI is like a storyteller who loves to make things up to make the story sound better. In science, making things up is dangerous.
  • The "Plagiarism" Risk: If the AI writes the paper, who owns it? The paper says the human must always be the boss. The AI is just the tool, like a hammer. You can't say the hammer built the house.
  • The "Bias" Risk: If the AI was trained on books written mostly by men from one country, it might think that's the only way to do science. We need to make sure the AI sees the whole library, not just one corner.

The Bottom Line

This paper is essentially saying: "AI is a fantastic co-pilot, but it cannot be the pilot."

Science is moving from a time where humans did everything alone, to a time where humans and robots work together. The robot handles the heavy lifting, the speed, and the data crunching. The human provides the judgment, the ethics, and the "spark" of true discovery. If we use these tools wisely, we can solve big problems faster. If we let the robot drive without a human in the seat, we might crash into a wall of fake science.