30 Comments
User's avatar
Michael House's avatar

What will slow research isn’t AI. It’s the flood of preprints being treated like peer-reviewed work across AI and computer science. Right now, an undergrad with a Canva poster and a faculty sponsor can push out ten preprints in a semester and get them cited like they’ve reshaped the field. OSF allows researchers to delete preregistrations, which sounds harmless until it’s used to quietly erase bad or fraudulent work. If something gets flagged, it’s gone. No history, no accountability. That’s a perfect setup for bad actors.

And we still haven’t dealt with the reproducibility crisis. We didn’t fix it. We just buried it under buzzwords, hype, and career incentives. Simultaneously, we are using completely broken scientific metaphors to justify AI architectures. We’re still pretending spiking neurons are equivalent to RNNs. That synaptic noise is optimization. That the behavior of starving mice tells us how humans think. These comparisons aren’t science. They’re branding.

Research architectures are more expensive, more power-hungry, and more opaque than ever. Despite the lack of a clear path to profitability, AI continues to consume billions of dollars in funding. The hype keeps growing. Amplified work often prioritizes speed, clout, and marketability over real understanding.

AI isn't a threat to science. The hype is. The culture is around it. The people enabling it are.

Expand full comment
David A. Westbrook's avatar

These are all points with which I agree, but they rest on an idealization very common among (and flattering to) scientists. AI is treated as without social/economic/historical context, as is often done with weapons, and of course with discredited ideas, subsequently relegated to the status of "bad science." But the science we have is the science we have, though we might hope for better. There is no way to fund AI's development (training data, compute, etc.) without a degree of "hype" the "culture" and "people enabling." That said, it does seem we could do better on all fronts, have indeed done better with at least some prior transformative technologies. It's a really interesting question of contemporary intellectual history.

Expand full comment
Maah Pishanih's avatar

If we take a step further and come to the most important factor shaping culture, we will encounter the economic structures we establish. And now, back to Marx.

Expand full comment
David A. Westbrook's avatar

Yes, political economy is, roughly speaking, what I do, albeit not in particularly Marxist fashion. But the CS is discourse is astonishingly devoid of context, incentives, and the like. It's just speculation on what the technology might do, and how that might work out,. which are of course well-nigh irresistibly interesting.

Expand full comment
Renaud Gaudron's avatar

I share your frustration about preprints. I suspect that many people do not even know the difference between a preprint and an actual, peer-reviewed paper.

Expand full comment
Thomas DeWitt's avatar

This is a fantastic essay. I do think it is worth emphasizing (as you do point out) that there is a big difference between AI that outputs a solution and AI that outputs an explanation.

But as you point out, to favor explanation-producing-AI, the incentives need to be changed for that to become dominant. We should consider how this might be changed.

I also wrote something arguing explanation-producing and solution-producing AI are different in important ways: https://thomasddewitt.substack.com/p/the-new-ai-for-science-is-different

Expand full comment
Leif Hancox-Li's avatar

See this paper claiming that the results from Park et al are due to a bug in plotting. https://arxiv.org/abs/2402.14583

Expand full comment
Arvind Narayanan's avatar

Yes, this is one of many critiques of the paper. We acknowledge them in the essay and link to a discussion of the issues. The authors' response is broadly convincing to us, but nonetheless the finding should definitely be treated as preliminary, which is why we present five different lines of evidence, all seeming to point to a slowdown.

Expand full comment
Alex Tolley's avatar

One point that was not mentioned, but clearly affects the production process, is how NSF funding has changed. IDK if it has changed, but for years there were complaints that the NSF funded far to few "blue-sky" ideas and preferred to fund science experiments whose results were expected to confirm a hypothesis, and therefore deemed successful. Whether this applied to other funding agencies in the US and other countries, I have no knowledge, but if it were common it would help support your explanation.

In my experience, if a new technique proves useful, then the literature would subsequently be filled with experiments using this technique. During this period, other techniques to serve a similar role would be ignored. This is a phenomenon that happens in many endeavors, creating "fads" that last for a while, until progress slows, and new techniques and instruments are developed that work better and result in better explanations.

Expand full comment
WD Lindberg's avatar

Not just NSF funding. A great deal of research funding is commercial. Marketing research, how to sell better, how to advertise better, how to run a process incrementally better, etc. In other words it is applied research and is expected to develop a return. Large corporation research departments used to fund blue sky topics (GE, IBM, et al). This only occurs by accident now. How many of those papers in the above chart are just churn (e.g papers with little or no content but rather just an object to add to ones published count)? The applied R&D I'm familiar with, was only partially effective, by the time the conclusions were issued, everyone realized the question had been misunderstood. This was primarily due to inadequate liaison between the question asker and the researcher. So more papers written but little or no progress.

Expand full comment
Michael Spencer's avatar

Funding to National Defense and Data centers subsidies might actually inhibit funding to science. The hope that AI agents might accelerate scientific breakthroughs seems fairly unlikely to actually materialize. Overconfidence in AI especially American AI also leads to huge wastes of capital chasing the wrong approaches to real innovation and R&D.

Expand full comment
peter snowdon's avatar

In many fields I am familiar with (translation, document drafting...), wide adoption of AI seems to go hand in hand with rising thresholds of fault tolerance. Film distributors are no longer afraid to release films with inaccurate/unidiomatic subtitles, if the cost of producing those subtitles is a fraction of employing a human to write them (or even just to check them). This culture of use seems to undercut - to some extent - your claim that commercial AI has incentives to troubleshoot and preempt/prevent errors that AI in science does not (yet) have.

I wonder also how far a general culture of higher tolerance for lower quality/less reliability in all fields might affect social and political attitudes towards science - and science funding - in the longer term? Perhaps science's authority will be undermined, not just by the specific ways scientists use AI, but by an evolution in our culture's attitude to certainty and intellectual authority in general - an evolution already underway, and which makes the widescale deployment of AI possible, as much as AI may in turn accelerate it.

Expand full comment
Jeffrey Quackenbush's avatar

One reason that science is not producing much in the way of fundamental insights never gets discussed: many of the gaps in human knowledge are *conceptual* and progress depends on doing original observational work that leads to concepts that can generate new bodies of descriptive corpora. The representations and the data for this new knowledge simply don't exist at the moment, and the normal experimental process of science won't be possible until someone dreams it up. The idea that humans have observed and conceptualized all the foundational knowledge in the world is absurd.

This type of work is exactly what LLMs are not able to do.

Expand full comment
Matt Cook's avatar

AI is a rubber stamp on the conventional wisdom because that’s all it knows.

But is the use of AI really worse than the whole peer review system?

Nothing much has changed since Thomas Kuhn's landmark book.

In fact, whole fields are exactly following the course of the geocentric theory so well discussed in your essay.

And the more funded a field of study is, the worse the conventional wisdom.

Because funding is given to scientists who tow the conventional line.

For example, dark matter and other silliness patching up discredited theories that scientists are funded to “believe” while discoveries of numerous cosmological anomalies by Halton Arp are simply ignored.

Climate “science” is similarly quack theories in the guise of science because those get funding.

Cancer research too. We know cancer is more of a metabolic disease, and yet the money and the conventional wisdom is on the long discredited genetic mutation theory.

And always, a few lone outsiders inevitably are more correct than the conventional wisdom.

Galileo was an outsider.

Why do we have to wait decades and decades for the referees who advocate discredited conventional wisdom to die off — so that decent science can eventually be published?

What feeds it is the referee system and the publishing monopoly e.g. Elsevier et al.

The science publishing monopoly should be replaced by Substack where smart people can comment and argue for and against a paper in public, in real time, building on line reputations for rigor.

Expand full comment
werdnagreb's avatar

> Sloppy use of AI may help in the short run, but will hinder meaningful scientific achievement.

You mention this in the context of scientific research, but I think it is true of AI in general.

Using AI to vibe code or create some slop might get you some clicks or publish an app quickly, but without understanding what the AI created bad things can happen.

Expand full comment
Sai Krishna's avatar

The second section reminded of this paper in which a simple chi-square stat was computed wrong and yet it got published in Nature journal

https://medium.com/@saikrishna_17904/do-we-tilt-right-or-left-to-kiss-chi-square-test-847552007ac9

Expand full comment
Thaddeus McIlroy's avatar

It's interesting for me to run parallels between what's happening in STM publishing and developments I'm a part of in trade/consumer publishing.

In trade publishing we've never had to adopt the pretense that what we're publishing is going to advance science or knowledge -- only, we hope each time, to advance our earnings, both as authors and as publishers. BUT, of course, every nonfiction author dreams or hopes that they are making some sort of contribution -- there are few who would crow "I just rewrote seven previous bestsellers into an unoriginal new book." The want to note, at a minimum, "well, at least I added some commentary, based on my unique experience, increasing the relevance of the earlier books. Plus, I quoted from my recent consulting learnings."

Does this seem so very different from what you're describing?

Economics and technology have changed the incentives in traditional publishing of all kinds, whether a daily paper, a monthly magazine, textbooks, trade books or The Lancet.

We keep looking to legacy formats to somehow gently adapt and address the new world. Unhappy families all, they are failing to do so, each in their own way.

There needs to be a more dramatic reinvention of "publishing." I'm still hoping that AI might aid in that quest.

Expand full comment
creative genius's avatar

"The Catholic Church welcomed the printing press as a way of solidifying its authority by printing Bibles..." This not true, Catholic church initially was against translations and printing Bibles. In fact they burned William Tyndale. It is the Protestants who pioneered the translations and printing of Bibles. It is Bible that catapulted the modern education and creating of Western civilization's ethos. I would suggest reading Vishal Mangalwadi's book "This book changed everything"

Expand full comment
Dominic Caldwell's avatar

The “AI slows science” worry is real — but the bottleneck isn’t speed, it’s translation. Science isn’t one big machine with a single attention economy; it’s an archipelago of separate islands, each with its own incentives and language. Drop AI into that, and you get wildly different effects. Sometimes it entrenches bad paradigms. Sometimes it opens doors that were bolted shut. What matters isn’t whether a field adopts AI, but how it connects across those islands.

Our problem isn’t too many lanes on the highway. It’s the bridges — or lack of them — between the places where real insight lives. AI can amplify the noise, or it can serve as a live interpreter between worlds. Which path we take depends on human navigation, not the technology itself.

https://open.substack.com/pub/dominiccaldwell/p/no-ai-is-not-slowing-progress?r=69tyy6&utm_campaign=post&utm_medium=web&showWelcomeOnShare=false

Expand full comment
Renaud Gaudron's avatar

Great read, thanks for sharing! Personally, I disagree with your point that bugs in AI-generated code are difficult to detect and therefore dangerous for scientific progress. I'd argue that the coding standards of the best-in-class models are already far higher than those of most humans. In all likelihood, this will reduce the number of bugs and lead to higher quality research. Sure, some bugs willl slip through, but there would be many more without AI, and I don't believe that would result in better scientific progress.

Expand full comment
Ursula Maria Mayer's avatar

Thanks for this excellent essay - I haven’t come across such a comprehensive and well-researched take on this topic before.

While my early work was in academia contributing to massively parallel multi-physics and engineering simulation codes, I’ve spent the past few years building large-scale AI systems and cloud infrastructure for clients across various industries.

You're absolutely right that industry has, in many cases, developed more mature software engineering practices and - often, though not always - faces stronger incentives to deliver functioning, production-grade software. Introducing these software engineering best practices into scientific work would certainly help improve the quality and reliability of AI code in academia. However, as pressure increases also in industry to release fast, software quality often suffers as well just as in research.

The hope that AI tools will accelerate software development - mirroring similar expectations in science - is widespread among many engineering managers. Media narratives often go even further, predicting that AI will soon replace software engineers altogether :-)

But recent evidence tells a different story - similar to your observations in in science. A study by the AI benchmarking non-profit METR (Blog: https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/; Paper: https://arxiv.org/abs/2507.09089) showed a similar production-progress paradox. When experienced developers used early 2025 AI coding tools (Cursor Pro, 3.5/3.7 Sonnet), they actually took 19% longer to complete tasks - despite expecting a 24% speed-up. Even more telling, after completing the slower work, they still believed the AI had helped and reported a perceived 20% speed-up. The gap between perception and reality here is remarkable.

This aligns closely with my own experience: AI tools are genuinely helpful when you're exploring new libraries, prototyping, or ramping up on a new tech stack. But as complexity increases - whether in the AI model itself, the codebase, or the surrounding infrastructure - the usefulness of these tools drops off rapidly. In some cases, they become more of a distraction than a help, especially for experienced AI engineers.

It will be fascinating to watch how these tools evolve in the coming years, how we choose to work with them - and whether and how their role in software engineering, scientific research, and other complex domains matures.

Expand full comment