On Using LLMs in Research
Updated: 26 Apr 2026
I started using LLMs more frequently at the beginning of the year, when they began to feel useful for answering technical questions. I only started using them seriously in my research around March. By then, they had become so good at programming that delegating some of that work to machines felt economical rather than merely convenient.
For programming, LLMs are useful for several reasons.
- They reduce repetitive code.
- There are diminishing returns to typing the same imports, headers, boilerplate, and function definitions again and again.
- After several years of programming, the pleasure comes less from the act of writing code and more from whether the code is correct, elegant, and well suited to the problem.
- They make it easier to focus on hypothesis generation and testing [1].
- When implementation becomes cheaper, one can test ideas more quickly, discard weak hypotheses earlier, and kill one’s darlings without spending months trapped in debugging.
-
They point toward a new workflow for computational science [2]:
hypothesis -> code -> result -> revised hypothesis -> revised experiment- This loop can, in principle, be handled by agents.
- I have not fully tried that yet, partly because I still derive satisfaction from controlling the flow myself rather than watching and waiting.
For learning science (physics/chemistry/mathematics), the barrier to entry has become much lower.
- It is now easier to learn only what you need, exactly when you need it.
- That can be a good strategy, especially when the goal is to solve a concrete problem.
- At the same time, creative work still depends on gathering knowledge from diverse fields.
- LLMs can accelerate that process too, but they should not replace the broader habit of intellectual wandering.
Literature review has also become easier.
- The main risk is that LLMs can funnel attention into a narrow region of the literature [6].
- They are good at following the center of a topic, but less reliable for finding what sits at the edges: neglected ideas, strange exceptions, and papers outside the usual citation paths [3].
The goal of ARC-AGI is a promising step toward a more general problem solver: one capable not only of recognizing patterns, but of analyzing and generating new abstract concepts [4], [5].
References
- Peng, S., Kalliamvakou, E., Cihon, P., & Demirer, M. (2023). The impact of AI on developer productivity. Microsoft Research.
- Lu, C., Lu, C., Lange, R. T., Foerster, J., Clune, J., & Ha, D. (2024). The AI Scientist. arXiv.
- Azamfirei, R., Kudchadkar, S. R., & Fackler, J. (2023). Large language models and the perils of their hallucinations. Critical Care, 27, 120.
- Chollet, F. (2019). On the Measure of Intelligence. arXiv.
- ARC Prize. ARC-AGI-1.
- Hao, Q., Xu, F., Li, Y., & Evans, J. (2026). Artificial intelligence tools expand scientists’ impact but contract science’s focus. Nature, 649(8099), 1237-1243.