Can AI Conduct Research: A Pragmatic Experiment

Read the full article See related articles

Discuss this preprint

Start a discussion What are Sciety discussions?

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

This study evaluated the capability of three prominent Large Language Model (LLM) AI tools—ChatGPT, Copilot, and Claude—to independently conduct a complete research process. The investigation involved generating hypothetical data (“silicon samples”), performing inductive thematic analysis, and composing research reports. Findings reveal that while AI can produce outputs resembling research products, the quality varies significantly, often lacking depth, accuracy, and synthesis. Notably, AI-generated data and analyses tend to be predictable, with issues such as hallucinated references and misquoted data, underscoring the necessity of human oversight. The study highlights both the potential and current limitations of AI in autonomous research, emphasizing that human researchers remain essential for producing high-quality, impactful scholarly work.

Article activity feed