Human Researchers are Superior to Large Language Models in Writing a Systematic Review in a Comparative Multitask Assessment
Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
Background The capability of Large Language Models (LLMs) to support and facilitate research activities has sparked growing interest in their integration into scientific workflows. This paper aims to evaluate and compare against human researchers the performance of 6 different LLMs in conducting the various tasks necessary to produce a systematic literature review. Methods The evaluation of the 6 LLMs was split into 3 tasks: literature search, article screening and selection (task 1); data extraction and analysis (task 2); final paper drafting (task 3). Their results were compared with a human-produced systematic review on the same topic, serving as reference standard. The evaluation was repeated on two rounds to evaluate reproducibility and improvements of LLMs over time. Results Out of the 18 scientific articles to be extracted from the literature for task 1, the best LLM managed to identify 13. Data extraction and analysis for task 2 was only partially accurate and cumbersome. The full papers generated by LLMs for task 3 were short and uninspiring, often not fully adhering to the standard template for a systematic review. Conclusion Currently, LLMs are not capable of independently conducting a scientific systematic review. However, their capabilities are advancing rapidly, and, with an appropriate supervision they can provide valuable support throughout the review process.