Benchmarking Multimodal Large Language Models for Forensic Science and Medicine: A Comprehensive Dataset and Evaluation Framework

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

Background

Multimodal large language models (MLLMs) have demonstrated substantial progress in medical and legal domains in recent years; however, their capabilities from the lens of forensic science—a field that is at the intersection of complex medical reasoning and legal interpretation, with conclusions critiqued by judicial scrutiny—remains largely unexplored. Forensic medicine uniquely depends on the accurate integration of often ambiguous text and visual information, yet systematic evaluations of MLLMs in this setting are lacking.

Methods

We conducted a comprehensive benchmarking study of eleven state-of-the-art MLLMs, including proprietary (GPT-4o, Claude 4 Sonnet, Gemini 2.5 Flash) and open-source (Llama 4, Qwen 2.5-VL) models. Models were evaluated on 847 examination-style forensic questions drawn from various academic literature, case studies, and clinical assessments, covering nine forensic subdomains. Both text-only and image-based questions were included. Model performance was assessed using direct and chain-of-thought prompting, with automated scoring verified through manual revision.

Results

Performance improved consistently with newer model generations. Chain-of-thought prompting improved accuracy on text-based and choice-based tasks for most models, though this trend did not hold for image-based and open-ended questions. Visual reasoning and complex inference tasks revealed persistent limitations, with models underperforming in image interpretation and nuanced forensic scenarios. Model performance remained stable across forensic subdomains, suggesting topic type alone did not drive variability.

Conclusions

MLLMs show emerging potential for forensic education and structured assessments, particularly for reinforcing factual knowledge. However, their limitations in visual reasoning, open-ended interpretation, and forensic judgment preclude independent application in live forensic practice. Future efforts should prioritize the development of multimodal forensic datasets, domain-targeted fine-tuning, and task-aware prompting to improve reliability and generalizability. These findings provide the first systematic baseline for MLLM performance in forensic science and inform pathways for their cautious integration into medico-legal workflows.

Article activity feed