FormalizerInsideLLM: A Constraint-Based Reasoning Framework for Large Language Models via Axiomatic Control
Discuss this preprint
Start a discussion What are Sciety discussions?Listed in
This article is not in any list yet, why not save it to one of your lists.Abstract
We introduce FormalizerInsideLLM, a constraint-based reasoning framework designed to rigorously enforce symbolic logic constraints within large language models (LLMs). By explicitly limiting inference to human-defined axiom sets, our framework reliably eliminates hallucinations, enabling verifiable formal reasoning and experimentation with undecidable or axiom-dependent statements. Through comparative evaluations with GPT-3.5, GPT-4o, and Gemini Pro across number theory, geometry, and combinatorics, we demonstrate that GPT-4o and Gemini Pro significantly outperform GPT-3.5 in adhering to symbolic constraints, reliably identifying undecidable problems, and correctly deriving provable statements.