[ACL 2025] Are Your LLMs Capable of Stable Reasoning?
-
Updated
Aug 5, 2025 - Python
[ACL 2025] Are Your LLMs Capable of Stable Reasoning?
Ranking Large Language Models using the Principle of Least Action! Built during my time at Knit Space, Hubbali under the guidance Prof. Prakash Hegade.
Add a description, image, and links to the large-language-model-evaluation topic page so that developers can more easily learn about it.
To associate your repository with the large-language-model-evaluation topic, visit your repo's landing page and select "manage topics."