Skip to content

Releases: Gaganv882/llm-eval-analysis

llm-eval-analysis v1.4.6-beta.3

04 May 13:16
Compare
Choose a tag to compare

This release, version 1.4.6-beta.3, introduces enhancements to the multi-metric evaluation framework for human-bot dialogues. It now supports additional datasets and improves the analysis capabilities for LLMs like Claude and GPT-4o. Users can expect more accurate insights and streamlined performance in their evaluations.