Skip to content

llm-eval-analysis v1.4.6-beta.3

Latest
Compare
Choose a tag to compare
@Gaganv882 Gaganv882 released this 04 May 13:16
· 1 commit to main since this release

This release, version 1.4.6-beta.3, introduces enhancements to the multi-metric evaluation framework for human-bot dialogues. It now supports additional datasets and improves the analysis capabilities for LLMs like Claude and GPT-4o. Users can expect more accurate insights and streamlined performance in their evaluations.