MMLU Performance #67
w013nad
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I spent some time editing your MMLU test script to compare performance at different quantization levels. A full write-up may be found here.
https://www.reddit.com/r/LocalLLaMA/comments/16tgzzk/exllamav2_performance_with_different_quantization/
If you're interested, I can try to merge my script with yours (I'm unfamiliar with Git). Feel free to use this chart if it's useful.
Full results.
https://docs.google.com/spreadsheets/d/1MFmHDpqcf7CP_EYnwl1QsUP0KhS6jA8x1JYuwm3NH0U/edit?usp=sharing
Beta Was this translation helpful? Give feedback.
All reactions