Hi, on a single 4090 GPU with 24GB memory, the following command will cause out-of-memory. ```bash python main.py mmlu --model_name llama --model_path huggyllama/llama-7b ``` After that, I try executing the command on A100-40GB, the `nvidia-smi` result is <img width="718" alt="image" src="https://github.com/declare-lab/instruct-eval/assets/119666897/2f25c828-f372-46b9-acc8-b7f5a5832b11"> It seems that neither 4090/3090 with 24GB memory or V100 with 32GB memory cannot test Llama-7B on mmlu under above command. So how to evaluate llama-7b on mmlu on 24GB or 32GB GPU? any more options to enable? Thanks