In the reference implementation of Llama3.1-405B, temperature is [set to 1](https://github.com/mlcommons/inference/blob/7bf59976b5f4eb7c5b8f30a88af832e028028446/language/llama3.1-405b/SUT_VLLM.py#L71). Is this intentional? Normally, temperature should be set to zero for outputs for be more deterministic. Can an optimized implementation use a different temperature?