-
Notifications
You must be signed in to change notification settings - Fork 545
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Llama3.1-405B non-zero temperature #2116
Comments
N.B: We are aware of an open vLLM issue, due to which setting temperature to zero still results in non-determinism. Maybe we will need to recalibrate reference accuracy for the next round (v5.1). |
Inference WG 18/Feb/2025: multiple parties run the reference implementation and obtained identical results. Maybe it's a by-product of |
Am I alone in thinking that by "non-determinism" the authors of this article actually mean "determinism"? Surely, "the randomness is minimized" should mean "near determinism"? |
It's like flammable vs inflammable. |
In the reference implementation of Llama3.1-405B, temperature is set to 1. Is this intentional?
Normally, temperature should be set to zero for outputs for be more deterministic. Can an optimized implementation use a different temperature?
The text was updated successfully, but these errors were encountered: