-
Notifications
You must be signed in to change notification settings - Fork 9.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bug: Model isn't loading #9563
Comments
I'm also facing same issue. Tried updating to 3803 but didn't help. |
By any chance you are migrating from Ollama? If not where did you get your model files? |
I downloaded GGUF from huggingface: https://huggingface.co/bartowski/Phi-3.5-mini-instruct-GGUF |
I tried to run this exactly model: https://huggingface.co/bartowski/Phi-3.5-mini-instruct-GGUF/blob/main/Phi-3.5-mini-instruct-IQ2_M.gguf on UBUNTU and I got the problem of memory: ggml/src/ggml.c:438: fatal error probably the memory allocator is "over allocing" memory. I tried a bigger model (in parameters) https://huggingface.co/TheBloke/Llama-2-13B-GGUF/blob/main/llama-2-13b.Q2_K.gguf and the model ran normally with almost no extra memory |
@salocinrevenge add @ggerganov we keep getting bug reports because people don't realize that they cannot use the full context of current models. Should we revert this change? |
Did anyone find a fix for the original issue? |
I found out what was causing the issue. I was using the avx2 version while my CPU only supports avx. Changing it fixed it. It would be great if the program could inform the user of unsupported version. |
What happened?
llama3.1 isn't loading at all. I get following in the terminal and the program just quits.:
Name and Version
./llama-cli --version
version: 3787 (6026da5)
built with MSVC 19.29.30154.0 for x64
What operating system are you seeing the problem on?
Windows
Relevant log output
The text was updated successfully, but these errors were encountered: