Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bug: Model isn't loading #9563

Open
iladshyan opened this issue Sep 20, 2024 · 7 comments
Open

Bug: Model isn't loading #9563

iladshyan opened this issue Sep 20, 2024 · 7 comments
Labels
bug-unconfirmed high severity Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)

Comments

@iladshyan
Copy link

What happened?

llama3.1 isn't loading at all. I get following in the terminal and the program just quits.:

./llama-cli -m "C:\<path>\llama3.1.gguf" -p "The world is a place where"
build: 3787 (6026da52) with MSVC 19.29.30154.0 for x64
main: llama backend init
main: load the model and apply lora adapter, if any

Name and Version

./llama-cli --version
version: 3787 (6026da5)
built with MSVC 19.29.30154.0 for x64

What operating system are you seeing the problem on?

Windows

Relevant log output

build: 3787 (6026da52) with MSVC 19.29.30154.0 for x64
main: llama backend init
main: load the model and apply lora adapter, if any
@iladshyan iladshyan added bug-unconfirmed high severity Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow) labels Sep 20, 2024
@ninadakolekar
Copy link

I'm also facing same issue. Tried updating to 3803 but didn't help.

@iladshyan
Copy link
Author

I'm also facing same issue. Tried updating to 3803 but didn't help.

By any chance you are migrating from Ollama? If not where did you get your model files?

@ninadakolekar
Copy link

ninadakolekar commented Sep 23, 2024

I downloaded GGUF from huggingface: https://huggingface.co/bartowski/Phi-3.5-mini-instruct-GGUF

@github-actions github-actions bot added the stale label Oct 24, 2024
@salocinrevenge
Copy link

I tried to run this exactly model: https://huggingface.co/bartowski/Phi-3.5-mini-instruct-GGUF/blob/main/Phi-3.5-mini-instruct-IQ2_M.gguf on UBUNTU and I got the problem of memory:

ggml/src/ggml.c:438: fatal error
ggml_aligned_malloc: insufficient memory (attempted to allocate 49152,00 MB)

probably the memory allocator is "over allocing" memory.

I tried a bigger model (in parameters) https://huggingface.co/TheBloke/Llama-2-13B-GGUF/blob/main/llama-2-13b.Q2_K.gguf and the model ran normally with almost no extra memory

@slaren
Copy link
Collaborator

slaren commented Nov 1, 2024

@salocinrevenge add -c 1024 to the command line to use a smaller context size.

@ggerganov we keep getting bug reports because people don't realize that they cannot use the full context of current models. Should we revert this change?

@iladshyan
Copy link
Author

Did anyone find a fix for the original issue?

@iladshyan
Copy link
Author

I found out what was causing the issue. I was using the avx2 version while my CPU only supports avx. Changing it fixed it. It would be great if the program could inform the user of unsupported version.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug-unconfirmed high severity Used to report high severity bugs in llama.cpp (Malfunctioning hinder important workflow)
Projects
None yet
Development

No branches or pull requests

4 participants