Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support for older AMD CPUs? #19

Open
rwalle opened this issue Mar 22, 2025 · 2 comments
Open

Support for older AMD CPUs? #19

rwalle opened this issue Mar 22, 2025 · 2 comments
Assignees
Labels
p0 high priority

Comments

@rwalle
Copy link

rwalle commented Mar 22, 2025

According to project description, this only supports computers with the latest "Ryzen AI" CPUs.

Is there any blocker to support older CPUs without an NPU? Can we run LLMs with iGPU only?

Very few people have the latest generation of PCs, and I doubt anyone is going to update their perfectly fine computer purchased one year ago just to use LLM on consumer hardware.

I understand that there is a performance benefit of having the NPU, as mentioned in FAQ:

The iGPU can also handle LLMs but will consume more power and might have slower inference times compared to the NPU.

But hey, we all understand that we are not looking for the best performance on consumer hardware (otherwise I would be renting some H100s). If it runs on a machine with 680M or 780M, it is better than not being able to run it at all, despite being slower. This project could provide a meaningful alternative to Ollama, since it does not support Vulkan (ollama/ollama#5059). (I am aware that llama.cpp supports Vulkan by itself, but having a GUI is still nice.)

(Alternatively, if someone could create a fork or point me to where the restrictions are, I'll be happy to take a look.)

@rwalle rwalle changed the title Support for older CPUs? Support for older AMD CPUs? Mar 22, 2025
@kovtcharov
Copy link
Collaborator

kovtcharov commented Mar 23, 2025

Hi @rwalle, thanks for bringing this up, we've heard this request from several folks so we are actively working on enabling/supporting older systems on CPU and iGPU. I will report here once it becomes available, I expect in a few weeks.

@kovtcharov kovtcharov added the p0 high priority label Mar 23, 2025
@kovtcharov kovtcharov self-assigned this Mar 25, 2025
@kovtcharov
Copy link
Collaborator

kovtcharov commented Apr 1, 2025

You may want to checkout LM Studio in the meanwhile, they have a UI+llama.cpp/Vulkan support.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
p0 high priority
Projects
None yet
Development

No branches or pull requests

2 participants