Skip to content

Failed to download or register the following EPs: QNNExecutionProvider #393

@martinskuta

Description

@martinskuta

I am running the latest foundry version 0.8.117.0 and it doesn't load the QNNExecutionProvider so I cannot use the models on NPU.

There is not much info in the logs apart from:

2026-01-23 17:32:34.290 +01:00 [INF] Provider QNNExecutionProvider download and ensuring ready attempt: Failure
2026-01-23 17:32:34.293 +01:00 [INF] Download attempt for QNNExecutionProvider unsuccessful. Skipping all EP downloads. 
2026-01-23 17:32:34.293 +01:00 [INF] Finished attempt to autoregister certified EPs; finished in 1392ms
2026-01-23 17:32:34.293 +01:00 [INF] Failed to download or register the following EPs: QNNExecutionProvider. Will try installing again later.

My machine is Surface Pro 11, with Snapdragon X1E80100 running Windows 11 Pro Build 26220

When I run foundry model ls I see only models that can run on CPU:

Alias                          Device     Task           File Size    License      Model ID
-----------------------------------------------------------------------------------------------
phi-4                          CPU        chat           10.16 GB     MIT          Phi-4-generic-cpu:1
----------------------------------------------------------------------------------------------------------
phi-3.5-mini                   CPU        chat           2.53 GB      MIT          Phi-3.5-mini-instruct-generic-cpu:1
--------------------------------------------------------------------------------------------------------------------------
phi-3-mini-128k                CPU        chat           2.54 GB      MIT          Phi-3-mini-128k-instruct-generic-cpu:2
-----------------------------------------------------------------------------------------------------------------------------
phi-3-mini-4k                  CPU        chat           2.53 GB      MIT          Phi-3-mini-4k-instruct-generic-cpu:2
---------------------------------------------------------------------------------------------------------------------------
mistral-7b-v0.2                CPU        chat           4.07 GB      apache-2.0   mistralai-Mistral-7B-Instruct-v0-2-generic-cpu:2
---------------------------------------------------------------------------------------------------------------------------------------
deepseek-r1-14b                CPU        chat           11.51 GB     MIT          deepseek-r1-distill-qwen-14b-generic-cpu:3
---------------------------------------------------------------------------------------------------------------------------------
deepseek-r1-7b                 CPU        chat           6.43 GB      MIT          deepseek-r1-distill-qwen-7b-generic-cpu:3
--------------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-0.5b             CPU        chat, tools    0.80 GB      apache-2.0   qwen2.5-coder-0.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------------
phi-4-mini-reasoning           CPU        chat           4.52 GB      MIT          Phi-4-mini-reasoning-generic-cpu:3
-------------------------------------------------------------------------------------------------------------------------
qwen2.5-0.5b                   CPU        chat, tools    0.80 GB      apache-2.0   qwen2.5-0.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------
qwen2.5-1.5b                   CPU        chat, tools    1.78 GB      apache-2.0   qwen2.5-1.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-1.5b             CPU        chat, tools    1.78 GB      apache-2.0   qwen2.5-coder-1.5b-instruct-generic-cpu:4
--------------------------------------------------------------------------------------------------------------------------------
phi-4-mini                     CPU        chat, tools    4.80 GB      MIT          Phi-4-mini-instruct-generic-cpu:5
------------------------------------------------------------------------------------------------------------------------
qwen2.5-14b                    CPU        chat, tools    11.06 GB     apache-2.0   qwen2.5-14b-instruct-generic-cpu:4
-------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-14b              CPU        chat, tools    11.06 GB     apache-2.0   qwen2.5-coder-14b-instruct-generic-cpu:4
-------------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-7b               CPU        chat, tools    6.16 GB      apache-2.0   qwen2.5-coder-7b-instruct-generic-cpu:4
------------------------------------------------------------------------------------------------------------------------------
qwen2.5-7b                     CPU        chat, tools    6.16 GB      apache-2.0   qwen2.5-7b-instruct-generic-cpu:4
------------------------------------------------------------------------------------------------------------------------
gpt-oss-20b                    CPU        chat           12.26 GB     MIT          gpt-oss-20b-generic-cpu:1

When I install older version of the foundry winget install --id=Microsoft.FoundryLocal -v "0.6.87.59034" -e it seems to work:

Alias                          Device     Task               File Size    License      Model ID
-----------------------------------------------------------------------------------------------
phi-4                          CPU        chat-completion    10.16 GB     MIT          Phi-4-generic-cpu
--------------------------------------------------------------------------------------------------------
phi-3.5-mini                   NPU        chat-completion    2.78 GB      MIT          phi-3.5-mini-instruct-qnn-npu
                               CPU        chat-completion    2.53 GB      MIT          Phi-3.5-mini-instruct-generic-cpu
------------------------------------------------------------------------------------------------------------------------
deepseek-r1-14b                NPU        chat-completion    7.12 GB      MIT          deepseek-r1-distill-qwen-14b-qnn-npu
---------------------------------------------------------------------------------------------------------------------------
deepseek-r1-7b                 NPU        chat-completion    3.71 GB      MIT          deepseek-r1-distill-qwen-7b-qnn-npu
--------------------------------------------------------------------------------------------------------------------------
phi-3-mini-128k                NPU        chat-completion    2.78 GB      MIT          phi-3-mini-128k-instruct-qnn-npu
                               CPU        chat-completion    2.54 GB      MIT          Phi-3-mini-128k-instruct-generic-cpu
---------------------------------------------------------------------------------------------------------------------------
phi-3-mini-4k                  NPU        chat-completion    2.78 GB      MIT          phi-3-mini-4k-instruct-qnn-npu
                               CPU        chat-completion    2.53 GB      MIT          Phi-3-mini-4k-instruct-generic-cpu
-------------------------------------------------------------------------------------------------------------------------
mistral-7b-v0.2                CPU        chat-completion    4.07 GB      apache-2.0   mistralai-Mistral-7B-Instruct-v0-2-generic-cpu
-------------------------------------------------------------------------------------------------------------------------------------
whisper-base                   CPU        automatic-speech-recognition 0.37 GB      apache-2.0   openai-whisper-base-generic-cpu
----------------------------------------------------------------------------------------------------------------------
whisper-medium                 CPU        automatic-speech-recognition 3.05 GB      apache-2.0   openai-whisper-medium-generic-cpu
------------------------------------------------------------------------------------------------------------------------
whisper-small                  CPU        automatic-speech-recognition 1.05 GB      apache-2.0   openai-whisper-small-generic-cpu
-----------------------------------------------------------------------------------------------------------------------
whisper-tiny                   CPU        automatic-speech-recognition 0.22 GB      apache-2.0   openai-whisper-tiny-generic-cpu
----------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-0.5b             CPU        chat-completion    0.80 GB      apache-2.0   qwen2.5-coder-0.5b-instruct-generic-cpu
------------------------------------------------------------------------------------------------------------------------------
phi-4-mini-reasoning           CPU        chat-completion    4.52 GB      MIT          Phi-4-mini-reasoning-generic-cpu
-----------------------------------------------------------------------------------------------------------------------
qwen2.5-0.5b                   CPU        chat-completion    0.80 GB      apache-2.0   qwen2.5-0.5b-instruct-generic-cpu
------------------------------------------------------------------------------------------------------------------------
qwen2.5-1.5b                   NPU        chat-completion    2.78 GB      MIT          qwen2.5-1.5b-instruct-qnn-npu
                               CPU        chat-completion    1.78 GB      apache-2.0   qwen2.5-1.5b-instruct-generic-cpu
------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-1.5b             CPU        chat-completion    1.78 GB      apache-2.0   qwen2.5-coder-1.5b-instruct-generic-cpu
------------------------------------------------------------------------------------------------------------------------------
qwen2.5-14b                    CPU        chat-completion    11.06 GB     apache-2.0   qwen2.5-14b-instruct-generic-cpu
-----------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-14b              CPU        chat-completion    11.06 GB     apache-2.0   qwen2.5-coder-14b-instruct-generic-cpu
-----------------------------------------------------------------------------------------------------------------------------
qwen2.5-coder-7b               CPU        chat-completion    6.16 GB      apache-2.0   qwen2.5-coder-7b-instruct-generic-cpu
----------------------------------------------------------------------------------------------------------------------------
qwen2.5-7b                     NPU        chat-completion    2.78 GB      MIT          qwen2.5-7b-instruct-qnn-npu
                               CPU        chat-completion    6.16 GB      apache-2.0   qwen2.5-7b-instruct-generic-cpu
----------------------------------------------------------------------------------------------------------------------
whisper-large-v3-turbo         CPU        automatic-speech-recognition 8.79 GB      apache-2.0   openai-whisper-large-v3-turbo-generic-cpu
--------------------------------------------------------------------------------------------------------------------------------
gpt-oss-20b                    CPU        chat-completion    12.26 GB     MIT          gpt-oss-20b-generic-cpu

I have a tingle that it is related to the fact that I have windows workload execution provider that comes with windows, because if I run Get-AppxPackage -AllUsers "*.EP.*" | Select-Object -ExpandProperty PackageFullName I get this:

Windows.Workload.EP.Qualcomm.QNN.1.8_1.8.27.0_arm64__8wekyb3d8bbwe
WindowsWorkload.EP.Qualcomm.QNN.1.8_1.8.31.0_arm64__8wekyb3d8bbwe

but there is no way to remove those, because running Get-AppxPackage -AllUsers "WindowsWorkload.EP*" | Remove-AppxPackage as admin or as AUTHORITY/SYSTEM seems to start something in the terminal, but it disappears immediatelly without any result.

Any help would be appreciated. Thx.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions