-
Notifications
You must be signed in to change notification settings - Fork 146
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Not able to run Exclude-Embed models #1163
Comments
This is a known issue inside ONNX Runtime GenAI. The onnxruntime-genai/src/models/multi_modal_vision_model.cpp Lines 114 to 123 in 41c2543
The class that manages these embeddings can be found here. However, user access to the class has not been enabled currently. onnxruntime-genai/src/models/embeddings.cpp Lines 19 to 22 in 41c2543
Similar to how the generator object contains an |
Thanks for the response. Any possibility of getting this implemented anytime soon? |
We are currently preparing for a stable release so this will have to wait until after the release is completed. If you need this feature sooner or you would like to help add it, please feel free to open a PR to contribute as well. |
Describe the bug
I have taken a model which is supported by this library, and have removed the embeddings from the model using exclude-embeds option in builder.py. Now when i try to run the model, it is still asking for input_ids instead of input_embeds.
To Reproduce
Steps to reproduce the behavior:
Qwen/Qwen2.5-0.5 model converted by giving exclude-embeds option and run inference in java.
Desktop (please complete the following information):
Can someone share idea on how to make this work?
The text was updated successfully, but these errors were encountered: