-
Notifications
You must be signed in to change notification settings - Fork 2k
Implement CoT no-op for reasoning models #8375
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
This is AMAZING. Thank you @FireMasterK !! I think we'll need to think (no pun intended) about reasoning models like Qwen3 that expose their reasoning in the string... What does LiteLLM do for that. I guess it says Also I guess we should still return a |
c8b1c4e
to
b4e570b
Compare
I don't think they support telling whether a model displays its CoT or not, which is unfortunate. It also seems like LiteLLM doesn't have enough mappings for reasoning models (eg. Qwen3 models are not there), which means we can't solely rely on LiteLLM telling if a model supports reasoning or not.
Yes, but some model providers providers summarize the CoT, what should we do about these? I have added the parsing of |
b4e570b
to
2181050
Compare
2181050
to
a6fd573
Compare
a6fd573
to
903e7e4
Compare
I fixed the serialization issues, so all tests now pass! Let me know what I should do about #8375 (comment). I'm happy to create another PR or add them in this PR. Whichever you prefer. |
litellm.supports_reasoning
.Additional context: https://x.com/DSPyOSS/status/1931886536221724848