Skip to content

Conversation

@noahvandal
Copy link

@noahvandal noahvandal commented Jan 18, 2026

I added support for the Groq API and the Cerebras SDK, to enable a much higher throughput of RLM inference.

I did note that the gpt-oss series models did not perform that well, as there were errors on the manner in which it was trying to call tools, instead of using the REPL. The other models such as llama-3.3-70b-versatile, et. al., performed fine.
{NOTE}: trying out the gpt-oss on Cerebras worked just fine; this seems to be a Groq issue.

@noahvandal noahvandal changed the title Adding support for Groq API compatibility to have much higher response feat: Adding support for Groq API compatibility to have much higher response Jan 18, 2026
@noahvandal noahvandal changed the title feat: Adding support for Groq API compatibility to have much higher response feat: Adding support for Groq API, Cerebras SDK, compatibility to have much higher response Jan 18, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant