-
Notifications
You must be signed in to change notification settings - Fork 2.3k
Description
Problem (one or two sentences)
When using GLM-4.6 through an OpenAI-compatible custom endpoint in Roo Code, the model does not generate thinking tokens. The requests appear to be missing the required "thinking": {"type": "enabled"} parameter that GLM-4.6 needs to activate its reasoning mode.
Context (who is affected and when)
This affects users who are using GLM-4.6 with custom OpenAI-compatible endpoints (not Z.ai or OpenRouter providers) in Roo Code latest versions (3.28.12-3.28.15). The issue occurs consistently with every request since GLM-4.6's recent release (September 2025).
Reproduction steps
-
Configure Roo Code to use an OpenAI-compatible custom endpoint
-
Set the model to GLM-4.6 (glm-4.6)
-
Send any coding request to the model
-
Observe the response from GLM-4.6
-
Notice that no thinking tokens or reasoning blocks are included in the output
Expected result
GLM-4.6 should generate thinking tokens showing its step-by-step reasoning process before providing the final answer, similar to how it works with Z.ai native provider or OpenRouter. The thinking blocks should be visible in the Roo Code interface showing the model's internal reasoning.
Actual result
GLM-4.6 responds immediately without any thinking tokens or reasoning blocks. The model provides direct answers without showing the intermediate reasoning steps, even though GLM-4.6 is designed as a thinking model.
Variations tried (optional)
No response
App Version
3.28.15
API Provider (optional)
OpenAI Compatible
Model Used (optional)
GLM 4.6
Roo Code Task Links (optional)
No response
Relevant logs or errors (optional)
Metadata
Metadata
Assignees
Labels
Type
Projects
Status