Update Qwen3 235B B300 Configs to match Qwen3 B200 Configs#2669
Update Qwen3 235B B300 Configs to match Qwen3 B200 Configs#2669rhmukundan wants to merge 3 commits intomainfrom
Conversation
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
|
No actionable comments were generated in the recent review. 🎉 ℹ️ Recent review info⚙️ Run configurationConfiguration used: Path: .coderabbit.yaml Review profile: CHILL Plan: Pro Run ID: 📒 Files selected for processing (1)
📝 WalkthroughWalkthroughThis PR modifies QWEN3 workload configurations by disabling moe_a2a_overlap across two configuration variants and removing an explicit virtual_pipeline_model_parallel_size setting from the FP8_CS_V1 configuration. Changes
Estimated code review effort🎯 1 (Trivial) | ⏱️ ~3 minutes Possibly related PRs
Suggested labels
Suggested reviewers
🚥 Pre-merge checks | ✅ 4✅ Passed checks (4 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing Touches
🧪 Generate unit tests (beta)
Tip Try Coding Plans. Let us write the prompt for your AI agent so you can ship faster (with fewer bugs). Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
| expert_model_parallel_size=8, | ||
| global_batch_size=1024, | ||
| moe_a2a_overlap=True, | ||
| moe_a2a_overlap=False, |
There was a problem hiding this comment.
explain why in the pr desciption?
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
Summary by CodeRabbit
Note: This release contains only internal configuration adjustments with no user-facing changes.