Skip to content

Update Qwen3 235B B300 Configs to match Qwen3 B200 Configs#2669

Open
rhmukundan wants to merge 3 commits intomainfrom
rmukundan/match_qwen3_235_b300_b200_configs
Open

Update Qwen3 235B B300 Configs to match Qwen3 B200 Configs#2669
rhmukundan wants to merge 3 commits intomainfrom
rmukundan/match_qwen3_235_b300_b200_configs

Conversation

@rhmukundan
Copy link
Contributor

@rhmukundan rhmukundan commented Mar 5, 2026

Summary by CodeRabbit

  • Chores
    • Updated configuration settings for model workload optimization.

Note: This release contains only internal configuration adjustments with no user-facing changes.

Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
@copy-pr-bot
Copy link

copy-pr-bot bot commented Mar 5, 2026

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

@rhmukundan rhmukundan self-assigned this Mar 5, 2026
@rhmukundan rhmukundan marked this pull request as ready for review March 5, 2026 21:50
@coderabbitai
Copy link
Contributor

coderabbitai bot commented Mar 5, 2026

No actionable comments were generated in the recent review. 🎉

ℹ️ Recent review info
⚙️ Run configuration

Configuration used: Path: .coderabbit.yaml

Review profile: CHILL

Plan: Pro

Run ID: ebcb689e-d287-4108-aa45-34de7ea40655

📥 Commits

Reviewing files that changed from the base of the PR and between b730ab6 and a759cdf.

📒 Files selected for processing (1)
  • scripts/performance/configs/qwen/qwen3_workload_base_configs.py

📝 Walkthrough

Walkthrough

This PR modifies QWEN3 workload configurations by disabling moe_a2a_overlap across two configuration variants and removing an explicit virtual_pipeline_model_parallel_size setting from the FP8_CS_V1 configuration.

Changes

Cohort / File(s) Summary
QWEN3 Workload Configuration
scripts/performance/configs/qwen/qwen3_workload_base_configs.py
Disabled moe_a2a_overlap (changed from True to False) in QWEN3_235B_A22B_PRETRAIN_CONFIG_B300_BF16_V1 and QWEN3_235B_A22B_PRETRAIN_CONFIG_B300_FP8_CS_V1 configurations. Removed virtual_pipeline_model_parallel_size=4 from FP8_CS_V1 variant.

Estimated code review effort

🎯 1 (Trivial) | ⏱️ ~3 minutes

Possibly related PRs

Suggested labels

performance

Suggested reviewers

  • malay-nagda
  • dingqingy-nv
  • ko3n1g
🚥 Pre-merge checks | ✅ 4
✅ Passed checks (4 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title accurately describes the main change: updating Qwen3 235B B300 configurations to match B200 configurations, which aligns with the modifications to moe_a2a_overlap and virtual_pipeline_model_parallel_size settings.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
Test Results For Major Changes ✅ Passed Configuration-only changes standardizing B300 to match B200 parameters; minor adjustments to parallelism settings without code logic modifications.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
  • 📝 Generate docstrings (stacked PR)
  • 📝 Generate docstrings (commit on current branch)
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch rmukundan/match_qwen3_235_b300_b200_configs

Tip

Try Coding Plans. Let us write the prompt for your AI agent so you can ship faster (with fewer bugs).
Share your feedback on Discord.


Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@dingqingy-nv dingqingy-nv added performance performance/release Performance items related with NeMo release r0.3.0 Cherry-pick label for r0.3.0 release branch labels Mar 5, 2026
expert_model_parallel_size=8,
global_batch_size=1024,
moe_a2a_overlap=True,
moe_a2a_overlap=False,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

explain why in the pr desciption?

yaoyu-33
yaoyu-33 previously approved these changes Mar 6, 2026
Signed-off-by: Raghav Hrishikeshan Mukundan <rmukundan@nvidia.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

performance/release Performance items related with NeMo release performance r0.3.0 Cherry-pick label for r0.3.0 release branch

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants