Skip to content

Conversation

@JRD971000
Copy link

@JRD971000 JRD971000 commented Oct 27, 2025

What does this PR do?

Type of change: New feature

Overview: Support pruning num_moe_experts and moe_shared_expert_intermediate_size in mcore_minitron pruning

Usage

# Add a code snippet demonstrating how to use this

Testing

Before your PR is "Ready for review"

  • Make sure you read and follow Contributor guidelines and your commits are signed.
  • Is this change backward compatible?: Yes
  • Did you write any new necessary tests?: Yes/No
  • Did you add or update any necessary documentation?: Yes/No
  • Did you update Changelog?: Yes/No

@JRD971000 JRD971000 self-assigned this Oct 27, 2025
@copy-pr-bot
Copy link

copy-pr-bot bot commented Oct 27, 2025

This pull request requires additional validation before any workflows can run on NVIDIA's runners.

Pull request vetters can view their responsibilities here.

Contributors can view more details about this message here.

@kevalmorabia97 kevalmorabia97 changed the title Alit/moe dev Add MoE pruning support in Minitron Oct 27, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants