-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Roadmap] DeepSpeed Roadmap Q1 2025 #6946
Comments
Will Multi-Token Prediction mentioned in Deepseek V3 be added to the roadmap Q1? |
need FP8 training deepseek-MOE |
Plug-in support for the different accelerators |
@hijeffwu - could you clarify more on what you're requesting? Different accelerators are already supported in DeepSpeed. |
My idea is as follows: The current process for adding support for a new accelerator card involves creating a new Since the primary difference in AI chip vendors' support for DeepSpeed lies in kernel implementations, would it be possible to use "deepspeed-kernels" as the unified kernel code backend for DeepSpeed, while retaining only Python code in the main DeepSpeed repository? This approach could be like Megatron-LM + Apex + TransformerEngine, thereby making DeepSpeed more adaptable to diverse AI chip backends. Key points in this proposal:
This architecture aligns with observed practices in adapting DeepSpeed to non-NVIDIA hardware . |
This is a living document! For each item here, we intend to link the PR/issue for discussion.
This is DeepSpeed's first attempt at a public roadmap and will be updated with additional details.
The text was updated successfully, but these errors were encountered: