You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This discussion was converted from issue #1206 on October 23, 2024 21:05.
Heading
Bold
Italic
Quote
Code
Link
Numbered list
Unordered list
Task list
Attach files
Mention
Reference
Menu
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
Following the GPT Pretraining section in the
Megatron-LM
repo, we are able to successfully train a model usingMegatron-LM
I saw pointers on how to convert from HF to
nemo
. For example, this conversion script convert_llama_hf_to_nemo.pyHowever I did not see any examples of converting a ckpt saved using the
Megatron-LM
tonemo
format . Are there any examples for this?p.s. I am thinking of doing this conversion to
nemo
, so I can use tools likeNemo-Aligner
for post-trainingBeta Was this translation helpful? Give feedback.
All reactions