-
Notifications
You must be signed in to change notification settings - Fork 212
Open
Labels
area:ckptCheckpoint conversion, loading, export, and save pathsCheckpoint conversion, loading, export, and save pathsarea:modelModel implementations and HF bridge logicModel implementations and HF bridge logicbugSomething isn't workingSomething isn't workingcommunity-requestneeds-follow-upIssue needs follow-upIssue needs follow-up
Description
Description
Nemotron Nano v2 models (nvidia/NVIDIA-Nemotron-Nano-9B-v2, nvidia/NVIDIA-Nemotron-Nano-12B-v2) fail during convert_checkpoints.py import at the tokenizer-save step:
AttributeError: 'bool' object has no attribute 'chat_template'
self.tokenizer is False instead of a HuggingFace tokenizer object.
Root Cause
Three factors combine:
- The
nvcr.io/nvidia/nemo:26.02container's megatron-core (core_r0.16.0) defaultsHuggingFaceTokenizer.__init__touse_fast=False - Nemotron Nano v2 only ships a fast tokenizer (
tokenizer.json, notokenizer.model), soAutoTokenizer.from_pretrained(..., use_fast=False)silently returnsFalse NemotronHBridgedoes not implementget_hf_tokenizer_kwargs()to override this default
GLM45VBridge already solves this exact problem by implementing get_hf_tokenizer_kwargs() returning {"use_fast": True}.
Fix
Add get_hf_tokenizer_kwargs to NemotronHBridge, following the GLM45VBridge pattern. The call site in auto_bridge.py:768 already checks for this method via hasattr — no other changes needed.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
area:ckptCheckpoint conversion, loading, export, and save pathsCheckpoint conversion, loading, export, and save pathsarea:modelModel implementations and HF bridge logicModel implementations and HF bridge logicbugSomething isn't workingSomething isn't workingcommunity-requestneeds-follow-upIssue needs follow-upIssue needs follow-up