Skip to content

Commit 50b10b8

Browse files
Remove unused eos_id from model args
1 parent faa2122 commit 50b10b8

File tree

6 files changed

+0
-6
lines changed

6 files changed

+0
-6
lines changed

torchtitan/experiments/deterministic_vllm_rl/models/qwen3/model_vllm_compat.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -288,7 +288,6 @@ def __init__(self, model_args: Qwen3ModelArgs):
288288
self.model_args = model_args
289289
self.vocab_size = model_args.vocab_size
290290
self.n_layers = model_args.n_layers
291-
self.eos_id = model_args.eos_id
292291
self.head_dim = model_args.head_dim
293292

294293
self.tok_embeddings = nn.Embedding(model_args.vocab_size, model_args.dim)

torchtitan/experiments/deterministic_vllm_rl/simple_rl.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -332,7 +332,6 @@ def load_model(checkpoint_path: str, model_path: str, use_vllm_compat: bool = Tr
332332
max_seq_len=getattr(hf_config, "max_position_embeddings", 32768),
333333
qk_norm=True,
334334
depth_init=True,
335-
eos_id=getattr(hf_config, "eos_token_id", 151645),
336335
)
337336

338337
# state_dict is in standard TorchTitan format (w1, w2, w3)

torchtitan/experiments/transformers_backend/model/args.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,6 @@ class HFTransformerModelArgs(PretrainedConfig, BaseModelArgs):
5454
"n_kv_heads": "num_key_value_heads",
5555
"norm_eps": "rms_norm_eps",
5656
"max_seq_len": "max_position_embeddings",
57-
"eos_id": "eos_token_id",
5857
}
5958
}
6059

torchtitan/models/llama3/model/args.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,6 @@ class TransformerModelArgs(BaseModelArgs):
4545

4646
use_flex_attn: bool = False
4747
attn_mask_type: str = "causal"
48-
eos_id: int = 0
4948

5049
def update_from_config(self, job_config: JobConfig, **kwargs) -> None:
5150
seq_len = job_config.training.seq_len

torchtitan/models/qwen3/model/args.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,6 @@ class Qwen3ModelArgs(BaseModelArgs):
3838

3939
use_flex_attn: bool = False
4040
attn_mask_type: str = "causal"
41-
eos_id: int = 151645
4241

4342
enable_weight_tying: bool = False
4443

torchtitan/models/qwen3/model/model.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -384,7 +384,6 @@ def __init__(self, model_args: Qwen3ModelArgs):
384384
self.model_args = model_args
385385
self.vocab_size = model_args.vocab_size
386386
self.n_layers = model_args.n_layers
387-
self.eos_id = model_args.eos_id
388387
self.head_dim = model_args.head_dim
389388

390389
self.tok_embeddings = nn.Embedding(model_args.vocab_size, model_args.dim)

0 commit comments

Comments
 (0)