Skip to content

Commit f9323e3

Browse files
authored
fix rms norm import on non cuda device (deepspeedai#341)
1 parent b09c6a6 commit f9323e3

File tree

1 file changed

+2
-7
lines changed

1 file changed

+2
-7
lines changed

megatron/model/gpt_model.py

+2-7
Original file line numberDiff line numberDiff line change
@@ -14,16 +14,11 @@
1414
from .utils import init_method_normal
1515
from .utils import scaled_init_method_normal
1616

17-
from megatron.model import LayerNorm
17+
from megatron.model import LayerNorm, RMSNorm
1818
from .language_model import EmbeddingPipe
1919
from .transformer import ParallelTransformerLayerPipe, LMHeadPipe
2020
from deepspeed.pipe import PipelineModule, LayerSpec, TiedLayerSpec
2121

22-
try:
23-
from apex.normalization import MixedFusedRMSNorm
24-
except ImportError:
25-
MixedFusedRMSNorm = None
26-
2722
try:
2823
from deepspeed.checkpoint import (
2924
VOCABULARY_PARAMETER_PATTERNS,
@@ -290,7 +285,7 @@ def _to_float16(inputs):
290285
args.hidden_size,
291286
eps=args.layernorm_epsilon))
292287
else:
293-
self.specs.append(LayerSpec(MixedFusedRMSNorm, args.hidden_size, args.layernorm_epsilon))
288+
self.specs.append(LayerSpec(RMSNorm, args.hidden_size, args.layernorm_epsilon))
294289

295290
def _logits_helper(embedding, lm_output):
296291
"""A wrapper to massage inputs/outputs from pipeline. """

0 commit comments

Comments
 (0)