Skip to content

Commit 7b73061

Browse files
committed
fix
1 parent 4854f51 commit 7b73061

File tree

1 file changed

+3
-1
lines changed

1 file changed

+3
-1
lines changed

public/content/pretrain-llm-with-nvfp4/pretrain-llms-with-fp4-content.md

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,6 +15,8 @@ tags:
1515

1616
The growing scale of Large Language Models (LLMs) necessitates more efficient training methods. While 8-bit floating point (FP8) training is widely adopted, 4-bit floating point (FP4) formats offer further improvements in computational speed and memory usage. This guide provides a technical summary of **NVFP4**, a 4-bit format from NVIDIA, and the methodology required for its successful implementation in LLM pretraining.
1717

18+
**Architecture Note:** This guide is based on experiments with the **Mamba-Transformer** architecture, which combines Mamba state-space models and Transformer components.
19+
1820
## Background: Key Concepts in Numerical Precision
1921

2022
Before diving into NVFP4, it's essential to understand a few foundational concepts.
@@ -80,7 +82,7 @@ Achieving training outcomes comparable to FP8 requires a specific set of techniq
8082
Quantizing the entire model to FP4 can lead to divergence (model stops learning). A mixed-precision approach is crucial for stability.
8183

8284
![NVFP4 Quantized Linear Layer Compute Flow](/content/pretrain-llm-with-nvfp4/images/NVFP4_quantized_linear_layer_compute_flow.png)
83-
*Figure 5: Illustration of compute flow for a NVFP4 quantized linear layer. All GEMM operations quantize their inputs to NVFP4.*
85+
*Figure 5: Illustration of compute flow for a NVFP4 quantized linear layer. All GEMM operations quantize their inputs to NVFP4.* - understanding this image will require deeper analysis and detailed understanding of the paper
8486

8587

8688
**Implementation:**

0 commit comments

Comments
 (0)