Skip to content

Commit 7faa750

Browse files
Qkd004LittleHeroZZZX
authored andcommitted
[CodeStyle][Typos][L-6,L-8] Fix typo("lengths","likely") (#7616)
* fix:'lengths' and 'likely' * fix:'likely' * fix:'lengths'
1 parent fb10baf commit 7faa750

File tree

3 files changed

+2
-4
lines changed

3 files changed

+2
-4
lines changed

_typos.toml

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -69,11 +69,9 @@ intput = "intput"
6969
lable = "lable"
7070
learing = "learing"
7171
legth = "legth"
72-
lengthes = "lengthes"
7372
lenth = "lenth"
7473
leran = "leran"
7574
libary = "libary"
76-
likey = "likey"
7775
mantained = "mantained"
7876
matrics = "matrics"
7977
mdule = "mdule"

docs/design/dist_train/distributed_training_review.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ Synchronous training usually faces scalability and performance issues, if not ca
3030
Similar to asynchronous training, the benefit of synchronous training diminishes quickly. Depending on the models, increasing the number of trainers (effectively batch size) beyond a point won’t delivers faster converge time or better final model quality.
3131

3232
# Codistillation
33-
Codistillation is a technique that tries to scale the training further. A few training instance (each training instance can be distributed) are performed during the same period. Each training instance has extra losses that comes from the prediction of other training instances. (likey teacher and student) The training process converges faster and usually converge to a better model quality. [4]
33+
Codistillation is a technique that tries to scale the training further. A few training instance (each training instance can be distributed) are performed during the same period. Each training instance has extra losses that comes from the prediction of other training instances. (likely teacher and student) The training process converges faster and usually converge to a better model quality. [4]
3434

3535

3636
# Reference

docs/design/network/deep_speech_2.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -127,7 +127,7 @@ Key ingredients about the layers:
127127
- **Data Layers**:
128128
- Frame sequences data of audio **spectrogram** (with FFT).
129129
- Token sequences data of **transcription** text (labels).
130-
- These two type of sequences do not have the same lengthes, thus a CTC-loss layer is required.
130+
- These two type of sequences do not have the same lengths, thus a CTC-loss layer is required.
131131
- **2D Convolution Layers**:
132132
- Not only temporal convolution, but also **frequency convolution**. Like a 2D image convolution, but with a variable dimension (i.e. temporal dimension).
133133
- With striding for only the first convlution layer.

0 commit comments

Comments
 (0)