Update README.md
Browse files
README.md
CHANGED
|
@@ -19,7 +19,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 19 |
|
| 20 |
# libri-alpha-0.85-Temp-1-processor-change
|
| 21 |
|
| 22 |
-
This model is a
|
| 23 |
It achieves the following results on the evaluation set:
|
| 24 |
- Loss: 78.4467
|
| 25 |
- Wer: 0.1153
|
|
@@ -37,6 +37,7 @@ More information needed
|
|
| 37 |
More information needed
|
| 38 |
|
| 39 |
## Training procedure
|
|
|
|
| 40 |
|
| 41 |
### Training hyperparameters
|
| 42 |
|
|
@@ -52,6 +53,9 @@ The following hyperparameters were used during training:
|
|
| 52 |
- lr_scheduler_warmup_steps: 500
|
| 53 |
- num_epochs: 30
|
| 54 |
- mixed_precision_training: Native AMP
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
### Training results
|
| 57 |
|
|
|
|
| 19 |
|
| 20 |
# libri-alpha-0.85-Temp-1-processor-change
|
| 21 |
|
| 22 |
+
This model is a distilled version of [Wav2vec2](https://huggingface.co/) on the 30% of the Librispeech-clean.100 dataset.
|
| 23 |
It achieves the following results on the evaluation set:
|
| 24 |
- Loss: 78.4467
|
| 25 |
- Wer: 0.1153
|
|
|
|
| 37 |
More information needed
|
| 38 |
|
| 39 |
## Training procedure
|
| 40 |
+
Knowledge distillation of Wav2vec2-base-960h teacher model with 6 attention layers for student model.
|
| 41 |
|
| 42 |
### Training hyperparameters
|
| 43 |
|
|
|
|
| 53 |
- lr_scheduler_warmup_steps: 500
|
| 54 |
- num_epochs: 30
|
| 55 |
- mixed_precision_training: Native AMP
|
| 56 |
+
- alpha: 0.75(ignore name of repo)
|
| 57 |
+
- temperature: 1
|
| 58 |
+
|
| 59 |
|
| 60 |
### Training results
|
| 61 |
|