Update README.md
Browse files
README.md
CHANGED
|
@@ -123,11 +123,13 @@ The evaluation code is available in the [ASR Benchmark repository](https://githu
|
|
| 123 |
WER was computed **without punctuation or uppercase letters** and datasets were cleaned.
|
| 124 |
The [SUMM-RE dataset](https://huggingface.co/datasets/linagora/SUMM-RE) is the only one used **exclusively for evaluation**, meaning neither model saw it during training.
|
| 125 |
|
| 126 |
-
Evaluations can be very long (especially for whisper) so we used a subset of the test split for most datasets:
|
| 127 |
-
- 15% of CommonVoice
|
| 128 |
-
- 33% of MultiLingual LibriSpeech
|
| 129 |
-
- 33% of SUMM-RE
|
| 130 |
-
- 33% of VoxPopuli
|
|
|
|
|
|
|
| 131 |
|
| 132 |

|
| 133 |
|
|
|
|
| 123 |
WER was computed **without punctuation or uppercase letters** and datasets were cleaned.
|
| 124 |
The [SUMM-RE dataset](https://huggingface.co/datasets/linagora/SUMM-RE) is the only one used **exclusively for evaluation**, meaning neither model saw it during training.
|
| 125 |
|
| 126 |
+
Evaluations can be very long (especially for whisper) so we selected only segments with a duration over 1 second and used a subset of the test split for most datasets:
|
| 127 |
+
- 15% of CommonVoice: 2424 rows (3.9h)
|
| 128 |
+
- 33% of MultiLingual LibriSpeech: 800 rows (3.3h)
|
| 129 |
+
- 33% of SUMM-RE: 1004 rows (2h). We selected only segments above 4 seconds to ensure quality.
|
| 130 |
+
- 33% of VoxPopuli: 678 rows (1.6h)
|
| 131 |
+
- Multilingual TEDx: 972 rows (1.5h)
|
| 132 |
+
- 50% of our internal Youtube corpus: 956 rows (1h)
|
| 133 |
|
| 134 |

|
| 135 |
|