guynich's picture
Upload README.md with huggingface_hub
62e2c8e verified
---
language:
- en
pretty_name: librispeech_asr_test_vad
tags:
- speech
license: cc-by-4.0
task_categories:
- text-classification
---
Voice Activity Detection (VAD) Test Dataset
This dataset is based on the `test.clean` and `test.other` splits from the
[librispeech_asr](https://huggingface.co/datasets/openslr/librispeech_asr)
corpus. It includes two binary labels:
- **speech**: Indicates presence of speech ([0, 1]), computed using a dynamic threshold method with background noise estimation and smoothing.
- **confidence**: A post-processing flag to optionally correct transient dropouts in speech. It is set to 1 by default, but switches to 0 for up to ~0.1 seconds (3 chunks of audio) following a transition from speech to silence. Approximately 7% of the `speech` labels in this dataset are `confidence` 0. The remaining 93% are `confidence` 1 enabling VAD testing.
The dataset has minimal background noise, making it suitable for mixing with
external noise samples to test VAD robustness.
## Example data
A plot for an example showing audio samples and the `speech` feature.
<img src="assets/test_other_item_02.png" alt="Example from test.other"/>
The example below shows brief dropouts in the `speech` feature during natural
short pauses of quiet in the talker's speech. Since some VAD models may react
more slowly, the `confidence` feature offers a way to optionally ignore these
transient droputs when evaluating performance.
<img src="assets/test_clean_item_02.png" alt="Example from test.other"/>
# Example usage of dataset
The VAD model under test must support processing a chunk size of 512 audio
samples at 16000 Hz generating a prediction for each `speech` feature.
```console
import datasets
import numpy as np
from sklearn.metrics import roc_auc_score
dataset = datasets.load_dataset("guynich/librispeech_asr_test_vad")
audio = dataset["test.clean"][0]["audio"]["array"]
speech = dataset["test.clean"][0]["speech"]
# Compute voice activity probabilities
speech_probs = vad_model(audio)
# Add test code here
roc_auc = roc_auc_score(speech, speech_probs)
```
In practice you would run the AUC computation across the entire test split.
## Ignore transient dropouts
The `confidence` values can be used to filter the data. Removing zero confidence
values excludes 6.8% of the dataset and causes numerical increase in
computed precision. This compensates for slower moving voice activity decisions
as encountered in real-world applications.
```console
confidence = dataset["test.clean"][0]["confidence"]
speech_array = np.array(speech)
speech_probs_array = np.array(speech_probs)
roc_auc_confidence = roc_auc_score(
speech_array[np.array(confidence) == 1],
speech_probs_array[np.array(confidence) == 1],
)
```
# Model evaluation example
Example AUC plots computed for
[Silero VAD](https://github.com/snakers4/silero-vad?tab=readme-ov-file)
model with `test.clean` split.
<img src="assets/roc_test_clean.png" alt="Example from test.clean with Silero-VAD"/>
Precision values are increased when data is sliced by `confidence` values.
These low-confidence `speech` labels are flagged rather than removed, allowing
users to either exclude them (as shown here) or handle them with other methods.
<img src="assets/roc_test_clean_exclude_low_confidence.png" alt="Example from test.clean with Silero-VAD"/>
# License Information
This derivative dataset retains the same license as the source dataset
[librispeech_asr](https://huggingface.co/datasets/openslr/librispeech_asr).
[CC BY 4.0](https://creativecommons.org/licenses/by/4.0/)
# Citation Information
Labels contributed by Guy Nicholson were added to the following dataset.
```
@inproceedings{panayotov2015librispeech,
title={Librispeech: an ASR corpus based on public domain audio books},
author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},
booktitle={Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on},
pages={5206--5210},
year={2015},
organization={IEEE}
}
```