|
|
--- |
|
|
language: |
|
|
- en |
|
|
pretty_name: librispeech_asr_test_vad |
|
|
tags: |
|
|
- speech |
|
|
license: cc-by-4.0 |
|
|
task_categories: |
|
|
- text-classification |
|
|
--- |
|
|
|
|
|
# librispeech_asr_test_vad |
|
|
|
|
|
A dataset for testing voice activity detection (VAD). |
|
|
|
|
|
This dataset uses test splits [`test.clean`, `test.other`] extracted |
|
|
from the |
|
|
[`librispeech_asr` dataset](https://huggingface.co/datasets/openslr/librispeech_asr). |
|
|
|
|
|
There are two additional features. |
|
|
|
|
|
1. Binary classification of speech activity, called `speech`. These binary values [0, 1] were computed from speech audio samples using a dynamic threshold method with background noise estimation and smoothing. |
|
|
|
|
|
2. Binary classification of confidence, called `confidence`. These binary values [0, 1] are computed as follows. The default confidence is 1. After a `speech` transition from 1 to 0 the confidence is set to 0 up to a maximum of three 0s in `speech` (approximately 0.1 second). This can be used to correct for temporary blips in the `speech` feature and unknown decay in the method under test. |
|
|
|
|
|
This test dataset has little background noise thus enables mixing with noise |
|
|
samples to assess voice activity detection robustness. |
|
|
|
|
|
## Example data |
|
|
|
|
|
A plot for an example showing audio samples and the `speech` feature. |
|
|
|
|
|
<img src="assets/test_other_item_02.png" alt="Example from test.other"/> |
|
|
|
|
|
The following example demonstrates short zero blips in the `speech` feature for |
|
|
valid short pauses in the talker's speech. However a VAD model under test may |
|
|
have slower reaction time. The `confidence` feature provides an optional means |
|
|
for reducing the impact of these short zero blips when computing metrics for a |
|
|
method under test. |
|
|
|
|
|
<img src="assets/test_clean_item_02.png" alt="Example from test.other"/> |
|
|
|
|
|
# Example usage of dataset |
|
|
|
|
|
The model under test must support processing a chunk size of 512 audio samples |
|
|
at 16000 Hz generating a prediction for each `speech` feature. |
|
|
|
|
|
```console |
|
|
import datasets |
|
|
import numpy as np |
|
|
from sklearn.metrics import roc_auc_score |
|
|
|
|
|
dataset = datasets.load_dataset("guynich/librispeech_asr_test_vad") |
|
|
|
|
|
audio = dataset["test.clean"][0]["audio"]["array"] |
|
|
speech = dataset["test.clean"][0]["speech"] |
|
|
|
|
|
# Compute probabilities from model under test (block size 512). |
|
|
speech_probs = model_under_test(audio) |
|
|
|
|
|
# Add test code here such as AUC metrics. |
|
|
# In practice you would run this across the entire test split. |
|
|
roc_auc = roc_auc_score(speech, speech_probs) |
|
|
``` |
|
|
|
|
|
The confidence values can be used to slice the data. This removes 6.8% of the |
|
|
entire dataset `speech` features and removing these low confidence values |
|
|
increases precision. |
|
|
```console |
|
|
confidence = dataset["test.clean"][0]["confidence"] |
|
|
|
|
|
speech_array = np.array(speech) |
|
|
speech_probs_array = np.array(speech_probs) |
|
|
|
|
|
roc_auc_confidence = roc_auc_score( |
|
|
speech_array[np.array(confidence) == 1], |
|
|
speech_probs_array[np.array(confidence) == 1], |
|
|
) |
|
|
``` |
|
|
|
|
|
# Silero-VAD model testing |
|
|
|
|
|
Example AUC plots computed for Silero-VAD model model with `test.clean` split. |
|
|
|
|
|
<img src="assets/roc_test_clean.png" alt="Example from test.clean with Silero-VAD"/> |
|
|
|
|
|
Precision values are increased when data is sliced by confidence values. |
|
|
|
|
|
<img src="assets/roc_test_clean_exclude_low_confidence.png" alt="Example from test.clean with Silero-VAD"/> |
|
|
|
|
|
# License Information |
|
|
|
|
|
This dataset retains the same license as the source dataset. |
|
|
|
|
|
[CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) |
|
|
|
|
|
|
|
|
# Citation Information |
|
|
|
|
|
Derived from this dataset. |
|
|
``` |
|
|
@inproceedings{panayotov2015librispeech, |
|
|
title={Librispeech: an ASR corpus based on public domain audio books}, |
|
|
author={Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev}, |
|
|
booktitle={Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on}, |
|
|
pages={5206--5210}, |
|
|
year={2015}, |
|
|
organization={IEEE} |
|
|
} |
|
|
``` |
|
|
|