xingjian-bosonai commited on
Commit
a30e8ac
·
verified ·
1 Parent(s): 1bd2a55

Update README.md

Browse files

Add majority of README content.

Files changed (1) hide show
  1. README.md +146 -3
README.md CHANGED
@@ -1,3 +1,146 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ - zh
6
+ - de
7
+ - ko
8
+ ---
9
+
10
+ # HiggsAudio-V2: Redefining Expressiveness in Audio Generation
11
+
12
+ We are open-sourcing HiggsAudio-V2, a powerful audio foundation model pretrained on over 10 million hours of audio data and a diverse set of text data.
13
+ Despite having no post-training or fine-tuning, HiggsAudio-V2 excels in expressive audio generation, thanks to its deep language and acoustic understanding.
14
+
15
+ On [EmergentTTS-Eval](https://github.com/boson-ai/emergenttts-eval-public), the model achieves win rates of **75.7%** and **55.7%** over "gpt-4o-mini-tts" on the "Emotions" and "Questions" categories, respectively. It also obtains state-of-the-art performance on traditional TTS benchmarks like Seed-TTS Eval and Emotional Speech Dataset (ESD). Moreover, the model demonstrates capabilities rarely seen in previous systems, including automatic prosody adaptation during narration, zero-shot generation of natural multi-speaker dialogues in multiple languages, melodic humming with the cloned voice, and simultaneous generation of speech and background music.
16
+ Check our open-source repository https://github.com/boson-ai/higgs-audio for more details.
17
+
18
+
19
+ <p>
20
+ <img src="./emergent-tts-emotions-win-rate.png" width=900>
21
+ </p>
22
+
23
+
24
+ ## Technical Details
25
+
26
+ <p>
27
+ <img src="./higgs_audio_v2_architecture_combined.png" width=900>
28
+ </p>
29
+
30
+ HiggsAudio-V2 adopts the "generation variant" depicted in the architecture figure above. Its strong performance is driven by three key technical innovations:
31
+
32
+ - We developed an automated annotation pipeline that leverages multiple ASR models, sound event classification models, and our in-house audio understanding model. Using this pipeline, we cleaned and annotated 10 million hours audio data, which we refer to as AudioVerse. The in-house understanding model is finetuned on top of Higgs Audio V1 Understanding, which adopts the "understanding variant" shown in the architecture figure.
33
+ - We trained a unified audio tokenizer from scratch that captures both semantic and acoustic features.
34
+ - We proposed the DualFFN architecture, which enhances the LLM’s ability to model acoustics tokens with minimal computational overhead.
35
+
36
+
37
+ ### Audio Tokenizer
38
+
39
+ <p>
40
+ <img src="./higgs_audio_tokenizer_architecture.png" width=900>
41
+ </p>
42
+
43
+ We introduce a new discretized audio tokenizer that runs at just 25 frames per second while keeping—or even improving—audio quality compared to tokenizers with twice the bitrate.
44
+ Our model is the first to train on 24 kHz data covering speech, music, and sound events in one unified system.
45
+ It also uses a simple non-diffusion encoder/decoder for fast, batch inference. It achieves state-of-the-art performance in semantic and acoustic evaluations.
46
+ Check https://huggingface.co/bosonai/higgs-audio-v2-tokenizer-staging for more information about the tokenizer.
47
+
48
+ ### Model Architecture -- Dual FFN
49
+
50
+ HiggsAudio-V2 is built on top of [Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B). To enhance the model’s ability to process audio tokens,
51
+ we incorporate the "DualFFN" architecture as an audio adapter.
52
+ DualFFN acts as an audio-specific expert, boosting the LLM's performance with minimal computational overhead.
53
+ Our implementation preserves 91% of the original LLM’s training speed with the inclusion of DualFFN, which has 2.2B parameters.
54
+ Thus, the total number of parameter for HiggsAudio-V2 is 3.6B (LLM) + 2.2B (Audio Dual FFN), and it has the same training / inference FLOPs as Llama-3.2-3B.
55
+ Ablation study shows that the model equipped with DualFFN consistently outperforms its counterpart in terms of word error rate (WER) and speaker similarity.
56
+ See [Higgs-Audio Architecture Blog](https://github.com/boson-ai/higgs-audio/tech_blogs/ARCHITECTURE_BLOG.md) for more information.
57
+
58
+
59
+ ## Evaluation
60
+
61
+ Here's the performance of HiggsAudio-V2 on four benchmarks, [Seed-TTS Eval](https://github.com/BytedanceSpeech/seed-tts-eval), [Emotional Speech Dataset (ESD)](https://paperswithcode.com/dataset/esd), [EmergentTTS-Eval](https://arxiv.org/abs/2505.23009), and Multi-speaker Eval:
62
+
63
+ #### Seed-TTS Eval & ESD
64
+
65
+ We prompt HiggsAudio-V2 with `<ref_text, ref_audio, text>` for zero-shot TTS. We adopt the standard evaluation metric in Seed-TTS Eval and ESD.
66
+
67
+ | | SeedTTS-Eval| | ESD | |
68
+ |------------------------------|--------|--------|---------|-------------------|
69
+ | | WER ↓ | SIM ↑ | WER ↓ | SIM (emo2vec) ↑ |
70
+ | Cosyvoice2 | 2.28 | 65.49 | 2.71 | 80.48 |
71
+ | Qwen2.5-omni† | 2.33 | 64.10 | - | - |
72
+ | ElevenLabs Multilingual V2 | **1.43** | 50.00 | 1.66 | 65.87 |
73
+ | HiggsAudio V1 | 2.18 | 66.27 | **1.49** | 82.84 |
74
+ | HiggsAudio V2 (base) | 2.44 | **67.70** | 1.78 | **86.13** |
75
+
76
+
77
+ #### EmergentTTS-Eval ("Emotions" and "Questions")
78
+
79
+ Following the [EmergentTTS-Eval Paper](https://arxiv.org/abs/2505.23009), we report the win-rate over "gpt-4o-mini-tts" with the "alloy" voice. For HiggsAudio-V2, we report its performance by shallow-cloning the ["belinda"](examples/voice_prompts/belinda.wav) voice.
80
+
81
+ | Model | Emotions (%) ↑ | Questions (%) ↑ |
82
+ |------------------------------------|--------------|----------------|
83
+ | HiggsAudio-V2 (base) | **75.71%** | **55.71%** |
84
+ | [gpt-4o-audio-preview†](https://platform.openai.com/docs/models/gpt-4o-audio-preview) | 61.64% | 47.85% |
85
+ | [Hume.AI](https://www.hume.ai/research) | 61.60% | 43.21% |
86
+ | **BASELINE:** [gpt-4o-mini-tts](https://platform.openai.com/docs/models/gpt-4o-mini-tts) | 50.00% | 50.00% |
87
+ | [Qwen 2.5 Omni†](https://github.com/QwenLM/Qwen2.5-Omni) | 41.60% | 51.78% |
88
+ | [minimax/speech-02-hd](https://replicate.com/minimax/speech-02-hd) | 40.86% | 47.32% |
89
+ | [ElevenLabs Multilingual v2](https://elevenlabs.io/blog/eleven-multilingual-v2) | 30.35% | 39.46% |
90
+ | [DeepGram Aura-2](https://deepgram.com/learn/introducing-aura-2-enterprise-text-to-speech) | 29.28% | 48.21% |
91
+ | [Sesame csm-1B](https://github.com/SesameAILabs/csm) | 15.96% | 31.78% |
92
+
93
+ <sup><sub>'†' means using the strong-prompting method described in the paper.</sub></sup>
94
+
95
+
96
+ #### Multi-speaker Eval
97
+
98
+ We also designed a multi-speaker evaluation benchmark to evaluate the capability of HiggsAudio-V2 for multi-speaker dialog generation. The benchmark contains three subsets
99
+
100
+ - `two-speaker-conversation`: 1000 synthetic dialogues involving two speakers. It contains two reference audio clips to evaluate the model’s ability in double voice cloning.
101
+ - `small talk`: 250 synthetic dialogues characterized by short utterances and a limited number of turns (4–6). It also contains two reference audio clips to test double voice cloning, though the dialogues are shorter and simpler than those in two-speaker-conversation.
102
+ - `small talk (no ref)`: 250 synthetic dialogues, also with short utterances and 4–6 turns. Unlike the other subsets, it does not include reference audio and is designed to evaluate the model’s ability to automatically assign appropriate voices to speakers.
103
+
104
+
105
+ We evaluate the word-error-rate (WER) and the geometric mean between intra-speaker similarity and inter-speaker dis-similarity on these three subsets. Other than HiggsAudio-V2, we also evaluated [MoonCast](https://github.com/jzq2000/MoonCast) and [nari-labs/dia](https://github.com/nari-labs/dia). Results are summarized in the following table. See more details in the [multi-speaker evaluation blog](PLACEHOLDER).
106
+
107
+ | | two-speaker-conversation | |small talk | | small talk (no ref) | |
108
+ | ---------------------------------------------- | -------------- | ------------------ | ---------- | -------------- | ------------------- | -------------- |
109
+ | | WER ↓ | Mean Sim & Dis-sim ↑ | WER ↓ | Mean Sim & Dis-sim ↑ | WER ↓ | Mean Sim & Dis-sim ↑ |
110
+ | [MoonCast](https://github.com/jzq2000/MoonCast) | 38.77 | 46.02 | **8.33** | 63.68 | 24.65 | 53.94 |
111
+ | [nari-labs/dia](https://github.com/nari-labs/dia) | \- | \- | 17.62 | 63.15 | 19.46 | **61.14** |
112
+ | HiggsAudio-V2 (base) | **18.88** | **51.95** | 11.89 | **67.92** | **14.65** | 55.28 |
113
+
114
+
115
+
116
+ ## Get Started
117
+
118
+ You need to first install the [higgs-audio codebase](https://github.com/boson-ai/higgs-audio):
119
+
120
+ ```bash
121
+ git clone https://github.com/boson-ai/higgs-audio.git
122
+
123
+ cd higgs-audio
124
+ python3 -m venv higgs_audio_env
125
+ source higgs_audio_env/bin/activate
126
+ pip install -r requirements.txt
127
+ pip install -e .
128
+ ```
129
+
130
+ Afterwards, you can launch generation examples via the `generation.py` provided in the repository.
131
+
132
+ ```bash
133
+
134
+ python3 examples/generation.py \
135
+ --transcript examples/transcript/single_speaker/en_basic.txt \
136
+ --ref_audio belinda \
137
+ --seed 12345
138
+
139
+ ```
140
+
141
+ Alternatively, here's a python script that you can try to convert text to speech after you have installed higgs-audio.
142
+
143
+ ## License
144
+
145
+ TBA
146
+