openvoice plugin
Browse files
.ipynb_checkpoints/README-checkpoint.md
CHANGED
|
@@ -23,7 +23,7 @@ For more details, please check our interspeech paper: [DreamVoice](https://arxiv
|
|
| 23 |
To listen to demos and download dataset, please check dreamvoice's homepage: [Homepage](https://haidog-yaqub.github.io/dreamvoice_demo/)
|
| 24 |
|
| 25 |
|
| 26 |
-
#
|
| 27 |
|
| 28 |
To load the models, you need to install packages:
|
| 29 |
|
|
@@ -91,6 +91,24 @@ gen_audio2, sr = dreamvoice.simplevc('examples/test2.wav', use_spk_cache=True)
|
|
| 91 |
dreamvoice.save_audio('gen2.wav', gen_audio2, sr)
|
| 92 |
```
|
| 93 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 94 |
- End-to-end DreamVoice VC Model
|
| 95 |
|
| 96 |
```python
|
|
|
|
| 23 |
To listen to demos and download dataset, please check dreamvoice's homepage: [Homepage](https://haidog-yaqub.github.io/dreamvoice_demo/)
|
| 24 |
|
| 25 |
|
| 26 |
+
# How to Use
|
| 27 |
|
| 28 |
To load the models, you need to install packages:
|
| 29 |
|
|
|
|
| 91 |
dreamvoice.save_audio('gen2.wav', gen_audio2, sr)
|
| 92 |
```
|
| 93 |
|
| 94 |
+
# Training Guide
|
| 95 |
+
|
| 96 |
+
1. download VCTK and LibriTTS-R
|
| 97 |
+
2. extract speaker embeddings and cache in local path:
|
| 98 |
+
```
|
| 99 |
+
# example code
|
| 100 |
+
dreamvoice/train_utils/prepare/prepare_se.py
|
| 101 |
+
```
|
| 102 |
+
3. modify trainning config and train your dreamvoice plugin:
|
| 103 |
+
```
|
| 104 |
+
# example code
|
| 105 |
+
cd dreamvoice/train_utils/src
|
| 106 |
+
accelerate launch train.py
|
| 107 |
+
```
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
# Extra Features
|
| 111 |
+
|
| 112 |
- End-to-end DreamVoice VC Model
|
| 113 |
|
| 114 |
```python
|
.ipynb_checkpoints/update-checkpoint.py
DELETED
|
@@ -1,26 +0,0 @@
|
|
| 1 |
-
from huggingface_hub import HfApi, HfFolder, Repository
|
| 2 |
-
|
| 3 |
-
# Authenticate with Hugging Face Hub
|
| 4 |
-
api = HfApi()
|
| 5 |
-
token = HfFolder.get_token()
|
| 6 |
-
if token is None:
|
| 7 |
-
raise ValueError("Hugging Face token is not set. Please authenticate first.")
|
| 8 |
-
|
| 9 |
-
# Create or clone the private repository using the token
|
| 10 |
-
repo = Repository(
|
| 11 |
-
local_dir="DreamVoice",
|
| 12 |
-
clone_from="myshell-ai/DreamVoice",
|
| 13 |
-
use_auth_token=token
|
| 14 |
-
)
|
| 15 |
-
|
| 16 |
-
repo.lfs_track(".pt")
|
| 17 |
-
repo.lfs_track(".png")
|
| 18 |
-
|
| 19 |
-
# Add all changes to git
|
| 20 |
-
repo.git_add()
|
| 21 |
-
|
| 22 |
-
# Commit the changes with a message
|
| 23 |
-
repo.git_commit("Initial commit")
|
| 24 |
-
|
| 25 |
-
# Push the changes to the remote repository
|
| 26 |
-
repo.git_push()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
CHANGED
|
@@ -23,7 +23,7 @@ For more details, please check our interspeech paper: [DreamVoice](https://arxiv
|
|
| 23 |
To listen to demos and download dataset, please check dreamvoice's homepage: [Homepage](https://haidog-yaqub.github.io/dreamvoice_demo/)
|
| 24 |
|
| 25 |
|
| 26 |
-
#
|
| 27 |
|
| 28 |
To load the models, you need to install packages:
|
| 29 |
|
|
@@ -91,6 +91,24 @@ gen_audio2, sr = dreamvoice.simplevc('examples/test2.wav', use_spk_cache=True)
|
|
| 91 |
dreamvoice.save_audio('gen2.wav', gen_audio2, sr)
|
| 92 |
```
|
| 93 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 94 |
- End-to-end DreamVoice VC Model
|
| 95 |
|
| 96 |
```python
|
|
|
|
| 23 |
To listen to demos and download dataset, please check dreamvoice's homepage: [Homepage](https://haidog-yaqub.github.io/dreamvoice_demo/)
|
| 24 |
|
| 25 |
|
| 26 |
+
# How to Use
|
| 27 |
|
| 28 |
To load the models, you need to install packages:
|
| 29 |
|
|
|
|
| 91 |
dreamvoice.save_audio('gen2.wav', gen_audio2, sr)
|
| 92 |
```
|
| 93 |
|
| 94 |
+
# Training Guide
|
| 95 |
+
|
| 96 |
+
1. download VCTK and LibriTTS-R
|
| 97 |
+
2. extract speaker embeddings and cache in local path:
|
| 98 |
+
```
|
| 99 |
+
# example code
|
| 100 |
+
dreamvoice/train_utils/prepare/prepare_se.py
|
| 101 |
+
```
|
| 102 |
+
3. modify trainning config and train your dreamvoice plugin:
|
| 103 |
+
```
|
| 104 |
+
# example code
|
| 105 |
+
cd dreamvoice/train_utils/src
|
| 106 |
+
accelerate launch train.py
|
| 107 |
+
```
|
| 108 |
+
|
| 109 |
+
|
| 110 |
+
# Extra Features
|
| 111 |
+
|
| 112 |
- End-to-end DreamVoice VC Model
|
| 113 |
|
| 114 |
```python
|
dreamvoice/train_utils/prepare/prepare_se.py
CHANGED
|
@@ -92,10 +92,10 @@ if __name__ == '__main__':
|
|
| 92 |
model = ToneColorConverter(f'{ckpt_converter}/config.json', device=device)
|
| 93 |
model.load_ckpt(f'{ckpt_converter}/checkpoint.pth')
|
| 94 |
|
| 95 |
-
input_folder = '/
|
| 96 |
output_folder = 'spk/VCTK-Corpus/'
|
| 97 |
process_audio_folder(input_folder, output_folder, model, device)
|
| 98 |
|
| 99 |
-
input_folder = '/
|
| 100 |
output_folder = 'spk/LibriTTS-R/train-clean-360/'
|
| 101 |
process_audio_folder(input_folder, output_folder, model, device)
|
|
|
|
| 92 |
model = ToneColorConverter(f'{ckpt_converter}/config.json', device=device)
|
| 93 |
model.load_ckpt(f'{ckpt_converter}/checkpoint.pth')
|
| 94 |
|
| 95 |
+
input_folder = 'your_path/VCTK-Corpus/'
|
| 96 |
output_folder = 'spk/VCTK-Corpus/'
|
| 97 |
process_audio_folder(input_folder, output_folder, model, device)
|
| 98 |
|
| 99 |
+
input_folder = 'your_path/LibriTTS-R/train-clean-360'
|
| 100 |
output_folder = 'spk/LibriTTS-R/train-clean-360/'
|
| 101 |
process_audio_folder(input_folder, output_folder, model, device)
|
dreamvoice/train_utils/src/{train_light.py → train.py}
RENAMED
|
File without changes
|