adricl commited on
Commit
18dc639
·
verified ·
1 Parent(s): 1aece6b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -1
README.md CHANGED
@@ -9,11 +9,36 @@ tags:
9
  - music
10
  - instrument
11
  pipeline_tag: audio-to-audio
 
 
 
12
  ---
13
 
 
14
 
15
  This model is trained from scratch using tokenized midi music.
16
  I have trained a MidiTok tokeniser (REMI) and its made by spliting multi-track midi into a single track.
17
 
18
  We then trained in on a small dataset.
19
- Its using the Mistral model that has been cut down quite a bit.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - music
10
  - instrument
11
  pipeline_tag: audio-to-audio
12
+ model-index:
13
+ - name: Mistral_MidiTok_Transformer_Single_Instrument_Small
14
+ results: []
15
  ---
16
 
17
+ # Mistral_MidiTok_Transformer_Single_Instrument_Small
18
 
19
  This model is trained from scratch using tokenized midi music.
20
  I have trained a MidiTok tokeniser (REMI) and its made by spliting multi-track midi into a single track.
21
 
22
  We then trained in on a small dataset.
23
+ Its using the Mistral model that has been cut down quite a bit.
24
+
25
+ ### Training hyperparameters
26
+
27
+ The following hyperparameters were used during training:
28
+ - learning_rate: 0.0001
29
+ - train_batch_size: 30
30
+ - eval_batch_size: 30
31
+ - seed: 444
32
+ - gradient_accumulation_steps: 3
33
+ - total_train_batch_size: 90
34
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
35
+ - lr_scheduler_type: cosine_with_restarts
36
+ - lr_scheduler_warmup_ratio: 0.3
37
+ - training_steps: 20000
38
+
39
+ ### Framework versions
40
+
41
+ - Transformers 4.46.2
42
+ - Pytorch 2.1.0+cu121
43
+ - Datasets 3.1.0
44
+ - Tokenizers 0.20.3