cipherpy commited on
Commit
0d16027
·
verified ·
1 Parent(s): 5eb849c

Training in progress, epoch 1

Browse files
README.md CHANGED
@@ -34,11 +34,11 @@ This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.15.2
38
- - Transformers: 4.50.0.dev0
39
- - Pytorch: 2.5.1
40
- - Datasets: 3.3.2
41
- - Tokenizers: 0.21.1
42
 
43
  ## Citations
44
 
@@ -49,7 +49,7 @@ Cite TRL as:
49
  ```bibtex
50
  @misc{vonwerra2022trl,
51
  title = {{TRL: Transformer Reinforcement Learning}},
52
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
53
  year = 2020,
54
  journal = {GitHub repository},
55
  publisher = {GitHub},
 
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.20.0
38
+ - Transformers: 4.54.1
39
+ - Pytorch: 2.7.1
40
+ - Datasets: 4.0.0
41
+ - Tokenizers: 0.21.4
42
 
43
  ## Citations
44
 
 
49
  ```bibtex
50
  @misc{vonwerra2022trl,
51
  title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
  year = 2020,
54
  journal = {GitHub repository},
55
  publisher = {GitHub},
adapter_config.json CHANGED
@@ -3,6 +3,7 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "google/gemma-3-4b-pt",
5
  "bias": "none",
 
6
  "eva_config": null,
7
  "exclude_modules": null,
8
  "fan_in_fan_out": false,
@@ -22,23 +23,25 @@
22
  "embed_tokens"
23
  ],
24
  "peft_type": "LORA",
 
25
  "r": 16,
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
 
29
  "down_proj",
30
- "out_proj",
31
- "o_proj",
32
  "gate_proj",
33
- "fc1",
 
34
  "v_proj",
35
- "up_proj",
36
  "k_proj",
37
- "q_proj",
38
- "fc2",
39
- "lm_head"
40
  ],
41
  "task_type": "CAUSAL_LM",
 
42
  "use_dora": false,
 
43
  "use_rslora": false
44
  }
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "google/gemma-3-4b-pt",
5
  "bias": "none",
6
+ "corda_config": null,
7
  "eva_config": null,
8
  "exclude_modules": null,
9
  "fan_in_fan_out": false,
 
23
  "embed_tokens"
24
  ],
25
  "peft_type": "LORA",
26
+ "qalora_group_size": 16,
27
  "r": 16,
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
+ "fc2",
32
  "down_proj",
 
 
33
  "gate_proj",
34
+ "o_proj",
35
+ "out_proj",
36
  "v_proj",
 
37
  "k_proj",
38
+ "up_proj",
39
+ "fc1",
40
+ "q_proj"
41
  ],
42
  "task_type": "CAUSAL_LM",
43
+ "trainable_token_indices": null,
44
  "use_dora": false,
45
+ "use_qalora": false,
46
  "use_rslora": false
47
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1861a30196eeaa272e703d4cfc770fee78893db97e178123fa3b1d72977be637
3
- size 2839124552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:892b2560b10d88a7c938d11c14b4c5d2c9cf4957439871ce591edeaaaf3ae14f
3
+ size 2839126480
chat_template.jinja ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {{ bos_token }}
2
+ {%- if messages[0]['role'] == 'system' -%}
3
+ {%- if messages[0]['content'] is string -%}
4
+ {%- set first_user_prefix = messages[0]['content'] + '
5
+
6
+ ' -%}
7
+ {%- else -%}
8
+ {%- set first_user_prefix = messages[0]['content'][0]['text'] + '
9
+
10
+ ' -%}
11
+ {%- endif -%}
12
+ {%- set loop_messages = messages[1:] -%}
13
+ {%- else -%}
14
+ {%- set first_user_prefix = "" -%}
15
+ {%- set loop_messages = messages -%}
16
+ {%- endif -%}
17
+ {%- for message in loop_messages -%}
18
+ {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}
19
+ {{ raise_exception("Conversation roles must alternate user/assistant/user/assistant/...") }}
20
+ {%- endif -%}
21
+ {%- if (message['role'] == 'assistant') -%}
22
+ {%- set role = "model" -%}
23
+ {%- else -%}
24
+ {%- set role = message['role'] -%}
25
+ {%- endif -%}
26
+ {{ '<start_of_turn>' + role + '
27
+ ' + (first_user_prefix if loop.first else "") }}
28
+ {%- if message['content'] is string -%}
29
+ {{ message['content'] | trim }}
30
+ {%- elif message['content'] is iterable -%}
31
+ {%- for item in message['content'] -%}
32
+ {%- if item['type'] == 'image' -%}
33
+ {{ '<start_of_image>' }}
34
+ {%- elif item['type'] == 'text' -%}
35
+ {{ item['text'] | trim }}
36
+ {%- endif -%}
37
+ {%- endfor -%}
38
+ {%- else -%}
39
+ {{ raise_exception("Invalid content type") }}
40
+ {%- endif -%}
41
+ {{ '<end_of_turn>
42
+ ' }}
43
+ {%- endfor -%}
44
+ {%- if add_generation_prompt -%}
45
+ {{'<start_of_turn>model
46
+ '}}
47
+ {%- endif -%}
runs/Jul31_15-24-50_gpu-dom-cmlre/events.out.tfevents.1753955693.gpu-dom-cmlre.1424966.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7db3fcdf360fd0d39fcd77c97f0db5a433b76168e449ef74ad9ccb37c1760560
3
+ size 7846
runs/Jul31_15-26-21_gpu-dom-cmlre/events.out.tfevents.1753955784.gpu-dom-cmlre.1426754.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcbb73f29ca35425d34c6ec3eabdafb6f799db2ffc96fdf062e7228a9b63f46e
3
+ size 7846
runs/Jul31_15-29-42_gpu-dom-cmlre/events.out.tfevents.1753955984.gpu-dom-cmlre.1428687.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:845843bb7e7a161b6928e428e9364a1ae01c422df518ca98d25bcb04eff766ea
3
+ size 7846
runs/Jul31_15-31-28_gpu-dom-cmlre/events.out.tfevents.1753956090.gpu-dom-cmlre.1429792.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a1ff35cd50926a29f93553562719bc12495296066245d5d98bbc7d8cbf3132d
3
+ size 118930
tokenizer_config.json CHANGED
@@ -51325,7 +51325,6 @@
51325
  },
51326
  "boi_token": "<start_of_image>",
51327
  "bos_token": "<bos>",
51328
- "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- endif -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n",
51329
  "clean_up_tokenization_spaces": false,
51330
  "eoi_token": "<end_of_image>",
51331
  "eos_token": "<eos>",
 
51325
  },
51326
  "boi_token": "<start_of_image>",
51327
  "bos_token": "<bos>",
 
51328
  "clean_up_tokenization_spaces": false,
51329
  "eoi_token": "<end_of_image>",
51330
  "eos_token": "<eos>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c3f21fcd695d1af2f3abaa74b859d72062324a61db29d36060a886342ba157e1
3
- size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c3fe3d814b409ccd0ea571291503f248b6d25dc00cd4e90cf76deb5fed98e8f
3
+ size 6225