connectif-ddesotto commited on
Commit
65eb06e
·
verified ·
1 Parent(s): 8f5b4c5

connectif-ddesotto/google/gemma-3-12b-it-hermes-function-calling-with-think

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: google/gemma-3-12b-it
3
+ library_name: transformers
4
+ model_name: gemma-3-12b-it-hermes-function-calling-with-think
5
+ tags:
6
+ - generated_from_trainer
7
+ - sft
8
+ - trl
9
+ licence: license
10
+ ---
11
+
12
+ # Model Card for gemma-3-12b-it-hermes-function-calling-with-think
13
+
14
+ This model is a fine-tuned version of [google/gemma-3-12b-it](https://huggingface.co/google/gemma-3-12b-it).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
+
17
+ ## Quick start
18
+
19
+ ```python
20
+ from transformers import pipeline
21
+
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="connectif-ddesotto/gemma-3-12b-it-hermes-function-calling-with-think", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
26
+ ```
27
+
28
+ ## Training procedure
29
+
30
+
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
+ ### Framework versions
36
+
37
+ - TRL: 0.19.0
38
+ - Transformers: 4.53.0
39
+ - Pytorch: 2.7.1
40
+ - Datasets: 3.6.0
41
+ - Tokenizers: 0.21.2
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
adapter_config.json ADDED
@@ -0,0 +1,41 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "google/gemma-3-12b-it",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 64,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.05,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "r": 16,
24
+ "rank_pattern": {},
25
+ "revision": null,
26
+ "target_modules": [
27
+ "k_proj",
28
+ "up_proj",
29
+ "q_proj",
30
+ "embed_tokens",
31
+ "down_proj",
32
+ "v_proj",
33
+ "lm_head",
34
+ "gate_proj",
35
+ "o_proj"
36
+ ],
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_rslora": false
41
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a6bf72ac10fc539233b4280ae21f27474d3e52ac712b1942fa30619ce5fada8
3
+ size 4334672176
chat_template.jinja ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ {{ bos_token }}
3
+ {%- if messages[0]['role'] == 'system' -%}
4
+ {%- if messages[0]['content'] is string -%}
5
+ {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}
6
+ {%- else -%}
7
+ {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}
8
+ {%- endif -%}
9
+ {%- set loop_messages = messages[1:] -%}
10
+ {%- else -%}
11
+ {%- set first_user_prefix = "" -%}
12
+ {%- set loop_messages = messages -%}
13
+ {%- endif -%}
14
+ {%- for message in loop_messages -%}
15
+ {%- if (message['role'] == 'assistant') -%}
16
+ {%- set role = "model" -%}
17
+ {%- else -%}
18
+ {%- set role = message['role'] -%}
19
+ {%- endif -%}
20
+ {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else "") }}
21
+ {%- if message['content'] is string -%}
22
+ {{ message['content'] | trim }}
23
+ {%- elif message['content'] is iterable -%}
24
+ {%- for item in message['content'] -%}
25
+ {%- if item['type'] == 'image' -%}
26
+ {{ '<start_of_image>' }}
27
+ {%- elif item['type'] == 'text' -%}
28
+ {{ item['text'] | trim }}
29
+ {%- endif -%}
30
+ {%- endfor -%}
31
+ {%- else -%}
32
+ {{ raise_exception("Invalid content type") }}
33
+ {%- endif -%}
34
+ {{ '<end_of_turn>\n' }}
35
+ {%- endfor -%}
36
+ {%- if add_generation_prompt -%}
37
+ {{'<start_of_turn>model\n'}}
38
+ {%- endif -%}
runs/Jul01_11-56-05_3353a50ea6ea/events.out.tfevents.1751371098.3353a50ea6ea ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f01ee96e818e42088ae8795b94b739050c6a334c02e11bb0b120af14da586b1a
3
+ size 51886
special_tokens_map.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<tools>",
4
+ "</tools>",
5
+ "<think>",
6
+ "</think>",
7
+ "<tool_call>",
8
+ "</tool_call>",
9
+ "<tool_response>",
10
+ "</tool_response>",
11
+ "<pad>",
12
+ "<eos>"
13
+ ],
14
+ "boi_token": "<start_of_image>",
15
+ "bos_token": {
16
+ "content": "<bos>",
17
+ "lstrip": false,
18
+ "normalized": false,
19
+ "rstrip": false,
20
+ "single_word": false
21
+ },
22
+ "eoi_token": "<end_of_image>",
23
+ "eos_token": {
24
+ "content": "<eos>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "image_token": "<image_soft_token>",
31
+ "pad_token": "<pad>",
32
+ "unk_token": {
33
+ "content": "<unk>",
34
+ "lstrip": false,
35
+ "normalized": false,
36
+ "rstrip": false,
37
+ "single_word": false
38
+ }
39
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b725c35c59e9b553a5ba40416108694df696dcd075a6320decdab080b815917f
3
+ size 33386068
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab86b9b254ceac9595f85d0126d5e642e31e9150d7170857e68c31bc1bb5646f
3
+ size 6225