wangyichen25 commited on
Commit
b02afb7
·
verified ·
1 Parent(s): cb81f33

Training in progress, step 40

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: medgemma-27b-it-sft-lora-hyperkvasir
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - sft
 
9
  licence: license
10
  ---
11
 
@@ -27,18 +27,18 @@ print(output["generated_text"])
27
 
28
  ## Training procedure
29
 
30
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/wangyichen25-mayo-clinic/medgemma_it_27b_hyperkvasir/runs/4gtu1vqh)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
- - TRL: 0.21.0
38
- - Transformers: 4.55.0
39
- - Pytorch: 2.8.0
40
- - Datasets: 4.0.0
41
- - Tokenizers: 0.21.4
42
 
43
  ## Citations
44
 
 
4
  model_name: medgemma-27b-it-sft-lora-hyperkvasir
5
  tags:
6
  - generated_from_trainer
 
7
  - sft
8
+ - trl
9
  licence: license
10
  ---
11
 
 
27
 
28
  ## Training procedure
29
 
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/wangyichen25-mayo-clinic/medgemma_it_27b_hyperkvasir/runs/mi06jcd2)
31
 
32
 
33
  This model was trained with SFT.
34
 
35
  ### Framework versions
36
 
37
+ - TRL: 0.25.0
38
+ - Transformers: 4.57.1
39
+ - Pytorch: 2.9.0
40
+ - Datasets: 4.4.1
41
+ - Tokenizers: 0.22.1
42
 
43
  ## Citations
44
 
adapter_config.json CHANGED
@@ -28,16 +28,16 @@
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
- "gate_proj",
32
- "down_proj",
33
- "k_proj",
34
- "q_proj",
35
- "up_proj",
36
  "out_proj",
37
- "fc2",
38
  "o_proj",
39
- "fc1",
40
- "v_proj"
 
 
 
 
 
41
  ],
42
  "target_parameters": null,
43
  "task_type": "CAUSAL_LM",
 
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
+ "fc1",
 
 
 
 
32
  "out_proj",
 
33
  "o_proj",
34
+ "up_proj",
35
+ "fc2",
36
+ "down_proj",
37
+ "q_proj",
38
+ "v_proj",
39
+ "gate_proj",
40
+ "k_proj"
41
  ],
42
  "target_parameters": null,
43
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7e1cec513b0bfae8cdad246461483cfedd353c44b4bcd5a2ee0d75da2b5a4120
3
  size 6127553104
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33bade23015f4045ee0a8d0d679bae5c4dc74148e0f11343dcf69b439959847a
3
  size 6127553104
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fd81184c4386bdd5320f1754d4cda79540e3bb45d4e9eeffadfdb4c17e09fef2
3
- size 6353
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c665ff9710ba066622bdc47a0845adeeeb156957d33148906e62f67561245a3f
3
+ size 6481