zzigakovacic commited on
Commit
c7f8a5d
·
verified ·
1 Parent(s): 3c7185b

Training in progress, step 5

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "Qwen/Qwen2-VL-7B-Instruct",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -24,8 +24,8 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "q_proj",
28
- "v_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "trainable_token_indices": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-VL-3B-Instruct",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "v_proj",
28
+ "q_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2b3123c05c4b21f60216711a910f152ba58ab1420ac51a5599cff3e8667ecbf
3
- size 10108960
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15798576e10c74c1e8417ea55774bbea128dc7cc6c335a93897b8977e58e21b5
3
+ size 7391728
added_tokens.json CHANGED
@@ -1,7 +1,14 @@
1
  {
 
 
2
  "<|box_end|>": 151649,
3
  "<|box_start|>": 151648,
4
  "<|endoftext|>": 151643,
 
 
 
 
 
5
  "<|im_end|>": 151645,
6
  "<|im_start|>": 151644,
7
  "<|image_pad|>": 151655,
@@ -9,6 +16,7 @@
9
  "<|object_ref_start|>": 151646,
10
  "<|quad_end|>": 151651,
11
  "<|quad_start|>": 151650,
 
12
  "<|video_pad|>": 151656,
13
  "<|vision_end|>": 151653,
14
  "<|vision_pad|>": 151654,
 
1
  {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
  "<|box_end|>": 151649,
5
  "<|box_start|>": 151648,
6
  "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
  "<|im_end|>": 151645,
13
  "<|im_start|>": 151644,
14
  "<|image_pad|>": 151655,
 
16
  "<|object_ref_start|>": 151646,
17
  "<|quad_end|>": 151651,
18
  "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
  "<|video_pad|>": 151656,
21
  "<|vision_end|>": 151653,
22
  "<|vision_pad|>": 151654,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:88a3a6fcb80132f76da8aa40cdc3fccd7e5d8468ef15421f5b0c2715e85217d2
3
- size 11420538
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5eee858c5123a4279c3e1f7b81247343f356ac767940b2692a928ad929543214
3
+ size 11422063
tokenizer_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "151643": {
@@ -112,6 +113,70 @@
112
  "rstrip": false,
113
  "single_word": false,
114
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
115
  }
116
  },
117
  "additional_special_tokens": [
@@ -130,13 +195,13 @@
130
  "<|video_pad|>"
131
  ],
132
  "bos_token": null,
 
133
  "clean_up_tokenization_spaces": false,
134
  "eos_token": "<|im_end|>",
135
  "errors": "replace",
136
  "extra_special_tokens": {},
137
- "model_max_length": 32768,
138
  "pad_token": "<|endoftext|>",
139
- "padding_side": "left",
140
  "split_special_tokens": false,
141
  "tokenizer_class": "Qwen2Tokenizer",
142
  "unk_token": null
 
1
  {
2
+ "add_bos_token": false,
3
  "add_prefix_space": false,
4
  "added_tokens_decoder": {
5
  "151643": {
 
113
  "rstrip": false,
114
  "single_word": false,
115
  "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
  }
181
  },
182
  "additional_special_tokens": [
 
195
  "<|video_pad|>"
196
  ],
197
  "bos_token": null,
198
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}",
199
  "clean_up_tokenization_spaces": false,
200
  "eos_token": "<|im_end|>",
201
  "errors": "replace",
202
  "extra_special_tokens": {},
203
+ "model_max_length": 131072,
204
  "pad_token": "<|endoftext|>",
 
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eea158ad0cf7dd1db17b54f22954488f6c3fcb8e2d288ea7e6a9471e7501cdca
3
- size 5816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03b1da72e1f5d6d3faea94007f26cb2c97acfb4180b577447f0c1a268deec5f8
3
+ size 5752