Paradoxis commited on
Commit
e967700
·
verified ·
1 Parent(s): 28007a1

Training in progress, step 500

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: Qwen2.5-VL-3B-Instruct-Thinking
5
  tags:
6
  - generated_from_trainer
7
- - trl
8
  - grpo
 
9
  licence: license
10
  ---
11
 
 
4
  model_name: Qwen2.5-VL-3B-Instruct-Thinking
5
  tags:
6
  - generated_from_trainer
 
7
  - grpo
8
+ - trl
9
  licence: license
10
  ---
11
 
adapter_config.json CHANGED
@@ -25,13 +25,13 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
28
- "o_proj",
29
- "v_proj",
30
  "k_proj",
31
- "gate_proj",
32
  "up_proj",
33
- "down_proj",
34
- "q_proj"
 
 
 
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
 
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
 
28
  "k_proj",
 
29
  "up_proj",
30
+ "o_proj",
31
+ "gate_proj",
32
+ "v_proj",
33
+ "q_proj",
34
+ "down_proj"
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64f0dd95f25dacc6dfcc1de1a33c38aeae82bfac6b62fdec821fa0e6b1b93993
3
  size 74405904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a201eb79b883696f3308cfe084c98f161ba392b9f7a64c708ef71e68ce807a0a
3
  size 74405904
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9c672c52b27e6be4e079b4efe127c5f583aebaa9ce72ff9e263230f1e7f72d8e
3
  size 7185
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e736039d09a02f279c14a2b6fe16882e7396e8246b1637223973da8a0041ca6
3
  size 7185