PhongNgoGia commited on
Commit
9b8d5ff
·
verified ·
1 Parent(s): 0b890df

Training in progress, step 10

Browse files
adapter_config.json CHANGED
@@ -1,10 +1,8 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "PhongNgoGia/Qwen2.5-1.5B-Lora",
5
  "bias": "none",
6
- "eva_config": null,
7
- "exclude_modules": null,
8
  "fan_in_fan_out": false,
9
  "inference_mode": true,
10
  "init_lora_weights": true,
@@ -13,7 +11,6 @@
13
  "layers_to_transform": null,
14
  "loftq_config": {},
15
  "lora_alpha": 32,
16
- "lora_bias": false,
17
  "lora_dropout": 0.1,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
@@ -23,8 +20,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "q_proj",
27
- "v_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "PhongNgoGia/Qwen2.5-1.5B-Lora-1",
5
  "bias": "none",
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
11
  "layers_to_transform": null,
12
  "loftq_config": {},
13
  "lora_alpha": 32,
 
14
  "lora_dropout": 0.1,
15
  "megatron_config": null,
16
  "megatron_core": "megatron.core",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a2a8386d6fd6703420c1cdba19c2f32ccfb857f700555c73d25eac2ce380464
3
  size 4372840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f759e55776139fac07718f215134bb9995d1d9da43c9ab2c45d9b586a6d2462
3
  size 4372840
runs/Mar10_16-18-46_d0a90b366007/events.out.tfevents.1741623528.d0a90b366007.9089.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f3f2de57e1f6cfd2f25ab49c6f4e9fede491b52f671a57b8591c4c110169e83
3
+ size 5651
runs/Mar10_16-22-57_d0a90b366007/events.out.tfevents.1741623778.d0a90b366007.14598.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8a46d93e58ece3c672a3167fe30999878769c5b81c67cf3c6d152f7304e9426
3
+ size 5651
runs/Mar10_16-24-41_d0a90b366007/events.out.tfevents.1741623882.d0a90b366007.15004.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f23be90893223f2767cbf8cd01862f8274bcad4aa3ceff1e501b53bb40e9e98d
3
+ size 5650
runs/Mar10_16-27-43_d0a90b366007/events.out.tfevents.1741624064.d0a90b366007.15361.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:531c22ebcea50482680718f69874c2378406782ce30f2c8c4375efea97699383
3
+ size 5650
runs/Mar10_16-29-05_d0a90b366007/events.out.tfevents.1741624146.d0a90b366007.15685.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bedcc76a36f812ffaa3172e51c28d137ede23290225cdf548983d1e362803d5e
3
+ size 5650
runs/Mar10_16-31-00_d0a90b366007/events.out.tfevents.1741624261.d0a90b366007.16071.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8bc26ed550d36b3451b58a7d31115f48f38192fa813e471e59b3809d26b0fcd
3
+ size 6198
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:934257c5afe972f455de7119dceba39b38e1399a9c6e9beeab599b24a9804a3a
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6500660b2a49b282f851ef4a7f882d0909df887bfad06db1caec0ee96d1dc0ae
3
  size 5752