Darkester commited on
Commit
3ea75c6
·
verified ·
1 Parent(s): e59799a

Upload 14 files

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<Thought>": 50257
3
+ }
config.json CHANGED
@@ -5,6 +5,7 @@
5
  ],
6
  "attn_pdrop": 0.1,
7
  "bos_token_id": 1,
 
8
  "embd_pdrop": 0.1,
9
  "eos_token_id": 2,
10
  "gradient_checkpointing": false,
@@ -33,8 +34,7 @@
33
  "summary_proj_to_labels": true,
34
  "summary_type": "cls_index",
35
  "summary_use_proj": true,
36
- "torch_dtype": "float32",
37
- "transformers_version": "4.51.3",
38
  "use_cache": true,
39
- "vocab_size": 50264
40
  }
 
5
  ],
6
  "attn_pdrop": 0.1,
7
  "bos_token_id": 1,
8
+ "dtype": "float32",
9
  "embd_pdrop": 0.1,
10
  "eos_token_id": 2,
11
  "gradient_checkpointing": false,
 
34
  "summary_proj_to_labels": true,
35
  "summary_type": "cls_index",
36
  "summary_use_proj": true,
37
+ "transformers_version": "4.57.0",
 
38
  "use_cache": true,
39
+ "vocab_size": 50258
40
  }
generation_config.json CHANGED
@@ -1,7 +1,9 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
- "eos_token_id": 2,
 
 
5
  "pad_token_id": 0,
6
- "transformers_version": "4.51.3"
7
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
+ "eos_token_id": [
5
+ 2
6
+ ],
7
  "pad_token_id": 0,
8
+ "transformers_version": "4.57.0"
9
  }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b933eb783a99e61bb56c7166eb048e9e32ea03e570343980e5ffecb5dba32777
3
- size 500941440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d613e790cd7b996c586bd592a1629fd2523962a8d41b2d59811d21a00a02b19d
3
+ size 500923008
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18da2c9155cd816360d678fb26f921486eac0d676d4bdfa7ee845ae774c8febe
3
- size 1001976762
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5e8fb28ca2b025604a8bb6a0183839d986906f09aa51aae586a61d54bbaa4b2
3
+ size 1001942411
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d20fef646b31bd0da60e042930355f95bad6e9309ba53f239d47962584cee4ee
3
- size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68a252ed94a395c93ad12e059a0201f53477ef2c6cb571266a3603711265a034
3
+ size 14645
scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a9acc78a1984093d893ebbe2aa3e8c676c849c54208c1237140b3e76d9d7d85
3
- size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aaf72c95e5f10009a800f90c81f3898ad133ba4e6267316f252e283ab34c4c0e
3
+ size 1383
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:77d7bffa60dee30450bbde43b505ba5d1aeb72b9b39eb756f6197faaaea51f34
3
- size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b62a1567c6b47404cf0850ec3dd85b4092d190c27f4f873664e4a739a1a79063
3
+ size 1465
special_tokens_map.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<Thought>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ }
10
+ ],
11
+ "bos_token": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": true,
15
+ "rstrip": false,
16
+ "single_word": false
17
+ },
18
+ "eos_token": {
19
+ "content": "</s>",
20
+ "lstrip": false,
21
+ "normalized": true,
22
+ "rstrip": false,
23
+ "single_word": false
24
+ },
25
+ "mask_token": {
26
+ "content": "<mask>",
27
+ "lstrip": false,
28
+ "normalized": false,
29
+ "rstrip": false,
30
+ "single_word": false
31
+ },
32
+ "pad_token": {
33
+ "content": "<pad>",
34
+ "lstrip": false,
35
+ "normalized": true,
36
+ "rstrip": false,
37
+ "single_word": false
38
+ },
39
+ "unk_token": {
40
+ "content": "<unk>",
41
+ "lstrip": false,
42
+ "normalized": true,
43
+ "rstrip": false,
44
+ "single_word": false
45
+ }
46
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<pad>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "<unk>",
31
+ "lstrip": false,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "<mask>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "50257": {
46
+ "content": "<Thought>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ }
53
+ },
54
+ "additional_special_tokens": [
55
+ "<Thought>"
56
+ ],
57
+ "bos_token": "<s>",
58
+ "clean_up_tokenization_spaces": true,
59
+ "eos_token": "</s>",
60
+ "errors": "replace",
61
+ "extra_special_tokens": {},
62
+ "mask_token": "<mask>",
63
+ "model_max_length": 2048,
64
+ "pad_token": "<pad>",
65
+ "padding_side": "left",
66
+ "tokenizer_class": "GPT2Tokenizer",
67
+ "truncation_side": "left",
68
+ "trust_remote_code": false,
69
+ "unk_token": "<unk>",
70
+ "use_fast": true
71
+ }
trainer_state.json CHANGED
@@ -2,159 +2,82 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 1.088139281828074,
6
  "eval_steps": 500,
7
- "global_step": 2000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.0544069640914037,
14
- "grad_norm": 1.3827155828475952,
15
- "learning_rate": 4.913855640188611e-05,
16
- "loss": 2.226,
17
  "step": 100
18
  },
19
  {
20
- "epoch": 0.1088139281828074,
21
- "grad_norm": 0.7897017002105713,
22
- "learning_rate": 4.823177366702939e-05,
23
- "loss": 1.2196,
24
  "step": 200
25
  },
26
  {
27
- "epoch": 0.1632208922742111,
28
- "grad_norm": 1.1451177597045898,
29
- "learning_rate": 4.732499093217265e-05,
30
- "loss": 1.2182,
31
  "step": 300
32
  },
33
  {
34
- "epoch": 0.2176278563656148,
35
- "grad_norm": 1.9186509847640991,
36
- "learning_rate": 4.641820819731593e-05,
37
- "loss": 1.2724,
38
  "step": 400
39
  },
40
  {
41
- "epoch": 0.2720348204570185,
42
- "grad_norm": 1.4905147552490234,
43
- "learning_rate": 4.5511425462459194e-05,
44
- "loss": 1.1771,
45
  "step": 500
46
  },
47
  {
48
- "epoch": 0.3264417845484222,
49
- "grad_norm": 1.3085367679595947,
50
- "learning_rate": 4.460464272760247e-05,
51
- "loss": 1.1407,
52
  "step": 600
53
  },
54
  {
55
- "epoch": 0.3808487486398259,
56
- "grad_norm": 1.6310745477676392,
57
- "learning_rate": 4.3697859992745737e-05,
58
- "loss": 1.1146,
59
  "step": 700
60
  },
61
  {
62
- "epoch": 0.4352557127312296,
63
- "grad_norm": 1.0869531631469727,
64
- "learning_rate": 4.2791077257889015e-05,
65
- "loss": 1.1015,
66
  "step": 800
67
  },
68
  {
69
- "epoch": 0.4896626768226333,
70
- "grad_norm": 1.387512445449829,
71
- "learning_rate": 4.1884294523032286e-05,
72
- "loss": 1.0764,
73
  "step": 900
74
- },
75
- {
76
- "epoch": 0.544069640914037,
77
- "grad_norm": 1.8865373134613037,
78
- "learning_rate": 4.097751178817556e-05,
79
- "loss": 1.079,
80
- "step": 1000
81
- },
82
- {
83
- "epoch": 0.5984766050054406,
84
- "grad_norm": 1.1464284658432007,
85
- "learning_rate": 4.007072905331883e-05,
86
- "loss": 1.1797,
87
- "step": 1100
88
- },
89
- {
90
- "epoch": 0.6528835690968444,
91
- "grad_norm": 1.3149100542068481,
92
- "learning_rate": 3.91639463184621e-05,
93
- "loss": 1.0676,
94
- "step": 1200
95
- },
96
- {
97
- "epoch": 0.7072905331882481,
98
- "grad_norm": 1.3868759870529175,
99
- "learning_rate": 3.825716358360537e-05,
100
- "loss": 1.1711,
101
- "step": 1300
102
- },
103
- {
104
- "epoch": 0.7616974972796517,
105
- "grad_norm": 1.046326994895935,
106
- "learning_rate": 3.735038084874864e-05,
107
- "loss": 1.1291,
108
- "step": 1400
109
- },
110
- {
111
- "epoch": 0.8161044613710555,
112
- "grad_norm": 0.825319230556488,
113
- "learning_rate": 3.6443598113891914e-05,
114
- "loss": 1.0609,
115
- "step": 1500
116
- },
117
- {
118
- "epoch": 0.8705114254624592,
119
- "grad_norm": 1.6592735052108765,
120
- "learning_rate": 3.5536815379035185e-05,
121
- "loss": 1.0676,
122
- "step": 1600
123
- },
124
- {
125
- "epoch": 0.9249183895538629,
126
- "grad_norm": 1.2844988107681274,
127
- "learning_rate": 3.4630032644178456e-05,
128
- "loss": 1.062,
129
- "step": 1700
130
- },
131
- {
132
- "epoch": 0.9793253536452666,
133
- "grad_norm": 1.3953850269317627,
134
- "learning_rate": 3.372324990932173e-05,
135
- "loss": 1.0538,
136
- "step": 1800
137
- },
138
- {
139
- "epoch": 1.0337323177366704,
140
- "grad_norm": 1.2760506868362427,
141
- "learning_rate": 3.2816467174465e-05,
142
- "loss": 0.968,
143
- "step": 1900
144
- },
145
- {
146
- "epoch": 1.088139281828074,
147
- "grad_norm": 1.5799071788787842,
148
- "learning_rate": 3.190968443960827e-05,
149
- "loss": 0.9365,
150
- "step": 2000
151
  }
152
  ],
153
  "logging_steps": 100,
154
- "max_steps": 5514,
155
  "num_input_tokens_seen": 0,
156
  "num_train_epochs": 3,
157
- "save_steps": 500,
158
  "stateful_callbacks": {
159
  "TrainerControl": {
160
  "args": {
@@ -162,12 +85,12 @@
162
  "should_evaluate": false,
163
  "should_log": false,
164
  "should_save": true,
165
- "should_training_stop": false
166
  },
167
  "attributes": {}
168
  }
169
  },
170
- "total_flos": 2090336256000000.0,
171
  "train_batch_size": 4,
172
  "trial_name": null,
173
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
  "eval_steps": 500,
7
+ "global_step": 939,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.32,
14
+ "grad_norm": 0.8309588432312012,
15
+ "learning_rate": 4.472843450479233e-05,
16
+ "loss": 2.1268,
17
  "step": 100
18
  },
19
  {
20
+ "epoch": 0.64,
21
+ "grad_norm": 0.8111929893493652,
22
+ "learning_rate": 3.940362087326944e-05,
23
+ "loss": 1.9513,
24
  "step": 200
25
  },
26
  {
27
+ "epoch": 0.96,
28
+ "grad_norm": 0.7366995215415955,
29
+ "learning_rate": 3.407880724174654e-05,
30
+ "loss": 1.9131,
31
  "step": 300
32
  },
33
  {
34
+ "epoch": 1.2784,
35
+ "grad_norm": 0.7454879283905029,
36
+ "learning_rate": 2.8753993610223644e-05,
37
+ "loss": 1.8164,
38
  "step": 400
39
  },
40
  {
41
+ "epoch": 1.5984,
42
+ "grad_norm": 0.7343336939811707,
43
+ "learning_rate": 2.3429179978700748e-05,
44
+ "loss": 1.816,
45
  "step": 500
46
  },
47
  {
48
+ "epoch": 1.9184,
49
+ "grad_norm": 0.7088562250137329,
50
+ "learning_rate": 1.8104366347177852e-05,
51
+ "loss": 1.7918,
52
  "step": 600
53
  },
54
  {
55
+ "epoch": 2.2368,
56
+ "grad_norm": 0.7369393706321716,
57
+ "learning_rate": 1.2779552715654951e-05,
58
+ "loss": 1.7424,
59
  "step": 700
60
  },
61
  {
62
+ "epoch": 2.5568,
63
+ "grad_norm": 0.7260516285896301,
64
+ "learning_rate": 7.4547390841320565e-06,
65
+ "loss": 1.744,
66
  "step": 800
67
  },
68
  {
69
+ "epoch": 2.8768000000000002,
70
+ "grad_norm": 0.6936702728271484,
71
+ "learning_rate": 2.1299254526091587e-06,
72
+ "loss": 1.7355,
73
  "step": 900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
74
  }
75
  ],
76
  "logging_steps": 100,
77
+ "max_steps": 939,
78
  "num_input_tokens_seen": 0,
79
  "num_train_epochs": 3,
80
+ "save_steps": 5000,
81
  "stateful_callbacks": {
82
  "TrainerControl": {
83
  "args": {
 
85
  "should_evaluate": false,
86
  "should_log": false,
87
  "should_save": true,
88
+ "should_training_stop": true
89
  },
90
  "attributes": {}
91
  }
92
  },
93
+ "total_flos": 7838760960000000.0,
94
  "train_batch_size": 4,
95
  "trial_name": null,
96
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3b92fad9c09a73aae21441d2f394f747f6e1203d0cca82e74f610a1fc8e8dd1
3
- size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:247d11e1712ddd52718707e0c098c161c131deb0e10e431622dcb2919154098c
3
+ size 5777
vocab.json ADDED
The diff for this file is too large to render. See raw diff