ToastyPigeon commited on
Commit
70638f8
·
verified ·
1 Parent(s): cc54f39

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -239
README.md DELETED
@@ -1,239 +0,0 @@
1
- ---
2
- library_name: peft
3
- license: apache-2.0
4
- base_model: mistralai/Mistral-Small-24B-Instruct-2501
5
- tags:
6
- - axolotl
7
- - generated_from_trainer
8
- datasets:
9
- - ToastyPigeon/some-rp-extended
10
- - ToastyPigeon/gutenberg-sft
11
- - ToastyPigeon/fujin-filtered-instruct
12
- - ToastyPigeon/adventure-combined-no-slop-matches-instruct
13
- - ToastyPigeon/nifty-multi-filtered
14
- model-index:
15
- - name: new-ms-rp-test-ws
16
- results: []
17
- ---
18
-
19
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
20
- should probably proofread and complete it, then remove this comment. -->
21
-
22
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
23
- <details><summary>See axolotl config</summary>
24
-
25
- axolotl version: `0.6.0`
26
- ```yaml
27
- # git clone https://github.com/axolotl-ai-cloud/axolotl
28
- # cd axolotl
29
- # git checkout 844331005c1ef45430ff26b9f42f757dce6ee66a
30
- # pip3 install packaging ninja huggingface_hub[cli]
31
- # pip3 install -e '.[flash-attn,deepspeed]'
32
- # huggingface-cli login --token $hf_key && wandb login $wandb_key
33
- # python -m axolotl.cli.preprocess new-ms-rp-test.yml
34
- # accelerate launch -m axolotl.cli.train new-ms-rp-test.yml
35
- # python -m axolotl.cli.merge_lora new-ms-rp-test.yml
36
- # huggingface-cli upload ToastyPigeon/new-ms-rp-test train-workspace/merged . --exclude "*.md"
37
-
38
- # git clone https://github.com/axolotl-ai-cloud/axolotl && cd axolotl && pip3 install packaging ninja huggingface_hub[cli] && pip3 install cut-cross-entropy[transformers] && pip3 install -e '.[flash-attn,deepspeed]' && cd .. && huggingface-cli login --token $hf_key && wandb login $wandb_key
39
-
40
- # Model
41
- base_model: mistralai/Mistral-Small-24B-Instruct-2501
42
- model_type: AutoModelForCausalLM
43
- tokenizer_type: AutoTokenizer
44
-
45
- load_in_8bit: false
46
- load_in_4bit: true
47
- strict: false
48
- bf16: true
49
- fp16:
50
- tf32: false
51
- flash_attention: true
52
- special_tokens:
53
-
54
- # Output
55
- output_dir: ./train-workspace
56
- hub_model_id: ToastyPigeon/new-ms-rp-test-ws
57
- hub_strategy: "all_checkpoints"
58
- resume_from_checkpoint:
59
- saves_per_epoch: 2
60
-
61
- # Data
62
- sequence_len: 4096 # fits
63
- min_sample_len: 128
64
- dataset_prepared_path: last_run_prepared
65
- datasets:
66
- - path: ToastyPigeon/some-rp-extended
67
- type: chat_template
68
- field_messages: conversations
69
- message_field_role: from
70
- message_field_content: value
71
- - path: ToastyPigeon/gutenberg-sft
72
- type: chat_template
73
- field_messages: conversations
74
- message_field_role: from
75
- message_field_content: value
76
- - path: ToastyPigeon/fujin-filtered-instruct
77
- type: chat_template
78
- field_messages: conversations
79
- message_field_role: from
80
- message_field_content: value
81
- - path: ToastyPigeon/adventure-combined-no-slop-matches-instruct
82
- type: chat_template
83
- field_messages: conversations
84
- message_field_role: from
85
- message_field_content: value
86
- split: train[:1000]
87
- - path: ToastyPigeon/nifty-multi-filtered
88
- type: completion
89
- field: text
90
- split: train[:1000]
91
- warmup_steps: 5
92
- shuffle_merged_datasets: true
93
- sample_packing: true
94
- pad_to_sequence_len: true
95
- chat_template: jinja
96
- chat_template_jinja: "{%- set default_system_message = \"You are an unbiased, unfiltered, and unaligned AI.\" %}\n\n{{- bos_token }}\n\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set system_message = default_system_message %}\n {%- set loop_messages = messages %}\n{%- endif %}\n{{- '[SYSTEM_PROMPT]' + system_message + '[/SYSTEM_PROMPT]' }}\n\n{%- for message in loop_messages %}\n {%- if message['role'] == 'user' %}\n {{- '[INST]' + message['content'] + '[/INST]' }}\n {%- elif message['role'] == 'system' %}\n {{- '[SYSTEM_PROMPT]' + message['content'] + '[/SYSTEM_PROMPT]' }}\n {%- elif message['role'] == 'assistant' %}\n {{- message['content'] + eos_token }}\n {%- else %}\n {{- raise_exception('Only user, system and assistant roles are supported!') }}\n {%- endif %}\n{%- endfor %}"
97
-
98
- # Batching
99
- num_epochs: 1
100
- gradient_accumulation_steps: 1
101
- micro_batch_size: 4
102
- eval_batch_size: 1
103
-
104
- # Evaluation
105
- val_set_size: 100
106
- evals_per_epoch: 10
107
- eval_table_size:
108
- eval_max_new_tokens: 256
109
- eval_sample_packing: false
110
-
111
- save_safetensors: true
112
-
113
- # WandB
114
- wandb_project: MS-Rp-Test
115
- #wandb_entity:
116
-
117
- gradient_checkpointing: 'unsloth'
118
- #gradient_checkpointing_kwargs:
119
- # use_reentrant: false
120
-
121
- unsloth_cross_entropy_loss: true
122
- #unsloth_lora_mlp: true
123
- #unsloth_lora_qkv: true
124
- #unsloth_lora_o: true
125
-
126
- # LoRA
127
- adapter: qlora
128
- lora_model_dir:
129
- lora_r: 16
130
- lora_alpha: 32
131
- lora_dropout: 0.5
132
- lora_target_linear:
133
- lora_fan_in_fan_out:
134
- lora_target_modules:
135
- - gate_proj
136
- - down_proj
137
- - up_proj
138
- - q_proj
139
- - v_proj
140
- - k_proj
141
- - o_proj
142
- lora_modules_to_save:
143
-
144
- # Optimizer
145
- optimizer: paged_ademamix_8bit # adamw_8bit
146
- lr_scheduler: cosine
147
- learning_rate: 5e-5
148
- cosine_min_lr_ratio: 0.1
149
- weight_decay: 0.01
150
- max_grad_norm: 1.0
151
-
152
- # Misc
153
- train_on_inputs: false
154
- group_by_length: false
155
- early_stopping_patience:
156
- local_rank:
157
- logging_steps: 1
158
- xformers_attention:
159
- debug:
160
- deepspeed: /workspace/axolotl/deepspeed_configs/zero3_bf16.json # previously blank
161
- fsdp:
162
- fsdp_config:
163
-
164
- plugins:
165
- - axolotl.integrations.liger.LigerPlugin
166
- # - axolotl.integrations.cut_cross_entropy.CutCrossEntropyPlugin
167
- #cut_cross_entropy: true
168
- liger_rope: true
169
- liger_rms_norm: true
170
- liger_layer_norm: true
171
- liger_glu_activation: true
172
- liger_fused_linear_cross_entropy: true
173
-
174
- gc_steps: 10
175
- seed: 69
176
- ```
177
-
178
- </details><br>
179
-
180
- # new-ms-rp-test-ws
181
-
182
- This model is a fine-tuned version of [mistralai/Mistral-Small-24B-Instruct-2501](https://huggingface.co/mistralai/Mistral-Small-24B-Instruct-2501) on the ToastyPigeon/some-rp-extended, the ToastyPigeon/gutenberg-sft, the ToastyPigeon/fujin-filtered-instruct, the ToastyPigeon/adventure-combined-no-slop-matches-instruct and the ToastyPigeon/nifty-multi-filtered datasets.
183
- It achieves the following results on the evaluation set:
184
- - Loss: 1.9832
185
-
186
- ## Model description
187
-
188
- More information needed
189
-
190
- ## Intended uses & limitations
191
-
192
- More information needed
193
-
194
- ## Training and evaluation data
195
-
196
- More information needed
197
-
198
- ## Training procedure
199
-
200
- ### Training hyperparameters
201
-
202
- The following hyperparameters were used during training:
203
- - learning_rate: 5e-05
204
- - train_batch_size: 4
205
- - eval_batch_size: 1
206
- - seed: 69
207
- - distributed_type: multi-GPU
208
- - num_devices: 4
209
- - total_train_batch_size: 16
210
- - total_eval_batch_size: 4
211
- - optimizer: Use OptimizerNames.PAGED_ADEMAMIX_8BIT and the args are:
212
- No additional optimizer arguments
213
- - lr_scheduler_type: cosine
214
- - lr_scheduler_warmup_steps: 5
215
- - num_epochs: 1.0
216
-
217
- ### Training results
218
-
219
- | Training Loss | Epoch | Step | Validation Loss |
220
- |:-------------:|:------:|:----:|:---------------:|
221
- | 1.9624 | 0.0056 | 1 | 2.0950 |
222
- | 2.0035 | 0.1017 | 18 | 2.0137 |
223
- | 2.0876 | 0.2034 | 36 | 2.0019 |
224
- | 1.9243 | 0.3051 | 54 | 1.9959 |
225
- | 2.0775 | 0.4068 | 72 | 1.9916 |
226
- | 1.8296 | 0.5085 | 90 | 1.9892 |
227
- | 1.9782 | 0.6102 | 108 | 1.9868 |
228
- | 2.1268 | 0.7119 | 126 | 1.9848 |
229
- | 1.7833 | 0.8136 | 144 | 1.9841 |
230
- | 2.1287 | 0.9153 | 162 | 1.9832 |
231
-
232
-
233
- ### Framework versions
234
-
235
- - PEFT 0.14.0
236
- - Transformers 4.48.1
237
- - Pytorch 2.5.1+cu124
238
- - Datasets 3.2.0
239
- - Tokenizers 0.21.0