kernalkue commited on
Commit
2194d1b
·
0 Parent(s):

Initial commit: EXL3 6.0bpw quantization with proper LFS tracking

Browse files
.gitattributes ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
2
+ *.bin filter=lfs diff=lfs merge=lfs -text
3
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
4
+ *.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: [BruhzWater/Sapphira-L3.3-70b-0.1]
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+
8
+ ---
9
+ # Sapphira-L3.3-70b-0.1
10
+
11
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66ca56e62400073af3ad2972/CPUXeq81a9o0_ClXCcG68.png)
12
+
13
+ Storytelling and RP model with increased coherence, thanks to cogito-v2-preview-llama-70B.
14
+
15
+ iMatrix quants: https://huggingface.co/mradermacher/Sapphira-L3.3-70b-0.1-i1-GGUF
16
+
17
+ Static quants: https://huggingface.co/mradermacher/Sapphira-L3.3-70b-0.1-GGUF
18
+
19
+ Chat Template:
20
+ -
21
+ Llama3
22
+
23
+ Instruction Template:
24
+ -
25
+ Deep Cogito
26
+
27
+ Llama3
28
+
29
+ Sampler Settings
30
+ -
31
+
32
+ Starter:
33
+ ```
34
+ Temp: 1
35
+ Min_P: 0.02
36
+ Top_P: 1
37
+ ```
38
+
39
+ Experimental 1:
40
+ ```
41
+ Temp: .95 - 1.1
42
+ Min_P: .015 - .03
43
+ Top_P: .97 - .99
44
+ XTC_Threshold: .11
45
+ XTC_Probability: .15
46
+ ```
47
+
48
+ Experimental 2:
49
+ ```
50
+ Temp: .95 - 1.1
51
+ Min_P: .015 - .03
52
+ Top_P: 1
53
+ Typical_P: .99
54
+ XTC_Threshold: .11
55
+ XTC_Probability: .15
56
+ ```
57
+
58
+ ### Merge Method
59
+
60
+ This model was merged using the [Multi-SLERP](https://goddard.blog/posts/multislerp-wow-what-a-cool-idea) merge method using deepcogito--cogito-v2-preview-llama-70B as a base.
61
+
62
+ ### Models Merged
63
+
64
+ The following models were included in the merge:
65
+ * BruhzWater--Apocrypha-L3.3-70b-0.3
66
+ * BruhzWater--Serpents-Tongue-L3.3-70b-0.3
67
+
68
+ ### Configuration
69
+
70
+ The following YAML configuration was used to produce this model:
71
+
72
+ ```yaml
73
+ models:
74
+ - model: /workspace/cache/models--BruhzWater--Apocrypha-L3.3-70b-0.3/snapshots/3facb4c0a7b953ff34a5caa90976830bf82a84c2
75
+ parameters:
76
+ weight: [0.5]
77
+ - model: /workspace/cache/models--BruhzWater--Serpents-Tongue-L3.3-70b-0.3/snapshots/d007a7bcc7047d712abb2dfb6ad940fe03cd2047
78
+ parameters:
79
+ weight: [0.5]
80
+ base_model: /workspace/cache/models--deepcogito--cogito-v2-preview-llama-70B/snapshots/1e1d12e8eaebd6084a8dcf45ecdeaa2f4b8879ce
81
+ merge_method: multislerp
82
+ tokenizer:
83
+ source: base
84
+ chat_template: llama3
85
+ parameters:
86
+ normalize_weights: false
87
+ eps: 1e-9
88
+ pad_to_multiple_of: 8
89
+ int8_mask: true
90
+ dtype: bfloat16
91
+ ```
92
+
93
+ ### Instruct Template
94
+
95
+ Deep Cogito
96
+
97
+ ```
98
+ {{- '<|begin_of_text|>' }}
99
+ {%- if not tools is defined %}
100
+ {%- set tools = none %}
101
+ {%- endif %}
102
+ {%- if not enable_thinking is defined %}
103
+ {%- set enable_thinking = false %}
104
+ {%- endif %}
105
+ {#- This block extracts the system message, so we can slot it into the right place. #}
106
+ {%- if messages[0]['role'] == 'system' %}
107
+ {%- set system_message = messages[0]['content']|trim %}
108
+ {%- set messages = messages[1:] %}
109
+ {%- else %}
110
+ {%- set system_message = "" %}
111
+ {%- endif %}
112
+ {#- Set the system message. If enable_thinking is true, add the "Enable deep thinking subroutine." #}
113
+ {%- if enable_thinking %}
114
+ {%- if system_message != "" %}
115
+ {%- set system_message = "Enable deep thinking subroutine.
116
+
117
+ " ~ system_message %}
118
+ {%- else %}
119
+ {%- set system_message = "Enable deep thinking subroutine." %}
120
+ {%- endif %}
121
+ {%- endif %}
122
+ {#- Set the system message. In case there are tools present, add them to the system message. #}
123
+ {%- if tools is not none or system_message != '' %}
124
+ {{- "<|start_header_id|>system<|end_header_id|>
125
+
126
+ " }}
127
+ {{- system_message }}
128
+ {%- if tools is not none %}
129
+ {%- if system_message != "" %}
130
+ {{- "
131
+
132
+ " }}
133
+ {%- endif %}
134
+ {{- "Available Tools:
135
+ " }}
136
+ {%- for t in tools %}
137
+ {{- t | tojson(indent=4) }}
138
+ {{- "
139
+
140
+ " }}
141
+ {%- endfor %}
142
+ {%- endif %}
143
+ {{- "<|eot_id|>" }}
144
+ {%- endif %}
145
+
146
+ {#- Rest of the messages #}
147
+ {%- for message in messages %}
148
+ {#- The special cases are when the message is from a tool (via role ipython/tool/tool_results) or when the message is from the assistant, but has "tool_calls". If not, we add the message directly as usual. #}
149
+ {#- Case 1 - Usual, non tool related message. #}
150
+ {%- if not (message.role == "ipython" or message.role == "tool" or message.role == "tool_results" or (message.tool_calls is defined and message.tool_calls is not none)) %}
151
+ {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>
152
+
153
+ ' }}
154
+ {%- if message['content'] is string %}
155
+ {{- message['content'] | trim }}
156
+ {%- else %}
157
+ {%- for item in message['content'] %}
158
+ {%- if item.type == 'text' %}
159
+ {{- item.text | trim }}
160
+ {%- endif %}
161
+ {%- endfor %}
162
+ {%- endif %}
163
+ {{- '<|eot_id|>' }}
164
+
165
+ {#- Case 2 - the response is from the assistant, but has a tool call returned. The assistant may also have returned some content along with the tool call. #}
166
+ {%- elif message.tool_calls is defined and message.tool_calls is not none %}
167
+ {{- "<|start_header_id|>assistant<|end_header_id|>
168
+
169
+ " }}
170
+ {%- if message['content'] is string %}
171
+ {{- message['content'] | trim }}
172
+ {%- else %}
173
+ {%- for item in message['content'] %}
174
+ {%- if item.type == 'text' %}
175
+ {{- item.text | trim }}
176
+ {%- if item.text | trim != "" %}
177
+ {{- "
178
+
179
+ " }}
180
+ {%- endif %}
181
+ {%- endif %}
182
+ {%- endfor %}
183
+ {%- endif %}
184
+ {{- "[" }}
185
+ {%- for tool_call in message.tool_calls %}
186
+ {%- set out = tool_call.function|tojson %}
187
+ {%- if not tool_call.id is defined %}
188
+ {{- out }}
189
+ {%- else %}
190
+ {{- out[:-1] }}
191
+ {{- ', "id": "' + tool_call.id + '"}' }}
192
+ {%- endif %}
193
+ {%- if not loop.last %}
194
+ {{- ", " }}
195
+ {%- else %}
196
+ {{- "]<|eot_id|>" }}
197
+ {%- endif %}
198
+ {%- endfor %}
199
+
200
+ {#- Case 3 - the response is from a tool call. The tool call may have an id associated with it as well. If it does, we add it to the prompt. #}
201
+ {%- elif message.role == "ipython" or message["role"] == "tool_results" or message["role"] == "tool" %}
202
+ {{- "<|start_header_id|>ipython<|end_header_id|>
203
+
204
+ " }}
205
+ {%- if message.tool_call_id is defined and message.tool_call_id != '' %}
206
+ {{- '{"content": ' + (message.content | tojson) + ', "call_id": "' + message.tool_call_id + '"}' }}
207
+ {%- else %}
208
+ {{- '{"content": ' + (message.content | tojson) + '}' }}
209
+ {%- endif %}
210
+ {{- "<|eot_id|>" }}
211
+ {%- endif %}
212
+ {%- endfor %}
213
+ {%- if add_generation_prompt %}
214
+ {{- '<|start_header_id|>assistant<|end_header_id|>
215
+
216
+ ' }}
217
+ {%- endif %}
218
+ ```
chat_template.jinja ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {% set loop_messages = messages %}
2
+ {% for message in loop_messages %}
3
+ {% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}
4
+ {% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}
5
+ {{ content }}
6
+ {% endfor %}
7
+ {% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}
config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e6efa44dcc545093b52fa5bd9bbf39590dece6b3d064b47c81cf2273976e6fe
3
+ size 1173
mergekit_config.yml ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - model: /workspace/cache/models--BruhzWater--Apocrypha-L3.3-70b-0.3/snapshots/3facb4c0a7b953ff34a5caa90976830bf82a84c2
3
+ parameters:
4
+ weight: [0.5]
5
+ - model: /workspace/cache/models--BruhzWater--Serpents-Tongue-L3.3-70b-0.3/snapshots/d007a7bcc7047d712abb2dfb6ad940fe03cd2047
6
+ parameters:
7
+ weight: [0.5]
8
+ base_model: /workspace/cache/models--deepcogito--cogito-v2-preview-llama-70B/snapshots/1e1d12e8eaebd6084a8dcf45ecdeaa2f4b8879ce
9
+ merge_method: multislerp
10
+ tokenizer:
11
+ source: base
12
+ chat_template: llama3
13
+ parameters:
14
+ normalize_weights: false
15
+ eps: 1e-9
16
+ pad_to_multiple_of: 8
17
+ int8_mask: true
18
+ dtype: bfloat16
model-00001-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7eaa33849ca6a8fee52f475a19e79971adf6c43114e2eff65c36bd91d028db45
3
+ size 8522220432
model-00002-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9edbc8c5e1c9bd6268802804ec511b1f63fa7e53384b6b5cab4ad71ed5f9474d
3
+ size 8347135176
model-00003-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:729621ff1f1b63a7a63c97913ca50da15b005625b65f8388b2339b0a1139fed7
3
+ size 8347135176
model-00004-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b4eb9c9cb2bfd1963464b3ca2f02b6f760f59a2a52a67be34b3f4d1a8ae51e8
3
+ size 8347135176
model-00005-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac398889101b3cebd99d10ab496688af32ec8e5e54f9600cf631c8774992626
3
+ size 8347135176
model-00006-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:709decc56986c1cece30fe28b120cc467cab004317ee5a68dd3a5926bc8e13e0
3
+ size 8347135176
model-00007-of-00007.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5dddc8bc2bc5b3b014e3685700fca65d83406bf37a524525950d80679209cf1
3
+ size 3998731056
model.safetensors.index.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46b07d76d45fe0859f7175e29113f0afdf1e5b81396e00fb2cd61066c31c8f09
3
+ size 155376
quantization_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc888fd18a7c5e5d2caaf7251c9dd27872c6ed610cfb7f95b40d143d65c4d11f
3
+ size 593665
special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e708c3f5e64acf85bbe5ad01467a1248faadb73e83b41793087ecced586e8f
3
+ size 454
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2152033590c3bc3e5d4a66356b91f3ace338a57cf2357b68db871e9bc4337909
3
+ size 50569