kernalkue commited on
Commit
ed053bd
·
0 Parent(s):

Initial commit: 5.0bpw EXL3 quantization of Sapphira-L3.3-70b-0.1

Browse files
.gitattributes ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
2
+ *.bin filter=lfs diff=lfs merge=lfs -text
3
+ *.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,218 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: []
3
+ library_name: transformers
4
+ tags:
5
+ - mergekit
6
+ - merge
7
+
8
+ ---
9
+ # Sapphira-L3.3-70b-0.1
10
+
11
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/66ca56e62400073af3ad2972/CPUXeq81a9o0_ClXCcG68.png)
12
+
13
+ Storytelling and RP model with increased coherence, thanks to cogito-v2-preview-llama-70B.
14
+
15
+ iMatrix quants: https://huggingface.co/mradermacher/Sapphira-L3.3-70b-0.1-i1-GGUF
16
+
17
+ Static quants: https://huggingface.co/mradermacher/Sapphira-L3.3-70b-0.1-GGUF
18
+
19
+ Chat Template:
20
+ -
21
+ Llama3
22
+
23
+ Instruction Template:
24
+ -
25
+ Deep Cogito
26
+
27
+ Llama3
28
+
29
+ Sampler Settings
30
+ -
31
+
32
+ Starter:
33
+ ```
34
+ Temp: 1
35
+ Min_P: 0.02
36
+ Top_P: 1
37
+ ```
38
+
39
+ Experimental 1:
40
+ ```
41
+ Temp: .95 - 1.1
42
+ Min_P: .015 - .03
43
+ Top_P: .97 - .99
44
+ XTC_Threshold: .11
45
+ XTC_Probability: .15
46
+ ```
47
+
48
+ Experimental 2:
49
+ ```
50
+ Temp: .95 - 1.1
51
+ Min_P: .015 - .03
52
+ Top_P: 1
53
+ Typical_P: .99
54
+ XTC_Threshold: .11
55
+ XTC_Probability: .15
56
+ ```
57
+
58
+ ### Merge Method
59
+
60
+ This model was merged using the [Multi-SLERP](https://goddard.blog/posts/multislerp-wow-what-a-cool-idea) merge method using deepcogito--cogito-v2-preview-llama-70B as a base.
61
+
62
+ ### Models Merged
63
+
64
+ The following models were included in the merge:
65
+ * BruhzWater--Apocrypha-L3.3-70b-0.3
66
+ * BruhzWater--Serpents-Tongue-L3.3-70b-0.3
67
+
68
+ ### Configuration
69
+
70
+ The following YAML configuration was used to produce this model:
71
+
72
+ ```yaml
73
+ models:
74
+ - model: /workspace/cache/models--BruhzWater--Apocrypha-L3.3-70b-0.3/snapshots/3facb4c0a7b953ff34a5caa90976830bf82a84c2
75
+ parameters:
76
+ weight: [0.5]
77
+ - model: /workspace/cache/models--BruhzWater--Serpents-Tongue-L3.3-70b-0.3/snapshots/d007a7bcc7047d712abb2dfb6ad940fe03cd2047
78
+ parameters:
79
+ weight: [0.5]
80
+ base_model: /workspace/cache/models--deepcogito--cogito-v2-preview-llama-70B/snapshots/1e1d12e8eaebd6084a8dcf45ecdeaa2f4b8879ce
81
+ merge_method: multislerp
82
+ tokenizer:
83
+ source: base
84
+ chat_template: llama3
85
+ parameters:
86
+ normalize_weights: false
87
+ eps: 1e-9
88
+ pad_to_multiple_of: 8
89
+ int8_mask: true
90
+ dtype: bfloat16
91
+ ```
92
+
93
+ ### Instruct Template
94
+
95
+ Deep Cogito
96
+
97
+ ```
98
+ {{- '<|begin_of_text|>' }}
99
+ {%- if not tools is defined %}
100
+ {%- set tools = none %}
101
+ {%- endif %}
102
+ {%- if not enable_thinking is defined %}
103
+ {%- set enable_thinking = false %}
104
+ {%- endif %}
105
+ {#- This block extracts the system message, so we can slot it into the right place. #}
106
+ {%- if messages[0]['role'] == 'system' %}
107
+ {%- set system_message = messages[0]['content']|trim %}
108
+ {%- set messages = messages[1:] %}
109
+ {%- else %}
110
+ {%- set system_message = "" %}
111
+ {%- endif %}
112
+ {#- Set the system message. If enable_thinking is true, add the "Enable deep thinking subroutine." #}
113
+ {%- if enable_thinking %}
114
+ {%- if system_message != "" %}
115
+ {%- set system_message = "Enable deep thinking subroutine.
116
+
117
+ " ~ system_message %}
118
+ {%- else %}
119
+ {%- set system_message = "Enable deep thinking subroutine." %}
120
+ {%- endif %}
121
+ {%- endif %}
122
+ {#- Set the system message. In case there are tools present, add them to the system message. #}
123
+ {%- if tools is not none or system_message != '' %}
124
+ {{- "<|start_header_id|>system<|end_header_id|>
125
+
126
+ " }}
127
+ {{- system_message }}
128
+ {%- if tools is not none %}
129
+ {%- if system_message != "" %}
130
+ {{- "
131
+
132
+ " }}
133
+ {%- endif %}
134
+ {{- "Available Tools:
135
+ " }}
136
+ {%- for t in tools %}
137
+ {{- t | tojson(indent=4) }}
138
+ {{- "
139
+
140
+ " }}
141
+ {%- endfor %}
142
+ {%- endif %}
143
+ {{- "<|eot_id|>" }}
144
+ {%- endif %}
145
+
146
+ {#- Rest of the messages #}
147
+ {%- for message in messages %}
148
+ {#- The special cases are when the message is from a tool (via role ipython/tool/tool_results) or when the message is from the assistant, but has "tool_calls". If not, we add the message directly as usual. #}
149
+ {#- Case 1 - Usual, non tool related message. #}
150
+ {%- if not (message.role == "ipython" or message.role == "tool" or message.role == "tool_results" or (message.tool_calls is defined and message.tool_calls is not none)) %}
151
+ {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>
152
+
153
+ ' }}
154
+ {%- if message['content'] is string %}
155
+ {{- message['content'] | trim }}
156
+ {%- else %}
157
+ {%- for item in message['content'] %}
158
+ {%- if item.type == 'text' %}
159
+ {{- item.text | trim }}
160
+ {%- endif %}
161
+ {%- endfor %}
162
+ {%- endif %}
163
+ {{- '<|eot_id|>' }}
164
+
165
+ {#- Case 2 - the response is from the assistant, but has a tool call returned. The assistant may also have returned some content along with the tool call. #}
166
+ {%- elif message.tool_calls is defined and message.tool_calls is not none %}
167
+ {{- "<|start_header_id|>assistant<|end_header_id|>
168
+
169
+ " }}
170
+ {%- if message['content'] is string %}
171
+ {{- message['content'] | trim }}
172
+ {%- else %}
173
+ {%- for item in message['content'] %}
174
+ {%- if item.type == 'text' %}
175
+ {{- item.text | trim }}
176
+ {%- if item.text | trim != "" %}
177
+ {{- "
178
+
179
+ " }}
180
+ {%- endif %}
181
+ {%- endif %}
182
+ {%- endfor %}
183
+ {%- endif %}
184
+ {{- "[" }}
185
+ {%- for tool_call in message.tool_calls %}
186
+ {%- set out = tool_call.function|tojson %}
187
+ {%- if not tool_call.id is defined %}
188
+ {{- out }}
189
+ {%- else %}
190
+ {{- out[:-1] }}
191
+ {{- ', "id": "' + tool_call.id + '"}' }}
192
+ {%- endif %}
193
+ {%- if not loop.last %}
194
+ {{- ", " }}
195
+ {%- else %}
196
+ {{- "]<|eot_id|>" }}
197
+ {%- endif %}
198
+ {%- endfor %}
199
+
200
+ {#- Case 3 - the response is from a tool call. The tool call may have an id associated with it as well. If it does, we add it to the prompt. #}
201
+ {%- elif message.role == "ipython" or message["role"] == "tool_results" or message["role"] == "tool" %}
202
+ {{- "<|start_header_id|>ipython<|end_header_id|>
203
+
204
+ " }}
205
+ {%- if message.tool_call_id is defined and message.tool_call_id != '' %}
206
+ {{- '{"content": ' + (message.content | tojson) + ', "call_id": "' + message.tool_call_id + '"}' }}
207
+ {%- else %}
208
+ {{- '{"content": ' + (message.content | tojson) + '}' }}
209
+ {%- endif %}
210
+ {{- "<|eot_id|>" }}
211
+ {%- endif %}
212
+ {%- endfor %}
213
+ {%- if add_generation_prompt %}
214
+ {{- '<|start_header_id|>assistant<|end_header_id|>
215
+
216
+ ' }}
217
+ {%- endif %}
218
+ ```
chat_template.jinja ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {% set loop_messages = messages %}
2
+ {% for message in loop_messages %}
3
+ {% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}
4
+ {% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}
5
+ {{ content }}
6
+ {% endfor %}
7
+ {% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}
config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59eef339bbb977ca3cfd57e6e4bf48f1af90f045a2d412f77086bfc95a53a86e
3
+ size 1173
mergekit_config.yml ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models:
2
+ - model: /workspace/cache/models--BruhzWater--Apocrypha-L3.3-70b-0.3/snapshots/3facb4c0a7b953ff34a5caa90976830bf82a84c2
3
+ parameters:
4
+ weight: [0.5]
5
+ - model: /workspace/cache/models--BruhzWater--Serpents-Tongue-L3.3-70b-0.3/snapshots/d007a7bcc7047d712abb2dfb6ad940fe03cd2047
6
+ parameters:
7
+ weight: [0.5]
8
+ base_model: /workspace/cache/models--deepcogito--cogito-v2-preview-llama-70B/snapshots/1e1d12e8eaebd6084a8dcf45ecdeaa2f4b8879ce
9
+ merge_method: multislerp
10
+ tokenizer:
11
+ source: base
12
+ chat_template: llama3
13
+ parameters:
14
+ normalize_weights: false
15
+ eps: 1e-9
16
+ pad_to_multiple_of: 8
17
+ int8_mask: true
18
+ dtype: bfloat16
model-00001-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03b2f6e74d135ea5ba6026e2429e41d17ee2d9824d4a25e1886139981d12f61a
3
+ size 8522938256
model-00002-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51d1a9aaa20fbd1631ab2f9df752cd587df88dcd7ab61182edc440f795774f2d
3
+ size 8562121120
model-00003-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:881f1bd964f8aba51ea83503e7806c557600e05db3db86260448ee5f2a1a3537
3
+ size 8562121120
model-00004-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42921f05804dd418b6df892fffbaaf568d7fefcab2cc524b45b0e410ec8865ab
3
+ size 8562121120
model-00005-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3c605e62b53c2718ab08eaa1bacb8ed34309819401a8b861ec69ada5944ef0f
3
+ size 8562121120
model-00006-of-00006.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4aea4ae38ae19a0a39e49be24845f374553f6013af38b597d6e3b2f44ac90b8
3
+ size 2928824720
model.safetensors.index.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73cb2e515f909a784b79d373740573006696d88b7d7729e3659d12e6c0fb1a57
3
+ size 155376
quantization_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13417a19df521eabc8ee56b0c0c7c03dcf2ea36598e987e5055781251e49787d
3
+ size 593665
special_tokens_map.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94e708c3f5e64acf85bbe5ad01467a1248faadb73e83b41793087ecced586e8f
3
+ size 454
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
tokenizer_config.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2152033590c3bc3e5d4a66356b91f3ace338a57cf2357b68db871e9bc4337909
3
+ size 50569