Text Generation
GGUF
English
reasoning
thinking
uncensored
gated
mixture of experts
Mixture of Experts
8x3B
Llama 3.2 MOE
128k context
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
float32
swearing
rp
horror
mergekit
llama-3
llama-3.2
Update README.md
Browse files
README.md
CHANGED
|
@@ -56,7 +56,8 @@ models into one massive powerhouse at 18.4B parameters (equal to 24B - 8 X 3B).
|
|
| 56 |
|
| 57 |
This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
|
| 58 |
|
| 59 |
-
This model is also "gated", contains a master reasoning model (this can be turned on/off)
|
|
|
|
| 60 |
|
| 61 |
The "gated" strucure means the "reasoning model" is re-inforced by the other 7 models in the MOE during reasoning, and then during
|
| 62 |
output generation / non-reasoning the non-reasoning model(s) take control.
|
|
|
|
| 56 |
|
| 57 |
This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
|
| 58 |
|
| 59 |
+
This model is also "gated", contains a master reasoning model (this can be turned on/off), was built at float32 (32 bit) precision
|
| 60 |
+
and quants have the output tensor at Q8_0, with a few choice quants at f16 (16 bit) and a Q8_0 with f32 (32 bit).
|
| 61 |
|
| 62 |
The "gated" strucure means the "reasoning model" is re-inforced by the other 7 models in the MOE during reasoning, and then during
|
| 63 |
output generation / non-reasoning the non-reasoning model(s) take control.
|