DavidAU commited on
Commit
1f19862
·
verified ·
1 Parent(s): 6fd5c35

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -3
README.md CHANGED
@@ -56,13 +56,18 @@ models into one massive powerhouse at 18.4B parameters (equal to 24B - 8 X 3B).
56
 
57
  This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
58
 
59
- This model is also "gated", contains a master reasoning model (this can be turned on/off) and was built at float32 precision.
60
 
61
- This is the "reasoning / gated version" of this model:
 
 
 
 
 
62
 
63
  [ https://huggingface.co/DavidAU/Llama-3.2-8X3B-MOE-Dark-Champion-Instruct-uncensored-abliterated-18.4B-GGUF ]
64
 
65
- (193 likes // 447,000+ downloads to date. )
66
 
67
  And it is fast: 50+ t/s (2 experts) on a low end 16GB card, IQ4XS.
68
 
 
56
 
57
  This model's instruction following, and output generation for creative writing, prose, fiction and role play are exceptional.
58
 
59
+ This model is also "gated", contains a master reasoning model (this can be turned on/off) and was built at float32 (32 bit) precision.
60
 
61
+ The "gated" strucure means the "reasoning model" is re-inforced by the other 7 models in the MOE during reasoning, and then during
62
+ output generation / non-reasoning the non-reasoning model(s) take control.
63
+
64
+ Also, with "gating" you can directly access/control the model(s) you want to use during instruction following and generation.
65
+
66
+ This model is the "reasoning / gated version" of this model:
67
 
68
  [ https://huggingface.co/DavidAU/Llama-3.2-8X3B-MOE-Dark-Champion-Instruct-uncensored-abliterated-18.4B-GGUF ]
69
 
70
+ (193 likes // 447,000+ downloads to date. (Hugging face only shows last 30 days at the repo) )
71
 
72
  And it is fast: 50+ t/s (2 experts) on a low end 16GB card, IQ4XS.
73