Update README.md
Browse files
README.md
CHANGED
|
@@ -24,10 +24,12 @@ base_model: cerebras/Qwen3-Coder-REAP-25B-A3B
|
|
| 24 |
|
| 25 |
# Qwen3-Coder-REAP-25B-A3B-qx65x-hi-mlx
|
| 26 |
|
| 27 |
-
This version of the Deckard(qx) formula uses embeddings at 6 bit, along with the head and select attention paths.
|
| 28 |
|
| 29 |
The model is quantized with group size 32(hi).
|
| 30 |
|
|
|
|
|
|
|
| 31 |
This is an update from the model: [Qwen3-Coder-REAP-25B-A3B-qx64-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Coder-REAP-25B-A3B-qx64-hi-mlx) that uses the base and embeddings at 4 bit.
|
| 32 |
|
| 33 |
Metrics coming soon.
|
|
|
|
| 24 |
|
| 25 |
# Qwen3-Coder-REAP-25B-A3B-qx65x-hi-mlx
|
| 26 |
|
| 27 |
+
This version of the Deckard(qx) formula uses embeddings at 6 bit, along with the head and select attention paths, leaving the rest at 5 bit.
|
| 28 |
|
| 29 |
The model is quantized with group size 32(hi).
|
| 30 |
|
| 31 |
+
It is aimed as a mid-range quant with a quality approaching q8, that would run comfortably on a smaller Mac.
|
| 32 |
+
|
| 33 |
This is an update from the model: [Qwen3-Coder-REAP-25B-A3B-qx64-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Coder-REAP-25B-A3B-qx64-hi-mlx) that uses the base and embeddings at 4 bit.
|
| 34 |
|
| 35 |
Metrics coming soon.
|