This is a MXFP4_MOE quantization of the model Qwen3-Coder-REAP-246B-A35B

Original model: https://huggingface.co/cerebras/Qwen3-Coder-REAP-246B-A35B-FP8

The model was originally in FP8, which limits its precision.
I attempted to apply MXFP4 quantization after converting the model to FP32, but the quality degradation from the initial FP8 quantization cannot be fully reversed.

Downloads last month
242
GGUF
Model size
246B params
Architecture
qwen3moe
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for noctrex/Qwen3-Coder-REAP-246B-A35B-MXFP4_MOE-GGUF

Collection including noctrex/Qwen3-Coder-REAP-246B-A35B-MXFP4_MOE-GGUF