This 4bit W4A16 model has been quantized using GPTQModel.

ARC_Challenge and MMLU evals pending.

Downloads last month
70
Safetensors
Model size
357B params
Tensor type
BF16
F16
I32
F32
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for ModelCloud/GLM-4.6-GPTQMODEL-W4A16-v2

Base model

zai-org/GLM-4.6
Quantized
(32)
this model