For inference with sglang and kt-kernel: https://lmsys.org/blog/2025-10-22-KTransformers/

This version is packed specifically for NUMA tensor parallel = 4

Downloads last month
22
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for CPU-Hybrid-MoE/DeepSeek-V3-0324-CPU-NUMA4-AMXINT8

Finetuned
(70)
this model