Trained from anthracite-core/Mistral-Small-3.2-24B-Instruct-2506-ChatML for convenience (no Pixtral compatibility needed). No vision adapter currently.

Test model trained at 16k context on 50M tokens of long-form human writing (mostly books).

Haven't tested yet but regular Tekken v7 instruct will work and samplers are probably the same as you'd use for 3.2 Instruct.

Downloads last month
1
Safetensors
Model size
24B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Burnt-Toast/ms3.2-24b-longform