Christian Otto Stelter PRO
stelterlab
AI & ML interests
None yet
Recent Activity
liked a model 8 days ago
Sehyo/Qwen3.5-35B-A3B-NVFP4 new activity 11 days ago
stelterlab/Qwen3-30B-A3B-Instruct-2507-AWQ:Qwen3.5-35B-A3B AWQ quant planned? new activity 25 days ago
RedHatAI/Qwen3.5-397B-A17B-FP8-dynamic:Which transformer version did you use?Organizations
None yet
Qwen3.5-35B-A3B AWQ quant planned?
2
#1 opened 15 days ago
by
amidwestnoob
Which transformer version did you use?
#3 opened 25 days ago
by
stelterlab
tokenizer_config.json missing chat_template field (tool calling broken without workaround)
1
#1 opened 26 days ago
by
seanthomaswilliams
Updated tokenizer_config.json now w/ chat_template included
#2 opened 26 days ago
by
stelterlab
NVFP4 / AWQ Quants or llm-compressor recipe
🤗 1
1
#1 opened 4 months ago
by
stelterlab
vLLM v0.11.1 seems to work, but v0.11.2 fails
👍❤️ 2
9
#3 opened 4 months ago
by
stelterlab
Error when running in VLLM
👍 2
21
#1 opened 6 months ago
by
d8rt8v
Unable to run the model in VLLM: KeyError: 'layers.14.mlp.gate.qweight'
3
#1 opened 8 months ago
by
fredericodeveloper
Rope Scaling pre-applied?
6
#1 opened 8 months ago
by
the1dv
AWQ version
👍 14
13
#8 opened 9 months ago
by
celsowm
How did you use auto-round to quantize?
3
#4 opened 9 months ago
by
stelterlab
please update to Mistral-Small-3.2-24B-Instruct-2506
1
#5 opened 9 months ago
by
celsowm
Tool Calling issue with stelterlab/Mistral-Small-24B-Instruct-2501-AWQ
1
#4 opened 10 months ago
by
sbhatt765
Do you any plan to quantize the Qwen3-30B-A3B-AWQ model?
1
#2 opened 11 months ago
by
Jeanxx
Could you share the code script that convert the original Qwen3-8B to Qwen3-8B-AWQ?
1
#1 opened 11 months ago
by
wenmin-wu
gpqa diamond
1
#1 opened 11 months ago
by
madferit421
How to reduce "Think" responses when using vLLM for inference?
1
#1 opened 11 months ago
by
rjsng0904
Really good work
🔥 2
11
#1 opened about 1 year ago
by
divmgl
FP8 Dynamic/W8A16 Quants Please
4
#44 opened 12 months ago
by
rjmehta
i love you
1
#1 opened 12 months ago
by
nisten