Help me run it with llama.cpp and 128K context size
#46
by
JLouisBiz
- opened
I have tried to run it as:
echo "Starting LOCAL model: $MODEL"
$LLAMA_SERVER --jinja -fa on -c "$CONTEXT" -ngl "$NGL" -v \
--log-timestamps --host "$HOST" \
--rope-scaling yarn --yarn-orig-ctx 65536 \
-m "$MODEL" >> "$LOG" 2>&1 &
yet in llama.cpp I can just see this:
Is there way to run SmolLM3-3B with 128K and with llama.cpp?
