Help me run it with llama.cpp and 128K context size

#46
by JLouisBiz - opened

I have tried to run it as:

    echo "Starting LOCAL model: $MODEL"
    $LLAMA_SERVER --jinja -fa on -c "$CONTEXT" -ngl "$NGL" -v \
          --log-timestamps --host "$HOST" \
          --rope-scaling yarn --yarn-orig-ctx 65536 \
          -m "$MODEL" >> "$LOG" 2>&1 &

yet in llama.cpp I can just see this:

2025-12-27_13-15

Is there way to run SmolLM3-3B with 128K and with llama.cpp?

Sign up or log in to comment