This is a gguf version of Apertus-8B
To use it, please :
- clone the llama.cpp repository: https://github.com/ggml-org/llama.cpp
- add Apertus-8B-Instruct.jinja inside the folder llama.cpp\models\templates
- go to llama.cpp\build\bin
- run the following command: llama-cli -m --chat-template-file ../../models/templates/Apertus-8B-Instruct.jinja -i --color -n 512 -c 4096 --jinja
I used WSL to have it running.
- Downloads last month
- 102
Hardware compatibility
Log In
to view the estimation
16-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support