metadata
license: mit
base_model:
- inclusionAI/Ring-mini-2.0
Introduction
Use https://github.com/im0qianqian/llama.cpp to quantize.
For model inference, please download our release package from this url https://github.com/im0qianqian/llama.cpp/releases .
Quick start
# Use a local model file
llama-cli -m my_model.gguf
# Launch OpenAI-compatible API server
llama-server -m my_model.gguf
Demo
PR
Let's look forward to the following PR being merged:

