File size: 906 Bytes
d9c86db e058bac d9c86db e058bac d9c86db e058bac d9c86db e058bac |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
---
license: mit
base_model:
- inclusionAI/Ring-mini-2.0
---
## Introduction
Use https://github.com/im0qianqian/llama.cpp to quantize.
For model inference, please download our release package from this url https://github.com/im0qianqian/llama.cpp/releases .
## Quick start
```bash
# Use a local model file
llama-cli -m my_model.gguf
# Launch OpenAI-compatible API server
llama-server -m my_model.gguf
```
## Demo


## PR
Let's look forward to the following PR being merged:
- [#16063 model : add BailingMoeV2 support](https://github.com/ggml-org/llama.cpp/pull/16063)
- [#16028 Add support for Ling v2](https://github.com/ggml-org/llama.cpp/pull/16028)
|