Update README.md
Browse files
README.md
CHANGED
|
@@ -21,6 +21,14 @@ library_name: transformers
|
|
| 21 |
|
| 22 |
This model is an 8-bit quantized model for [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2) using [AutoGPTQ](https://github.com/AutoGPTQ/AutoGPTQ). The quantized version is 2x smaller than the original model and thus requires less memory and provides faster inference.
|
| 23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
See [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2) for details about model architecture and data.
|
| 25 |
|
| 26 |
* **Contributors**
|
|
@@ -28,6 +36,10 @@ See [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.
|
|
| 28 |
- [Xinqi Chen](https://huggingface.co/Keely0419)
|
| 29 |
- [Kei Sawada](https://huggingface.co/keisawada)
|
| 30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
---
|
| 32 |
|
| 33 |
# Benchmarking
|
|
@@ -44,7 +56,7 @@ See [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.
|
|
| 44 |
| [Qwen/QwQ-32B](https://huggingface.co/Qwen/QwQ-32B) | 76.12 | 8.58 | 8.25
|
| 45 |
| [rinna/qwq-bakeneko-32b](https://huggingface.co/rinna/qwq-bakeneko-32b) | 78.31 | 8.81 | 8.52
|
| 46 |
|
| 47 |
-
For detailed benchmarking results, please refer to [rinna's LM benchmark page](https://rinnakk.github.io/research/benchmarks/lm/index.html).
|
| 48 |
|
| 49 |
---
|
| 50 |
|
|
|
|
| 21 |
|
| 22 |
This model is an 8-bit quantized model for [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2) using [AutoGPTQ](https://github.com/AutoGPTQ/AutoGPTQ). The quantized version is 2x smaller than the original model and thus requires less memory and provides faster inference.
|
| 23 |
|
| 24 |
+
| Model Type | Model Name
|
| 25 |
+
| :- | :-
|
| 26 |
+
| Japanese Continual Pre-Training Model | Qwen2.5 Bakeneko 32B [[HF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b)
|
| 27 |
+
| Instruction-Tuning Model | Qwen2.5 Bakeneko 32B Instruct [[HF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct)[[AWQ]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-awq)[[GGUF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-gguf)[[GPTQ int8]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-gptq-int4)
|
| 28 |
+
| DeepSeek R1 Distill Qwen2.5 Merged Reasoning Model | DeepSeek R1 Distill Qwen2.5 Bakeneko 32B [[HF]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b)[[AWQ]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-awq)[[GGUF]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-gguf)[[GPTQ int8]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/deepseek-r1-distill-qwen2.5-bakeneko-32b-gptq-int4)
|
| 29 |
+
| QwQ Merged Reasoning Model | QwQ Bakeneko 32B [[HF]](https://huggingface.co/rinna/qwq-bakeneko-32b)[[AWQ]](https://huggingface.co/rinna/qwq-bakeneko-32b-awq)[[GGUF]](https://huggingface.co/rinna/qwq-bakeneko-32b-gguf)[[GPTQ int8]](https://huggingface.co/rinna/qwq-bakeneko-32b-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/qwq-bakeneko-32b-gptq-int4)
|
| 30 |
+
| QwQ Bakeneko Merged Instruction-Tuning Model | Qwen2.5 Bakeneko 32B Instruct V2 [[HF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2)[[AWQ]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-awq)[[GGUF]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-gguf)[[GPTQ int8]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-gptq-int8)[[GPTQ int4]](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2-gptq-int4)
|
| 31 |
+
|
| 32 |
See [rinna/qwen2.5-bakeneko-32b-instruct-v2](https://huggingface.co/rinna/qwen2.5-bakeneko-32b-instruct-v2) for details about model architecture and data.
|
| 33 |
|
| 34 |
* **Contributors**
|
|
|
|
| 36 |
- [Xinqi Chen](https://huggingface.co/Keely0419)
|
| 37 |
- [Kei Sawada](https://huggingface.co/keisawada)
|
| 38 |
|
| 39 |
+
* **Release date**
|
| 40 |
+
|
| 41 |
+
February 19, 2025
|
| 42 |
+
|
| 43 |
---
|
| 44 |
|
| 45 |
# Benchmarking
|
|
|
|
| 56 |
| [Qwen/QwQ-32B](https://huggingface.co/Qwen/QwQ-32B) | 76.12 | 8.58 | 8.25
|
| 57 |
| [rinna/qwq-bakeneko-32b](https://huggingface.co/rinna/qwq-bakeneko-32b) | 78.31 | 8.81 | 8.52
|
| 58 |
|
| 59 |
+
For detailed benchmarking results, please refer to [rinna's LM benchmark page (Sheet 20250319)](https://rinnakk.github.io/research/benchmarks/lm/index.html).
|
| 60 |
|
| 61 |
---
|
| 62 |
|