feihu.hf
commited on
Commit
·
7c0a8dc
1
Parent(s):
2ebdfe8
update readme
Browse files- README.md +8 -1
- figures/benchmark.jpg +0 -0
README.md
CHANGED
|
@@ -19,6 +19,11 @@ tags:
|
|
| 19 |
|
| 20 |
QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
|
| 21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
**This repo contains the QwQ 32B model**, which has the following features:
|
| 23 |
- Type: Causal Language Models
|
| 24 |
- Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
|
|
@@ -31,6 +36,8 @@ QwQ is the reasoning model of the Qwen series. Compared with conventional instru
|
|
| 31 |
|
| 32 |
**Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
|
| 33 |
|
|
|
|
|
|
|
| 34 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
|
| 35 |
|
| 36 |
## Requirements
|
|
@@ -126,7 +133,7 @@ If you find our work helpful, feel free to give us a cite.
|
|
| 126 |
|
| 127 |
```
|
| 128 |
@misc{qwq32b,
|
| 129 |
-
title = {
|
| 130 |
url = {https://qwenlm.github.io/blog/qwq-32b/},
|
| 131 |
author = {Qwen Team},
|
| 132 |
month = {March},
|
|
|
|
| 19 |
|
| 20 |
QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.
|
| 21 |
|
| 22 |
+
<p align="center">
|
| 23 |
+
<img width="100%" src="figures/benchmark.jpg">
|
| 24 |
+
</p>
|
| 25 |
+
|
| 26 |
+
|
| 27 |
**This repo contains the QwQ 32B model**, which has the following features:
|
| 28 |
- Type: Causal Language Models
|
| 29 |
- Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
|
|
|
|
| 36 |
|
| 37 |
**Note:** For the best experience, please review the [usage guidelines](#usage-guidelines) before deploying QwQ models.
|
| 38 |
|
| 39 |
+
You can try our [demo](https://huggingface.co/spaces/Qwen/QwQ-32B-Demo) or access QwQ models via [QwenChat](https://chat.qwen.ai).
|
| 40 |
+
|
| 41 |
For more details, please refer to our [blog](https://qwenlm.github.io/blog/qwq-32b/), [GitHub](https://github.com/QwenLM/Qwen2.5), and [Documentation](https://qwen.readthedocs.io/en/latest/).
|
| 42 |
|
| 43 |
## Requirements
|
|
|
|
| 133 |
|
| 134 |
```
|
| 135 |
@misc{qwq32b,
|
| 136 |
+
title = {QwQ-32B: The Power of Scaling RL},
|
| 137 |
url = {https://qwenlm.github.io/blog/qwq-32b/},
|
| 138 |
author = {Qwen Team},
|
| 139 |
month = {March},
|
figures/benchmark.jpg
ADDED
|