Add pipeline tag and usage examples
#4
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,7 +1,9 @@
|
|
| 1 |
---
|
|
|
|
|
|
|
|
|
|
| 2 |
library_name: transformers
|
| 3 |
license: apache-2.0
|
| 4 |
-
base_model: Qwen/Qwen2.5-32B-Instruct
|
| 5 |
tags:
|
| 6 |
- llama-factory
|
| 7 |
- full
|
|
@@ -9,8 +11,7 @@ tags:
|
|
| 9 |
model-index:
|
| 10 |
- name: OpenThinker2-32B
|
| 11 |
results: []
|
| 12 |
-
|
| 13 |
-
- open-thoughts/OpenThoughts2-1M
|
| 14 |
---
|
| 15 |
|
| 16 |
<p align="center">
|
|
@@ -20,12 +21,11 @@ datasets:
|
|
| 20 |
> [!NOTE]
|
| 21 |
> We have released a paper for OpenThoughts! See our paper [here](https://arxiv.org/abs/2506.04178).
|
| 22 |
|
| 23 |
-
# OpenThinker2-32B
|
| 24 |
|
| 25 |
-
|
| 26 |
-
[OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
|
| 27 |
|
| 28 |
-
The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
|
| 29 |
This model improves upon our previous [OpenThinker-32B](https://huggingface.co/open-thoughts/OpenThinker-32B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
|
| 30 |
The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
|
| 31 |
|
|
@@ -38,6 +38,40 @@ The numbers reported in the table below are evaluated with our open-source tool
|
|
| 38 |
| [S1.1-32B](https://huggingface.co/simplescaling/s1.1-32B) | β
| 59.3 | 42.7 | 91.5 | 87.4 | 62.0 | 58.7 |
|
| 39 |
|
| 40 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
## Data
|
| 42 |
|
| 43 |
This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
|
|
@@ -85,6 +119,7 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
|
|
| 85 |
- π [OpenThoughts Paper](https://arxiv.org/abs/2506.04178)
|
| 86 |
- π [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
|
| 87 |
- π» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
|
|
|
|
| 88 |
- π§ [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
|
| 89 |
- π€ [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
|
| 90 |
- π€ [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.
|
|
@@ -100,4 +135,4 @@ More info can be found in our repository: [https://github.com/open-thoughts/open
|
|
| 100 |
primaryClass={cs.LG},
|
| 101 |
url={https://arxiv.org/abs/2506.04178},
|
| 102 |
}
|
| 103 |
-
```
|
|
|
|
| 1 |
---
|
| 2 |
+
base_model: Qwen/Qwen2.5-32B-Instruct
|
| 3 |
+
datasets:
|
| 4 |
+
- open-thoughts/OpenThoughts2-1M
|
| 5 |
library_name: transformers
|
| 6 |
license: apache-2.0
|
|
|
|
| 7 |
tags:
|
| 8 |
- llama-factory
|
| 9 |
- full
|
|
|
|
| 11 |
model-index:
|
| 12 |
- name: OpenThinker2-32B
|
| 13 |
results: []
|
| 14 |
+
pipeline_tag: text-generation
|
|
|
|
| 15 |
---
|
| 16 |
|
| 17 |
<p align="center">
|
|
|
|
| 21 |
> [!NOTE]
|
| 22 |
> We have released a paper for OpenThoughts! See our paper [here](https://arxiv.org/abs/2506.04178).
|
| 23 |
|
| 24 |
+
# OpenThinker2-32B: A Powerful Open-Data Reasoning Model
|
| 25 |
|
| 26 |
+
OpenThinker2-32B is a fine-tuned version of [Qwen/Qwen2.5-32B-Instruct](https://huggingface.co/Qwen/Qwen2.5-32B-Instruct) trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset. This model represents a significant advancement in open-source reasoning capabilities, achieving competitive performance with closed-source models on various benchmarks. OpenThinker2-32B excels at complex reasoning tasks involving math, code, and general knowledge, demonstrating its versatility and robustness.
|
|
|
|
| 27 |
|
| 28 |
+
The [OpenThinker2-32B](https://huggingface.co/open-thoughts/OpenThinker2-32B) model is the highest performing open-data model.
|
| 29 |
This model improves upon our previous [OpenThinker-32B](https://huggingface.co/open-thoughts/OpenThinker-32B) model, which was trained on 114k examples from [OpenThoughts-114k](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
|
| 30 |
The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
|
| 31 |
|
|
|
|
| 38 |
| [S1.1-32B](https://huggingface.co/simplescaling/s1.1-32B) | β
| 59.3 | 42.7 | 91.5 | 87.4 | 62.0 | 58.7 |
|
| 39 |
|
| 40 |
|
| 41 |
+
## Usage Examples
|
| 42 |
+
|
| 43 |
+
This model can be easily used with the Hugging Face `pipeline` API for text generation.
|
| 44 |
+
|
| 45 |
+
**Example 1: Simple Text Generation**
|
| 46 |
+
|
| 47 |
+
```python
|
| 48 |
+
from transformers import pipeline
|
| 49 |
+
|
| 50 |
+
generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
|
| 51 |
+
result = generator("Once upon a time,", max_length=50)
|
| 52 |
+
print(result[0]['generated_text'])
|
| 53 |
+
```
|
| 54 |
+
|
| 55 |
+
**Example 2: Controlling the Length of the Generated Text**
|
| 56 |
+
|
| 57 |
+
```python
|
| 58 |
+
from transformers import pipeline
|
| 59 |
+
|
| 60 |
+
generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
|
| 61 |
+
result = generator("The quick brown fox jumps over the lazy dog.", max_length=100)
|
| 62 |
+
print(result[0]['generated_text'])
|
| 63 |
+
```
|
| 64 |
+
|
| 65 |
+
**Example 3: Setting the Temperature**
|
| 66 |
+
|
| 67 |
+
```python
|
| 68 |
+
from transformers import pipeline
|
| 69 |
+
|
| 70 |
+
generator = pipeline('text-generation', model='open-thoughts/OpenThinker2-32B')
|
| 71 |
+
result = generator("Write a short poem about nature:", max_length=50, temperature=0.7)
|
| 72 |
+
print(result[0]['generated_text'])
|
| 73 |
+
```
|
| 74 |
+
|
| 75 |
## Data
|
| 76 |
|
| 77 |
This model was trained on the [OpenThoughts2-1M](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M) dataset.
|
|
|
|
| 119 |
- π [OpenThoughts Paper](https://arxiv.org/abs/2506.04178)
|
| 120 |
- π [OpenThoughts2 and OpenThinker2 Blog Post](https://www.open-thoughts.ai/blog/thinkagain)
|
| 121 |
- π» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
|
| 122 |
+
- π [Project Page](https://openthoughts.ai)
|
| 123 |
- π§ [OpenThoughts2-1M dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts2-1M)
|
| 124 |
- π€ [OpenThinker2-7B model](https://huggingface.co/open-thoughts/OpenThinker2-7B)
|
| 125 |
- π€ [OpenThinker2-32B model](https://huggingface.co/open-thoughts/OpenThinker2-32B) - this model.
|
|
|
|
| 135 |
primaryClass={cs.LG},
|
| 136 |
url={https://arxiv.org/abs/2506.04178},
|
| 137 |
}
|
| 138 |
+
```
|