Files changed (1) hide show
  1. README.md +118 -104
README.md CHANGED
@@ -1,104 +1,118 @@
1
- ---
2
- library_name: transformers
3
- license: apache-2.0
4
- base_model: Qwen/Qwen2.5-7B-Instruct
5
- tags:
6
- - llama-factory
7
- - full
8
- - generated_from_trainer
9
- model-index:
10
- - name: OpenThinker-7B
11
- results: []
12
- datasets:
13
- - open-thoughts/open-thoughts-114k
14
- ---
15
-
16
- <p align="center">
17
- <img src="https://huggingface.co/datasets/open-thoughts/open-thoughts-114k/resolve/main/open_thoughts.png" width="50%">
18
- </p>
19
-
20
- # OpenThinker-7B
21
-
22
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the
23
- [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k) dataset.
24
-
25
- The dataset is derived by distilling DeepSeek-R1 using the [data pipeline available on github](https://github.com/open-thoughts/open-thoughts).
26
- More info about the dataset can be found on the dataset card at [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
27
-
28
- This model improves upon the [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B), which used 17k examples ([Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)).
29
- The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
30
-
31
- | | AIME24 | MATH500 | GPQA-Diamond | LCBv2 Easy | LCBv2 Medium | LCBv2 Hard | LCBv2 All |
32
- | --------------------------- | -------- | ------- | ------------ | ----------- | ------------- | ----------- | ---------- |
33
- | OpenThinker-7B | 31.3 | 83.0 | 42.4 | 75.3 | 28.6 | 6.5 | 39.9 |
34
- | Bespoke-Stratos-7B | 22.7 | 79.6 | 38.9 | 71.4 | 25.2 | 0.8 | 35.8 |
35
- | DeepSeek-R1-Distill-Qwen-7B | 60 | 88.2 | 46.9 | 79.7 | 45.1 | 14.6 | 50.1 |
36
- | gpt-4o-0513 | 8.7 | 75.8 | 46.5 | 87.4 | 42.7 | 8.9 | 50.5 |
37
- | o1-mini | 64 | 85.6 | 60 | 92.8 | 74.7 | 39.8 | 72.8 |
38
-
39
- We are fully open-source. Our [model weights](https://huggingface.co/open-thoughts), [datasets](https://huggingface.co/open-thoughts), [data generation code](https://github.com/open-thoughts/open-thoughts), [evaluation code](https://github.com/mlfoundations/Evalchemy), and [training code](https://github.com/hiyouga/LLaMA-Factory) are all publicly available.
40
-
41
- | | Open Weights | Open Data | Open Code |
42
- |--|--------------|-----------| --------- |
43
- |OpenThinker-7B|βœ…|[βœ…](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k)|[βœ…](https://github.com/open-thoughts/open-thoughts) |
44
- |Bespoke-Stratos-7B|βœ…|[βœ…](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)|[βœ…](https://github.com/bespokelabsai/curator/tree/main/examples/bespoke-stratos-data-generation)|
45
- |DeepSeek-R1-Distill-Qwen-7B|βœ…|❌|❌|
46
- |gpt-4o-0513|❌|❌|❌|❌|
47
- |o1-mini|❌|❌|❌|❌|
48
-
49
-
50
- ## Intended uses & limitations
51
-
52
- Apache 2.0 License
53
-
54
-
55
- ## Training procedure
56
-
57
- We used four 8xH100 nodes to train the model for 20 hours.
58
-
59
- ### Training hyperparameters
60
-
61
- The following hyperparameters were used during training:
62
- - learning_rate: 1e-05
63
- - train_batch_size: 1
64
- - eval_batch_size: 8
65
- - seed: 42
66
- - distributed_type: multi-GPU
67
- - num_devices: 32
68
- - gradient_accumulation_steps: 3
69
- - total_train_batch_size: 96
70
- - total_eval_batch_size: 256
71
- - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
72
- - lr_scheduler_type: cosine
73
- - lr_scheduler_warmup_ratio: 0.1
74
- - num_epochs: 3.0
75
-
76
- ### Framework versions
77
-
78
- - Transformers 4.46.1
79
- - Pytorch 2.3.0
80
- - Datasets 3.1.0
81
- - Tokenizers 0.20.3
82
-
83
- More info can be found in our repository: [https://github.com/open-thoughts/open-thoughts](https://github.com/open-thoughts/open-thoughts).
84
-
85
- # Citation
86
- ```
87
- @misc{openthoughts,
88
- author = {Team, OpenThoughts},
89
- month = jan,
90
- title = {{Open Thoughts}},
91
- howpublished = {https://open-thoughts.ai},
92
- year = {2025}
93
- }
94
- ```
95
-
96
- # Links
97
- - πŸ“Š [Open Thoughts Launch Blog Post](https://www.open-thoughts.ai/blog/launch)
98
- - πŸ’» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
99
- - 🧠 [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k)
100
- - πŸ€– [OpenThinker-7B model](https://huggingface.co/open-thoughts/OpenThinker-7B) - this model.
101
- - πŸ“Š [Bespoke-Stratos Blog Post](https://www.bespokelabs.ai/blog/bespoke-stratos-the-unreasonable-effectiveness-of-reasoning-distillation)
102
- - 🧠 [Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)
103
- - πŸ€– [Bespoke-Stratos-32B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-32B)
104
- - πŸ€– [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen2.5-7B-Instruct
5
+ tags:
6
+ - llama-factory
7
+ - full
8
+ - generated_from_trainer
9
+ datasets:
10
+ - open-thoughts/open-thoughts-114k
11
+ language:
12
+ - zho
13
+ - eng
14
+ - fra
15
+ - spa
16
+ - por
17
+ - deu
18
+ - ita
19
+ - rus
20
+ - jpn
21
+ - kor
22
+ - vie
23
+ - tha
24
+ - ara
25
+ model-index:
26
+ - name: OpenThinker-7B
27
+ results: []
28
+ ---
29
+
30
+ <p align="center">
31
+ <img src="https://huggingface.co/datasets/open-thoughts/open-thoughts-114k/resolve/main/open_thoughts.png" width="50%">
32
+ </p>
33
+
34
+ # OpenThinker-7B
35
+
36
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) on the
37
+ [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k) dataset.
38
+
39
+ The dataset is derived by distilling DeepSeek-R1 using the [data pipeline available on github](https://github.com/open-thoughts/open-thoughts).
40
+ More info about the dataset can be found on the dataset card at [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/open-thoughts-114k).
41
+
42
+ This model improves upon the [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B), which used 17k examples ([Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)).
43
+ The numbers reported in the table below are evaluated with our open-source tool [Evalchemy](https://github.com/mlfoundations/Evalchemy).
44
+
45
+ | | AIME24 | MATH500 | GPQA-Diamond | LCBv2 Easy | LCBv2 Medium | LCBv2 Hard | LCBv2 All |
46
+ | --------------------------- | -------- | ------- | ------------ | ----------- | ------------- | ----------- | ---------- |
47
+ | OpenThinker-7B | 31.3 | 83.0 | 42.4 | 75.3 | 28.6 | 6.5 | 39.9 |
48
+ | Bespoke-Stratos-7B | 22.7 | 79.6 | 38.9 | 71.4 | 25.2 | 0.8 | 35.8 |
49
+ | DeepSeek-R1-Distill-Qwen-7B | 60 | 88.2 | 46.9 | 79.7 | 45.1 | 14.6 | 50.1 |
50
+ | gpt-4o-0513 | 8.7 | 75.8 | 46.5 | 87.4 | 42.7 | 8.9 | 50.5 |
51
+ | o1-mini | 64 | 85.6 | 60 | 92.8 | 74.7 | 39.8 | 72.8 |
52
+
53
+ We are fully open-source. Our [model weights](https://huggingface.co/open-thoughts), [datasets](https://huggingface.co/open-thoughts), [data generation code](https://github.com/open-thoughts/open-thoughts), [evaluation code](https://github.com/mlfoundations/Evalchemy), and [training code](https://github.com/hiyouga/LLaMA-Factory) are all publicly available.
54
+
55
+ | | Open Weights | Open Data | Open Code |
56
+ |--|--------------|-----------| --------- |
57
+ |OpenThinker-7B|βœ…|[βœ…](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k)|[βœ…](https://github.com/open-thoughts/open-thoughts) |
58
+ |Bespoke-Stratos-7B|βœ…|[βœ…](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)|[βœ…](https://github.com/bespokelabsai/curator/tree/main/examples/bespoke-stratos-data-generation)|
59
+ |DeepSeek-R1-Distill-Qwen-7B|βœ…|❌|❌|
60
+ |gpt-4o-0513|❌|❌|❌|❌|
61
+ |o1-mini|❌|❌|❌|❌|
62
+
63
+
64
+ ## Intended uses & limitations
65
+
66
+ Apache 2.0 License
67
+
68
+
69
+ ## Training procedure
70
+
71
+ We used four 8xH100 nodes to train the model for 20 hours.
72
+
73
+ ### Training hyperparameters
74
+
75
+ The following hyperparameters were used during training:
76
+ - learning_rate: 1e-05
77
+ - train_batch_size: 1
78
+ - eval_batch_size: 8
79
+ - seed: 42
80
+ - distributed_type: multi-GPU
81
+ - num_devices: 32
82
+ - gradient_accumulation_steps: 3
83
+ - total_train_batch_size: 96
84
+ - total_eval_batch_size: 256
85
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
86
+ - lr_scheduler_type: cosine
87
+ - lr_scheduler_warmup_ratio: 0.1
88
+ - num_epochs: 3.0
89
+
90
+ ### Framework versions
91
+
92
+ - Transformers 4.46.1
93
+ - Pytorch 2.3.0
94
+ - Datasets 3.1.0
95
+ - Tokenizers 0.20.3
96
+
97
+ More info can be found in our repository: [https://github.com/open-thoughts/open-thoughts](https://github.com/open-thoughts/open-thoughts).
98
+
99
+ # Citation
100
+ ```
101
+ @misc{openthoughts,
102
+ author = {Team, OpenThoughts},
103
+ month = jan,
104
+ title = {{Open Thoughts}},
105
+ howpublished = {https://open-thoughts.ai},
106
+ year = {2025}
107
+ }
108
+ ```
109
+
110
+ # Links
111
+ - πŸ“Š [Open Thoughts Launch Blog Post](https://www.open-thoughts.ai/blog/launch)
112
+ - πŸ’» [Open Thoughts GitHub Repository](https://github.com/open-thoughts/open-thoughts)
113
+ - 🧠 [OpenThoughts-114k dataset](https://huggingface.co/datasets/open-thoughts/OpenThoughts-114k)
114
+ - πŸ€– [OpenThinker-7B model](https://huggingface.co/open-thoughts/OpenThinker-7B) - this model.
115
+ - πŸ“Š [Bespoke-Stratos Blog Post](https://www.bespokelabs.ai/blog/bespoke-stratos-the-unreasonable-effectiveness-of-reasoning-distillation)
116
+ - 🧠 [Bespoke-Stratos-17k dataset](https://huggingface.co/datasets/bespokelabs/Bespoke-Stratos-17k)
117
+ - πŸ€– [Bespoke-Stratos-32B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-32B)
118
+ - πŸ€– [Bespoke-Stratos-7B model](https://huggingface.co/bespokelabs/Bespoke-Stratos-7B)