Update README.md
Browse files
README.md
CHANGED
@@ -10,11 +10,11 @@ tags:
|
|
10 |
- pytorch
|
11 |
---
|
12 |
|
13 |
-
# OpenCodeReasoning-
|
14 |
|
15 |
## Model Overview
|
16 |
|
17 |
-
OpenCodeReasoning-
|
18 |
It is a reasoning model that is post trained for reasoning while code generation. The model supports a context length of 32K tokens.
|
19 |
|
20 |
This model is ready for commercial use.
|
@@ -49,7 +49,7 @@ Developers designing AI-powered code generation applications. Also suitable for
|
|
49 |
|
50 |
## Intended use
|
51 |
|
52 |
-
OpenCodeReasoning-
|
53 |
|
54 |
## Input
|
55 |
- **Input Type:** Text
|
@@ -87,7 +87,7 @@ Example:
|
|
87 |
import torch
|
88 |
import transformers
|
89 |
|
90 |
-
model_id = "nvidia/OpenCodeReasoning-
|
91 |
model_kwargs = {"torch_dtype": torch.bfloat16, "trust_remote_code": True, "device_map": "auto"}
|
92 |
tokenizer = transformers.AutoTokenizer.from_pretrained(model_id)
|
93 |
tokenizer.pad_token_id = tokenizer.eos_token_id
|
@@ -139,7 +139,7 @@ Data Labeling for Evaluation Datasets:
|
|
139 |
| Models | Pass@1 |
|
140 |
|:--------------|:------------|
|
141 |
| R1-Distill-Qwen-7B | 37.6 |
|
142 |
-
| OpenCodeReasoning-
|
143 |
|
144 |
User Prompt Template (without starter code):
|
145 |
|
|
|
10 |
- pytorch
|
11 |
---
|
12 |
|
13 |
+
# OpenCodeReasoning-7B
|
14 |
|
15 |
## Model Overview
|
16 |
|
17 |
+
OpenCodeReasoning-7B is a large language model (LLM) which is a derivative of [Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) (AKA the *reference model*).
|
18 |
It is a reasoning model that is post trained for reasoning while code generation. The model supports a context length of 32K tokens.
|
19 |
|
20 |
This model is ready for commercial use.
|
|
|
49 |
|
50 |
## Intended use
|
51 |
|
52 |
+
OpenCodeReasoning-7B is a competitive code generation focused reasoning and chat model intended to be used in English.
|
53 |
|
54 |
## Input
|
55 |
- **Input Type:** Text
|
|
|
87 |
import torch
|
88 |
import transformers
|
89 |
|
90 |
+
model_id = "nvidia/OpenCodeReasoning-7B"
|
91 |
model_kwargs = {"torch_dtype": torch.bfloat16, "trust_remote_code": True, "device_map": "auto"}
|
92 |
tokenizer = transformers.AutoTokenizer.from_pretrained(model_id)
|
93 |
tokenizer.pad_token_id = tokenizer.eos_token_id
|
|
|
139 |
| Models | Pass@1 |
|
140 |
|:--------------|:------------|
|
141 |
| R1-Distill-Qwen-7B | 37.6 |
|
142 |
+
| OpenCodeReasoning-7B | 51.3 |
|
143 |
|
144 |
User Prompt Template (without starter code):
|
145 |
|