wasiuddina commited on
Commit
54fb6ef
·
verified ·
1 Parent(s): 8478d18

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -10,11 +10,11 @@ tags:
10
  - pytorch
11
  ---
12
 
13
- # OpenCodeReasoning-Distill-Qwen-7B-Instruct
14
 
15
  ## Model Overview
16
 
17
- OpenCodeReasoning-Distill-Qwen-7B-Instruct is a large language model (LLM) which is a derivative of [Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) (AKA the *reference model*).
18
  It is a reasoning model that is post trained for reasoning while code generation. The model supports a context length of 32K tokens.
19
 
20
  This model is ready for commercial use.
@@ -49,7 +49,7 @@ Developers designing AI-powered code generation applications. Also suitable for
49
 
50
  ## Intended use
51
 
52
- OpenCodeReasoning-Distill-Qwen-7B-Instruct is a competitive code generation focused reasoning and chat model intended to be used in English.
53
 
54
  ## Input
55
  - **Input Type:** Text
@@ -87,7 +87,7 @@ Example:
87
  import torch
88
  import transformers
89
 
90
- model_id = "nvidia/OpenCodeReasoning-Distill-Qwen-7B-Instruct"
91
  model_kwargs = {"torch_dtype": torch.bfloat16, "trust_remote_code": True, "device_map": "auto"}
92
  tokenizer = transformers.AutoTokenizer.from_pretrained(model_id)
93
  tokenizer.pad_token_id = tokenizer.eos_token_id
@@ -139,7 +139,7 @@ Data Labeling for Evaluation Datasets:
139
  | Models | Pass@1 |
140
  |:--------------|:------------|
141
  | R1-Distill-Qwen-7B | 37.6 |
142
- | OpenCodeReasoning-Distill-Qwen-7B-Instruct | 51.3 |
143
 
144
  User Prompt Template (without starter code):
145
 
 
10
  - pytorch
11
  ---
12
 
13
+ # OpenCodeReasoning-7B
14
 
15
  ## Model Overview
16
 
17
+ OpenCodeReasoning-7B is a large language model (LLM) which is a derivative of [Qwen2.5-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) (AKA the *reference model*).
18
  It is a reasoning model that is post trained for reasoning while code generation. The model supports a context length of 32K tokens.
19
 
20
  This model is ready for commercial use.
 
49
 
50
  ## Intended use
51
 
52
+ OpenCodeReasoning-7B is a competitive code generation focused reasoning and chat model intended to be used in English.
53
 
54
  ## Input
55
  - **Input Type:** Text
 
87
  import torch
88
  import transformers
89
 
90
+ model_id = "nvidia/OpenCodeReasoning-7B"
91
  model_kwargs = {"torch_dtype": torch.bfloat16, "trust_remote_code": True, "device_map": "auto"}
92
  tokenizer = transformers.AutoTokenizer.from_pretrained(model_id)
93
  tokenizer.pad_token_id = tokenizer.eos_token_id
 
139
  | Models | Pass@1 |
140
  |:--------------|:------------|
141
  | R1-Distill-Qwen-7B | 37.6 |
142
+ | OpenCodeReasoning-7B | 51.3 |
143
 
144
  User Prompt Template (without starter code):
145