Gustavo de Rosa
commited on
Commit
·
0f18025
1
Parent(s):
d8d3b44
chore(root): Adds top_k information even if 50 is already the default.
Browse files- README.md +2 -1
- generation_config.json +1 -0
README.md
CHANGED
@@ -56,7 +56,7 @@ library_name: transformers
|
|
56 |
## Usage
|
57 |
|
58 |
> [!IMPORTANT]
|
59 |
-
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
60 |
|
61 |
*Phi-4-reasoning-plus has shown strong performance on reasoning-intensive tasks. In our experiments, we extended its maximum number of tokens to 64k, and it handled longer sequences with promising results, maintaining coherence and logical consistency over extended inputs. This makes it a compelling option to explore for tasks that require deep, multi-step reasoning or extensive context.*
|
62 |
|
@@ -90,6 +90,7 @@ outputs = model.generate(
|
|
90 |
inputs.to(model.device),
|
91 |
max_new_tokens=4096,
|
92 |
temperature=0.8,
|
|
|
93 |
top_p=0.95,
|
94 |
do_sample=True,
|
95 |
)
|
|
|
56 |
## Usage
|
57 |
|
58 |
> [!IMPORTANT]
|
59 |
+
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_k=50`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
60 |
|
61 |
*Phi-4-reasoning-plus has shown strong performance on reasoning-intensive tasks. In our experiments, we extended its maximum number of tokens to 64k, and it handled longer sequences with promising results, maintaining coherence and logical consistency over extended inputs. This makes it a compelling option to explore for tasks that require deep, multi-step reasoning or extensive context.*
|
62 |
|
|
|
90 |
inputs.to(model.device),
|
91 |
max_new_tokens=4096,
|
92 |
temperature=0.8,
|
93 |
+
top_k=50,
|
94 |
top_p=0.95,
|
95 |
do_sample=True,
|
96 |
)
|
generation_config.json
CHANGED
@@ -5,6 +5,7 @@
|
|
5 |
"eos_token_id": 100265,
|
6 |
"pad_token_id": 100349,
|
7 |
"temperature": 0.8,
|
|
|
8 |
"top_p": 0.95,
|
9 |
"transformers_version": "4.51.1"
|
10 |
}
|
|
|
5 |
"eos_token_id": 100265,
|
6 |
"pad_token_id": 100349,
|
7 |
"temperature": 0.8,
|
8 |
+
"top_k": 50,
|
9 |
"top_p": 0.95,
|
10 |
"transformers_version": "4.51.1"
|
11 |
}
|