Gustavo de Rosa commited on
Commit
0f18025
·
1 Parent(s): d8d3b44

chore(root): Adds top_k information even if 50 is already the default.

Browse files
Files changed (2) hide show
  1. README.md +2 -1
  2. generation_config.json +1 -0
README.md CHANGED
@@ -56,7 +56,7 @@ library_name: transformers
56
  ## Usage
57
 
58
  > [!IMPORTANT]
59
- > To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
60
 
61
  *Phi-4-reasoning-plus has shown strong performance on reasoning-intensive tasks. In our experiments, we extended its maximum number of tokens to 64k, and it handled longer sequences with promising results, maintaining coherence and logical consistency over extended inputs. This makes it a compelling option to explore for tasks that require deep, multi-step reasoning or extensive context.*
62
 
@@ -90,6 +90,7 @@ outputs = model.generate(
90
  inputs.to(model.device),
91
  max_new_tokens=4096,
92
  temperature=0.8,
 
93
  top_p=0.95,
94
  do_sample=True,
95
  )
 
56
  ## Usage
57
 
58
  > [!IMPORTANT]
59
+ > To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_k=50`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
60
 
61
  *Phi-4-reasoning-plus has shown strong performance on reasoning-intensive tasks. In our experiments, we extended its maximum number of tokens to 64k, and it handled longer sequences with promising results, maintaining coherence and logical consistency over extended inputs. This makes it a compelling option to explore for tasks that require deep, multi-step reasoning or extensive context.*
62
 
 
90
  inputs.to(model.device),
91
  max_new_tokens=4096,
92
  temperature=0.8,
93
+ top_k=50,
94
  top_p=0.95,
95
  do_sample=True,
96
  )
generation_config.json CHANGED
@@ -5,6 +5,7 @@
5
  "eos_token_id": 100265,
6
  "pad_token_id": 100349,
7
  "temperature": 0.8,
 
8
  "top_p": 0.95,
9
  "transformers_version": "4.51.1"
10
  }
 
5
  "eos_token_id": 100265,
6
  "pad_token_id": 100349,
7
  "temperature": 0.8,
8
+ "top_k": 50,
9
  "top_p": 0.95,
10
  "transformers_version": "4.51.1"
11
  }