davanstrien HF Staff commited on
Commit
6311bc5
·
1 Parent(s): c2b4647

Update model version and parameters in generate-responses.py

Browse files
Files changed (1) hide show
  1. generate-responses.py +19 -12
generate-responses.py CHANGED
@@ -5,7 +5,7 @@
5
  # "huggingface-hub[hf_transfer]",
6
  # "torch",
7
  # "transformers",
8
- # "vllm",
9
  # ]
10
  #
11
  # ///
@@ -150,7 +150,7 @@ uv run https://huggingface.co/datasets/uv-scripts/vllm/raw/main/generate-respons
150
  def main(
151
  src_dataset_hub_id: str,
152
  output_dataset_hub_id: str,
153
- model_id: str = "Qwen/Qwen3-30B-A3B-Instruct-2507-FP8",
154
  messages_column: str = "messages",
155
  output_column: str = "response",
156
  temperature: float = 0.7,
@@ -160,6 +160,7 @@ def main(
160
  max_tokens: int = 16384,
161
  repetition_penalty: float = 1.0,
162
  gpu_memory_utilization: float = 0.90,
 
163
  tensor_parallel_size: Optional[int] = None,
164
  hf_token: Optional[str] = None,
165
  ):
@@ -179,6 +180,7 @@ def main(
179
  max_tokens: Maximum tokens to generate
180
  repetition_penalty: Repetition penalty parameter
181
  gpu_memory_utilization: GPU memory utilization factor
 
182
  tensor_parallel_size: Number of GPUs to use (auto-detect if None)
183
  hf_token: Hugging Face authentication token
184
  """
@@ -213,11 +215,16 @@ def main(
213
 
214
  # Initialize vLLM
215
  logger.info(f"Loading model: {model_id}")
216
- llm = LLM(
217
- model=model_id,
218
- tensor_parallel_size=tensor_parallel_size,
219
- gpu_memory_utilization=gpu_memory_utilization,
220
- )
 
 
 
 
 
221
 
222
  # Load tokenizer for chat template
223
  logger.info("Loading tokenizer...")
@@ -336,8 +343,8 @@ Examples:
336
  parser.add_argument(
337
  "--model-id",
338
  type=str,
339
- default="Qwen/Qwen3-30B-A3B-Instruct-2507-FP8",
340
- help="Model to use for generation (default: Qwen3-30B-A3B-Instruct-2507-FP8)",
341
  )
342
  parser.add_argument(
343
  "--messages-column",
@@ -390,8 +397,8 @@ Examples:
390
  parser.add_argument(
391
  "--gpu-memory-utilization",
392
  type=float,
393
- default=0.90,
394
- help="GPU memory utilization factor (default: 0.90)",
395
  )
396
  parser.add_argument(
397
  "--tensor-parallel-size",
@@ -439,7 +446,7 @@ Example HF Jobs command with multi-GPU:
439
  username/input-dataset \\
440
  username/output-dataset \\
441
  --messages-column messages \\
442
- --model-id Qwen/Qwen3-30B-A3B-Instruct-2507-FP8 \\
443
  --temperature 0.7 \\
444
  --max-tokens 16384
445
  """)
 
5
  # "huggingface-hub[hf_transfer]",
6
  # "torch",
7
  # "transformers",
8
+ # "vllm>=0.8.5",
9
  # ]
10
  #
11
  # ///
 
150
  def main(
151
  src_dataset_hub_id: str,
152
  output_dataset_hub_id: str,
153
+ model_id: str = "Qwen/Qwen3-30B-A3B-Instruct-2507",
154
  messages_column: str = "messages",
155
  output_column: str = "response",
156
  temperature: float = 0.7,
 
160
  max_tokens: int = 16384,
161
  repetition_penalty: float = 1.0,
162
  gpu_memory_utilization: float = 0.90,
163
+ max_model_len: Optional[int] = None,
164
  tensor_parallel_size: Optional[int] = None,
165
  hf_token: Optional[str] = None,
166
  ):
 
180
  max_tokens: Maximum tokens to generate
181
  repetition_penalty: Repetition penalty parameter
182
  gpu_memory_utilization: GPU memory utilization factor
183
+ max_model_len: Maximum model context length (None uses model default)
184
  tensor_parallel_size: Number of GPUs to use (auto-detect if None)
185
  hf_token: Hugging Face authentication token
186
  """
 
215
 
216
  # Initialize vLLM
217
  logger.info(f"Loading model: {model_id}")
218
+ vllm_kwargs = {
219
+ "model": model_id,
220
+ "tensor_parallel_size": tensor_parallel_size,
221
+ "gpu_memory_utilization": gpu_memory_utilization,
222
+ }
223
+ if max_model_len is not None:
224
+ vllm_kwargs["max_model_len"] = max_model_len
225
+ logger.info(f"Using max_model_len={max_model_len}")
226
+
227
+ llm = LLM(**vllm_kwargs)
228
 
229
  # Load tokenizer for chat template
230
  logger.info("Loading tokenizer...")
 
343
  parser.add_argument(
344
  "--model-id",
345
  type=str,
346
+ default="Qwen/Qwen3-30B-A3B-Instruct-2507",
347
+ help="Model to use for generation (default: Qwen3-30B-A3B-Instruct-2507)",
348
  )
349
  parser.add_argument(
350
  "--messages-column",
 
397
  parser.add_argument(
398
  "--gpu-memory-utilization",
399
  type=float,
400
+ default=0.95,
401
+ help="GPU memory utilization factor (default: 0.95)",
402
  )
403
  parser.add_argument(
404
  "--tensor-parallel-size",
 
446
  username/input-dataset \\
447
  username/output-dataset \\
448
  --messages-column messages \\
449
+ --model-id Qwen/Qwen3-30B-A3B-Instruct-2507 \\
450
  --temperature 0.7 \\
451
  --max-tokens 16384
452
  """)