Reality123b commited on
Commit
5205e2c
·
verified ·
1 Parent(s): 3003014

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -5,7 +5,7 @@ import torch
5
 
6
  class ModelInput(BaseModel):
7
  prompt: str
8
- max_new_tokens: int = 2048
9
 
10
  app = FastAPI()
11
 
@@ -13,13 +13,13 @@ app = FastAPI()
13
  generator = pipeline(
14
  "text-generation",
15
  model="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
16
- device=1 # Use CPU (change to device=0 for GPU)
17
  )
18
 
19
  # Create text streamer
20
  streamer = TextStreamer(generator.tokenizer, skip_prompt=True)
21
 
22
- def generate_response(prompt: str, max_new_tokens: int = 2048):
23
  try:
24
  messages = [{"role": "user", "content": prompt}]
25
  output = generator(messages, max_new_tokens=max_new_tokens, do_sample=False, streamer=streamer)
 
5
 
6
  class ModelInput(BaseModel):
7
  prompt: str
8
+ max_new_tokens: int = 64000
9
 
10
  app = FastAPI()
11
 
 
13
  generator = pipeline(
14
  "text-generation",
15
  model="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B",
16
+ device=0 # Use CPU (change to device=0 for GPU)
17
  )
18
 
19
  # Create text streamer
20
  streamer = TextStreamer(generator.tokenizer, skip_prompt=True)
21
 
22
+ def generate_response(prompt: str, max_new_tokens: int = 64000):
23
  try:
24
  messages = [{"role": "user", "content": prompt}]
25
  output = generator(messages, max_new_tokens=max_new_tokens, do_sample=False, streamer=streamer)