muryshev commited on
Commit
96cde50
·
verified ·
1 Parent(s): 39ad741

Update: input tokens are on the same device as a model

Browse files
Files changed (1) hide show
  1. model.py +1 -1
model.py CHANGED
@@ -18,7 +18,7 @@ async def get_answer_from_llm(question: str = None):
18
  messages = [{"role": "user", "content": f"{question}"}]
19
  input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
20
  gen_tokens = model.generate(
21
- input_ids,
22
  max_new_tokens=100,
23
  do_sample=True,
24
  temperature=0.3,
 
18
  messages = [{"role": "user", "content": f"{question}"}]
19
  input_ids = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
20
  gen_tokens = model.generate(
21
+ input_ids.to(device),
22
  max_new_tokens=100,
23
  do_sample=True,
24
  temperature=0.3,