lukisko commited on
Commit
3bd22b1
·
verified ·
1 Parent(s): 5aa627d

try to have continos generation

Browse files
Files changed (1) hide show
  1. app.py +13 -10
app.py CHANGED
@@ -6,14 +6,17 @@ from huggingface_hub import login
6
 
7
  login(os.getenv('HF_LOGIN'))
8
 
9
- #pipe = pipeline("text-generation", model="utter-project/EuroLLM-9B-Instruct")
10
- pipe = pipeline("text-generation", model="utter-project/EuroLLM-1.7B-Instruct")
11
- textik = st.text_area('enter some input!')
12
- message = [
13
- {"role": "user", "content": textik},
14
- ]
15
 
16
- if textik:
17
- out = pipe(message, max_new_tokens=1024)[0]['generated_text'][1]
18
- st.write('## '+out['role'])
19
- st.write(out['content'])
 
 
 
 
 
 
6
 
7
  login(os.getenv('HF_LOGIN'))
8
 
9
+ model_id = "utter-project/EuroLLM-1.7B-Instruct"
10
+ tokenizer = AutoTokenizer.from_pretrained(model_id, torch_dtype=torch.bfloat16)
11
+ model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)
12
+ model.generation_config.pad_token_id = tokenizer.pad_token_id
 
 
13
 
14
+ inner = st.text_area('enter some input!')
15
+ text = '<|im_start|><|im_end|>user\n'+inner+'<|im_end|>\n<|im_start|>assistant\n'
16
+
17
+ inputs = tokenizer(text, return_tensors="pt")
18
+ outputs = inputs['input_ids']
19
+ if inner:
20
+ while (not torch.any(outputs[0][-token_step_size:] == 4)):
21
+ outputs = model.generate(input_ids=outputs, attention_mask=torch.ones_like(outputs),max_new_tokens=token_step_size)
22
+ st.write(tokenizer.decode(outputs[0][-token_step_size:], skip_special_tokens=True))#, end=' ', flush=True)