Spaces:
Running
Running
try to have continos generation
Browse files
app.py
CHANGED
@@ -6,14 +6,17 @@ from huggingface_hub import login
|
|
6 |
|
7 |
login(os.getenv('HF_LOGIN'))
|
8 |
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
{"role": "user", "content": textik},
|
14 |
-
]
|
15 |
|
16 |
-
|
17 |
-
|
18 |
-
|
19 |
-
|
|
|
|
|
|
|
|
|
|
|
|
6 |
|
7 |
login(os.getenv('HF_LOGIN'))
|
8 |
|
9 |
+
model_id = "utter-project/EuroLLM-1.7B-Instruct"
|
10 |
+
tokenizer = AutoTokenizer.from_pretrained(model_id, torch_dtype=torch.bfloat16)
|
11 |
+
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16)
|
12 |
+
model.generation_config.pad_token_id = tokenizer.pad_token_id
|
|
|
|
|
13 |
|
14 |
+
inner = st.text_area('enter some input!')
|
15 |
+
text = '<|im_start|><|im_end|>user\n'+inner+'<|im_end|>\n<|im_start|>assistant\n'
|
16 |
+
|
17 |
+
inputs = tokenizer(text, return_tensors="pt")
|
18 |
+
outputs = inputs['input_ids']
|
19 |
+
if inner:
|
20 |
+
while (not torch.any(outputs[0][-token_step_size:] == 4)):
|
21 |
+
outputs = model.generate(input_ids=outputs, attention_mask=torch.ones_like(outputs),max_new_tokens=token_step_size)
|
22 |
+
st.write(tokenizer.decode(outputs[0][-token_step_size:], skip_special_tokens=True))#, end=' ', flush=True)
|