Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -104,8 +104,12 @@ class TweetDatasetProcessor:
|
|
104 |
**Only generate the tweet. Do not include analysis, explanation, or any other content.**
|
105 |
"""
|
106 |
|
107 |
-
|
108 |
-
|
|
|
|
|
|
|
|
|
109 |
generated_tweet = self.tokenizer.decode(output[0], skip_special_tokens=True).strip()
|
110 |
|
111 |
return generated_tweet
|
|
|
104 |
**Only generate the tweet. Do not include analysis, explanation, or any other content.**
|
105 |
"""
|
106 |
|
107 |
+
inputs = self.tokenizer(prompt, return_tensors='pt', truncation=True, padding=True)
|
108 |
+
input_ids = inputs['input_ids']
|
109 |
+
attention_mask = inputs['attention_mask']
|
110 |
+
|
111 |
+
pad_token_id = self.tokenizer.eos_token_id
|
112 |
+
output = self.model.generate(input_ids, attention_mask=attention_mask, max_length=150, num_return_sequences=1, temperature=1.0, pad_token_id=pad_token_id)
|
113 |
generated_tweet = self.tokenizer.decode(output[0], skip_special_tokens=True).strip()
|
114 |
|
115 |
return generated_tweet
|