Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -7,7 +7,12 @@ model = GPT2LMHeadModel.from_pretrained("gpt2")
|
|
7 |
tokenizer = GPT2Tokenizer.from_pretrained("gpt2")
|
8 |
|
9 |
def predict_fake_news(text):
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
11 |
output = model.generate(input_ids, max_length=50, num_return_sequences=1, pad_token_id=tokenizer.eos_token_id, attention_mask=input_ids)
|
12 |
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
13 |
|
|
|
7 |
tokenizer = GPT2Tokenizer.from_pretrained("gpt2")
|
8 |
|
9 |
def predict_fake_news(text):
|
10 |
+
input_text = text
|
11 |
+
|
12 |
+
input_ids = tokenizer.encode(input_text, return_tensors='pt', add_special_tokens=True)
|
13 |
+
|
14 |
+
input_ids = input_ids[:, :1024]
|
15 |
+
|
16 |
output = model.generate(input_ids, max_length=50, num_return_sequences=1, pad_token_id=tokenizer.eos_token_id, attention_mask=input_ids)
|
17 |
generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
|
18 |
|