Spaces:
Sleeping
Sleeping
s-a-malik
commited on
Commit
·
32936b7
1
Parent(s):
f89d8b2
tidy
Browse files
app.py
CHANGED
|
@@ -78,17 +78,10 @@ def generate(
|
|
| 78 |
# Generate without threading
|
| 79 |
with torch.no_grad():
|
| 80 |
outputs = model.generate(**generation_kwargs)
|
| 81 |
-
print(outputs.sequences.shape, input_ids.shape)
|
| 82 |
generated_tokens = outputs.sequences[0, input_ids.shape[1]:]
|
| 83 |
-
print("Generated tokens:", generated_tokens, generated_tokens.shape)
|
| 84 |
generated_text = tokenizer.decode(generated_tokens, skip_special_tokens=True)
|
| 85 |
-
print("Generated text:", generated_text)
|
| 86 |
# hidden states
|
| 87 |
hidden = outputs.hidden_states # list of tensors, one for each token, then (batch size, sequence length, hidden size)
|
| 88 |
-
print(len(hidden))
|
| 89 |
-
print(len(hidden[1])) # layers
|
| 90 |
-
print(hidden[1][0].shape) # (sequence length, hidden size)
|
| 91 |
-
# stack token embeddings
|
| 92 |
|
| 93 |
# TODO do this loop on the fly instead of waiting for the whole generation
|
| 94 |
highlighted_text = ""
|
|
|
|
| 78 |
# Generate without threading
|
| 79 |
with torch.no_grad():
|
| 80 |
outputs = model.generate(**generation_kwargs)
|
|
|
|
| 81 |
generated_tokens = outputs.sequences[0, input_ids.shape[1]:]
|
|
|
|
| 82 |
generated_text = tokenizer.decode(generated_tokens, skip_special_tokens=True)
|
|
|
|
| 83 |
# hidden states
|
| 84 |
hidden = outputs.hidden_states # list of tensors, one for each token, then (batch size, sequence length, hidden size)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
|
| 86 |
# TODO do this loop on the fly instead of waiting for the whole generation
|
| 87 |
highlighted_text = ""
|