Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -17,14 +17,16 @@ tokenizer_names = [
|
|
17 |
# Load all the tokenizers
|
18 |
tokenizers = {name: AutoTokenizer.from_pretrained(name) for name in tokenizer_names}
|
19 |
|
20 |
-
# Function to generate responses using different tokenizers
|
21 |
def generate_responses(prompt):
|
22 |
responses = {}
|
23 |
for name, tokenizer in tokenizers.items():
|
24 |
-
|
25 |
-
|
26 |
-
|
27 |
-
|
|
|
|
|
|
|
28 |
return responses
|
29 |
|
30 |
# Gradio interface setup
|
@@ -39,3 +41,4 @@ interface = gr.Interface(
|
|
39 |
# Launch the Gradio interface
|
40 |
interface.launch()
|
41 |
|
|
|
|
17 |
# Load all the tokenizers
|
18 |
tokenizers = {name: AutoTokenizer.from_pretrained(name) for name in tokenizer_names}
|
19 |
|
|
|
20 |
def generate_responses(prompt):
|
21 |
responses = {}
|
22 |
for name, tokenizer in tokenizers.items():
|
23 |
+
try:
|
24 |
+
inputs = tokenizer(prompt, return_tensors="pt")
|
25 |
+
outputs = model.generate(**inputs)
|
26 |
+
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
27 |
+
responses[name] = response
|
28 |
+
except Exception as e:
|
29 |
+
responses[name] = f"Error: {str(e)}"
|
30 |
return responses
|
31 |
|
32 |
# Gradio interface setup
|
|
|
41 |
# Launch the Gradio interface
|
42 |
interface.launch()
|
43 |
|
44 |
+
|