Spaces:
Sleeping
Sleeping
update app
Browse files
app.py
CHANGED
@@ -1,43 +1,24 @@
|
|
1 |
-
from transformers import AutoTokenizer, AutoModelForCausalLM
|
2 |
-
import torch
|
3 |
import gradio as gr
|
|
|
4 |
from gtts import gTTS
|
5 |
-
from langdetect import detect
|
6 |
import tempfile
|
|
|
7 |
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float32)
|
12 |
|
13 |
def respond(user_input):
|
14 |
if not user_input:
|
15 |
-
return "Please
|
16 |
|
17 |
# Detect language
|
18 |
-
|
|
|
|
|
|
|
19 |
|
20 |
-
#
|
21 |
prompt = f"[INST] {user_input} [/INST]"
|
22 |
inputs = tokenizer(prompt, return_tensors="pt")
|
23 |
-
outputs = model.generate(**inputs,
|
24 |
-
|
25 |
-
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
26 |
-
|
27 |
-
# Save voice response to a temporary file (Hugging Face compatible)
|
28 |
-
tts = gTTS(text=response, lang='hi' if detected_lang == 'hi' else 'en')
|
29 |
-
temp_audio = tempfile.NamedTemporaryFile(delete=False, suffix=".mp3")
|
30 |
-
tts.save(temp_audio.name)
|
31 |
-
|
32 |
-
return response, temp_audio.name
|
33 |
-
|
34 |
-
iface = gr.Interface(
|
35 |
-
fn=respond,
|
36 |
-
inputs=gr.Textbox(lines=2, placeholder="Ask me anything..."),
|
37 |
-
outputs=[gr.Textbox(label="TeachMe Says"), gr.Audio(label="Voice", autoplay=True)],
|
38 |
-
title="TeachMe - Your Smart Tutor",
|
39 |
-
description="Light AI bot with Hindi + English voice support."
|
40 |
-
)
|
41 |
-
|
42 |
-
if __name__ == "__main__":
|
43 |
-
iface.launch()
|
|
|
|
|
|
|
1 |
import gradio as gr
|
2 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
3 |
from gtts import gTTS
|
|
|
4 |
import tempfile
|
5 |
+
from langdetect import detect
|
6 |
|
7 |
+
# Load tokenizer and model
|
8 |
+
tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-medium")
|
9 |
+
model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-medium")
|
|
|
10 |
|
11 |
def respond(user_input):
|
12 |
if not user_input:
|
13 |
+
return "Please enter a question.", None
|
14 |
|
15 |
# Detect language
|
16 |
+
try:
|
17 |
+
detected_lang = detect(user_input)
|
18 |
+
except:
|
19 |
+
detected_lang = "en"
|
20 |
|
21 |
+
# Generate AI response
|
22 |
prompt = f"[INST] {user_input} [/INST]"
|
23 |
inputs = tokenizer(prompt, return_tensors="pt")
|
24 |
+
outputs = model.generate(**inputs, max_new_**
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|