seemamourya210 commited on
Commit
477e877
·
verified ·
1 Parent(s): 2fd398a

update app

Browse files
Files changed (1) hide show
  1. app.py +12 -31
app.py CHANGED
@@ -1,43 +1,24 @@
1
- from transformers import AutoTokenizer, AutoModelForCausalLM
2
- import torch
3
  import gradio as gr
 
4
  from gtts import gTTS
5
- from langdetect import detect
6
  import tempfile
 
7
 
8
- model_name = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
9
-
10
- tokenizer = AutoTokenizer.from_pretrained(model_name)
11
- model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float32)
12
 
13
  def respond(user_input):
14
  if not user_input:
15
- return "Please ask something.", None
16
 
17
  # Detect language
18
- detected_lang = detect(user_input)
 
 
 
19
 
20
- # Prepare prompt
21
  prompt = f"[INST] {user_input} [/INST]"
22
  inputs = tokenizer(prompt, return_tensors="pt")
23
- outputs = model.generate(**inputs, max_new_tokens=256, pad_token_id=tokenizer.eos_token_id)
24
-
25
- response = tokenizer.decode(outputs[0], skip_special_tokens=True)
26
-
27
- # Save voice response to a temporary file (Hugging Face compatible)
28
- tts = gTTS(text=response, lang='hi' if detected_lang == 'hi' else 'en')
29
- temp_audio = tempfile.NamedTemporaryFile(delete=False, suffix=".mp3")
30
- tts.save(temp_audio.name)
31
-
32
- return response, temp_audio.name
33
-
34
- iface = gr.Interface(
35
- fn=respond,
36
- inputs=gr.Textbox(lines=2, placeholder="Ask me anything..."),
37
- outputs=[gr.Textbox(label="TeachMe Says"), gr.Audio(label="Voice", autoplay=True)],
38
- title="TeachMe - Your Smart Tutor",
39
- description="Light AI bot with Hindi + English voice support."
40
- )
41
-
42
- if __name__ == "__main__":
43
- iface.launch()
 
 
 
1
  import gradio as gr
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM
3
  from gtts import gTTS
 
4
  import tempfile
5
+ from langdetect import detect
6
 
7
+ # Load tokenizer and model
8
+ tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-medium")
9
+ model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-medium")
 
10
 
11
  def respond(user_input):
12
  if not user_input:
13
+ return "Please enter a question.", None
14
 
15
  # Detect language
16
+ try:
17
+ detected_lang = detect(user_input)
18
+ except:
19
+ detected_lang = "en"
20
 
21
+ # Generate AI response
22
  prompt = f"[INST] {user_input} [/INST]"
23
  inputs = tokenizer(prompt, return_tensors="pt")
24
+ outputs = model.generate(**inputs, max_new_**