Tech-Meld commited on
Commit
4a48aba
·
verified ·
1 Parent(s): c40ae5e

Create app.py

Browse files
Files changed (1) hide show
  1. app.py +31 -0
app.py ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer
4
+
5
+ @gr.cache(allow_output_mutation=True)
6
+ def load_model():
7
+ model_id = "Tech-Meld/Hajax_Chat_1.0"
8
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
9
+ model = AutoModelForCausalLM.from_pretrained(model_id)
10
+ return model, tokenizer
11
+
12
+ def get_response(input_text, model, tokenizer):
13
+ inputs = tokenizer.encode(input_text + tokenizer.eos_token, return_tensors='pt')
14
+ outputs = model.generate(inputs, max_length=1000, pad_token_id=tokenizer.eos_token_id)
15
+ response = tokenizer.decode(outputs[:, inputs.shape[-1]:][0], skip_special_tokens=True)
16
+ return response
17
+
18
+ model, tokenizer = load_model()
19
+
20
+ def chat(input_text):
21
+ response = get_response(input_text, model, tokenizer)
22
+ return response
23
+
24
+ iface = gr.Interface(
25
+ chat,
26
+ "text",
27
+ "text",
28
+ title="Chat with Hajax",
29
+ description="Type your message and press Enter to chat with the AI.",
30
+ )
31
+ iface.launch()