Deepan13 commited on
Commit
d4417c8
·
1 Parent(s): 9d1ffbc

Add CodeLlama demo app

Browse files
Files changed (2) hide show
  1. app.py +24 -0
  2. requirements.txt +3 -0
app.py ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
+
4
+ model_id = "codellama/CodeLlama-7b-hf"
5
+
6
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
7
+ model = AutoModelForCausalLM.from_pretrained(
8
+ model_id,
9
+ device_map="auto"
10
+ )
11
+
12
+ pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
13
+
14
+ def generate_response(prompt):
15
+ output = pipe(prompt, max_length=512, do_sample=True, temperature=0.3)[0]['generated_text']
16
+ return output
17
+
18
+ gr.Interface(
19
+ fn=generate_response,
20
+ inputs=gr.Textbox(lines=5, label="Your prompt"),
21
+ outputs=gr.Textbox(label="Code Llama response"),
22
+ title="Code Llama Demo",
23
+ description="Ask questions or request code snippets!"
24
+ ).launch()
requirements.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ transformers
2
+ gradio
3
+ accelerate