suriya7 commited on
Commit
8fba5ff
·
verified ·
1 Parent(s): a4ba610

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -4
app.py CHANGED
@@ -3,15 +3,14 @@ import pprint
3
  import subprocess
4
  from threading import Thread
5
  from transformers import AutoTokenizer, TextIteratorStreamer
6
-
7
 
8
  result = subprocess.run(["lscpu"], text=True, capture_output=True)
9
  pprint.pprint(result.stdout)
10
 
11
 
12
- checkpoint = "suriya7/Gemma-2b-SFT"
13
- tokenizer = AutoTokenizer.from_pretrained(checkpoint)
14
- model = AutoModelForSeq2SeqLM.from_pretrained(checkpoint)
15
 
16
 
17
  def run_generation(user_text, top_p, temperature, top_k, max_new_tokens):
 
3
  import subprocess
4
  from threading import Thread
5
  from transformers import AutoTokenizer, TextIteratorStreamer
6
+ from transformers import AutoTokenizer, AutoModelForCausalLM
7
 
8
  result = subprocess.run(["lscpu"], text=True, capture_output=True)
9
  pprint.pprint(result.stdout)
10
 
11
 
12
+ tokenizer = AutoTokenizer.from_pretrained("suriya7/Gemma-2b-SFT")
13
+ model = AutoModelForCausalLM.from_pretrained("suriya7/Gemma-2b-SFT")
 
14
 
15
 
16
  def run_generation(user_text, top_p, temperature, top_k, max_new_tokens):