DesiredName commited on
Commit
12c0528
·
verified ·
1 Parent(s): 0d75ad9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -9
app.py CHANGED
@@ -1,18 +1,16 @@
1
  from fastapi import FastAPI
2
  import uvicorn
 
3
 
4
  model_name = "DavidAU/Llama-3.2-4X3B-MOE-Hell-California-Uncensored-10B-GGUF"
5
 
6
- from transformers import AutoModel, AutoTokenizer, TextStreamer
7
- import torch
8
-
9
  # Load model and tokenizer
10
- tokenizer = AutoTokenizer.from_pretrained(model_name)
11
- model = AutoModel.from_pretrained(
12
- model_name,
13
- device_map="auto",
14
- trust_remote_code=True
15
- )
16
 
17
  def llama2_chat(prompt):
18
  inputs = tokenizer(input_text, return_tensors="pt", padding=True, truncation=True)
 
1
  from fastapi import FastAPI
2
  import uvicorn
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM
4
 
5
  model_name = "DavidAU/Llama-3.2-4X3B-MOE-Hell-California-Uncensored-10B-GGUF"
6
 
 
 
 
7
  # Load model and tokenizer
8
+
9
+ model_id = "TheBloke/TinyLlama-1.1B-Chat-v1.0-GGUF"
10
+ filename = "tinyllama-1.1b-chat-v1.0.Q6_K.gguf"
11
+
12
+ tokenizer = AutoTokenizer.from_pretrained(model_id, gguf_file=filename)
13
+ model = AutoModelForCausalLM.from_pretrained(model_id, gguf_file=filename)
14
 
15
  def llama2_chat(prompt):
16
  inputs = tokenizer(input_text, return_tensors="pt", padding=True, truncation=True)