Cran-May commited on
Commit
7eb86aa
·
1 Parent(s): d9e334f

Update model.py

Browse files
Files changed (1) hide show
  1. model.py +3 -10
model.py CHANGED
@@ -1,17 +1,10 @@
1
 
2
  from typing import Iterator
3
 
 
4
 
5
-
6
- model_id = 'xuqinyang/baichuan-13b-chat-ggml-int4'
7
-
8
- from huggingface_hub import snapshot_download,hf_hub_download
9
- #旧
10
- #snapshot_download(model_id, local_dir="./",revision="7f71a8abefa7b2eede3f74ce0564abe5fbe6874a")
11
- snapshot_download(model_id, local_dir="./",revision="b2414a0ceee68fe09c99ace44446cfc9a1c52b08")
12
- hf_hub_download(repo_id="baichuan-inc/Baichuan-13B-Chat",local_dir="./", filename="tokenizer.model")
13
- from llama_cpp import Llama
14
- llm = Llama(model_path="./ggml-model-q4_0.bin", n_ctx=4096,seed=-1)
15
 
16
  def run(message: str,
17
  chat_history: list[tuple[str, str]],
 
1
 
2
  from typing import Iterator
3
 
4
+ from ctransformers import AutoModelForCausalLM
5
 
6
+ # Set gpu_layers to the number of layers to offload to GPU. Set to 0 if no GPU acceleration is available on your system.
7
+ llm = AutoModelForCausalLM.from_pretrained("OpenBuddy/openbuddy-gguf", model_file="openbuddy-mistral-7b-v13.1-Q3_K.gguf", model_type="mistral", gpu_layers=0)
 
 
 
 
 
 
 
 
8
 
9
  def run(message: str,
10
  chat_history: list[tuple[str, str]],