asadsandhu commited on
Commit
2833445
·
1 Parent(s): e531b46

code updated.

Browse files
Files changed (1) hide show
  1. app.py +9 -6
app.py CHANGED
@@ -3,7 +3,7 @@ import pandas as pd
3
  import faiss
4
  import torch
5
  import numpy as np
6
-
7
  from sentence_transformers import SentenceTransformer
8
  from transformers import AutoTokenizer, AutoModelForCausalLM
9
 
@@ -20,13 +20,16 @@ embedding_model = SentenceTransformer("all-MiniLM-L6-v2")
20
  # ===============================
21
  # Load LLM on CPU
22
  # ===============================
23
- model_id = "PrunaAI/BioMistral-7B-bnb-8bit-smashed"
24
- tokenizer = AutoTokenizer.from_pretrained("BioMistral/BioMistral-7B")
 
 
25
  model = AutoModelForCausalLM.from_pretrained(
26
  model_id,
27
- trust_remote_code=True,
28
- device_map=None, # CPU only
29
- )
 
30
  tokenizer.pad_token = tokenizer.eos_token
31
 
32
  # ===============================
 
3
  import faiss
4
  import torch
5
  import numpy as np
6
+ from accelerate import init_empty_weights, load_checkpoint_and_dispatch
7
  from sentence_transformers import SentenceTransformer
8
  from transformers import AutoTokenizer, AutoModelForCausalLM
9
 
 
20
  # ===============================
21
  # Load LLM on CPU
22
  # ===============================
23
+ model_id = "BioMistral/BioMistral-7B"
24
+
25
+ print(f"Loading tokenizer and model: {model_id}")
26
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
27
  model = AutoModelForCausalLM.from_pretrained(
28
  model_id,
29
+ torch_dtype=torch.float16,
30
+ low_cpu_mem_usage=True,
31
+ )# .to("cpu")
32
+
33
  tokenizer.pad_token = tokenizer.eos_token
34
 
35
  # ===============================