kas1 commited on
Commit
c009192
·
1 Parent(s): 5fb9513

Override quantization settings and remove bitsandbytes

Browse files
Files changed (1) hide show
  1. app.py +10 -2
app.py CHANGED
@@ -5,11 +5,19 @@ import accelerate
5
  print("Accelerate version:", accelerate.__version__)
6
 
7
  # Load the original model
8
- original_model = AutoModelForCausalLM.from_pretrained("unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit",load_in_4bit=False)
 
 
 
 
9
  original_tokenizer = AutoTokenizer.from_pretrained("unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit")
10
 
11
  # Load the fine-tuned model
12
- fine_tuned_model = AutoModelForCausalLM.from_pretrained("kas1/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit-John1",load_in_4bit=False)
 
 
 
 
13
  fine_tuned_tokenizer = AutoTokenizer.from_pretrained("kas1/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit-John1")
14
 
15
  # Function to generate responses from both models
 
5
  print("Accelerate version:", accelerate.__version__)
6
 
7
  # Load the original model
8
+ original_model = AutoModelForCausalLM.from_pretrained(
9
+ "unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit",
10
+ load_in_4bit=False, # Disable 4-bit quantization
11
+ quantization_config=None # Explicitly disable quantization
12
+ )
13
  original_tokenizer = AutoTokenizer.from_pretrained("unsloth/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit")
14
 
15
  # Load the fine-tuned model
16
+ fine_tuned_model = AutoModelForCausalLM.from_pretrained(
17
+ "kas1/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit-John1",
18
+ load_in_4bit=False, # Disable 4-bit quantization
19
+ quantization_config=None # Explicitly disable quantization
20
+ )
21
  fine_tuned_tokenizer = AutoTokenizer.from_pretrained("kas1/DeepSeek-R1-Distill-Llama-8B-unsloth-bnb-4bit-John1")
22
 
23
  # Function to generate responses from both models