ddededstger commited on
Commit
5215f12
Β·
verified Β·
1 Parent(s): c7a8f24

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -3
app.py CHANGED
@@ -1,9 +1,10 @@
1
  import gradio as gr
2
  from peft import AutoPeftModelForCausalLM
3
  from transformers import AutoTokenizer
4
- from huggingface_hub import login
5
  import torch
6
- import os # For env vars
 
7
 
8
  # Login using secret (secure, no hardcode)
9
  login(os.environ["HF_TOKEN"])
@@ -14,7 +15,18 @@ base_model_id = "mistralai/Ministral-8B-Instruct-2410"
14
  MAX_LENGTH = 32768
15
  DEVICE = "cuda" if torch.cuda.is_available() else "cpu" # Use GPU if available (ZeroGPU on HF)
16
 
17
- model = AutoPeftModelForCausalLM.from_pretrained(model_id)
 
 
 
 
 
 
 
 
 
 
 
18
  tokenizer = AutoTokenizer.from_pretrained(base_model_id)
19
  model = model.to(DEVICE)
20
  model.eval()
 
1
  import gradio as gr
2
  from peft import AutoPeftModelForCausalLM
3
  from transformers import AutoTokenizer
4
+ from huggingface_hub import login, snapshot_download
5
  import torch
6
+ import os
7
+ import json
8
 
9
  # Login using secret (secure, no hardcode)
10
  login(os.environ["HF_TOKEN"])
 
15
  MAX_LENGTH = 32768
16
  DEVICE = "cuda" if torch.cuda.is_available() else "cpu" # Use GPU if available (ZeroGPU on HF)
17
 
18
+ # Workaround: Download model files, edit adapter_config.json to remove 'model_type'
19
+ local_dir = snapshot_download(repo_id=model_id)
20
+ config_path = os.path.join(local_dir, "adapter_config.json")
21
+ with open(config_path, 'r') as f:
22
+ config = json.load(f)
23
+ if 'model_type' in config:
24
+ del config['model_type']
25
+ with open(config_path, 'w') as f:
26
+ json.dump(config, f)
27
+
28
+ # Now load the model from modified local dir
29
+ model = AutoPeftModelForCausalLM.from_pretrained(local_dir)
30
  tokenizer = AutoTokenizer.from_pretrained(base_model_id)
31
  model = model.to(DEVICE)
32
  model.eval()