Ujeshhh commited on
Commit
b6881b4
·
verified ·
1 Parent(s): b6a69cb

Update analysis.py

Browse files
Files changed (1) hide show
  1. analysis.py +9 -14
analysis.py CHANGED
@@ -1,25 +1,20 @@
1
- from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM
2
- from huggingface_hub import login
3
- import os
4
  import torch
 
5
 
6
- # ✅ Get API token from environment variable
7
  HF_TOKEN = os.getenv("HF_TOKEN")
 
8
 
9
- # Authenticate with Hugging Face (without exposing the token in code)
10
- login(HF_TOKEN)
11
-
12
- # ✅ Load Model Efficiently
13
- MODEL_NAME = "tiiuae/falcon-7b-instruct"
14
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, token=HF_TOKEN)
 
 
15
  model = AutoModelForCausalLM.from_pretrained(
16
  MODEL_NAME,
17
- token=HF_TOKEN,
18
- device_map="auto",
19
- torch_dtype=torch.float16,
20
- load_in_8bit=True
21
  )
22
-
23
  # ✅ Create LLM Pipeline
24
  llm_pipeline = pipeline("text-generation", model=model, tokenizer=tokenizer, device_map="auto")
25
 
 
1
+ from transformers import AutoModelForCausalLM, AutoTokenizer
 
 
2
  import torch
3
+ import os
4
 
 
5
  HF_TOKEN = os.getenv("HF_TOKEN")
6
+ MODEL_NAME = "mistralai/Mistral-7B-Instruct-v0.1-GGUF" # Use a smaller model
7
 
8
+ # Load tokenizer
 
 
 
 
9
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, token=HF_TOKEN)
10
+
11
+ # Load model (use torch.float16 if on GPU, otherwise use torch.float32 for CPU)
12
  model = AutoModelForCausalLM.from_pretrained(
13
  MODEL_NAME,
14
+ torch_dtype=torch.float32, # Change to float16 if running on GPU
15
+ device_map="auto", # Uses CPU if no GPU is available
16
+ token=HF_TOKEN
 
17
  )
 
18
  # ✅ Create LLM Pipeline
19
  llm_pipeline = pipeline("text-generation", model=model, tokenizer=tokenizer, device_map="auto")
20