chrisvoncsefalvay commited on
Commit
0015607
·
1 Parent(s): 810bbae

Use TinyLlama tokenizer as fallback for corrupted tokenizer

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -9,10 +9,13 @@ DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
9
  # Initialize model and tokenizer
10
  print(f"Loading model {MODEL_ID}...")
11
  try:
12
- tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, use_fast=False)
 
13
  except Exception as e:
14
- print(f"Failed to load tokenizer with use_fast=False, trying with use_fast=True: {e}")
15
- tokenizer = AutoTokenizer.from_pretrained(MODEL_ID, use_fast=True, trust_remote_code=True)
 
 
16
 
17
  model = AutoModelForCausalLM.from_pretrained(
18
  MODEL_ID,
 
9
  # Initialize model and tokenizer
10
  print(f"Loading model {MODEL_ID}...")
11
  try:
12
+ # Try loading the tokenizer normally
13
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
14
  except Exception as e:
15
+ print(f"Failed to load tokenizer from {MODEL_ID}: {e}")
16
+ print("Using fallback tokenizer from base model...")
17
+ # Use the base model's tokenizer as fallback
18
+ tokenizer = AutoTokenizer.from_pretrained("TinyLlama/TinyLlama-1.1B-Chat-v1.0")
19
 
20
  model = AutoModelForCausalLM.from_pretrained(
21
  MODEL_ID,