cpt core 4
Browse files
scripts/cpt_core_model_4.py
CHANGED
|
@@ -34,16 +34,16 @@ model, tokenizer = FastLanguageModel.from_pretrained(
|
|
| 34 |
|
| 35 |
model = FastLanguageModel.get_peft_model(
|
| 36 |
model,
|
| 37 |
-
|
| 38 |
-
r=16, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
| 39 |
target_modules=[
|
| 40 |
"q_proj", "k_proj", "v_proj", "o_proj",
|
| 41 |
"gate_proj",
|
| 42 |
"up_proj", "down_proj",
|
| 43 |
"embed_tokens", "lm_head",
|
| 44 |
],
|
| 45 |
-
|
| 46 |
-
lora_alpha=16,
|
| 47 |
lora_dropout=0, # Supports any, but = 0 is optimized
|
| 48 |
bias='none', # Supports any, but = "none" is optimized
|
| 49 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|
|
|
|
| 34 |
|
| 35 |
model = FastLanguageModel.get_peft_model(
|
| 36 |
model,
|
| 37 |
+
r=256, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
| 38 |
+
# r=16, # Choose any number > 0 ! Suggested 8, 16, 32, 64, 128
|
| 39 |
target_modules=[
|
| 40 |
"q_proj", "k_proj", "v_proj", "o_proj",
|
| 41 |
"gate_proj",
|
| 42 |
"up_proj", "down_proj",
|
| 43 |
"embed_tokens", "lm_head",
|
| 44 |
],
|
| 45 |
+
lora_alpha=32,
|
| 46 |
+
# lora_alpha=16,
|
| 47 |
lora_dropout=0, # Supports any, but = 0 is optimized
|
| 48 |
bias='none', # Supports any, but = "none" is optimized
|
| 49 |
# [NEW] "unsloth" uses 30% less VRAM, fits 2x larger batch sizes!
|