winglian commited on
Commit
283f87f
·
1 Parent(s): 322a1e2

add README and axolotl config

Browse files
Files changed (2) hide show
  1. README.md +6 -0
  2. configs/stable-manticore-lora.yml +103 -0
README.md ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ ---
2
+ ---
3
+
4
+ # StableLManticore 7B
5
+
6
+
configs/stable-manticore-lora.yml ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: stabilityai/stablelm-base-alpha-7b
2
+ base_model_config: stabilityai/stablelm-base-alpha-7b
3
+ model_type: GPTNeoXForCausalLM
4
+ load_in_8bit: false
5
+ datasets:
6
+ - path: winglian/evals
7
+ data_files:
8
+ - hf/ARC-Challenge.jsonl
9
+ - hf/ARC-Easy.jsonl
10
+ - mmlu/abstract_algebra.jsonl
11
+ - mmlu/conceptual_physics.jsonl
12
+ - mmlu/formal_logic.jsonl
13
+ - mmlu/high_school_physics.jsonl
14
+ - mmlu/logical_fallacies.jsonl
15
+ type: explainchoice
16
+ - path: winglian/evals
17
+ data_files:
18
+ - openai/tldr.jsonl
19
+ type: summarizetldr
20
+ - path: winglian/evals
21
+ data_files:
22
+ - hellaswag/hellaswag-concise.jsonl
23
+ type: concisechoice
24
+ - path: metaeval/ScienceQA_text_only
25
+ type: concisechoice
26
+ - path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
27
+ type: alpaca
28
+ - path: ehartford/wizard_vicuna_70k_unfiltered
29
+ type: sharegpt
30
+ - path: winglian/chatlogs-en-cleaned
31
+ data_files:
32
+ - sharegpt_cleaned.jsonl
33
+ type: sharegpt
34
+ - path: teknium/GPT4-LLM-Cleaned
35
+ type: alpaca
36
+ - path: teknium/GPTeacher-General-Instruct
37
+ data_files: gpt4-instruct-similarity-0.6-dataset.json
38
+ type: gpteacher
39
+ - path: QingyiSi/Alpaca-CoT
40
+ data_files:
41
+ - Chain-of-Thought/formatted_cot_data/aqua_train.json
42
+ - Chain-of-Thought/formatted_cot_data/creak_train.json
43
+ - Chain-of-Thought/formatted_cot_data/ecqa_train.json
44
+ - Chain-of-Thought/formatted_cot_data/esnli_train.json
45
+ - Chain-of-Thought/formatted_cot_data/gsm8k_train.json
46
+ - Chain-of-Thought/formatted_cot_data/qasc_train.json
47
+ - Chain-of-Thought/formatted_cot_data/qed_train.json
48
+ - Chain-of-Thought/formatted_cot_data/sensemaking_train.json
49
+ - Chain-of-Thought/formatted_cot_data/strategyqa_train.json
50
+ - GPTeacher/Roleplay/formatted_roleplay-similarity_0.6-instruct-dataset.json
51
+ type: alpaca
52
+ dataset_prepared_path: last_run_prepared
53
+ val_set_size: 0.02
54
+ adapter: lora
55
+ lora_model_dir:
56
+ sequence_len: 4096
57
+ max_packed_sequence_len: 2048
58
+ lora_r: 8
59
+ lora_alpha: 16
60
+ lora_dropout: 0.05
61
+ lora_target_modules:
62
+ - query_key_value
63
+ lora_fan_in_fan_out: true
64
+ wandb_project: stable-manticore-7b
65
+ wandb_watch:
66
+ wandb_run_id:
67
+ wandb_log_model:
68
+ output_dir: ./stable-manticore-7b-lora
69
+ batch_size: 42
70
+ micro_batch_size: 2
71
+ num_epochs: 4
72
+ optimizer: adamw_bnb_8bit
73
+ torchdistx_path:
74
+ lr_scheduler: cosine
75
+ learning_rate: 0.000003
76
+ train_on_inputs: false
77
+ group_by_length: false
78
+ bf16: false
79
+ fp16: true
80
+ tf32: true
81
+ gradient_checkpointing: false
82
+ early_stopping_patience:
83
+ resume_from_checkpoint:
84
+ local_rank:
85
+ logging_steps: 1
86
+ xformers_attention:
87
+ flash_attention:
88
+ gptq_groupsize:
89
+ gptq_model_v1:
90
+ warmup_steps: 322
91
+ eval_steps: 108
92
+ save_steps:
93
+ debug:
94
+ deepspeed:
95
+ weight_decay: 0
96
+ fsdp:
97
+ fsdp_config:
98
+ special_tokens:
99
+ pad_token: "<|padding|>"
100
+ bos_token: "<|endoftext|>"
101
+ eos_token: "<|endoftext|>"
102
+ unk_token: "<|endoftext|>"
103
+