PSM24 commited on
Commit
d6a5634
·
verified ·
1 Parent(s): 7246afb

Upload 4 files

Browse files
adapter_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
+ "bias": "none",
6
+ "eva_config": null,
7
+ "exclude_modules": null,
8
+ "fan_in_fan_out": false,
9
+ "inference_mode": false,
10
+ "init_lora_weights": true,
11
+ "layer_replication": null,
12
+ "layers_pattern": null,
13
+ "layers_to_transform": null,
14
+ "loftq_config": {},
15
+ "lora_alpha": 16,
16
+ "lora_bias": false,
17
+ "lora_dropout": 0.05,
18
+ "megatron_config": null,
19
+ "megatron_core": "megatron.core",
20
+ "modules_to_save": null,
21
+ "peft_type": "LORA",
22
+ "r": 8,
23
+ "rank_pattern": {},
24
+ "revision": null,
25
+ "target_modules": [
26
+ "q_proj",
27
+ "down_proj",
28
+ "gate_proj",
29
+ "k_proj",
30
+ "up_proj",
31
+ "v_proj",
32
+ "o_proj"
33
+ ],
34
+ "task_type": "CAUSAL_LM",
35
+ "use_dora": false,
36
+ "use_rslora": false
37
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da5758ab6f71c06c74597d971bd53d759a228f64aaebdd706fa210ab4d297c98
3
+ size 210685048
stats.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"world_size": 1, "epochs": 5, "steps": 11976, "seqs": 37730, "tokens": 312391120, "last_epoch_steps": 0, "last_epoch_seqs": 0, "last_epoch_tokens": 0, "wandb_run_id": null, "loss_ema": 0.26658365699635356, "loss_sum": 14.137018859386444, "eval_losses_avg": [NaN, NaN, NaN, NaN, NaN, NaN, 7.156456470489502, NaN, NaN, NaN, NaN, NaN, 7.434809684753418, NaN, NaN, NaN, NaN, 7.742526054382324, NaN, NaN, NaN, 5.851824760437012, 6.587878704071045, NaN, NaN, 1.814159631729126, 6.935285568237305, NaN, NaN, NaN, NaN, 4.725756645202637, NaN, NaN, 3.7440850734710693, NaN, 3.624691963195801, NaN, 1.7548154592514038, NaN, NaN, 5.833590507507324, NaN, NaN, NaN, NaN, 7.577816486358643, NaN, 6.335320949554443, 6.516195774078369, 7.621451377868652, 5.837831020355225, 4.262629985809326, 3.242569923400879, NaN, NaN, NaN, NaN, 4.948696136474609, 4.459714412689209, NaN, 6.048461437225342, NaN, NaN, NaN, NaN, 5.848104476928711, 1.582127571105957, NaN, NaN, 6.735745906829834, 4.996155261993408, NaN, NaN, 5.9206156730651855, NaN, NaN, NaN, NaN, NaN, 6.825109004974365, NaN, NaN, NaN, 3.210527181625366, 2.8243603706359863, 5.8360595703125, NaN, 1.6897790431976318, NaN, 3.706303119659424, 6.706119060516357, NaN, NaN, NaN, NaN, 5.9158196449279785, 2.2692787647247314, NaN, NaN, 6.438658237457275, 2.098296642303467, 4.838111877441406, 6.066786766052246, 5.627449989318848, NaN, 2.2828876972198486, 2.210242509841919, NaN, NaN, NaN, NaN, 6.5944600105285645, 2.014805316925049, NaN, NaN, NaN, NaN, NaN, 4.841887950897217, NaN, 6.457818031311035, NaN, 5.662258148193359, NaN, 6.584174633026123, NaN, NaN, NaN, NaN, NaN, 5.804415702819824, 7.476558208465576, 7.281716346740723, NaN, 3.204416275024414, 5.887225151062012, NaN, 5.522943019866943, NaN, 7.468421936035156, 3.586181163787842, NaN, NaN, NaN, NaN, 7.412368297576904, NaN, NaN, 3.865264892578125, NaN, NaN, 4.811367034912109, 6.739513874053955, NaN, NaN, 4.651363372802734, NaN, NaN, NaN, NaN, NaN, NaN, 7.889192581176758, 3.7398288249969482, NaN, 5.875000953674316, NaN, NaN, NaN, NaN, NaN, NaN, 7.1523590087890625, NaN, 6.695902347564697, 5.453124523162842, 1.8444628715515137, 4.324471950531006, 6.565804958343506, NaN, NaN, 6.797978401184082, NaN, 6.775160789489746, 7.728910446166992, NaN, NaN, NaN, 6.11285400390625, NaN, NaN, NaN, 3.4552664756774902, 3.961127996444702, NaN, NaN, NaN, NaN, NaN, 3.383535146713257, 3.2812461853027344, NaN, NaN, NaN, NaN, 3.7107677459716797, NaN, 3.079761266708374, NaN, NaN, NaN, 5.857424259185791, 6.921466827392578, NaN, 5.034664630889893, NaN, NaN, NaN, NaN, 8.119081497192383, NaN, 2.086627721786499, 3.6019527912139893, NaN, 4.365927219390869, 4.202207088470459, NaN, 1.821563482284546, 4.347193717956543, 3.9450297355651855, 7.101864337921143, NaN, 5.642711162567139, 6.04410982131958, NaN, NaN, 2.851240396499634, NaN, NaN, NaN, NaN, NaN, NaN, NaN, 5.0581889152526855, NaN, NaN, 1.9336364269256592, NaN, NaN, 2.5444180965423584, 2.34977650642395, NaN, NaN, NaN, NaN, 3.9221155643463135, NaN, NaN, 6.586892604827881, 6.406312465667725, NaN, 1.7575258016586304, NaN, NaN, 1.4857674837112427, 6.958829402923584, NaN, NaN, 1.9712440967559814, NaN, 7.137136459350586, 2.428687810897827, NaN, NaN, NaN, NaN, 1.9834805727005005, 2.677837610244751, NaN, 5.523686408996582, 3.976874589920044, NaN, NaN, NaN, 7.4216413497924805, NaN, NaN, 4.17341423034668, 1.8640849590301514, NaN, NaN, 6.510715961456299, NaN, NaN, NaN, NaN, 6.541986465454102, NaN, NaN, NaN, NaN, 3.9177701473236084, NaN, NaN, 7.017882347106934, 4.7805352210998535, 4.128536701202393, 4.443770885467529, 7.234183311462402, NaN, NaN, NaN, NaN, NaN, NaN, 6.367973327636719, NaN, NaN, NaN, NaN, NaN, NaN, NaN, 2.6809616088867188, NaN, 4.459456443786621, NaN, NaN, NaN, NaN, 4.71559476852417, NaN, NaN, NaN, 2.4511547088623047, NaN, NaN, NaN, NaN, 7.343827247619629, NaN, 5.857483386993408, NaN, 5.850103855133057, 7.4540534019470215, NaN, NaN, 5.598409652709961, NaN, 3.399038553237915, 6.792981147766113, 6.3928399085998535, 6.303090572357178, NaN, 8.005901336669922, 5.948835372924805, NaN, NaN, NaN, 2.822648048400879, 3.848792314529419, 3.956167221069336, NaN, NaN, NaN, NaN, NaN, 2.3499913215637207, NaN, 6.240464210510254, NaN, NaN, NaN, NaN, 6.249919891357422, 3.86247181892395, 3.054460287094116, 7.166861057281494, 2.142493486404419, NaN, 2.244842529296875, 3.6683707237243652, NaN, 1.7975513935089111, NaN, 2.279144763946533, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, 2.8672080039978027, 4.498049736022949, NaN, NaN, NaN, 7.648446083068848, NaN, NaN, 4.9051008224487305, NaN, NaN, NaN, NaN, NaN, 3.0258944034576416, 3.585089683532715, 6.460017204284668, NaN, NaN, 6.5293354988098145, 6.140143394470215, NaN, 4.457757472991943, NaN, NaN, 8.47806453704834, 7.952311038970947, NaN, NaN, NaN, 7.237600326538086, NaN, NaN, NaN, NaN, NaN, NaN, 6.0425639152526855, NaN, 7.073198318481445, 3.2806644439697266, NaN, NaN, 5.480556011199951, NaN, 8.39113998413086, NaN, 4.732751369476318, 4.552467346191406, NaN, NaN, NaN, NaN, 5.539732933044434, NaN, 5.0976128578186035, NaN, NaN, 6.1718244552612305, NaN, NaN, NaN, 3.4206156730651855, 2.1956121921539307, 5.323253154754639, 6.857846736907959, NaN, NaN, 4.001266956329346, 2.6679840087890625, NaN, 3.9468014240264893, NaN, NaN, NaN, 7.4508209228515625, NaN, NaN, 3.024857997894287, NaN, 4.3979082107543945, 2.2977216243743896, NaN, NaN, NaN, NaN, NaN, 2.127079486846924, NaN, NaN, NaN, NaN, NaN, NaN, NaN, 5.54118013381958, NaN, 2.714792251586914, NaN, 5.991085052490234, NaN, NaN, NaN, 6.292999744415283, 2.680255174636841, 7.158297061920166, NaN, NaN, 6.5720601081848145, NaN, NaN, 3.576815366744995, NaN, 6.570065498352051, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, NaN, 2.1660032272338867, NaN, 5.6654791831970215, NaN, 4.350648880004883, NaN, NaN, NaN, NaN, 1.8763904571533203, NaN, NaN, 6.22961950302124, 1.902146577835083, NaN, NaN, 6.660125255584717, 6.744833946228027, 6.078064918518066, NaN, 2.7205452919006348, NaN, 7.640829086303711, NaN, NaN, 6.176510810852051, NaN, 4.165894031524658, 2.06752610206604, NaN, NaN, NaN, 5.184463977813721, NaN, NaN, NaN, 2.989898681640625, NaN, NaN, 5.65305233001709, 7.673363208770752, NaN, NaN, NaN, NaN, NaN, 2.1629552841186523, NaN, NaN, NaN, NaN, 7.604574680328369, 8.14582633972168, NaN, NaN, NaN, 1.819821834564209, 7.274652481079102, 4.688459396362305, NaN, NaN, NaN, NaN, NaN, NaN, 3.8662045001983643, 4.224400997161865, NaN, NaN, 3.9641048908233643, NaN, NaN, NaN, NaN, NaN, NaN]}
train_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"base_model_dir": "/llm-downloader-destination/base/fireworks/qwen2p5-72b-instruct/hf", "output_model_dir": "gs://fireworks-artifacts-mattrhorn27-f729ee-ab71a7/sftj-zyxu3xhf/8196e4/grwen-3-72b/checkpoint", "checkpoint_dir": "/dev/shm/checkpoints", "gcs_checkpoint_dir": "gs://fireworks-fine-tuning-checkpoints/sftj-mattrhorn27-f729ee-zyxu3xhf/checkpoints", "max_checkpoints_to_keep": 1, "checkpoint_interval": 3600, "train": true, "learning_rate": 0.0001, "learning_rate_warmup_steps": 0, "grad_accum_steps": 1, "epochs": 5, "early_stop": false, "seed": 42, "dataset_dir": "/mnt/staging/dataset", "eval_auto_carveout": true, "eval_dataset_dir": null, "train_limit": null, "max_context_len": 32768, "batch_size": 32768, "min_evals_per_epoch": 4, "precision": null, "status_file": "gs://fireworks-fine-tuning-job-status/sftj-mattrhorn27-f729ee-zyxu3xhf", "billing_file": "gs://fireworks-fine-tuning-metadata/sftj-mattrhorn27-f729ee-zyxu3xhf/billing", "wandb": false, "wandb_entity": null, "wandb_api_key": null, "wandb_project": null, "wandb_run_id": null, "profile": null, "weight_sharding": null, "activation_sharding": null, "empty_weights": false, "peft_addon_dir": null, "lora_rank": 8, "template_kind": "conversation", "template": null, "eval_train_ratio": 0.02}