manu commited on
Commit
1a0227a
·
1 Parent(s): 959e8ae

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "mock_training_run/llama_configs/config_small.json",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -7,13 +7,13 @@
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
  "hidden_act": "silu",
10
- "hidden_size": 768,
11
  "initializer_range": 0.02,
12
- "intermediate_size": 768,
13
  "max_position_embeddings": 2048,
14
  "model_type": "llama",
15
  "num_attention_heads": 12,
16
- "num_hidden_layers": 6,
17
  "num_key_value_heads": 12,
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-05,
@@ -21,7 +21,7 @@
21
  "rope_theta": 10000.0,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "float32",
24
- "transformers_version": "4.34.0",
25
  "use_cache": true,
26
  "vocab_size": 32000
27
  }
 
1
  {
2
+ "_name_or_path": "mock_training_run/llama_configs/config.json",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
7
  "bos_token_id": 1,
8
  "eos_token_id": 2,
9
  "hidden_act": "silu",
10
+ "hidden_size": 1536,
11
  "initializer_range": 0.02,
12
+ "intermediate_size": 4128,
13
  "max_position_embeddings": 2048,
14
  "model_type": "llama",
15
  "num_attention_heads": 12,
16
+ "num_hidden_layers": 12,
17
  "num_key_value_heads": 12,
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-05,
 
21
  "rope_theta": 10000.0,
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "float32",
24
+ "transformers_version": "4.34.1",
25
  "use_cache": true,
26
  "vocab_size": 32000
27
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1da14be436c5653de52c247db34f495c449d5ea0f3a226a9b60c3b6e21700e4e
3
- size 295757212
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64f8a1e11e54b0d43b7c22ed5519bbae53d71b7c58852996389c4c264db6a28c
3
+ size 1759438985
special_tokens_map.json CHANGED
@@ -1,111 +1,23 @@
1
  {
2
- "additional_special_tokens": [
3
- "<unk>",
4
- "<s>",
5
- "</s>",
6
- "<pad>",
7
- "<extra_id_0>",
8
- "<extra_id_1>",
9
- "<extra_id_2>",
10
- "<extra_id_3>",
11
- "<extra_id_4>",
12
- "<extra_id_5>",
13
- "<extra_id_6>",
14
- "<extra_id_7>",
15
- "<extra_id_8>",
16
- "<extra_id_9>",
17
- "<extra_id_10>",
18
- "<extra_id_11>",
19
- "<extra_id_12>",
20
- "<extra_id_13>",
21
- "<extra_id_14>",
22
- "<extra_id_15>",
23
- "<extra_id_16>",
24
- "<extra_id_17>",
25
- "<extra_id_18>",
26
- "<extra_id_19>",
27
- "<extra_id_20>",
28
- "<extra_id_21>",
29
- "<extra_id_22>",
30
- "<extra_id_23>",
31
- "<extra_id_24>",
32
- "<extra_id_25>",
33
- "<extra_id_26>",
34
- "<extra_id_27>",
35
- "<extra_id_28>",
36
- "<extra_id_29>",
37
- "<extra_id_30>",
38
- "<extra_id_31>",
39
- "<extra_id_32>",
40
- "<extra_id_33>",
41
- "<extra_id_34>",
42
- "<extra_id_35>",
43
- "<extra_id_36>",
44
- "<extra_id_37>",
45
- "<extra_id_38>",
46
- "<extra_id_39>",
47
- "<extra_id_40>",
48
- "<extra_id_41>",
49
- "<extra_id_42>",
50
- "<extra_id_43>",
51
- "<extra_id_44>",
52
- "<extra_id_45>",
53
- "<extra_id_46>",
54
- "<extra_id_47>",
55
- "<extra_id_48>",
56
- "<extra_id_49>",
57
- "<extra_id_50>",
58
- "<extra_id_51>",
59
- "<extra_id_52>",
60
- "<extra_id_53>",
61
- "<extra_id_54>",
62
- "<extra_id_55>",
63
- "<extra_id_56>",
64
- "<extra_id_57>",
65
- "<extra_id_58>",
66
- "<extra_id_59>",
67
- "<extra_id_60>",
68
- "<extra_id_61>",
69
- "<extra_id_62>",
70
- "<extra_id_63>",
71
- "<extra_id_64>",
72
- "<extra_id_65>",
73
- "<extra_id_66>",
74
- "<extra_id_67>",
75
- "<extra_id_68>",
76
- "<extra_id_69>",
77
- "<extra_id_70>",
78
- "<extra_id_71>",
79
- "<extra_id_72>",
80
- "<extra_id_73>",
81
- "<extra_id_74>",
82
- "<extra_id_75>",
83
- "<extra_id_76>",
84
- "<extra_id_77>",
85
- "<extra_id_78>",
86
- "<extra_id_79>",
87
- "<extra_id_80>",
88
- "<extra_id_81>",
89
- "<extra_id_82>",
90
- "<extra_id_83>",
91
- "<extra_id_84>",
92
- "<extra_id_85>",
93
- "<extra_id_86>",
94
- "<extra_id_87>",
95
- "<extra_id_88>",
96
- "<extra_id_89>",
97
- "<extra_id_90>",
98
- "<extra_id_91>",
99
- "<extra_id_92>",
100
- "<extra_id_93>",
101
- "<extra_id_94>",
102
- "<extra_id_95>",
103
- "<extra_id_96>",
104
- "<extra_id_97>",
105
- "<extra_id_98>",
106
- "<extra_id_99>"
107
- ],
108
- "bos_token": "<s>",
109
- "eos_token": "</s>",
110
- "unk_token": "<unk>"
111
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  }
tokenizer_config.json CHANGED
@@ -833,112 +833,7 @@
833
  "special": true
834
  }
835
  },
836
- "additional_special_tokens": [
837
- "<unk>",
838
- "<s>",
839
- "</s>",
840
- "<pad>",
841
- "<extra_id_0>",
842
- "<extra_id_1>",
843
- "<extra_id_2>",
844
- "<extra_id_3>",
845
- "<extra_id_4>",
846
- "<extra_id_5>",
847
- "<extra_id_6>",
848
- "<extra_id_7>",
849
- "<extra_id_8>",
850
- "<extra_id_9>",
851
- "<extra_id_10>",
852
- "<extra_id_11>",
853
- "<extra_id_12>",
854
- "<extra_id_13>",
855
- "<extra_id_14>",
856
- "<extra_id_15>",
857
- "<extra_id_16>",
858
- "<extra_id_17>",
859
- "<extra_id_18>",
860
- "<extra_id_19>",
861
- "<extra_id_20>",
862
- "<extra_id_21>",
863
- "<extra_id_22>",
864
- "<extra_id_23>",
865
- "<extra_id_24>",
866
- "<extra_id_25>",
867
- "<extra_id_26>",
868
- "<extra_id_27>",
869
- "<extra_id_28>",
870
- "<extra_id_29>",
871
- "<extra_id_30>",
872
- "<extra_id_31>",
873
- "<extra_id_32>",
874
- "<extra_id_33>",
875
- "<extra_id_34>",
876
- "<extra_id_35>",
877
- "<extra_id_36>",
878
- "<extra_id_37>",
879
- "<extra_id_38>",
880
- "<extra_id_39>",
881
- "<extra_id_40>",
882
- "<extra_id_41>",
883
- "<extra_id_42>",
884
- "<extra_id_43>",
885
- "<extra_id_44>",
886
- "<extra_id_45>",
887
- "<extra_id_46>",
888
- "<extra_id_47>",
889
- "<extra_id_48>",
890
- "<extra_id_49>",
891
- "<extra_id_50>",
892
- "<extra_id_51>",
893
- "<extra_id_52>",
894
- "<extra_id_53>",
895
- "<extra_id_54>",
896
- "<extra_id_55>",
897
- "<extra_id_56>",
898
- "<extra_id_57>",
899
- "<extra_id_58>",
900
- "<extra_id_59>",
901
- "<extra_id_60>",
902
- "<extra_id_61>",
903
- "<extra_id_62>",
904
- "<extra_id_63>",
905
- "<extra_id_64>",
906
- "<extra_id_65>",
907
- "<extra_id_66>",
908
- "<extra_id_67>",
909
- "<extra_id_68>",
910
- "<extra_id_69>",
911
- "<extra_id_70>",
912
- "<extra_id_71>",
913
- "<extra_id_72>",
914
- "<extra_id_73>",
915
- "<extra_id_74>",
916
- "<extra_id_75>",
917
- "<extra_id_76>",
918
- "<extra_id_77>",
919
- "<extra_id_78>",
920
- "<extra_id_79>",
921
- "<extra_id_80>",
922
- "<extra_id_81>",
923
- "<extra_id_82>",
924
- "<extra_id_83>",
925
- "<extra_id_84>",
926
- "<extra_id_85>",
927
- "<extra_id_86>",
928
- "<extra_id_87>",
929
- "<extra_id_88>",
930
- "<extra_id_89>",
931
- "<extra_id_90>",
932
- "<extra_id_91>",
933
- "<extra_id_92>",
934
- "<extra_id_93>",
935
- "<extra_id_94>",
936
- "<extra_id_95>",
937
- "<extra_id_96>",
938
- "<extra_id_97>",
939
- "<extra_id_98>",
940
- "<extra_id_99>"
941
- ],
942
  "bos_token": "<s>",
943
  "clean_up_tokenization_spaces": false,
944
  "eos_token": "</s>",
 
833
  "special": true
834
  }
835
  },
836
+ "additional_special_tokens": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
837
  "bos_token": "<s>",
838
  "clean_up_tokenization_spaces": false,
839
  "eos_token": "</s>",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff15deee9c5aa2fe39c0f9564f410e6476913d80ebfbafdf4665ea6528c30433
3
  size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fbf9539d6349c2e5eeebbb25ddc0a48bae2c0f6e8afdb68d382fb72d99a025e
3
  size 4091