tensor-tech commited on
Commit
f0c76f1
·
verified ·
1 Parent(s): 8a4307e

Run 5. Outer Step 50. Inner Step 22.

Browse files
Files changed (1) hide show
  1. config.json +22 -23
config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "all_reduce_scores": {
3
  "0": "NON_PARTICIPATING",
4
  "1": "SUCCESS",
@@ -24,7 +25,7 @@
24
  "117": "NON_PARTICIPATING",
25
  "118": "NON_PARTICIPATING",
26
  "119": "NON_PARTICIPATING",
27
- "12": "NON_PARTICIPATING",
28
  "120": "NON_PARTICIPATING",
29
  "121": "NON_PARTICIPATING",
30
  "122": "NON_PARTICIPATING",
@@ -54,7 +55,7 @@
54
  "144": "NON_PARTICIPATING",
55
  "145": "NON_PARTICIPATING",
56
  "146": "NON_PARTICIPATING",
57
- "147": "NON_PARTICIPATING",
58
  "148": "NON_PARTICIPATING",
59
  "149": "NON_PARTICIPATING",
60
  "15": "SUCCESS",
@@ -67,13 +68,13 @@
67
  "156": "NON_PARTICIPATING",
68
  "157": "NON_PARTICIPATING",
69
  "158": "NON_PARTICIPATING",
70
- "159": "SUCCESS",
71
  "16": "SUCCESS",
72
  "160": "NON_PARTICIPATING",
73
  "161": "NON_PARTICIPATING",
74
  "162": "NON_PARTICIPATING",
75
  "163": "NON_PARTICIPATING",
76
- "164": "SUCCESS",
77
  "165": "NON_PARTICIPATING",
78
  "166": "NON_PARTICIPATING",
79
  "167": "NON_PARTICIPATING",
@@ -92,18 +93,18 @@
92
  "179": "NON_PARTICIPATING",
93
  "18": "SUCCESS",
94
  "180": "NON_PARTICIPATING",
95
- "181": "SUCCESS",
96
  "182": "NON_PARTICIPATING",
97
  "183": "NON_PARTICIPATING",
98
  "184": "NON_PARTICIPATING",
99
  "185": "NON_PARTICIPATING",
100
  "186": "NON_PARTICIPATING",
101
- "187": "NON_PARTICIPATING",
102
  "188": "NON_PARTICIPATING",
103
  "189": "NON_PARTICIPATING",
104
  "19": "SUCCESS",
105
  "190": "NON_PARTICIPATING",
106
- "191": "SUCCESS",
107
  "192": "NON_PARTICIPATING",
108
  "193": "NON_PARTICIPATING",
109
  "194": "NON_PARTICIPATING",
@@ -111,10 +112,10 @@
111
  "196": "NON_PARTICIPATING",
112
  "197": "NON_PARTICIPATING",
113
  "198": "NON_PARTICIPATING",
114
- "199": "NON_PARTICIPATING",
115
  "2": "SUCCESS",
116
  "20": "SUCCESS",
117
- "200": "NON_PARTICIPATING",
118
  "201": "NON_PARTICIPATING",
119
  "202": "NON_PARTICIPATING",
120
  "203": "NON_PARTICIPATING",
@@ -125,7 +126,7 @@
125
  "208": "NON_PARTICIPATING",
126
  "209": "NON_PARTICIPATING",
127
  "21": "SUCCESS",
128
- "210": "NON_PARTICIPATING",
129
  "211": "NON_PARTICIPATING",
130
  "212": "NON_PARTICIPATING",
131
  "213": "NON_PARTICIPATING",
@@ -139,14 +140,14 @@
139
  "220": "NON_PARTICIPATING",
140
  "221": "NON_PARTICIPATING",
141
  "222": "NON_PARTICIPATING",
142
- "223": "SUCCESS",
143
  "224": "NON_PARTICIPATING",
144
  "225": "NON_PARTICIPATING",
145
  "226": "NON_PARTICIPATING",
146
  "227": "NON_PARTICIPATING",
147
  "228": "NON_PARTICIPATING",
148
  "229": "NON_PARTICIPATING",
149
- "23": "NON_PARTICIPATING",
150
  "230": "NON_PARTICIPATING",
151
  "231": "NON_PARTICIPATING",
152
  "232": "NON_PARTICIPATING",
@@ -164,7 +165,7 @@
164
  "243": "NON_PARTICIPATING",
165
  "244": "NON_PARTICIPATING",
166
  "245": "NON_PARTICIPATING",
167
- "246": "SUCCESS",
168
  "247": "NON_PARTICIPATING",
169
  "248": "NON_PARTICIPATING",
170
  "249": "NON_PARTICIPATING",
@@ -178,7 +179,7 @@
178
  "26": "NON_PARTICIPATING",
179
  "27": "SUCCESS",
180
  "28": "SUCCESS",
181
- "29": "NON_PARTICIPATING",
182
  "3": "SUCCESS",
183
  "30": "SUCCESS",
184
  "31": "SUCCESS",
@@ -226,12 +227,12 @@
226
  "7": "SUCCESS",
227
  "70": "SUCCESS",
228
  "71": "SUCCESS",
229
- "72": "SUCCESS",
230
  "73": "SUCCESS",
231
  "74": "SUCCESS",
232
  "75": "SUCCESS",
233
  "76": "SUCCESS",
234
- "77": "SUCCESS",
235
  "78": "SUCCESS",
236
  "79": "SUCCESS",
237
  "8": "SUCCESS",
@@ -252,7 +253,7 @@
252
  "93": "SUCCESS",
253
  "94": "SUCCESS",
254
  "95": "SUCCESS",
255
- "96": "NON_PARTICIPATING",
256
  "97": "SUCCESS",
257
  "98": "SUCCESS",
258
  "99": "SUCCESS"
@@ -263,18 +264,16 @@
263
  "attention_bias": false,
264
  "attention_dropout": 0.0,
265
  "block_list": [
266
- 5759413,
267
- 5759450
268
  ],
269
  "bos_token_id": 1,
270
  "eos_token_id": 2,
271
- "head_dim": 64,
272
  "hidden_act": "silu",
273
  "hidden_size": 2048,
274
  "initializer_range": 0.02,
275
- "inner_step": 65,
276
  "intermediate_size": 5632,
277
- "last_allreduce_block": 5744479,
278
  "max_position_embeddings": 2048,
279
  "mlp_bias": false,
280
  "model_type": "llama",
@@ -287,7 +286,7 @@
287
  "rope_theta": 10000.0,
288
  "tie_word_embeddings": false,
289
  "torch_dtype": "float32",
290
- "transformers_version": "4.52.4",
291
  "use_cache": false,
292
  "vocab_size": 32000
293
  }
 
1
  {
2
+ "_name_or_path": "unrented5443/sn38-v2-5",
3
  "all_reduce_scores": {
4
  "0": "NON_PARTICIPATING",
5
  "1": "SUCCESS",
 
25
  "117": "NON_PARTICIPATING",
26
  "118": "NON_PARTICIPATING",
27
  "119": "NON_PARTICIPATING",
28
+ "12": "SUCCESS",
29
  "120": "NON_PARTICIPATING",
30
  "121": "NON_PARTICIPATING",
31
  "122": "NON_PARTICIPATING",
 
55
  "144": "NON_PARTICIPATING",
56
  "145": "NON_PARTICIPATING",
57
  "146": "NON_PARTICIPATING",
58
+ "147": "SUCCESS",
59
  "148": "NON_PARTICIPATING",
60
  "149": "NON_PARTICIPATING",
61
  "15": "SUCCESS",
 
68
  "156": "NON_PARTICIPATING",
69
  "157": "NON_PARTICIPATING",
70
  "158": "NON_PARTICIPATING",
71
+ "159": "NON_PARTICIPATING",
72
  "16": "SUCCESS",
73
  "160": "NON_PARTICIPATING",
74
  "161": "NON_PARTICIPATING",
75
  "162": "NON_PARTICIPATING",
76
  "163": "NON_PARTICIPATING",
77
+ "164": "NON_PARTICIPATING",
78
  "165": "NON_PARTICIPATING",
79
  "166": "NON_PARTICIPATING",
80
  "167": "NON_PARTICIPATING",
 
93
  "179": "NON_PARTICIPATING",
94
  "18": "SUCCESS",
95
  "180": "NON_PARTICIPATING",
96
+ "181": "NON_PARTICIPATING",
97
  "182": "NON_PARTICIPATING",
98
  "183": "NON_PARTICIPATING",
99
  "184": "NON_PARTICIPATING",
100
  "185": "NON_PARTICIPATING",
101
  "186": "NON_PARTICIPATING",
102
+ "187": "SUCCESS",
103
  "188": "NON_PARTICIPATING",
104
  "189": "NON_PARTICIPATING",
105
  "19": "SUCCESS",
106
  "190": "NON_PARTICIPATING",
107
+ "191": "NON_PARTICIPATING",
108
  "192": "NON_PARTICIPATING",
109
  "193": "NON_PARTICIPATING",
110
  "194": "NON_PARTICIPATING",
 
112
  "196": "NON_PARTICIPATING",
113
  "197": "NON_PARTICIPATING",
114
  "198": "NON_PARTICIPATING",
115
+ "199": "SUCCESS",
116
  "2": "SUCCESS",
117
  "20": "SUCCESS",
118
+ "200": "SUCCESS",
119
  "201": "NON_PARTICIPATING",
120
  "202": "NON_PARTICIPATING",
121
  "203": "NON_PARTICIPATING",
 
126
  "208": "NON_PARTICIPATING",
127
  "209": "NON_PARTICIPATING",
128
  "21": "SUCCESS",
129
+ "210": "SUCCESS",
130
  "211": "NON_PARTICIPATING",
131
  "212": "NON_PARTICIPATING",
132
  "213": "NON_PARTICIPATING",
 
140
  "220": "NON_PARTICIPATING",
141
  "221": "NON_PARTICIPATING",
142
  "222": "NON_PARTICIPATING",
143
+ "223": "NON_PARTICIPATING",
144
  "224": "NON_PARTICIPATING",
145
  "225": "NON_PARTICIPATING",
146
  "226": "NON_PARTICIPATING",
147
  "227": "NON_PARTICIPATING",
148
  "228": "NON_PARTICIPATING",
149
  "229": "NON_PARTICIPATING",
150
+ "23": "SUCCESS",
151
  "230": "NON_PARTICIPATING",
152
  "231": "NON_PARTICIPATING",
153
  "232": "NON_PARTICIPATING",
 
165
  "243": "NON_PARTICIPATING",
166
  "244": "NON_PARTICIPATING",
167
  "245": "NON_PARTICIPATING",
168
+ "246": "NON_PARTICIPATING",
169
  "247": "NON_PARTICIPATING",
170
  "248": "NON_PARTICIPATING",
171
  "249": "NON_PARTICIPATING",
 
179
  "26": "NON_PARTICIPATING",
180
  "27": "SUCCESS",
181
  "28": "SUCCESS",
182
+ "29": "SUCCESS",
183
  "3": "SUCCESS",
184
  "30": "SUCCESS",
185
  "31": "SUCCESS",
 
227
  "7": "SUCCESS",
228
  "70": "SUCCESS",
229
  "71": "SUCCESS",
230
+ "72": "NON_PARTICIPATING",
231
  "73": "SUCCESS",
232
  "74": "SUCCESS",
233
  "75": "SUCCESS",
234
  "76": "SUCCESS",
235
+ "77": "NON_PARTICIPATING",
236
  "78": "SUCCESS",
237
  "79": "SUCCESS",
238
  "8": "SUCCESS",
 
253
  "93": "SUCCESS",
254
  "94": "SUCCESS",
255
  "95": "SUCCESS",
256
+ "96": "SUCCESS",
257
  "97": "SUCCESS",
258
  "98": "SUCCESS",
259
  "99": "SUCCESS"
 
264
  "attention_bias": false,
265
  "attention_dropout": 0.0,
266
  "block_list": [
267
+ 5774607
 
268
  ],
269
  "bos_token_id": 1,
270
  "eos_token_id": 2,
 
271
  "hidden_act": "silu",
272
  "hidden_size": 2048,
273
  "initializer_range": 0.02,
274
+ "inner_step": 23,
275
  "intermediate_size": 5632,
276
+ "last_allreduce_block": 5770468,
277
  "max_position_embeddings": 2048,
278
  "mlp_bias": false,
279
  "model_type": "llama",
 
286
  "rope_theta": 10000.0,
287
  "tie_word_embeddings": false,
288
  "torch_dtype": "float32",
289
+ "transformers_version": "4.39.3",
290
  "use_cache": false,
291
  "vocab_size": 32000
292
  }