TimeRobber commited on
Commit
a19f67e
·
1 Parent(s): 0089441

Upload part 67

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +39 -0
  2. checkpoint_1007000/state.param_states.decoder.layers_20.pre_cross_attention_layer_norm.scale.v/.zarray +1 -0
  3. checkpoint_1007000/state.param_states.decoder.layers_20.pre_cross_attention_layer_norm.scale.v/0 +0 -0
  4. checkpoint_1007000/state.param_states.decoder.layers_21.pre_self_attention_layer_norm.scale.v/.zarray +1 -0
  5. checkpoint_1007000/state.param_states.decoder.layers_21.pre_self_attention_layer_norm.scale.v/0 +0 -0
  6. checkpoint_1007000/state.param_states.encoder.layers_11.pre_attention_layer_norm.scale.v/.zarray +1 -0
  7. checkpoint_1007000/state.param_states.encoder.layers_11.pre_attention_layer_norm.scale.v/0 +0 -0
  8. checkpoint_1007000/state.param_states.encoder.layers_2.pre_mlp_layer_norm.scale.v/.zarray +1 -0
  9. checkpoint_1007000/state.param_states.encoder.layers_2.pre_mlp_layer_norm.scale.v/0 +0 -0
  10. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray +1 -0
  11. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp13576605543653617079.~1663960300175972~ +1 -0
  12. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp16598975374831553922.~1663960300351113~ +1 -0
  13. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp3080446702152752525.~1663960300425415~ +1 -0
  14. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp5307430905264549681.~1663960300685176~ +1 -0
  15. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp7512294191037663151.~1663960300308492~ +1 -0
  16. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp9638125748092535728.~1663960300609443~ +1 -0
  17. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/0.0 +3 -0
  18. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/1.0 +3 -0
  19. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/2.0 +3 -0
  20. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/3.0 +3 -0
  21. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/4.0 +3 -0
  22. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/5.0 +3 -0
  23. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/6.0 +3 -0
  24. checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/7.0 +3 -0
  25. checkpoint_1007000/target.decoder.layers_16.pre_mlp_layer_norm.scale/.zarray +1 -0
  26. checkpoint_1007000/target.decoder.layers_16.pre_mlp_layer_norm.scale/0 +0 -0
  27. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray +1 -0
  28. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp18309613063477522090.~1663960300610154~ +1 -0
  29. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp3877979065970436294.~1663960300865015~ +1 -0
  30. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp4282457207427269042.~1663960300683999~ +1 -0
  31. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp7407717715040280912.~1663960300886846~ +1 -0
  32. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp845119067367046232.~1663960300500731~ +1 -0
  33. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp847583154931984749.~1663960300727687~ +1 -0
  34. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.0 +3 -0
  35. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.1 +3 -0
  36. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.2 +3 -0
  37. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.3 +3 -0
  38. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.5 +3 -0
  39. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.6 +3 -0
  40. checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.7 +3 -0
  41. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray +1 -0
  42. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp14563675033902146399.~1663960300423552~ +1 -0
  43. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp14860002699643837038.~1663960300688575~ +1 -0
  44. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp18393025719903268069.~1663960300728767~ +1 -0
  45. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp1932148472918231926.~1663960300176739~ +1 -0
  46. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp370052280185407334.~1663960300131801~ +1 -0
  47. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp4057197573727006811.~1663960300546976~ +1 -0
  48. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.0 +3 -0
  49. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.1 +3 -0
  50. checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.2 +3 -0
.gitattributes CHANGED
@@ -3331,3 +3331,42 @@ checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.5 filter=lfs diff=
3331
  checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
3332
  checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
3333
  checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3331
  checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
3332
  checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
3333
  checkpoint_1007000/target.encoder.layers_21.mlp.wi_0.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
3334
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
3335
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
3336
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
3337
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
3338
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
3339
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
3340
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
3341
+ checkpoint_1007000/target.encoder.layers_8.attention.key.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
3342
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
3343
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
3344
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
3345
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
3346
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
3347
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
3348
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
3349
+ checkpoint_1007000/target.decoder.layers_8.encoder_decoder_attention.value.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
3350
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/6.0 filter=lfs diff=lfs merge=lfs -text
3351
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/1.0 filter=lfs diff=lfs merge=lfs -text
3352
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/3.0 filter=lfs diff=lfs merge=lfs -text
3353
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
3354
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/7.0 filter=lfs diff=lfs merge=lfs -text
3355
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/5.0 filter=lfs diff=lfs merge=lfs -text
3356
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/2.0 filter=lfs diff=lfs merge=lfs -text
3357
+ checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/4.0 filter=lfs diff=lfs merge=lfs -text
3358
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
3359
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
3360
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
3361
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
3362
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
3363
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
3364
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
3365
+ checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.4 filter=lfs diff=lfs merge=lfs -text
3366
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.1 filter=lfs diff=lfs merge=lfs -text
3367
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.7 filter=lfs diff=lfs merge=lfs -text
3368
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.0 filter=lfs diff=lfs merge=lfs -text
3369
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.6 filter=lfs diff=lfs merge=lfs -text
3370
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.5 filter=lfs diff=lfs merge=lfs -text
3371
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.3 filter=lfs diff=lfs merge=lfs -text
3372
+ checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.2 filter=lfs diff=lfs merge=lfs -text
checkpoint_1007000/state.param_states.decoder.layers_20.pre_cross_attention_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.decoder.layers_20.pre_cross_attention_layer_norm.scale.v/0 ADDED
Binary file (14.5 kB). View file
 
checkpoint_1007000/state.param_states.decoder.layers_21.pre_self_attention_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.decoder.layers_21.pre_self_attention_layer_norm.scale.v/0 ADDED
Binary file (14.5 kB). View file
 
checkpoint_1007000/state.param_states.encoder.layers_11.pre_attention_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.encoder.layers_11.pre_attention_layer_norm.scale.v/0 ADDED
Binary file (15.3 kB). View file
 
checkpoint_1007000/state.param_states.encoder.layers_2.pre_mlp_layer_norm.scale.v/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/state.param_states.encoder.layers_2.pre_mlp_layer_norm.scale.v/0 ADDED
Binary file (15.3 kB). View file
 
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp13576605543653617079.~1663960300175972~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp16598975374831553922.~1663960300351113~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp3080446702152752525.~1663960300425415~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp5307430905264549681.~1663960300685176~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp7512294191037663151.~1663960300308492~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/.zarray.__tmp9638125748092535728.~1663960300609443~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[512,4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2c2b743b28ca7b57a7a70cec9aefe36f2901ac4c6fcb4512ec09ba8158fdec1
3
+ size 7820233
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/1.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9f260fa420d9a3be5079c7c8fbe846b033f0927c07f0730242da496395b2eb74
3
+ size 7818789
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/2.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67b1047232f6fe35c74011810cf339d8458bc15a787069e269b04bfa747f735f
3
+ size 7820307
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/3.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:548717a35fa9bff5e805cd799d63832fb556947eef7aa5c78bbbc5d7fbb074f8
3
+ size 7819499
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/4.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcfabac988acb16b75603def277c0a4c718e2158666e472761ef1777a6a90775
3
+ size 7820654
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/5.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4c5187fac27d42e483cf46e57da249395480b92ac2e1411cff2da305936d5f80
3
+ size 7820605
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/6.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:071aa82e113413dc7e82cbaa6659766f5436b06656fc38e6c056665679a82318
3
+ size 7821430
checkpoint_1007000/target.decoder.layers_14.self_attention.out.kernel/7.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82d73cd1be03e903fef4ea41457bbcd24cec872b925c56188e79eecc41d8e7d8
3
+ size 7820373
checkpoint_1007000/target.decoder.layers_16.pre_mlp_layer_norm.scale/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_16.pre_mlp_layer_norm.scale/0 ADDED
Binary file (14.3 kB). View file
 
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp18309613063477522090.~1663960300610154~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp3877979065970436294.~1663960300865015~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp4282457207427269042.~1663960300683999~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp7407717715040280912.~1663960300886846~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp845119067367046232.~1663960300500731~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/.zarray.__tmp847583154931984749.~1663960300727687~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:438ddd6ec441ba4139512d87fcb089060f8d4be607aed3d85d21bf29376e6a94
3
+ size 7805623
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2559ba51354c9a638c25bfc66c7e77643076c00a93f2feedea532e7cf9a5928b
3
+ size 7811335
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c9b45a7435dc5d9cf2ae85e840db58c90282d9d19d59a0f6a10d4b391119818
3
+ size 7806237
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:134b087ecbc788d473f4ccfcd9b50ec1b8632002f586efcd956c952b67b01575
3
+ size 7807242
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:58ac8244e38a7afff6f80c7bbdc238d9dcf76b643958ff5df847f1c863ecf5ec
3
+ size 7804549
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf2ae1fb2e64fbb44432fed0846598982a9a1694d1ad0b59c42873fc8020ba79
3
+ size 7814492
checkpoint_1007000/target.decoder.layers_19.self_attention.query.kernel/0.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81a3001499c32e2e8f04accac69e1a80a0738f4940ad7184653a80f63a9f31b4
3
+ size 7805911
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp14563675033902146399.~1663960300423552~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp14860002699643837038.~1663960300688575~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp18393025719903268069.~1663960300728767~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp1932148472918231926.~1663960300176739~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp370052280185407334.~1663960300131801~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/.zarray.__tmp4057197573727006811.~1663960300546976~ ADDED
@@ -0,0 +1 @@
 
 
1
+ {"chunks":[4096,512],"compressor":{"id":"gzip","level":1},"dimension_separator":".","dtype":"<f4","fill_value":null,"filters":null,"order":"C","shape":[4096,4096],"zarr_format":2}
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e50f618bd1b62aae2a1e2ca7c46eda9dccf8a9df49c604aabf47ed8e5ee2fba
3
+ size 7815970
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b03d467b4768d3028171fb2e0d0bf8e1dd4d5092cead4bb6cf9f3f0d487b6634
3
+ size 7817190
checkpoint_1007000/target.decoder.layers_5.encoder_decoder_attention.value.kernel/0.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:510e1c2f362c9b63d17317b206650e3ec74e4438b844d4b5aee7a90eff1648b1
3
+ size 7816753