stas commited on
Commit
40806d1
·
1 Parent(s): 3a91f3f
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. exp12/000-module.3.input_layernorm/events.out.tfevents.1638184517.r7i5n5.1015419.4 +3 -0
  2. exp12/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1638184518.r7i5n5.1015419.12 +3 -0
  3. exp12/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1638184518.r7i5n5.1015419.11 +3 -0
  4. exp12/000-module.3.mlp/events.out.tfevents.1638184518.r7i5n5.1015419.13 +3 -0
  5. exp12/000-module.3.post_attention_layernorm/events.out.tfevents.1638184518.r7i5n5.1015419.10 +3 -0
  6. exp12/000-module.3.self_attention.attention_dropout/events.out.tfevents.1638184518.r7i5n5.1015419.7 +3 -0
  7. exp12/000-module.3.self_attention.dense/events.out.tfevents.1638184518.r7i5n5.1015419.8 +3 -0
  8. exp12/000-module.3.self_attention.query_key_value/events.out.tfevents.1638184518.r7i5n5.1015419.5 +3 -0
  9. exp12/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1638184518.r7i5n5.1015419.6 +3 -0
  10. exp12/000-module.3.self_attention/events.out.tfevents.1638184518.r7i5n5.1015419.9 +3 -0
  11. exp12/000-module.3/events.out.tfevents.1638184519.r7i5n5.1015419.14 +3 -0
  12. exp12/000-module.4.input_layernorm/events.out.tfevents.1638184519.r7i5n5.1015419.15 +3 -0
  13. exp12/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1638184519.r7i5n5.1015419.23 +3 -0
  14. exp12/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1638184519.r7i5n5.1015419.22 +3 -0
  15. exp12/000-module.4.mlp/events.out.tfevents.1638184519.r7i5n5.1015419.24 +3 -0
  16. exp12/000-module.4.post_attention_layernorm/events.out.tfevents.1638184519.r7i5n5.1015419.21 +3 -0
  17. exp12/000-module.4.self_attention.attention_dropout/events.out.tfevents.1638184519.r7i5n5.1015419.18 +3 -0
  18. exp12/000-module.4.self_attention.dense/events.out.tfevents.1638184519.r7i5n5.1015419.19 +3 -0
  19. exp12/000-module.4.self_attention.query_key_value/events.out.tfevents.1638184519.r7i5n5.1015419.16 +3 -0
  20. exp12/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1638184519.r7i5n5.1015419.17 +3 -0
  21. exp12/000-module.4.self_attention/events.out.tfevents.1638184519.r7i5n5.1015419.20 +3 -0
  22. exp12/000-module.4/events.out.tfevents.1638184519.r7i5n5.1015419.25 +3 -0
  23. exp12/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1638184517.r7i5n5.1015419.2 +3 -0
  24. exp12/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1638184517.r7i5n5.1015419.1 +3 -0
  25. exp12/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1638184517.r7i5n5.1015419.0 +3 -0
  26. exp12/000-module.tied_modules.embed/events.out.tfevents.1638184517.r7i5n5.1015419.3 +3 -0
  27. exp12/000-module/events.out.tfevents.1638184519.r7i5n5.1015419.26 +3 -0
  28. exp12/028-module.17.input_layernorm/events.out.tfevents.1638184529.r8i4n3.750609.0 +3 -0
  29. exp12/028-module.17.mlp.dense_4h_to_h/events.out.tfevents.1638184531.r8i4n3.750609.8 +3 -0
  30. exp12/028-module.17.mlp.dense_h_to_4h/events.out.tfevents.1638184530.r8i4n3.750609.7 +3 -0
  31. exp12/028-module.17.mlp/events.out.tfevents.1638184531.r8i4n3.750609.9 +3 -0
  32. exp12/028-module.17.post_attention_layernorm/events.out.tfevents.1638184530.r8i4n3.750609.6 +3 -0
  33. exp12/028-module.17.self_attention.attention_dropout/events.out.tfevents.1638184530.r8i4n3.750609.3 +3 -0
  34. exp12/028-module.17.self_attention.dense/events.out.tfevents.1638184530.r8i4n3.750609.4 +3 -0
  35. exp12/028-module.17.self_attention.query_key_value/events.out.tfevents.1638184530.r8i4n3.750609.1 +3 -0
  36. exp12/028-module.17.self_attention.scale_mask_softmax/events.out.tfevents.1638184530.r8i4n3.750609.2 +3 -0
  37. exp12/028-module.17.self_attention/events.out.tfevents.1638184530.r8i4n3.750609.5 +3 -0
  38. exp12/028-module.17/events.out.tfevents.1638184531.r8i4n3.750609.10 +3 -0
  39. exp12/028-module.18.input_layernorm/events.out.tfevents.1638184531.r8i4n3.750609.11 +3 -0
  40. exp12/028-module.18.mlp.dense_4h_to_h/events.out.tfevents.1638184531.r8i4n3.750609.19 +3 -0
  41. exp12/028-module.18.mlp.dense_h_to_4h/events.out.tfevents.1638184531.r8i4n3.750609.18 +3 -0
  42. exp12/028-module.18.mlp/events.out.tfevents.1638184531.r8i4n3.750609.20 +3 -0
  43. exp12/028-module.18.post_attention_layernorm/events.out.tfevents.1638184531.r8i4n3.750609.17 +3 -0
  44. exp12/028-module.18.self_attention.attention_dropout/events.out.tfevents.1638184531.r8i4n3.750609.14 +3 -0
  45. exp12/028-module.18.self_attention.dense/events.out.tfevents.1638184531.r8i4n3.750609.15 +3 -0
  46. exp12/028-module.18.self_attention.query_key_value/events.out.tfevents.1638184531.r8i4n3.750609.12 +3 -0
  47. exp12/028-module.18.self_attention.scale_mask_softmax/events.out.tfevents.1638184531.r8i4n3.750609.13 +3 -0
  48. exp12/028-module.18.self_attention/events.out.tfevents.1638184531.r8i4n3.750609.16 +3 -0
  49. exp12/028-module.18/events.out.tfevents.1638184531.r8i4n3.750609.21 +3 -0
  50. exp12/028-module/events.out.tfevents.1638184531.r8i4n3.750609.22 +3 -0
exp12/000-module.3.input_layernorm/events.out.tfevents.1638184517.r7i5n5.1015419.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4265affaa9b9677109be376ff06d4364f416f478630f749ec34fdc0ca2b2790
3
+ size 35370760
exp12/000-module.3.mlp.dense_4h_to_h/events.out.tfevents.1638184518.r7i5n5.1015419.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7973118fbc6a129c7e175593410940968ed565ad639188f335d0e0d05a631068
3
+ size 35704840
exp12/000-module.3.mlp.dense_h_to_4h/events.out.tfevents.1638184518.r7i5n5.1015419.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff5dfede84fbc7577d808d21e56790920ca1ea18f45d99f63efc42f2ca4a3575
3
+ size 35704840
exp12/000-module.3.mlp/events.out.tfevents.1638184518.r7i5n5.1015419.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87925947159f3bf4c5b1fee643a67a5953a955100efd96f44179650eb2cdd727
3
+ size 20921800
exp12/000-module.3.post_attention_layernorm/events.out.tfevents.1638184518.r7i5n5.1015419.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0732acfaafec09259f42b2bfb3142a8b0d3932e27b570ed35c8f02b8450a99da
3
+ size 35370760
exp12/000-module.3.self_attention.attention_dropout/events.out.tfevents.1638184518.r7i5n5.1015419.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dec202a35ee813f16d42cbb682231654d076b5e7fe6a1985a05543ef06caf1fb
3
+ size 15409480
exp12/000-module.3.self_attention.dense/events.out.tfevents.1638184518.r7i5n5.1015419.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3ce4483881906b6ce44be657b52d8e0769c1ef022f56859edb65fb40c3824f7
3
+ size 35704840
exp12/000-module.3.self_attention.query_key_value/events.out.tfevents.1638184518.r7i5n5.1015419.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55d1d3b9e26a6991b1ad5b55e9bbec9eeec532cc61eae3132123efc2a63c0d22
3
+ size 33032200
exp12/000-module.3.self_attention.scale_mask_softmax/events.out.tfevents.1638184518.r7i5n5.1015419.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb519dfd4dacdd9da505f71f31b58cfef4f0c9b42697487e6cde7d109cff2b05
3
+ size 15409480
exp12/000-module.3.self_attention/events.out.tfevents.1638184518.r7i5n5.1015419.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08e0022a8c3961956f5520272f2e553e1c4897e39e55879926f3c0cdc4dfe10c
3
+ size 20921800
exp12/000-module.3/events.out.tfevents.1638184519.r7i5n5.1015419.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:542ad3100587cbbc225b2a762c7d5cb81ac6f0e5cd9baababdcef417d5dc43dc
3
+ size 20587720
exp12/000-module.4.input_layernorm/events.out.tfevents.1638184519.r7i5n5.1015419.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7dbd67936b3f26bec0c972fcefa8b3795886f5d006cba1d170ff44d68e3c00f
3
+ size 35370760
exp12/000-module.4.mlp.dense_4h_to_h/events.out.tfevents.1638184519.r7i5n5.1015419.23 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5d663ea401926d02f4eea32f090f939b9507bf1ac1ae7ef2ef8a1b851c793e6
3
+ size 35704840
exp12/000-module.4.mlp.dense_h_to_4h/events.out.tfevents.1638184519.r7i5n5.1015419.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:676bcd3d04bb3f23872919238f9f867ea088d15290806dbc3e280b8dfcc712d7
3
+ size 35704840
exp12/000-module.4.mlp/events.out.tfevents.1638184519.r7i5n5.1015419.24 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dff19f49f13dd42406d8583cecc3268c6e9d93e1b5fcc59947ffcc37e0b0e9e
3
+ size 20921800
exp12/000-module.4.post_attention_layernorm/events.out.tfevents.1638184519.r7i5n5.1015419.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e63b9ea30e09331b22440513ef3239c5d47040e1da4fc1df80c792710bc805f
3
+ size 35370760
exp12/000-module.4.self_attention.attention_dropout/events.out.tfevents.1638184519.r7i5n5.1015419.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daa09e9bf7d253b255f5a4c8201ca79140c0a241f945fae3756365b9480123fc
3
+ size 15409480
exp12/000-module.4.self_attention.dense/events.out.tfevents.1638184519.r7i5n5.1015419.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92ceac1e40a22e5be978898f6013035f4913e5dfc605e5a491ba929674538885
3
+ size 35704840
exp12/000-module.4.self_attention.query_key_value/events.out.tfevents.1638184519.r7i5n5.1015419.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:977f1d72cfc5a6077bfb5ac9b0d2b6aa6b6301fdff26c488659afa33886e901c
3
+ size 33032200
exp12/000-module.4.self_attention.scale_mask_softmax/events.out.tfevents.1638184519.r7i5n5.1015419.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:389c3d4c8c8b37aacd8a456d964239b7f134c0d51b463bf6fff82cd1a837af7b
3
+ size 15409480
exp12/000-module.4.self_attention/events.out.tfevents.1638184519.r7i5n5.1015419.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:351c864aab15e9e1bb8ac8abfc813bca2cd20978fd8da4804953cd19558be5dd
3
+ size 20921800
exp12/000-module.4/events.out.tfevents.1638184519.r7i5n5.1015419.25 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c741c91f0f7ed3a52f0e785f94a811097c9d6b5ce1dfbb5ebcf834cffb18080d
3
+ size 20587720
exp12/000-module.tied_modules.embed.embedding_dropout/events.out.tfevents.1638184517.r7i5n5.1015419.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48b0c8e881c2c7e8ddaf3825279c94c96382048b9d99345ec8c7579c25ae1006
3
+ size 10314760
exp12/000-module.tied_modules.embed.position_embeddings/events.out.tfevents.1638184517.r7i5n5.1015419.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8efd04f4fd9a3f55f51bc3849a163c1bbea8bef89f3ec5cc85d2afe09a5961c9
3
+ size 12736840
exp12/000-module.tied_modules.embed.word_embeddings/events.out.tfevents.1638184517.r7i5n5.1015419.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:580dc5cda6d87818ff1e5f16392552fa85957f6c7623694f2ed7469b2c7e9196
3
+ size 12736840
exp12/000-module.tied_modules.embed/events.out.tfevents.1638184517.r7i5n5.1015419.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4dc975c48ad12bac14a1ca7c95c4e33f5bfccb22d0c4f03c216ffcd792f9a52a
3
+ size 7725640
exp12/000-module/events.out.tfevents.1638184519.r7i5n5.1015419.26 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2157f3c4bbe5dd56adc03e265be20be851f8d9d1bd3134b2695b28e0f087a9ee
3
+ size 7725640
exp12/028-module.17.input_layernorm/events.out.tfevents.1638184529.r8i4n3.750609.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3939e8cce545197a665283c8c3855460885c4c7e533c37c38b5aa14899ef40a2
3
+ size 35370760
exp12/028-module.17.mlp.dense_4h_to_h/events.out.tfevents.1638184531.r8i4n3.750609.8 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15bd895dc7c17c892a9e55717b46110cf1f792f0fc3dd518fff13469642edf5e
3
+ size 35704840
exp12/028-module.17.mlp.dense_h_to_4h/events.out.tfevents.1638184530.r8i4n3.750609.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:310dcb9b68c486e783370d2152b7bda51a878b258093385c076148da46322471
3
+ size 35704840
exp12/028-module.17.mlp/events.out.tfevents.1638184531.r8i4n3.750609.9 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54f925d600d6ae0b5f9e33a1584121366c865648dca29e6ee0e52faf55eea883
3
+ size 20921800
exp12/028-module.17.post_attention_layernorm/events.out.tfevents.1638184530.r8i4n3.750609.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28f6544cfe6c5ffc973a2da3d725ca8f8502ca50e3fb1b531863bd6471194ea0
3
+ size 35370760
exp12/028-module.17.self_attention.attention_dropout/events.out.tfevents.1638184530.r8i4n3.750609.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a62e548acb1fb82afaf50af54ac3b269d03690c60b7ece0ab162b9df58634b0
3
+ size 15409480
exp12/028-module.17.self_attention.dense/events.out.tfevents.1638184530.r8i4n3.750609.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2f00f6e9078eeb7462a458c12028b3186fd61d8203ac32b96543d69a5d7042e
3
+ size 35704840
exp12/028-module.17.self_attention.query_key_value/events.out.tfevents.1638184530.r8i4n3.750609.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df0111f99a76fb5e767577a84564dbf23effa530fa751b935532535416cf5599
3
+ size 33032200
exp12/028-module.17.self_attention.scale_mask_softmax/events.out.tfevents.1638184530.r8i4n3.750609.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d8f7485afe5c697ce8271e1356864ccd8884d54a521d1ec6a2e875824bf05dc
3
+ size 15409480
exp12/028-module.17.self_attention/events.out.tfevents.1638184530.r8i4n3.750609.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b431ffa260ce07c3f75b5dad831f64f9f7fcec01209bc5d27314aaa37d250ec6
3
+ size 20921800
exp12/028-module.17/events.out.tfevents.1638184531.r8i4n3.750609.10 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dd318842c2b0d386009412103c2aa12f4475b74f58048e720ec8dafba1ee9dd
3
+ size 20587720
exp12/028-module.18.input_layernorm/events.out.tfevents.1638184531.r8i4n3.750609.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22d275585edb2382d1f98934099f12c315e01de84c1120270867854b5afa17eb
3
+ size 35370760
exp12/028-module.18.mlp.dense_4h_to_h/events.out.tfevents.1638184531.r8i4n3.750609.19 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:248fd0830e05838239b59d3944d6e28f879b6a28abd35021824403ac62a46be3
3
+ size 35704840
exp12/028-module.18.mlp.dense_h_to_4h/events.out.tfevents.1638184531.r8i4n3.750609.18 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c70c64c261aa4c69a2b1852d0d9dd10be01b42a275a6e3c19fe2561d8804838
3
+ size 35704840
exp12/028-module.18.mlp/events.out.tfevents.1638184531.r8i4n3.750609.20 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd4bd8af2c46f7654d53909b4a4cd09264d6716d23d4b2edc095bb782f4f37fa
3
+ size 20921800
exp12/028-module.18.post_attention_layernorm/events.out.tfevents.1638184531.r8i4n3.750609.17 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c0826bc8b7dd3687feb3a4e215996800aafd3a1d0ebadc02bdfbca235edb201
3
+ size 35370760
exp12/028-module.18.self_attention.attention_dropout/events.out.tfevents.1638184531.r8i4n3.750609.14 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91389e1c74e3da8b5e434bc0d54c8bd32f6f6bbda36397a8203990e4d21934b1
3
+ size 15409480
exp12/028-module.18.self_attention.dense/events.out.tfevents.1638184531.r8i4n3.750609.15 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f9bb1e171934123b0e4523cad309d01769acc66f8a4d817c8c597e9649cbeba
3
+ size 35704840
exp12/028-module.18.self_attention.query_key_value/events.out.tfevents.1638184531.r8i4n3.750609.12 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f127bac02f0088b18e390d2e014cda091be875e42f938afb74aa6859cd33c59b
3
+ size 33032200
exp12/028-module.18.self_attention.scale_mask_softmax/events.out.tfevents.1638184531.r8i4n3.750609.13 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b1543409ea96d2d4350d171942b6941dd58ce3d284cf6159dc017b1130c30b4
3
+ size 15409480
exp12/028-module.18.self_attention/events.out.tfevents.1638184531.r8i4n3.750609.16 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52264d9c62947bf19e04b03d853faa592293e2895a06c6e9f60e1136de04cad9
3
+ size 20921800
exp12/028-module.18/events.out.tfevents.1638184531.r8i4n3.750609.21 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffdb2f31ff4786de7fe941b60ab2a838ce853d262537640aa9bcb1ea2ab51585
3
+ size 20587720
exp12/028-module/events.out.tfevents.1638184531.r8i4n3.750609.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:220ca22e653c6a5cabd410b97b5953cf05c05865a83c8cbd40238e6c1de166d2
3
+ size 10314760