nev commited on
Commit
9460488
·
verified ·
1 Parent(s): fbb9a58

Add files using upload-large-folder tool

Browse files
Files changed (50) hide show
  1. layers.0.mlp/cfg.json +1 -0
  2. layers.0.mlp/sae.safetensors +3 -0
  3. layers.1.mlp/cfg.json +1 -0
  4. layers.1.mlp/sae.safetensors +3 -0
  5. layers.10.mlp/cfg.json +1 -0
  6. layers.10.mlp/sae.safetensors +3 -0
  7. layers.11.mlp/cfg.json +1 -0
  8. layers.11.mlp/sae.safetensors +3 -0
  9. layers.12.mlp/cfg.json +1 -0
  10. layers.12.mlp/sae.safetensors +3 -0
  11. layers.13.mlp/cfg.json +1 -0
  12. layers.13.mlp/sae.safetensors +3 -0
  13. layers.14.mlp/cfg.json +1 -0
  14. layers.14.mlp/sae.safetensors +3 -0
  15. layers.15.mlp/cfg.json +1 -0
  16. layers.15.mlp/sae.safetensors +3 -0
  17. layers.16.mlp/cfg.json +1 -0
  18. layers.16.mlp/sae.safetensors +3 -0
  19. layers.17.mlp/cfg.json +1 -0
  20. layers.17.mlp/sae.safetensors +3 -0
  21. layers.18.mlp/cfg.json +1 -0
  22. layers.18.mlp/sae.safetensors +3 -0
  23. layers.19.mlp/cfg.json +1 -0
  24. layers.19.mlp/sae.safetensors +3 -0
  25. layers.2.mlp/cfg.json +1 -0
  26. layers.2.mlp/sae.safetensors +3 -0
  27. layers.20.mlp/cfg.json +1 -0
  28. layers.20.mlp/sae.safetensors +3 -0
  29. layers.21.mlp/cfg.json +1 -0
  30. layers.21.mlp/sae.safetensors +3 -0
  31. layers.22.mlp/cfg.json +1 -0
  32. layers.22.mlp/sae.safetensors +3 -0
  33. layers.23.mlp/cfg.json +1 -0
  34. layers.23.mlp/sae.safetensors +3 -0
  35. layers.24.mlp/cfg.json +1 -0
  36. layers.24.mlp/sae.safetensors +3 -0
  37. layers.25.mlp/cfg.json +1 -0
  38. layers.25.mlp/sae.safetensors +3 -0
  39. layers.3.mlp/cfg.json +1 -0
  40. layers.3.mlp/sae.safetensors +3 -0
  41. layers.4.mlp/cfg.json +1 -0
  42. layers.4.mlp/sae.safetensors +3 -0
  43. layers.5.mlp/cfg.json +1 -0
  44. layers.6.mlp/cfg.json +1 -0
  45. layers.7.mlp/cfg.json +1 -0
  46. layers.7.mlp/sae.safetensors +3 -0
  47. layers.8.mlp/cfg.json +1 -0
  48. layers.8.mlp/sae.safetensors +3 -0
  49. layers.9.mlp/cfg.json +1 -0
  50. layers.9.mlp/sae.safetensors +3 -0
layers.0.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.0.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:deb9c9f23b888ae34b69c541d975fed1a5cb0650be11be9f2db0f7526e4aaa66
3
+ size 302130592
layers.1.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.1.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3fc8aab4b54383a5f9287be48d1aa63e642ae32f71df0050b30f487002abe26
3
+ size 302130592
layers.10.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.10.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66744f201f638344c4057d7235ed79f2e8ef5df075c74b43fa9f27f80204c294
3
+ size 302130592
layers.11.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.11.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1e27a8ebe9af484e38d57340a0fc664e2ff1b8b8f8e8c773ddbb83225491d2f
3
+ size 302130592
layers.12.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.12.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:063d895f7a720a3ea7d2e7893b02fa9a52abb54b8abb31c5b57402d909dde1d3
3
+ size 302130592
layers.13.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.13.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd2be42d675694960f3a6a82c24743d0767be848f6891295fd0b840a2c06fece
3
+ size 302130592
layers.14.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.14.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee4483de038614cceec8b37359fcbbc9dc675040335d6060f88f240468684f3e
3
+ size 302130592
layers.15.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.15.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b07a87ad38e1c8a201a7b8b1852d50af0fb600a437d8e1884092301c6ad66223
3
+ size 302130592
layers.16.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.16.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5230cfd2426b378e1712cf8ea1ec7f7aff66df476f0d4967be6ec55aa716387e
3
+ size 302130592
layers.17.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.17.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3c3b4c181dc2ffe4f8d9b91aebd021a7ae186a1a855caba9f2621ea3017765b5
3
+ size 302130592
layers.18.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.18.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:561fcb8e7e9b503b5403c115da03ca68dba084548778c4fde7fa28bd95abc0ab
3
+ size 302130592
layers.19.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.19.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6731e96560c0aa4d022209d1499c31a044dc86b009a3439737e1eba5a94d239
3
+ size 302130592
layers.2.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.2.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d1d6ebd269bdeb2fb4e65c60f762ae57a6c5e294a4e5649f81956c59fd7dec9
3
+ size 302130592
layers.20.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.20.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3853eed74bdfc9ed29ee4d6a204945844b5fb0ab80fef11a8741837c0bc541c1
3
+ size 302130592
layers.21.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.21.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a54d2b9a072578a40b936dff99ba4ac62c32ca04dd72f01d7f958364854c32bc
3
+ size 302130592
layers.22.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.22.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c44be5685436d4b1eb7c74441ca00374e2654ea7334b7438c309f94f26f1b59b
3
+ size 302130592
layers.23.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.23.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:321adadf14818d67933cf9d61cf3cdc23a602be61dc629d33d083b8c3c201a0c
3
+ size 302130592
layers.24.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.24.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1997f8f2e9a1e25063f3b52212bbddb6ffcd0c2e1af1e8a4cde623a929a0701
3
+ size 302130592
layers.25.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.25.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a560effba777eb0ad6adcf8f05af001d341f4ad8e21ab71248146cbf86e9a4c3
3
+ size 302130592
layers.3.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.3.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a93f3e301bb178499aa28025ed99c40e631c879ab38dfcd6dbc2baff24e2bf9
3
+ size 302130592
layers.4.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.4.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6410be1d6a58b68069e216a52f0bb131e742ece0f4ae9cfa3e9c7bb67107d6ee
3
+ size 302130592
layers.5.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.6.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.7.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.7.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a18b0e04f9e46be141a1a3fa7ad6bc84e6e1ff8d6a12c703292d43c99c4b960
3
+ size 302130592
layers.8.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.8.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c9f72ced268a011aa6cf5e5c3d0976f2597b1d31ebe88207de1ad4ee0cbff8
3
+ size 302130592
layers.9.mlp/cfg.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"activation": "topk", "expansion_factor": 32, "normalize_decoder": true, "num_latents": 16384, "k": 256, "multi_topk": false, "skip_connection": false, "transcode": true, "n_targets": 0, "n_sources": 0, "normalize_io": false, "divide_cross_layer": false, "train_post_encoder": true, "post_encoder_scale": false, "coalesce_topk": "none", "topk_coalesced": false, "d_in": 2304}
layers.9.mlp/sae.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7a97a3c603d8d5cc1d029e83ee50ac35da32fd1cefe49075eed7dcb35dd9142
3
+ size 302130592