| { |
| "MODEL": { |
| "num_hidden_layers": 24, |
| "hidden_size": 2048, |
| "num_attention_heads": 32, |
| "intermediate_size": 4096 |
| }, |
| "LAYER_MAPPINGS" : { |
| "word_embeddings": 1, |
| "transformer": [3, 24+2], |
| "final_layernorm": 28, |
| "final_word_embeddings": 29 |
| }, |
| "FULL_NAME_MAPPINGS": { |
| }, |
| "PARTIAL_NAME_MAPPINGS": { |
| "final_word_embeddings": { |
| "vocab_parallel_projection": "lm_head" |
| }, |
| "final_layernorm": { |
| "final_rmsnorm": "model.norm" |
| }, |
| "word_embeddings": { |
| "word_embeddings": "model.embed_tokens" |
| }, |
| "transformer": { |
| "dense_h_to_4h": "mlp.gate_proj", |
| "dense_4h_to_h": "mlp.down_proj", |
| "dense_h_to_4h_swiglu": "mlp.up_proj", |
| "post_attention_layernorm": "post_attention_layernorm", |
| "input_layernorm": "input_layernorm", |
| "dense": "self_attn.o_proj", |
| "query_key_value": {"query": "self_attn.q_proj", "key": "self_attn.k_proj", "value": "self_attn.v_proj"} |
| } |
| }, |
| "SPECIAL": { |
| "query_key_value": "attention_qkv" |
| } |
| } |
|
|