{ | |
"MODEL": { | |
"num_hidden_layers": 24, | |
"hidden_size": 2048, | |
"num_attention_heads": 32, | |
"intermediate_size": 4096 | |
"vocab_size": 50304 | |
}, | |
"LAYER_MAPPINGS" : { | |
"word_embeddings": 1, | |
"transformer": [3, 26], | |
"final_layernorm": 28, | |
"final_word_embeddings": 29 | |
}, | |
"FULL_NAME_MAPPINGS": { | |
}, | |
"PARTIAL_NAME_MAPPINGS": { | |
"final_word_embeddings": { | |
"vocab_parallel_projection": "lm_head" | |
}, | |
"final_layernorm": { | |
"final_rmsnorm": "model.norm" | |
}, | |
"word_embeddings": { | |
"word_embeddings": "model.embed_tokens" | |
}, | |
"transformer": { | |
"dense_h_to_4h": "mlp.gate_proj", | |
"dense_4h_to_h": "mlp.down_proj", | |
"dense_h_to_4h_swiglu": "mlp.up_proj", | |
"post_attention_layernorm": "post_attention_layernorm", | |
"input_layernorm": "input_layernorm", | |
"dense": "self_attn.o_proj", | |
"query_key_value": {"query": "self_attn.q_proj", "key": "self_attn.k_proj", "value": "self_attn.v_proj"} | |
} | |
}, | |
"SPECIAL": { | |
"query_key_value": "attention_qkv" | |
} | |
} | |