qgallouedec HF Staff commited on
Commit
4cb5a10
·
verified ·
1 Parent(s): 12de51b

Upload Gemma3ForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +4 -3
  2. model.safetensors +2 -2
config.json CHANGED
@@ -61,9 +61,9 @@
61
  ],
62
  "max_position_embeddings": 131072,
63
  "model_type": "gemma3_text",
64
- "num_attention_heads": 8,
65
  "num_hidden_layers": 2,
66
- "num_key_value_heads": 4,
67
  "query_pre_attn_scalar": 256,
68
  "rms_norm_eps": 1e-06,
69
  "rope_local_base_freq": 10000.0,
@@ -86,9 +86,10 @@
86
  "intermediate_size": 4304,
87
  "layer_norm_eps": 1e-06,
88
  "model_type": "siglip_vision_model",
89
- "num_attention_heads": 16,
90
  "num_channels": 3,
91
  "num_hidden_layers": 2,
 
92
  "patch_size": 14,
93
  "vision_use_head": false
94
  }
 
61
  ],
62
  "max_position_embeddings": 131072,
63
  "model_type": "gemma3_text",
64
+ "num_attention_heads": 4,
65
  "num_hidden_layers": 2,
66
+ "num_key_value_heads": 2,
67
  "query_pre_attn_scalar": 256,
68
  "rms_norm_eps": 1e-06,
69
  "rope_local_base_freq": 10000.0,
 
86
  "intermediate_size": 4304,
87
  "layer_norm_eps": 1e-06,
88
  "model_type": "siglip_vision_model",
89
+ "num_attention_heads": 4,
90
  "num_channels": 3,
91
  "num_hidden_layers": 2,
92
+ "num_key_value_heads": 2,
93
  "patch_size": 14,
94
  "vision_use_head": false
95
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:db80e8136185f45ba7732fd1f88b0fc7446ff8169fafdded5cdf9f57a7303b70
3
- size 22959728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:833414b03aca14b677c4b053c6619539f8d54e054ea52c362d5057ba75efeb6d
3
+ size 22566504