ngxson HF Staff commited on
Commit
208209f
·
verified ·
1 Parent(s): cb0aff4

Upload Gemma3ForConditionalGeneration

Browse files
Files changed (2) hide show
  1. config.json +6 -6
  2. model.safetensors +2 -2
config.json CHANGED
@@ -6,7 +6,7 @@
6
  "eoi_token_index": 256000,
7
  "image_token_index": 262144,
8
  "initializer_range": 0.02,
9
- "mm_tokens_per_image": 49,
10
  "model_type": "gemma3",
11
  "text_config": {
12
  "attention_bias": false,
@@ -16,13 +16,13 @@
16
  "final_logit_softcapping": null,
17
  "head_dim": 256,
18
  "hidden_activation": "gelu_pytorch_tanh",
19
- "hidden_size": 256,
20
  "initializer_range": 0.02,
21
  "intermediate_size": 512,
22
  "max_position_embeddings": 131072,
23
  "model_type": "gemma3_text",
24
  "num_attention_heads": 4,
25
- "num_hidden_layers": 4,
26
  "num_key_value_heads": 2,
27
  "query_pre_attn_scalar": 256,
28
  "rms_norm_eps": 1e-06,
@@ -40,15 +40,15 @@
40
  "vision_config": {
41
  "attention_dropout": 0.1,
42
  "hidden_act": "gelu_pytorch_tanh",
43
- "hidden_size": 256,
44
- "image_size": 224,
45
  "intermediate_size": 512,
46
  "layer_norm_eps": 1e-06,
47
  "model_type": "siglip_vision_model",
48
  "num_attention_heads": 4,
49
  "num_channels": 3,
50
  "num_hidden_layers": 4,
51
- "patch_size": 16,
52
  "torch_dtype": "float32"
53
  }
54
  }
 
6
  "eoi_token_index": 256000,
7
  "image_token_index": 262144,
8
  "initializer_range": 0.02,
9
+ "mm_tokens_per_image": 64,
10
  "model_type": "gemma3",
11
  "text_config": {
12
  "attention_bias": false,
 
16
  "final_logit_softcapping": null,
17
  "head_dim": 256,
18
  "hidden_activation": "gelu_pytorch_tanh",
19
+ "hidden_size": 128,
20
  "initializer_range": 0.02,
21
  "intermediate_size": 512,
22
  "max_position_embeddings": 131072,
23
  "model_type": "gemma3_text",
24
  "num_attention_heads": 4,
25
+ "num_hidden_layers": 8,
26
  "num_key_value_heads": 2,
27
  "query_pre_attn_scalar": 256,
28
  "rms_norm_eps": 1e-06,
 
40
  "vision_config": {
41
  "attention_dropout": 0.1,
42
  "hidden_act": "gelu_pytorch_tanh",
43
+ "hidden_size": 128,
44
+ "image_size": 32,
45
  "intermediate_size": 512,
46
  "layer_norm_eps": 1e-06,
47
  "model_type": "siglip_vision_model",
48
  "num_attention_heads": 4,
49
  "num_channels": 3,
50
  "num_hidden_layers": 4,
51
+ "patch_size": 2,
52
  "torch_dtype": "float32"
53
  }
54
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3e80680dee6c39af37585a46d337d682fa307a72b2d818483b845facabb59ce
3
- size 299212928
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4c464ac73780fbf2e223270fcf11fa2903e847c1f50d1e9c79688f166a80e67
3
+ size 157351984