Orion-zhen commited on
Commit
75d36f4
·
verified ·
1 Parent(s): c4ab8a7

Upload 2 files

Browse files
.gitattributes CHANGED
@@ -39,3 +39,4 @@ models/Phi-4-multimodal/tokenizer.json filter=lfs diff=lfs merge=lfs -text
39
  models/GLM-4-0414/tokenizer.json filter=lfs diff=lfs merge=lfs -text
40
  models/Mistral-Small-3.1-2503/tokenizer.json filter=lfs diff=lfs merge=lfs -text
41
  models/Command-A-03-2025/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
39
  models/GLM-4-0414/tokenizer.json filter=lfs diff=lfs merge=lfs -text
40
  models/Mistral-Small-3.1-2503/tokenizer.json filter=lfs diff=lfs merge=lfs -text
41
  models/Command-A-03-2025/tokenizer.json filter=lfs diff=lfs merge=lfs -text
42
+ models/Aya-expanse/tokenizer.json filter=lfs diff=lfs merge=lfs -text
models/Aya-expanse/config.json CHANGED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "CohereForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 5,
8
+ "eos_token_id": 255001,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "layer_norm_eps": 1e-05,
14
+ "logit_scale": 0.125,
15
+ "max_position_embeddings": 8192,
16
+ "model_type": "cohere",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 32,
19
+ "num_key_value_heads": 8,
20
+ "pad_token_id": 0,
21
+ "rope_theta": 10000,
22
+ "torch_dtype": "float16",
23
+ "transformers_version": "4.44.0",
24
+ "use_cache": true,
25
+ "use_qk_norm": false,
26
+ "vocab_size": 256000
27
+ }
models/Aya-expanse/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c69a7ea6c0927dfac8c349186ebcf0466a4723c21cbdb2e850cf559f0bee92b8
3
+ size 12777433