sgugger commited on
Commit
444f3b9
·
1 Parent(s): f6c0d72

Upload tiny models for AlbertForTokenClassification

Browse files
Files changed (6) hide show
  1. config.json +3 -3
  2. pytorch_model.bin +2 -2
  3. spiece.model +3 -0
  4. tf_model.h5 +2 -2
  5. tokenizer.json +0 -0
  6. tokenizer_config.json +1 -1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "temp/dummy/albert/AlbertForTokenClassification",
3
  "architectures": [
4
  "AlbertForTokenClassification"
5
  ],
@@ -23,7 +23,7 @@
23
  "pad_token_id": 0,
24
  "position_embedding_type": "absolute",
25
  "torch_dtype": "float32",
26
- "transformers_version": "4.25.0.dev0",
27
  "type_vocab_size": 16,
28
- "vocab_size": 1024
29
  }
 
1
  {
2
+ "_name_or_path": "tiny_models/albert/AlbertForTokenClassification",
3
  "architectures": [
4
  "AlbertForTokenClassification"
5
  ],
 
23
  "pad_token_id": 0,
24
  "position_embedding_type": "absolute",
25
  "torch_dtype": "float32",
26
+ "transformers_version": "4.28.0.dev0",
27
  "type_vocab_size": 16,
28
+ "vocab_size": 30000
29
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f2f2eb700f844cdac48c45208573364403765129caf6cef3a78bdbd00bdb7d9
3
- size 1050997
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89303599a83bf1205e75391e010904b193d3b8f924b1e376f5da848bd2429e84
3
+ size 15886709
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fefb02b667a6c5c2fe27602d28e5fb3428f66ab89c7d6f388e7c8d44a02d0336
3
+ size 760289
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:519942512fe2e697f93edf0a26b5ec64d0f9f4bd9027774f091962738fdcbdf2
3
- size 1148280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5050e6c5c2ad5abd64511bf85372273b451a2119ed7e96243ffaf6f05400b993
3
+ size 15983992
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -13,10 +13,10 @@
13
  "single_word": false
14
  },
15
  "model_max_length": 512,
16
- "name_or_path": "albert-xxlarge-v2",
17
  "pad_token": "<pad>",
18
  "remove_space": true,
19
  "sep_token": "[SEP]",
 
20
  "special_tokens_map_file": null,
21
  "tokenizer_class": "AlbertTokenizer",
22
  "unk_token": "<unk>"
 
13
  "single_word": false
14
  },
15
  "model_max_length": 512,
 
16
  "pad_token": "<pad>",
17
  "remove_space": true,
18
  "sep_token": "[SEP]",
19
+ "sp_model_kwargs": {},
20
  "special_tokens_map_file": null,
21
  "tokenizer_class": "AlbertTokenizer",
22
  "unk_token": "<unk>"