sgugger commited on
Commit
0280e4b
1 Parent(s): 8c6ac70

Upload tiny models for XLNetForQuestionAnsweringSimple

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "temp/dummy/xlnet/XLNetForQuestionAnsweringSimple",
3
  "architectures": [
4
  "XLNetForQuestionAnsweringSimple"
5
  ],
@@ -29,9 +29,9 @@
29
  "summary_type": "last",
30
  "summary_use_proj": true,
31
  "torch_dtype": "float32",
32
- "transformers_version": "4.25.0.dev0",
33
  "untie_r": true,
34
  "use_mems_eval": true,
35
  "use_mems_train": false,
36
- "vocab_size": 1024
37
  }
 
1
  {
2
+ "_name_or_path": "tiny_models/xlnet/XLNetForQuestionAnsweringSimple",
3
  "architectures": [
4
  "XLNetForQuestionAnsweringSimple"
5
  ],
 
29
  "summary_type": "last",
30
  "summary_use_proj": true,
31
  "torch_dtype": "float32",
32
+ "transformers_version": "4.28.0.dev0",
33
  "untie_r": true,
34
  "use_mems_eval": true,
35
  "use_mems_train": false,
36
+ "vocab_size": 32000
37
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5895a86faf39c49e59cb5577bbc05995614d235efc1854c74a6e97efbf3a1809
3
- size 434492
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:244b1175778e3cad744280ca0e79c5bdbedb2547da19a6d320fa7ccc23f82b1a
3
+ size 4399420
special_tokens_map.json CHANGED
@@ -9,7 +9,7 @@
9
  "mask_token": {
10
  "content": "<mask>",
11
  "lstrip": true,
12
- "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
9
  "mask_token": {
10
  "content": "<mask>",
11
  "lstrip": true,
12
+ "normalized": true,
13
  "rstrip": false,
14
  "single_word": false
15
  },
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a2e6ccaeed20cb6dcb836f49d8ef8965f407e48f2334cf7db6c52f08b5df136
3
- size 494176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:942758f334eba9a02b67363ba7af1a45a6c196e5b349a6387285a8876254df47
3
+ size 4459104
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -12,12 +12,11 @@
12
  "__type": "AddedToken",
13
  "content": "<mask>",
14
  "lstrip": true,
15
- "normalized": false,
16
  "rstrip": false,
17
  "single_word": false
18
  },
19
  "model_max_length": 1000000000000000019884624838656,
20
- "name_or_path": "temp/dummy/xlnet/processors",
21
  "pad_token": "<pad>",
22
  "remove_space": true,
23
  "sep_token": "<sep>",
 
12
  "__type": "AddedToken",
13
  "content": "<mask>",
14
  "lstrip": true,
15
+ "normalized": true,
16
  "rstrip": false,
17
  "single_word": false
18
  },
19
  "model_max_length": 1000000000000000019884624838656,
 
20
  "pad_token": "<pad>",
21
  "remove_space": true,
22
  "sep_token": "<sep>",