Update tiny models for T5ForQuestionAnswering

#21
Files changed (3) hide show
  1. config.json +1 -1
  2. pytorch_model.bin +1 -1
  3. tokenizer_config.json +1 -0
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "relative_attention_max_distance": 128,
24
  "relative_attention_num_buckets": 8,
25
  "torch_dtype": "float32",
26
- "transformers_version": "4.31.0.dev0",
27
  "use_cache": true,
28
  "vocab_size": 32100
29
  }
 
23
  "relative_attention_max_distance": 128,
24
  "relative_attention_num_buckets": 8,
25
  "torch_dtype": "float32",
26
+ "transformers_version": "4.32.0.dev0",
27
  "use_cache": true,
28
  "vocab_size": 32100
29
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2c65bb57b8b822ed8cd7f31dd2ada0b8790258644b5c2a97f3b69cc930330a4e
3
  size 4489613
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62d9182e8f57cdbf387ee03bc9ea1f8bb5c3c426f44c9ee3ef7616d167a3e19d
3
  size 4489613
tokenizer_config.json CHANGED
@@ -104,6 +104,7 @@
104
  "clean_up_tokenization_spaces": true,
105
  "eos_token": "</s>",
106
  "extra_ids": 100,
 
107
  "model_max_length": 512,
108
  "pad_token": "<pad>",
109
  "sp_model_kwargs": {},
 
104
  "clean_up_tokenization_spaces": true,
105
  "eos_token": "</s>",
106
  "extra_ids": 100,
107
+ "legacy": true,
108
  "model_max_length": 512,
109
  "pad_token": "<pad>",
110
  "sp_model_kwargs": {},