ydshieh HF Staff commited on
Commit
7c37bfb
·
1 Parent(s): b6133aa

Upload tiny models for TapasForSequenceClassification

Browse files
config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "temp/dummy/tapas/TapasForSequenceClassification",
3
+ "agg_temperature": 1.0,
4
+ "aggregation_labels": null,
5
+ "aggregation_loss_importance": 0.8,
6
+ "aggregation_loss_weight": 1.0,
7
+ "aggregation_temperature": 1.0,
8
+ "allow_empty_column_selection": false,
9
+ "answer_loss_cutoff": 100,
10
+ "answer_loss_importance": 0.001,
11
+ "architectures": [
12
+ "TapasForSequenceClassification"
13
+ ],
14
+ "attention_probs_dropout_prob": 0.1,
15
+ "average_approximation_function": "ratio",
16
+ "average_logits_per_cell": true,
17
+ "cell_selection_preference": 0.5,
18
+ "disable_per_token_loss": false,
19
+ "hidden_act": "gelu",
20
+ "hidden_dropout_prob": 0.1,
21
+ "hidden_size": 32,
22
+ "huber_loss_delta": 25.0,
23
+ "init_cell_selection_weights_to_zero": true,
24
+ "initializer_range": 0.02,
25
+ "intermediate_size": 37,
26
+ "layer_norm_eps": 1e-12,
27
+ "max_num_columns": 32,
28
+ "max_num_rows": 64,
29
+ "max_position_embeddings": 512,
30
+ "model_type": "tapas",
31
+ "no_aggregation_label_index": null,
32
+ "num_aggregation_labels": 4,
33
+ "num_attention_heads": 4,
34
+ "num_hidden_layers": 5,
35
+ "pad_token_id": 0,
36
+ "positive_label_weight": 10.0,
37
+ "positive_weight": 10.0,
38
+ "reset_position_index_per_cell": true,
39
+ "select_one_column": true,
40
+ "torch_dtype": "float32",
41
+ "transformers_version": "4.25.0.dev0",
42
+ "type_vocab_sizes": [
43
+ 3,
44
+ 256,
45
+ 256,
46
+ 2,
47
+ 256,
48
+ 256,
49
+ 10
50
+ ],
51
+ "use_answer_as_supervision": true,
52
+ "use_gumbel_for_agg": false,
53
+ "use_gumbel_for_aggregation": false,
54
+ "use_gumbel_for_cells": false,
55
+ "use_normalized_answer_loss": false,
56
+ "vocab_size": 30522
57
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:758fd91591b6ba7e434db1529b9e239472c652c47c31d5359b7c3373b910ab06
3
+ size 4276238
special_tokens_map.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[EMPTY]"
4
+ ],
5
+ "cls_token": "[CLS]",
6
+ "mask_token": "[MASK]",
7
+ "pad_token": "[PAD]",
8
+ "sep_token": "[SEP]",
9
+ "unk_token": "[UNK]"
10
+ }
tf_model.h5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bb68894fcd4008832a8649975e9ebf125178d60d5e27e77d8ce18556a4da694
3
+ size 4383664
tokenizer_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "[EMPTY]"
4
+ ],
5
+ "cell_trim_length": -1,
6
+ "cls_token": "[CLS]",
7
+ "do_basic_tokenize": true,
8
+ "do_lower_case": true,
9
+ "drop_rows_to_fit": false,
10
+ "empty_token": "[EMPTY]",
11
+ "mask_token": "[MASK]",
12
+ "max_column_id": null,
13
+ "max_question_length": null,
14
+ "max_row_id": null,
15
+ "min_question_length": null,
16
+ "model_max_length": 512,
17
+ "name_or_path": "google/tapas-base-finetuned-sqa",
18
+ "never_split": null,
19
+ "pad_token": "[PAD]",
20
+ "sep_token": "[SEP]",
21
+ "special_tokens_map_file": "/home/huggingface/.cache/huggingface/hub/models--google--tapas-base-finetuned-sqa/snapshots/81916d20eef75766aeae71b9487fd615017b0413/special_tokens_map.json",
22
+ "strip_accents": null,
23
+ "strip_column_names": false,
24
+ "tokenize_chinese_chars": true,
25
+ "tokenizer_class": "TapasTokenizer",
26
+ "unk_token": "[UNK]",
27
+ "update_answer_coordinates": false
28
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff