thesis release
Browse files- config.json +3 -2
- eval_results.txt +16 -16
- pytorch_model.bin +2 -2
- special_tokens_map.json +0 -0
- test_results.txt +20 -0
- tokenizer_config.json +1 -1
- training_args.bin +2 -2
- vocab.txt +0 -0
config.json
CHANGED
@@ -4,7 +4,7 @@
|
|
4 |
"BertForTokenClassification"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.1,
|
7 |
-
"
|
8 |
"hidden_act": "gelu",
|
9 |
"hidden_dropout_prob": 0.1,
|
10 |
"hidden_size": 768,
|
@@ -35,7 +35,8 @@
|
|
35 |
"num_hidden_layers": 12,
|
36 |
"pad_token_id": 0,
|
37 |
"position_embedding_type": "absolute",
|
38 |
-
"
|
|
|
39 |
"type_vocab_size": 2,
|
40 |
"use_cache": true,
|
41 |
"vocab_size": 28996
|
|
|
4 |
"BertForTokenClassification"
|
5 |
],
|
6 |
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"classifier_dropout": null,
|
8 |
"hidden_act": "gelu",
|
9 |
"hidden_dropout_prob": 0.1,
|
10 |
"hidden_size": 768,
|
|
|
35 |
"num_hidden_layers": 12,
|
36 |
"pad_token_id": 0,
|
37 |
"position_embedding_type": "absolute",
|
38 |
+
"torch_dtype": "float32",
|
39 |
+
"transformers_version": "4.17.0",
|
40 |
"type_vocab_size": 2,
|
41 |
"use_cache": true,
|
42 |
"vocab_size": 28996
|
eval_results.txt
CHANGED
@@ -1,20 +1,20 @@
|
|
1 |
-
eval_loss = 0.
|
2 |
-
eval_precision = 0.
|
3 |
-
eval_recall = 0.
|
4 |
-
eval_f1 = 0.
|
5 |
-
eval_accuracy = 0.
|
6 |
-
eval_matrix = [[
|
7 |
eval_classification_report = precision recall f1-score support
|
8 |
|
9 |
-
Anatomy 0.
|
10 |
-
Chemical 0.
|
11 |
-
Disease 0.
|
12 |
|
13 |
-
micro avg 0.85 0.
|
14 |
-
macro avg 0.
|
15 |
-
weighted avg 0.85 0.
|
16 |
|
17 |
-
eval_runtime =
|
18 |
-
eval_samples_per_second =
|
19 |
-
eval_steps_per_second =
|
20 |
-
epoch =
|
|
|
1 |
+
eval_loss = 0.17402765154838562
|
2 |
+
eval_precision = 0.8519104303671023
|
3 |
+
eval_recall = 0.8661870503597122
|
4 |
+
eval_f1 = 0.8589894242068155
|
5 |
+
eval_accuracy = 0.9750561746748678
|
6 |
+
eval_matrix = [[1841, 0, 31, 44, 0, 4, 219], [0, 4941, 10, 0, 36, 0, 360], [71, 17, 3565, 1, 0, 172, 420], [64, 0, 1, 1221, 0, 28, 102], [0, 25, 0, 0, 1476, 2, 245], [2, 3, 119, 35, 6, 2095, 463], [360, 282, 417, 227, 226, 404, 156701]]
|
7 |
eval_classification_report = precision recall f1-score support
|
8 |
|
9 |
+
Anatomy 0.76 0.84 0.80 2162
|
10 |
+
Chemical 0.92 0.92 0.92 5371
|
11 |
+
Disease 0.82 0.82 0.82 4282
|
12 |
|
13 |
+
micro avg 0.85 0.87 0.86 11815
|
14 |
+
macro avg 0.83 0.86 0.84 11815
|
15 |
+
weighted avg 0.85 0.87 0.86 11815
|
16 |
|
17 |
+
eval_runtime = 54.806
|
18 |
+
eval_samples_per_second = 36.273
|
19 |
+
eval_steps_per_second = 9.068
|
20 |
+
epoch = 15.0
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c0de00884e3be0723c9ae9e5d6880db22290f31d8eaddd7c82fedd8d93daa90
|
3 |
+
size 430984561
|
special_tokens_map.json
CHANGED
File without changes
|
test_results.txt
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
eval_loss = 0.17808516323566437
|
2 |
+
eval_precision = 0.8430268069222938
|
3 |
+
eval_recall = 0.855745384403417
|
4 |
+
eval_f1 = 0.8493384841543878
|
5 |
+
eval_accuracy = 0.975430554937475
|
6 |
+
eval_matrix = [[4007, 4, 58, 78, 0, 9, 460], [0, 4872, 9, 0, 41, 0, 456], [62, 14, 3706, 3, 0, 187, 452], [107, 0, 3, 2829, 0, 29, 275], [0, 23, 0, 0, 1398, 0, 207], [9, 3, 144, 28, 3, 2151, 399], [397, 414, 581, 253, 356, 457, 200226]]
|
7 |
+
eval_classification_report = precision recall f1-score support
|
8 |
+
|
9 |
+
Anatomy 0.84 0.85 0.84 4649
|
10 |
+
Chemical 0.90 0.89 0.89 5402
|
11 |
+
Disease 0.79 0.82 0.80 4465
|
12 |
+
|
13 |
+
micro avg 0.84 0.86 0.85 14516
|
14 |
+
macro avg 0.84 0.85 0.85 14516
|
15 |
+
weighted avg 0.84 0.86 0.85 14516
|
16 |
+
|
17 |
+
eval_runtime = 69.3988
|
18 |
+
eval_samples_per_second = 35.491
|
19 |
+
eval_steps_per_second = 8.876
|
20 |
+
epoch = 15.0
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
|
|
1 |
+
{"do_lower_case": false, "do_basic_tokenize": true, "never_split": null, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "special_tokens_map_file": null, "name_or_path": "dmis-lab/biobert-base-cased-v1.1", "tokenizer_class": "BertTokenizer"}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:289790c3ee17f28436f1993d68436e419d5800ae5288688b3ce9ca02d7b8f557
|
3 |
+
size 3055
|
vocab.txt
CHANGED
File without changes
|