Initial Commit
Browse files- config.json +42 -0
- eval_results.txt +20 -0
- pytorch_model.bin +3 -0
- special_tokens_map.json +1 -0
- tokenizer_config.json +1 -0
- training_args.bin +3 -0
- vocab.txt +0 -0
config.json
ADDED
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "dmis-lab/biobert-base-cased-v1.1",
|
3 |
+
"architectures": [
|
4 |
+
"BertForTokenClassification"
|
5 |
+
],
|
6 |
+
"attention_probs_dropout_prob": 0.1,
|
7 |
+
"gradient_checkpointing": false,
|
8 |
+
"hidden_act": "gelu",
|
9 |
+
"hidden_dropout_prob": 0.1,
|
10 |
+
"hidden_size": 768,
|
11 |
+
"id2label": {
|
12 |
+
"0": "B-Anatomy",
|
13 |
+
"1": "B-Chemical",
|
14 |
+
"2": "B-Disease",
|
15 |
+
"3": "I-Anatomy",
|
16 |
+
"4": "I-Chemical",
|
17 |
+
"5": "I-Disease",
|
18 |
+
"6": "O"
|
19 |
+
},
|
20 |
+
"initializer_range": 0.02,
|
21 |
+
"intermediate_size": 3072,
|
22 |
+
"label2id": {
|
23 |
+
"B-Anatomy": 0,
|
24 |
+
"B-Chemical": 1,
|
25 |
+
"B-Disease": 2,
|
26 |
+
"I-Anatomy": 3,
|
27 |
+
"I-Chemical": 4,
|
28 |
+
"I-Disease": 5,
|
29 |
+
"O": 6
|
30 |
+
},
|
31 |
+
"layer_norm_eps": 1e-12,
|
32 |
+
"max_position_embeddings": 512,
|
33 |
+
"model_type": "bert",
|
34 |
+
"num_attention_heads": 12,
|
35 |
+
"num_hidden_layers": 12,
|
36 |
+
"pad_token_id": 0,
|
37 |
+
"position_embedding_type": "absolute",
|
38 |
+
"transformers_version": "4.8.2",
|
39 |
+
"type_vocab_size": 2,
|
40 |
+
"use_cache": true,
|
41 |
+
"vocab_size": 28996
|
42 |
+
}
|
eval_results.txt
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
eval_loss = 0.1652524173259735
|
2 |
+
eval_precision = 0.8504558443325623
|
3 |
+
eval_recall = 0.8611187655001378
|
4 |
+
eval_f1 = 0.8557540905045526
|
5 |
+
eval_accuracy = 0.975697565751413
|
6 |
+
eval_matrix = [[3901, 4, 60, 102, 1, 10, 538], [0, 4949, 15, 0, 42, 1, 371], [56, 18, 3770, 1, 0, 207, 372], [101, 0, 2, 2779, 0, 26, 335], [0, 32, 0, 0, 1383, 0, 213], [6, 4, 113, 28, 4, 2247, 335], [354, 467, 574, 247, 369, 453, 200220]]
|
7 |
+
eval_classification_report = precision recall f1-score support
|
8 |
+
|
9 |
+
Anatomy 0.85 0.83 0.84 4649
|
10 |
+
Chemical 0.89 0.91 0.90 5402
|
11 |
+
Disease 0.80 0.84 0.82 4465
|
12 |
+
|
13 |
+
micro avg 0.85 0.86 0.86 14516
|
14 |
+
macro avg 0.85 0.86 0.85 14516
|
15 |
+
weighted avg 0.85 0.86 0.86 14516
|
16 |
+
|
17 |
+
eval_runtime = 16.1004
|
18 |
+
eval_samples_per_second = 152.977
|
19 |
+
eval_steps_per_second = 0.621
|
20 |
+
epoch = 35.0
|
pytorch_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:89ff1d03f6a557db79ca98a49d86d41def96fa8d586d91c968f19e6e8c8e0ecc
|
3 |
+
size 430983665
|
special_tokens_map.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80b5808a16224f2716b8ece4c7a86a0ef8a4b06c4cf6511871afb0dc3fb549ac
|
3 |
+
size 2607
|
vocab.txt
ADDED
The diff for this file is too large to render.
See raw diff
|
|