fagner commited on
Commit
2ff8852
·
1 Parent(s): 7016818

Initial Commit

Browse files
config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "dmis-lab/biobert-base-cased-v1.1",
3
+ "architectures": [
4
+ "BertForTokenClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "gradient_checkpointing": false,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "B-Anatomy",
13
+ "1": "B-Chemical",
14
+ "2": "B-Disease",
15
+ "3": "I-Anatomy",
16
+ "4": "I-Chemical",
17
+ "5": "I-Disease",
18
+ "6": "O"
19
+ },
20
+ "initializer_range": 0.02,
21
+ "intermediate_size": 3072,
22
+ "label2id": {
23
+ "B-Anatomy": 0,
24
+ "B-Chemical": 1,
25
+ "B-Disease": 2,
26
+ "I-Anatomy": 3,
27
+ "I-Chemical": 4,
28
+ "I-Disease": 5,
29
+ "O": 6
30
+ },
31
+ "layer_norm_eps": 1e-12,
32
+ "max_position_embeddings": 512,
33
+ "model_type": "bert",
34
+ "num_attention_heads": 12,
35
+ "num_hidden_layers": 12,
36
+ "pad_token_id": 0,
37
+ "position_embedding_type": "absolute",
38
+ "transformers_version": "4.8.2",
39
+ "type_vocab_size": 2,
40
+ "use_cache": true,
41
+ "vocab_size": 28996
42
+ }
eval_results.txt ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ eval_loss = 0.1652524173259735
2
+ eval_precision = 0.8504558443325623
3
+ eval_recall = 0.8611187655001378
4
+ eval_f1 = 0.8557540905045526
5
+ eval_accuracy = 0.975697565751413
6
+ eval_matrix = [[3901, 4, 60, 102, 1, 10, 538], [0, 4949, 15, 0, 42, 1, 371], [56, 18, 3770, 1, 0, 207, 372], [101, 0, 2, 2779, 0, 26, 335], [0, 32, 0, 0, 1383, 0, 213], [6, 4, 113, 28, 4, 2247, 335], [354, 467, 574, 247, 369, 453, 200220]]
7
+ eval_classification_report = precision recall f1-score support
8
+
9
+ Anatomy 0.85 0.83 0.84 4649
10
+ Chemical 0.89 0.91 0.90 5402
11
+ Disease 0.80 0.84 0.82 4465
12
+
13
+ micro avg 0.85 0.86 0.86 14516
14
+ macro avg 0.85 0.86 0.85 14516
15
+ weighted avg 0.85 0.86 0.86 14516
16
+
17
+ eval_runtime = 16.1004
18
+ eval_samples_per_second = 152.977
19
+ eval_steps_per_second = 0.621
20
+ epoch = 35.0
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89ff1d03f6a557db79ca98a49d86d41def96fa8d586d91c968f19e6e8c8e0ecc
3
+ size 430983665
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80b5808a16224f2716b8ece4c7a86a0ef8a4b06c4cf6511871afb0dc3fb549ac
3
+ size 2607
vocab.txt ADDED
The diff for this file is too large to render. See raw diff