|
{ |
|
"paper_id": "2020", |
|
"header": { |
|
"generated_with": "S2ORC 1.0.0", |
|
"date_generated": "2023-01-19T11:23:07.128224Z" |
|
}, |
|
"title": "BENNERD: A Neural Named Entity Linking System for COVID-19", |
|
"authors": [ |
|
{ |
|
"first": "Mohammad", |
|
"middle": [ |
|
"Golam" |
|
], |
|
"last": "Sohrab", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "Artificial Intelligence Research Center (AIRC)", |
|
"institution": "", |
|
"location": { |
|
"country": "Japan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Khoa", |
|
"middle": [ |
|
"N A" |
|
], |
|
"last": "Duong", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "Artificial Intelligence Research Center (AIRC)", |
|
"institution": "", |
|
"location": { |
|
"country": "Japan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Makoto", |
|
"middle": [], |
|
"last": "Miwa", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "Artificial Intelligence Research Center (AIRC)", |
|
"institution": "", |
|
"location": { |
|
"country": "Japan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Goran", |
|
"middle": [], |
|
"last": "Topi\u0107", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "Artificial Intelligence Research Center (AIRC)", |
|
"institution": "", |
|
"location": { |
|
"country": "Japan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
}, |
|
{ |
|
"first": "Hiroya", |
|
"middle": [], |
|
"last": "Takamura", |
|
"suffix": "", |
|
"affiliation": { |
|
"laboratory": "Artificial Intelligence Research Center (AIRC)", |
|
"institution": "", |
|
"location": { |
|
"country": "Japan" |
|
} |
|
}, |
|
"email": "[email protected]" |
|
} |
|
], |
|
"year": "", |
|
"venue": null, |
|
"identifiers": {}, |
|
"abstract": "We present a biomedical entity linking (EL) system BENNERD that detects named entities in text and links them to the unified medical language system (UMLS) knowledge base (KB) entries to facilitate the corona virus disease 2019 (COVID-19) research. BEN-NERD mainly covers biomedical domain, especially new entity types (e.g., coronavirus, viral proteins, immune responses) by addressing CORD-NER dataset. It includes several NLP tools to process biomedical texts including tokenization, flat and nested entity recognition, and candidate generation and ranking for EL that have been pre-trained using the CORD-NER corpus. To the best of our knowledge, this is the first attempt that addresses NER and EL on COVID-19-related entities, such as COVID-19 virus, potential vaccines, and spreading mechanism, that may benefit research on COVID-19. We release an online system to enable real-time entity annotation with linking for end users. We also release the manually annotated test set and CORD-NERD dataset for leveraging EL task. The BENNERD system is available at https://aistairc.github.io/BENNERD/.", |
|
"pdf_parse": { |
|
"paper_id": "2020", |
|
"_pdf_hash": "", |
|
"abstract": [ |
|
{ |
|
"text": "We present a biomedical entity linking (EL) system BENNERD that detects named entities in text and links them to the unified medical language system (UMLS) knowledge base (KB) entries to facilitate the corona virus disease 2019 (COVID-19) research. BEN-NERD mainly covers biomedical domain, especially new entity types (e.g., coronavirus, viral proteins, immune responses) by addressing CORD-NER dataset. It includes several NLP tools to process biomedical texts including tokenization, flat and nested entity recognition, and candidate generation and ranking for EL that have been pre-trained using the CORD-NER corpus. To the best of our knowledge, this is the first attempt that addresses NER and EL on COVID-19-related entities, such as COVID-19 virus, potential vaccines, and spreading mechanism, that may benefit research on COVID-19. We release an online system to enable real-time entity annotation with linking for end users. We also release the manually annotated test set and CORD-NERD dataset for leveraging EL task. The BENNERD system is available at https://aistairc.github.io/BENNERD/.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Abstract", |
|
"sec_num": null |
|
} |
|
], |
|
"body_text": [ |
|
{ |
|
"text": "In response to the coronavirus disease 2019 (COVID-19) for global research community to apply recent advances in natural language processing (NLP), COVID-19 Open Research Dataset (CORD-19) 1 is an emerging research challenge with a resource of over 181,000 scholarly articles that are related to the infectious disease COVID-19 caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). To facilitate COVID-19 studies, since NER is considered a fundamental step in text mining system, Xuan et al. (2020b) created CORD-NER dataset with comprehensive NE annotations. The annotations are based on distant or weak supervision. The dataset includes 29,500 documents from the CORD-19 corpus. The CORD-NER dataset gives a shed on NER, but they do not address linking task which is important to address COVID-19 research. For example, in the example sentence in Figure 1 , the mention SARS-CoV-2 needs to be disambiguated. Since the term SARS-CoV-2 in this sentence refers to a virus, it should be linked to an entry of a virus in the knowledge base, not to an entry of 'SARS-CoV-2 vaccination', which corresponds to therapeutic or preventive procedure to prevent a disease.", |
|
"cite_spans": [ |
|
{ |
|
"start": 498, |
|
"end": 517, |
|
"text": "Xuan et al. (2020b)", |
|
"ref_id": "BIBREF11" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 867, |
|
"end": 875, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "We present a BERT-based Exhaustive Neural Named Entity Recognition and Disambiguation (BENNERD) system. The system is composed of four models: NER model (Sohrab and Miwa, 2018) that enumerates all possible spans as potential entity mentions and classifies them into entity types, masked language model BERT (Devlin et al., 2019) , candidate generation model to find a list of candidate entities in the unified medical language system (UMLS) knowledge base (KB) for entity linking (EL) and candidate ranking model to disambiguate the entity for concept indexing. The BENNERD system provides a web interface to facilitate the process of text annotation and its disambiguation without any training for end users. In addition, we introduce Open Research Dataset for Named Entity Recognition and Disambiguation) dataset an extended version of CORD-NER as for leveraging EL task.", |
|
"cite_spans": [ |
|
{ |
|
"start": 307, |
|
"end": 328, |
|
"text": "(Devlin et al., 2019)", |
|
"ref_id": "BIBREF0" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Introduction", |
|
"sec_num": "1" |
|
}, |
|
{ |
|
"text": "The main objective of this work is to address recent pandemic of COVID-19 research. To facilitate COVID-19 studies, we introduce the BEN- NERD system that finds nested named entities and links them to a UMLS knowledge base (KB). BEN-NERD mainly comprises two platforms: a web interface and a back-end server. The overall workflow of the BENNERD system is illustrated in Figure 1 .", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 370, |
|
"end": 378, |
|
"text": "Figure 1", |
|
"ref_id": "FIGREF0" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "System Description", |
|
"sec_num": "2" |
|
}, |
|
{ |
|
"text": "The user interface of our BENNERD system is a web application with input panel, load a sample tab, annotation tab, gear box tab, and .TXT and .ANN tabs. Figure 2 shows an users' input interface of BENNERD. For a given text from users or loading a sample text from a sample list, the annotation tab will show the annotations with the text based on best NER-and EL-based training model. Figure 3 shows an example of text annotation based on the BENNERD's NER model. Different colors represent different entity types and, when the cursor floats over a coloured box representing an entity above text, the corresponding concept unique identifier (CUI) on the UMLS is shown. Figure 3 also shows an example where entity mention SARS-CoV-2 links to its corresponding CUI. Users can save the machine readable text in txt format and annotation files in the ann format where the ann annotation file provides standoff annotation output in brat (Stenetorp et al., 2012) 2 format.", |
|
"cite_spans": [ |
|
{ |
|
"start": 932, |
|
"end": 956, |
|
"text": "(Stenetorp et al., 2012)", |
|
"ref_id": "BIBREF8" |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 153, |
|
"end": 161, |
|
"text": "Figure 2", |
|
"ref_id": "FIGREF1" |
|
}, |
|
{ |
|
"start": 385, |
|
"end": 393, |
|
"text": "Figure 3", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 669, |
|
"end": 677, |
|
"text": "Figure 3", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "BENNERD Web Interface", |
|
"sec_num": "2.1" |
|
}, |
|
{ |
|
"text": "We provide a quick inside look of our BENNERD web interface (BWI). The data flow of BWI is presented as follows:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data Flow of Web Interface", |
|
"sec_num": "2.1.1" |
|
}, |
|
{ |
|
"text": "Server-side initialization (a) The BWI configuration, concept embeddings, and NER and EL models are loaded (b) GENIA sentence splitter and BERT basic tokenizer instances are initialized (c) When a text is submitted (a) The text is split into sentences and tokens (b) Token and sentence standoffs are identified (c) NER model is run on tokenized sentences (d) EL model is run on the result (e) The identified token spans are translated into text standoffs (f) The identified concepts' names are looked up in the UMLS database (g) A brat document is created (h) The brat document is translated into JSON, and sent to the client side (i) The brat visualizer renders the document", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Data Flow of Web Interface", |
|
"sec_num": "2.1.1" |
|
}, |
|
{ |
|
"text": "The BENNERD back-end implements a pipeline of tools (e.g., NER, EL), following the data flow described in Section 2.1.1. This section provides implementation details of our back-end modules for NER and EL.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "BENNERD Back-end", |
|
"sec_num": "2.2" |
|
}, |
|
{ |
|
"text": "We build the mention detection, a.k.a NER, based on the BERT model (Devlin et al., 2019) . The layer receives subword sequences and assigns contextual representations to the subwords via BERT. We denote a sentence by S = (x 1 , ..., x n ), where x i is the i-th word, and x i consists of s i subwords. This layer assigns a vector v i,j to the j-th subword of the i-th word. Then, we generate the vector embedding v i for each word x i by computing the unweighted average of its subword embeddings v i,j . We generate mention candidates based on the same idea as the span-based model (Lee et al., 2017; Sohrab and Miwa, 2018; Sohrab et al., 2019a,b) , in which all continuous word sequences are generated given a maximal size L x (span width). The representation x b,e \u2208 R dx for the span from the b-th word to the e-th word in a sentence is calculated from the embeddings of the first word, the last word, and the weighted average of all words in the span as follows:", |
|
"cite_spans": [ |
|
{ |
|
"start": 67, |
|
"end": 88, |
|
"text": "(Devlin et al., 2019)", |
|
"ref_id": "BIBREF0" |
|
}, |
|
{ |
|
"start": 583, |
|
"end": 601, |
|
"text": "(Lee et al., 2017;", |
|
"ref_id": "BIBREF3" |
|
}, |
|
{ |
|
"start": 602, |
|
"end": 624, |
|
"text": "Sohrab and Miwa, 2018;", |
|
"ref_id": "BIBREF5" |
|
}, |
|
{ |
|
"start": 625, |
|
"end": 648, |
|
"text": "Sohrab et al., 2019a,b)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Neural Named Entity Recognition", |
|
"sec_num": "2.2.1" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "x b,e = v b ; e i=b \u03b1 b,e,i v i ; v e ,", |
|
"eq_num": "(1)" |
|
} |
|
], |
|
"section": "Neural Named Entity Recognition", |
|
"sec_num": "2.2.1" |
|
}, |
|
{ |
|
"text": "where \u03b1 b,e,i denotes the attention value of the i-th word in a span from the b-th word to the e-th word, and [; ; ] denotes concatenation.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Neural Named Entity Recognition", |
|
"sec_num": "2.2.1" |
|
}, |
|
{ |
|
"text": "In our EL component, for every mention span x b,e of a concept in a document, we are supposed to identify its ID in the target KB. 3 Let us call the ID a concept unique identifier (CUI). The input is all predicted mention spans M = {m 1 , m 2 , . . . , m n }, where m i denotes the i-th mention and n denotes the total number of predicted mentions. The list of entity mentions {m i } i=1,...,n needs to be mapped to a list of corresponding CUIs {c i } i=1,...,n . We decompose EL into two subtasks: candidate generation and candidate ranking.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "Candidate Generation To find a list of candidate entities in KB to link with a given mention, we build a candidate generation layer adapting a dual-encoders model (Gillick et al., 2019) . Instead of normalizing entity definition to disambiguate entities, we simply normalize the semantic types in both mention and entity sides from UMLS. The representation of a mention m in a document by the semantic type t m , can be denoted as:", |
|
"cite_spans": [ |
|
{ |
|
"start": 163, |
|
"end": 185, |
|
"text": "(Gillick et al., 2019)", |
|
"ref_id": "BIBREF1" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "v m = [w m ; t m ] ,", |
|
"eq_num": "(2)" |
|
} |
|
], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "where t m \u2208 R dt m is the mention type embedding. For the entity (concept) side with semantic type information, the representation a e , and its entity type embedding t e \u2208 R dt e can be computed as:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "v e = [a e ; t e ] .", |
|
"eq_num": "(3)" |
|
} |
|
], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "We use cosine similarity to compute the similarity score between a mention m and an entity e and feed it into a linear layer (LL) to transform the score into an unbounded logit as:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "sim (m, e) = cos (v m , v e ) ,", |
|
"eq_num": "(4)" |
|
} |
|
], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "score (m, e) = LL(sim (m, e)).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "We employ the in-batch random negatives technique as described in the previous work (Gillick et al., 2019) . For evaluating the performance of the model during training, we use the in-batch recall@1 metric (Gillick et al., 2019) on the development set to track and save the best model. We calculate the embedding of each detected mention from the mention detection layer and each of all entities in KB and then using an approximate nearest neighbor search algorithm in Faiss (Johnson et al., 2019) to retrieve the top k entities as candidates for the ranking layer.", |
|
"cite_spans": [ |
|
{ |
|
"start": 84, |
|
"end": 106, |
|
"text": "(Gillick et al., 2019)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 206, |
|
"end": 228, |
|
"text": "(Gillick et al., 2019)", |
|
"ref_id": "BIBREF1" |
|
}, |
|
{ |
|
"start": 475, |
|
"end": 497, |
|
"text": "(Johnson et al., 2019)", |
|
"ref_id": "BIBREF2" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "Candidate Ranking The cosine similarity score in the candidate generation is insufficient to disambiguate the entities in which the correct entity should be assigned the highest score which is comparable from the k candidate entities. We employed a fully-connected neural network model to aim at ranking the entity candidate list to select the best entity linked to the mention. Given a mention m and a set of candidate entities {e 1 , e 2 , ..., e k }, we concatenate the embedding of m in Equation (2) with the embedding of each entity e i in Equation 3to form a vector v m,e i . Then the vector v m,e i is fed into a LL to compute the ranking score:", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "EQUATION", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [ |
|
{ |
|
"start": 0, |
|
"end": 8, |
|
"text": "EQUATION", |
|
"ref_id": "EQREF", |
|
"raw_str": "score(m, e i ) = LL(v m,e i ).", |
|
"eq_num": "(6)" |
|
} |
|
], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "The model is then trained using a softmax loss to maximize the score of the correct entity compared with other incorrect entities retrieved from the trained candidate generation model.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Entity Linking", |
|
"sec_num": "2.2.2" |
|
}, |
|
{ |
|
"text": "In this section, we evaluate our toolkit on CORD-NER and CORD-NERD datasets.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Experimental Settings", |
|
"sec_num": "3" |
|
}, |
|
{ |
|
"text": "We carry out our experiments on CORD-NER, a distant or weak supervision-based large-scale dataset that includes 29,500 documents, 2,533,485 sentences, and 10,388,642 mentions. In our experiment, CORD-NER covers 63 fine-grained entity types 4 . CORD-NER mainly supports four sources including 18 biomedical entity types 5 , 18 general entity types 6 , knowledge base entity types, and nine 7 seed-guided new entity types. We split the CORD-NER dataset into three subsets: train, development, and test, which respectively contain 20,000, 4,500, and 5,000 documents.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CORD-NER Dataset", |
|
"sec_num": "3.1" |
|
}, |
|
{ |
|
"text": "CORD-NER dataset comprises only NER task. To solve the EL task, we expand this dataset by leveraging a CUI for each mention in the CORD-NER dataset, we call this CORD-NERD dataset. We use the most recent UMLS version 2020AA release that includes coronavirus-related concepts. To create CORD-NERD dataset, we use a dictionary matching approach based on exact match using UMLS KB. 470, 248 mentions, among which 6, 794, 126 and 3, 676, 122 mentions are respectively present and absent in the UMLS. Therefore, the entity coverage ratio of CORD-NERD over the UMLS is 64.89%. We annotate the entity mentions that are not found in the UMLS with CUI LESS. To evaluate the EL performances on CORD-NERD, 302,166 mentions are assigned for 5,000 test set, we call this UMLS-based test set. The train and development sets of CORD-NERD dataset, we simply calls UMLS-based trainand UMLS-based dev-set respectively. Besides, we assigned a biologist to annotate 1,000 random sentences based on chemical, disease, and gene types to create a manually annotated test set. This test set includes 311 disease mentions for the NER task and 946 mentions 8 with their corresponding CUIs for the EL task.", |
|
"cite_spans": [ |
|
{ |
|
"start": 379, |
|
"end": 383, |
|
"text": "470,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 384, |
|
"end": 397, |
|
"text": "248 mentions,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 398, |
|
"end": 412, |
|
"text": "among which 6,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 413, |
|
"end": 417, |
|
"text": "794,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 418, |
|
"end": 428, |
|
"text": "126 and 3,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 429, |
|
"end": 433, |
|
"text": "676,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 434, |
|
"end": 437, |
|
"text": "122", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "CORD-NERD Dataset", |
|
"sec_num": "3.2" |
|
}, |
|
{ |
|
"text": "Each text and the corresponding annotation file are processed by BERT's basic tokenizer. After tokenization, each text and its corresponding annotation file was directly passed to the deep neural approach for mention detection and classification. Table 1 shows the performance of SciSpacy on CORD-NER dataset. In this table, the results are based on randomly picked 1,000 manually annotated sentences as the test set. Table 2 shows the performance comparison of our BENNERD with different pre-trained BERT models based on our test set. Since the manually annotated CORD-NER test set is not publicly available, we cannot directly compare our system performance. Instead, in Table 3 , we show the performance of gene, chemical, and disease based on our UMLS-based test set. Besides, in Table 4 , we also show the NER performances comparison of BENNERD with BC5CDR corpus-based SciSpacy model on the manually annotated disease entities.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 247, |
|
"end": 254, |
|
"text": "Table 1", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 418, |
|
"end": 425, |
|
"text": "Table 2", |
|
"ref_id": "TABREF2" |
|
}, |
|
{ |
|
"start": 673, |
|
"end": 680, |
|
"text": "Table 3", |
|
"ref_id": "TABREF3" |
|
}, |
|
{ |
|
"start": 784, |
|
"end": 791, |
|
"text": "Table 4", |
|
"ref_id": "TABREF4" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Data Prepossessing", |
|
"sec_num": "3.3" |
|
}, |
|
{ |
|
"text": "As we are the first to perform EL task on CORD-19 dataset, we present different scenarios to evaluate our candidate ranking performance. The results of EL are depicted in Table 5 . In this table, we evaluate our candidate ranking performances based on two experiment settings. In setting1, we train the CUIs based on manually annotated MedMention (Murty et al., 2018) dataset. In setting2, the BENNERD model is trained on automatically annotated CORD-NERD dataset. Table 5 also shows that our BENNERD model with setting2 is outperformed in compare to setting1 in every cases in terms of accuracy@ (1, 10, 20, 30, 40, 50) . Table 6 shows the EL performance on the manually annotated test set. In this table, it also shows that our system with setting2 is outperformed in compare to setting1. Besides, we also evaluate the manually annotated test set simply with string matching approach where the results of the top 10, 20, 30, 40 or 50 predictions for a gold candidate are unchanged.", |
|
"cite_spans": [ |
|
{ |
|
"start": 347, |
|
"end": 367, |
|
"text": "(Murty et al., 2018)", |
|
"ref_id": "BIBREF4" |
|
}, |
|
{ |
|
"start": 597, |
|
"end": 600, |
|
"text": "(1,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 601, |
|
"end": 604, |
|
"text": "10,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 605, |
|
"end": 608, |
|
"text": "20,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 609, |
|
"end": 612, |
|
"text": "30,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 613, |
|
"end": 616, |
|
"text": "40,", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 617, |
|
"end": 620, |
|
"text": "50)", |
|
"ref_id": null |
|
} |
|
], |
|
"ref_spans": [ |
|
{ |
|
"start": 171, |
|
"end": 178, |
|
"text": "Table 5", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 465, |
|
"end": 472, |
|
"text": "Table 5", |
|
"ref_id": null |
|
}, |
|
{ |
|
"start": 623, |
|
"end": 630, |
|
"text": "Table 6", |
|
"ref_id": "TABREF7" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Candidate Ranking Performance", |
|
"sec_num": "4.3" |
|
}, |
|
{ |
|
"text": "Finally, in Table 7 , we show the performance of nine new entity types discussed in Section 3.1 related to COVID-19 studies, which may benefit research on COVID-19 virus, spreading mechanism, and potential vaccines.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 12, |
|
"end": 19, |
|
"text": "Table 7", |
|
"ref_id": "TABREF8" |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Performances on COVID-19 Entity Types", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "Model UMLS-based Test set A@1 A@10 A@20 A@30 A@40 A@50 (%) Table 5 : EL performance on test set. We report Accuracy@n, where n = 1, 10, 20, 30, 40, 50. Accuracy@1, gold candidate was ranked highest. Accuracy@{10, 20, 30, 40, 50} indicates, gold candidate was in top 10, 20, 30, 40 or in 50 predictions of the ranker. Pred., NEs, and TP refers to predictions, named entities, and true positive respectively. Setting1 and 2 denotes model is trained on MEDMention and CORD-NERD datasets respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [ |
|
{ |
|
"start": 59, |
|
"end": 66, |
|
"text": "Table 5", |
|
"ref_id": null |
|
} |
|
], |
|
"eq_spans": [], |
|
"section": "Performances on COVID-19 Entity Types", |
|
"sec_num": "4.4" |
|
}, |
|
{ |
|
"text": "Manually Annotated Test set A@1 A@10 A@20 A@30 A@40 A@50 (%) ", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Model", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "To facilitate the biomedical text mining research on COVID-19, recently a few works have reported to address text mining tasks. Xuan et al. (2020b) created CORD-NER dataset with distant or weak supervision and reported first NER performances on different NER models. Motivated by this work, we presented a first web-based toolkit that addresses both NER and EL. In addition, we also extend the CORD-NER dataset to solve EL task. Xuan et al. (2020a) created EvidenceMiner system that retrieves sentence-level textual evidence from CORD-NER dataset. Tonia et al. (2020) developed an NLP pipeline to extract drug and vaccine information about SARS-CoV-2 and other viruses to help biomedical experts to easily track the latest scientific publications. To the best of our knowledge, this work is our first effort to solve both NER and EL models in a pipeline manner.", |
|
"cite_spans": [ |
|
{ |
|
"start": 128, |
|
"end": 147, |
|
"text": "Xuan et al. (2020b)", |
|
"ref_id": "BIBREF11" |
|
}, |
|
{ |
|
"start": 429, |
|
"end": 448, |
|
"text": "Xuan et al. (2020a)", |
|
"ref_id": "BIBREF10" |
|
}, |
|
{ |
|
"start": 548, |
|
"end": 567, |
|
"text": "Tonia et al. (2020)", |
|
"ref_id": "BIBREF9" |
|
} |
|
], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Related Work", |
|
"sec_num": "5" |
|
}, |
|
{ |
|
"text": "We presented the BENNERD system for entity linking, hoping that we can bring insights for the COVID-19 studies on making scientific discoveries. To the best of our knowledge, BENNERD represents the first web-based workflow of NER and EL for NLP research that addresses CORD-19 dataset that leads to create CORD-NERD dataset to facilitate COVID-19 work. The online system is available for meeting real-time extraction for end users. The BENNERD system is continually evolving; we will continue to improve the system as well as to implement new functions such as relation extraction to further facilitate COVID-19 research. We refer to visit https://aistairc.github.io/BENNERD/ to know more about BENNERD and CORD-NERD.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Conclusion", |
|
"sec_num": "6" |
|
}, |
|
{ |
|
"text": "https://brat.nlplab.org", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "We used the UMLS KB in the experiments.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
}, |
|
{ |
|
"text": "In the original CORD-NER paper(Xuan et al., 2020b), the authors reported 75 fine-grained entity types, but we found only 63 types.5 https://uofi.app.box.com/s/ k8pw7d5kozzpoum2jwfaqdaey1oij93x/file/ 6378663941866 https://spacy.io/api/annotation# named-entities 7 Coronavirus, Viral Protein, Livestock, Wildlife, Evolution, Physical Science, Substrate, Material, Immune Response 8 Among them, 38, 311, and 597 mentions are of chemical, disease, and gene entity types respectively.", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "", |
|
"sec_num": null |
|
} |
|
], |
|
"back_matter": [ |
|
{ |
|
"text": "We thank the anonymous reviewers for their valuable comments. This work is based on results obtained from a project commissioned by the Public/Private R&D Investment Strategic Expansion PrograM (PRISM).", |
|
"cite_spans": [], |
|
"ref_spans": [], |
|
"eq_spans": [], |
|
"section": "Acknowledgments", |
|
"sec_num": null |
|
} |
|
], |
|
"bib_entries": { |
|
"BIBREF0": { |
|
"ref_id": "b0", |
|
"title": "BERT: Pre-training of deep bidirectional transformers for language understanding", |
|
"authors": [ |
|
{ |
|
"first": "Jacob", |
|
"middle": [], |
|
"last": "Devlin", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Ming-Wei", |
|
"middle": [], |
|
"last": "Chang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kenton", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Kristina", |
|
"middle": [], |
|
"last": "Toutanova", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "4171--4186", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Tech- nologies, Volume 1 (Long and Short Papers), pages 4171-4186.", |
|
"links": null |
|
}, |
|
"BIBREF1": { |
|
"ref_id": "b1", |
|
"title": "Learning dense representations for entity retrieval", |
|
"authors": [ |
|
{ |
|
"first": "Daniel", |
|
"middle": [], |
|
"last": "Gillick", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sayali", |
|
"middle": [], |
|
"last": "Kulkarni", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Larry", |
|
"middle": [], |
|
"last": "Lansing", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Alessandro", |
|
"middle": [], |
|
"last": "Presta", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jason", |
|
"middle": [], |
|
"last": "Baldridge", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Eugene", |
|
"middle": [], |
|
"last": "Ie", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Diego", |
|
"middle": [], |
|
"last": "Garcia-Olano", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "528--537", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/K19-1049" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Daniel Gillick, Sayali Kulkarni, Larry Lansing, Alessandro Presta, Jason Baldridge, Eugene Ie, and Diego Garcia-Olano. 2019. Learning dense repre- sentations for entity retrieval. In Proceedings of the 23rd Conference on Computational Natural Lan- guage Learning (CoNLL), pages 528-537, Hong Kong, China. Association for Computational Lin- guistics.", |
|
"links": null |
|
}, |
|
"BIBREF2": { |
|
"ref_id": "b2", |
|
"title": "Billionscale similarity search with GPUs", |
|
"authors": [ |
|
{ |
|
"first": "J", |
|
"middle": [], |
|
"last": "Johnson", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "M", |
|
"middle": [], |
|
"last": "Douze", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "H", |
|
"middle": [], |
|
"last": "J\u00e9gou", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "IEEE Transactions on Big Data", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.1109/TBDATA.2019.2921572" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "J. Johnson, M. Douze, and H. J\u00e9gou. 2019. Billion- scale similarity search with GPUs. IEEE Transac- tions on Big Data.", |
|
"links": null |
|
}, |
|
"BIBREF3": { |
|
"ref_id": "b3", |
|
"title": "End-to-end neural coreference resolution", |
|
"authors": [ |
|
{ |
|
"first": "Kenton", |
|
"middle": [], |
|
"last": "Lee", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luheng", |
|
"middle": [], |
|
"last": "He", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Mike", |
|
"middle": [], |
|
"last": "Lewis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luke", |
|
"middle": [], |
|
"last": "Zettlemoyer", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2017, |
|
"venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "188--197", |
|
"other_ids": { |
|
"DOI": [ |
|
"10.18653/v1/D17-1018" |
|
] |
|
}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Kenton Lee, Luheng He, Mike Lewis, and Luke Zettle- moyer. 2017. End-to-end neural coreference reso- lution. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pages 188-197, Copenhagen, Denmark. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF4": { |
|
"ref_id": "b4", |
|
"title": "Hierarchical losses and new resources for fine-grained entity typing and linking", |
|
"authors": [ |
|
{ |
|
"first": "Shikhar", |
|
"middle": [], |
|
"last": "Murty", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Patrick", |
|
"middle": [], |
|
"last": "Verga", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Luke", |
|
"middle": [], |
|
"last": "Vilnis", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Irena", |
|
"middle": [], |
|
"last": "Radovanovic", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Andrew", |
|
"middle": [], |
|
"last": "Mccallum", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics", |
|
"volume": "1", |
|
"issue": "", |
|
"pages": "97--109", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Shikhar Murty, Patrick Verga, Luke Vilnis, Irena Radovanovic, and Andrew McCallum. 2018. Hier- archical losses and new resources for fine-grained entity typing and linking. In Proceedings of the 56th Annual Meeting of the Association for Compu- tational Linguistics (Volume 1: Long Papers), pages 97-109.", |
|
"links": null |
|
}, |
|
"BIBREF5": { |
|
"ref_id": "b5", |
|
"title": "Deep exhaustive model for nested named entity recognition", |
|
"authors": [ |
|
{ |
|
"first": "Golam", |
|
"middle": [], |
|
"last": "Mohammad", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Makoto", |
|
"middle": [], |
|
"last": "Sohrab", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Miwa", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2018, |
|
"venue": "Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "2843--2849", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mohammad Golam Sohrab and Makoto Miwa. 2018. Deep exhaustive model for nested named entity recognition. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Process- ing, pages 2843-2849, Brussels, Belgium. Associa- tion for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF6": { |
|
"ref_id": "b6", |
|
"title": "A neural pipeline approach for the pharmaconer shared task using contextual exhaustive models", |
|
"authors": [ |
|
{ |
|
"first": "Minh", |
|
"middle": [ |
|
"Thang" |
|
], |
|
"last": "Mohammad Golam Sohrab", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Makoto", |
|
"middle": [], |
|
"last": "Pham", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hiroya", |
|
"middle": [], |
|
"last": "Miwa", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Takamura", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of The 5th Workshop on BioNLP Open Shared Tasks", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "47--55", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mohammad Golam Sohrab, Minh Thang Pham, Makoto Miwa, and Hiroya Takamura. 2019a. A neu- ral pipeline approach for the pharmaconer shared task using contextual exhaustive models. In Pro- ceedings of The 5th Workshop on BioNLP Open Shared Tasks, pages 47-55.", |
|
"links": null |
|
}, |
|
"BIBREF7": { |
|
"ref_id": "b7", |
|
"title": "A generic neural exhaustive approach for entity recognition and sensitive span detect", |
|
"authors": [ |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Mohammad Golam Sohrab", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Minh", |
|
"middle": [], |
|
"last": "Pham", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Makoto", |
|
"middle": [], |
|
"last": "Thang", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "", |
|
"middle": [], |
|
"last": "Miwa", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2019, |
|
"venue": "Proceedings of the Iberian Languages Evaluation Forum", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "735--743", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Mohammad Golam Sohrab, Pham Minh Thang, and Makoto Miwa. 2019b. A generic neural exhaustive approach for entity recognition and sensitive span detect. In Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2019), pages 735-743, Span. IberLEF 2019.", |
|
"links": null |
|
}, |
|
"BIBREF8": { |
|
"ref_id": "b8", |
|
"title": "brat: a web-based tool for NLP-assisted text annotation", |
|
"authors": [ |
|
{ |
|
"first": "Pontus", |
|
"middle": [], |
|
"last": "Stenetorp", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sampo", |
|
"middle": [], |
|
"last": "Pyysalo", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Goran", |
|
"middle": [], |
|
"last": "Topi\u0107", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Tomoko", |
|
"middle": [], |
|
"last": "Ohta", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Sophia", |
|
"middle": [], |
|
"last": "Ananiadou", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Jun'ichi", |
|
"middle": [], |
|
"last": "Tsujii", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2012, |
|
"venue": "Proceedings of the Demonstrations at the 13th Conference of the European Chapter of the Association for Computational Linguistics", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "102--107", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Pontus Stenetorp, Sampo Pyysalo, Goran Topi\u0107, Tomoko Ohta, Sophia Ananiadou, and Jun'ichi Tsu- jii. 2012. brat: a web-based tool for NLP-assisted text annotation. In Proceedings of the Demonstra- tions at the 13th Conference of the European Chap- ter of the Association for Computational Linguistics, pages 102-107, Avignon, France. Association for Computational Linguistics.", |
|
"links": null |
|
}, |
|
"BIBREF9": { |
|
"ref_id": "b9", |
|
"title": "Nlp for extracting covid-19 drug and vaccine information from scientific literature", |
|
"authors": [ |
|
{ |
|
"first": "Korves", |
|
"middle": [], |
|
"last": "Tonia", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Peterson", |
|
"middle": [], |
|
"last": "Matthew", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Garay", |
|
"middle": [], |
|
"last": "Christopher", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Read", |
|
"middle": [], |
|
"last": "Tom", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chang", |
|
"middle": [], |
|
"last": "Wenling", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Bartlett", |
|
"middle": [], |
|
"last": "Marta", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Quattrochi", |
|
"middle": [], |
|
"last": "Lauren", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Hirschman", |
|
"middle": [], |
|
"last": "Lynette", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "Proceedings of the 28th Conference on Intelligent Systems for Moleculer Biology", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Korves Tonia, Peterson Matthew, Garay Christopher, Read Tom, Chang Wenling, Bartlett Marta, Quat- trochi Lauren, and Hirschman Lynette. 2020. Nlp for extracting covid-19 drug and vaccine infor- mation from scientific literature. In Proceedings of the 28th Conference on Intelligent Systems for Moleculer Biology.", |
|
"links": null |
|
}, |
|
"BIBREF10": { |
|
"ref_id": "b10", |
|
"title": "Automatic textual evidence mining in COVID-19 literature", |
|
"authors": [ |
|
{ |
|
"first": "Wang", |
|
"middle": [], |
|
"last": "Xuan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Liu", |
|
"middle": [], |
|
"last": "Weili", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Chauhan", |
|
"middle": [], |
|
"last": "Aabhas", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Guan", |
|
"middle": [], |
|
"last": "Yingjun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Han", |
|
"middle": [], |
|
"last": "Jiawei", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "2020 Intelligent Systems for Molecular Biology (ISMB'20)", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wang Xuan, Liu Weili, Chauhan Aabhas, Guan Yingjun, and Han Jiawei. 2020a. Automatic textual evidence mining in COVID-19 literature. 2020 In- telligent Systems for Molecular Biology (ISMB'20).", |
|
"links": null |
|
}, |
|
"BIBREF11": { |
|
"ref_id": "b11", |
|
"title": "Comprehensive named entity recognition on CORD-19 with distant or weak supervision", |
|
"authors": [ |
|
{ |
|
"first": "Wang", |
|
"middle": [], |
|
"last": "Xuan", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Song", |
|
"middle": [], |
|
"last": "Xiangchen", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Li", |
|
"middle": [], |
|
"last": "Bangzheng", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Guan", |
|
"middle": [], |
|
"last": "Yingjun", |
|
"suffix": "" |
|
}, |
|
{ |
|
"first": "Han", |
|
"middle": [], |
|
"last": "Jiawei", |
|
"suffix": "" |
|
} |
|
], |
|
"year": 2020, |
|
"venue": "", |
|
"volume": "", |
|
"issue": "", |
|
"pages": "", |
|
"other_ids": {}, |
|
"num": null, |
|
"urls": [], |
|
"raw_text": "Wang Xuan, Song Xiangchen, Li Bangzheng, Guan Yingjun, and Han Jiawei. 2020b. Comprehensive named entity recognition on CORD-19 with distant or weak supervision. 2020 Intelligent Systems for Molecular Biology (ISMB'20).", |
|
"links": null |
|
} |
|
}, |
|
"ref_entries": { |
|
"FIGREF0": { |
|
"text": "Workflow of BENNERD System", |
|
"uris": null, |
|
"num": null, |
|
"type_str": "figure" |
|
}, |
|
"FIGREF1": { |
|
"text": "BENNERD Users' Input Interface Figure 3: Entity Annotation and Linking with BEN-NERDConcept embeddings are indexed by Faiss(Johnson et al., 2019)", |
|
"uris": null, |
|
"num": null, |
|
"type_str": "figure" |
|
}, |
|
"TABREF1": { |
|
"html": null, |
|
"text": "51.86 64.69 80.31 59.65 68.46 CORD-NER System 82.14 74.68 72.23 82.93 75.22 78.89 75.73 68.42 71.89 Table 1: Performance comparison of baseline systems on three biomedical entity types in CORD-NER corpus.", |
|
"content": "<table><tr><td/><td>Gene</td><td/><td/><td>Chemical</td><td/><td/><td colspan=\"2\">Disease</td></tr><tr><td>P</td><td>R</td><td>F</td><td>P</td><td>R</td><td>F</td><td>P</td><td/><td>R F (%)</td></tr><tr><td colspan=\"6\">SciSpacy(BIONLP13CG) 91.48 82.06 86.51 64.66 39.81 49.28</td><td>8.11</td><td colspan=\"2\">2.75</td><td>4.11</td></tr><tr><td colspan=\"5\">SciSpacy(BC5CDR) -86.97 Model --Development set P R F</td><td>P</td><td colspan=\"2\">Test set R</td><td>F (%)</td></tr><tr><td colspan=\"9\">BENNERD + ClinicalCovid BERT (CCB) 84.62 86.43 85.52 82.83 83.23 83.03</td></tr><tr><td>BENNERD + SciBERT</td><td/><td colspan=\"7\">84.03 87.05 85.51 82.16 83.81 82.98</td></tr><tr><td>BENNERD + Covid BERT Base</td><td/><td colspan=\"7\">78.31 66.80 72.10 77.44 66.80 71.73</td></tr></table>", |
|
"num": null, |
|
"type_str": "table" |
|
}, |
|
"TABREF2": { |
|
"html": null, |
|
"text": "NER Performances using different pre-trained BERT models.", |
|
"content": "<table><tr><td>Model</td><td>P</td><td>Gene R</td><td>F</td><td>P</td><td>Chemical R</td><td>F</td><td>P</td><td>Disease R</td><td>F (%)</td></tr><tr><td colspan=\"10\">BENNERD + CCB 76.07 74.83 75.45 83.55 84.60 84.07 84.85 84.99 84.92</td></tr></table>", |
|
"num": null, |
|
"type_str": "table" |
|
}, |
|
"TABREF3": { |
|
"html": null, |
|
"text": "Performance comparison of BENNERD on three major biomedical entity types in CORD-NER corpus. CCB denotes ClinicalCovid BERT.", |
|
"content": "<table><tr><td>Model</td><td>P</td><td>R F (%)</td></tr><tr><td colspan=\"3\">SciSpacy(BC5CDR) 36.01 56.27 43.91</td></tr><tr><td>BENNERD</td><td colspan=\"2\">49.16 47.27 48.20</td></tr></table>", |
|
"num": null, |
|
"type_str": "table" |
|
}, |
|
"TABREF4": { |
|
"html": null, |
|
"text": "Performance comparison of BENNERD with pre-trained SciSpacy over the disease entity types on the manually annotated test set.", |
|
"content": "<table/>", |
|
"num": null, |
|
"type_str": "table" |
|
}, |
|
"TABREF5": { |
|
"html": null, |
|
"text": "BENNERD + NER's Pred. + Setting1 27.61 44.56 49.74 51.88 53.08 54.19 BENNERD + Gold NEs + Setting1 29.78 48.33 53.89 56.22 57.53 58.74 BENNERD + NER's TP + Setting1 30.31 48.91 54.60 56.95 58.27 59.49 BENNERD + NER's Pred. + Setting2 47.46 64.32 67.70 69.87 71.12", |
|
"content": "<table><tr><td/><td/><td>72.07</td></tr><tr><td>BENNERD + Gold NEs + Setting2</td><td>50.73 69.31 73.10 75.58 77.03</td><td>78.13</td></tr><tr><td>BENNERD + NER's TP + Setting2</td><td>53.90 73.06 76.90 79.36 80.79</td><td>81.87</td></tr></table>", |
|
"num": null, |
|
"type_str": "table" |
|
}, |
|
"TABREF7": { |
|
"html": null, |
|
"text": "EL performance on our manually annotated test set.", |
|
"content": "<table><tr><td>Model</td><td>UMLS-based Test set P R F (%)</td></tr><tr><td>Coronovirus</td><td>98.46 98.94 98.70</td></tr><tr><td>Viral Protein</td><td>89.39 91.09 90.23</td></tr><tr><td>Livestock</td><td>96.67 97.26 96.96</td></tr><tr><td>Wildlife</td><td>98.43 97.56 97.99</td></tr><tr><td>Evolution</td><td>97.16 98.46 97.80</td></tr><tr><td>Physical Science</td><td>96.80 93.08 94.90</td></tr><tr><td>Substrate</td><td>95.99 98.46 97.21</td></tr><tr><td>Material</td><td>94.80 90.46 92.58</td></tr><tr><td colspan=\"2\">Immune Response 97.29 99.42 98.35</td></tr></table>", |
|
"num": null, |
|
"type_str": "table" |
|
}, |
|
"TABREF8": { |
|
"html": null, |
|
"text": "Performances on nine types of COVID-19", |
|
"content": "<table/>", |
|
"num": null, |
|
"type_str": "table" |
|
} |
|
} |
|
} |
|
} |