location: Current position: Home >> Scientific Research >> Paper Publications

Extracting drug-drug interactions from texts with BioBERT and multiple entity-aware attentions

Hits:

Indexed by:Journal Papers

Date of Publication:2020-06-01

Journal:JOURNAL OF BIOMEDICAL INFORMATICS

Included Journals:SCIE

Volume:106

ISSN No.:1532-0464

Key Words:Drug-drug interactions; BioBERT; Entity-aware attention

Abstract:Drug-drug interactions (DDIs) extraction is one of the important tasks in the field of biomedical relation extraction, which plays an important role in the field of pharmacovigilance. Previous neural network based models have achieved good performance in DDIs extraction. However, most of the previous models did not make good use of the information of drug entity names, which can help to judge the relation between drugs. This is mainly because drug names are often very complex, leading to the fact that neural network models cannot understand their semantics directly. To address this issue, we propose a DDIs extraction model using multiple entity-aware attentions with various entity information. We use an output-modified bidirectional transformer (BioBERT) and a bidirectional gated recurrent unit layer (BiGRU) to obtain the vector representation of sentences. The vectors of drug description documents encoded by Doc2Vec are used as drug description information, which is an external knowledge to our model. Then we construct three different kinds of entity-aware attentions to get the sentence representations with entity information weighted, including attentions using the drug description information. The outputs of attention layers are concatenated and fed into a multi-layer perception layer. Finally, we get the result by a softmax classifier. The F-score is used to evaluate our model, which is also adopted by most previous DDIs extraction models. We evaluate our proposed model on the DDIExtraction 2013 corpus, which is the benchmark corpus of this domain, and achieves the state-of-the-art result (80.9% in F-score).

Pre One:Integrating Language Model and Reading Control Gate in BLSTM-CRF for Biomedical Named Entity Recognition

Next One:Multiple fragment-level interactive networks for answer selection