Journal of Shanghai University(Natural Science Edition) ›› 2021, Vol. 27 ›› Issue (5): 983-992.doi: 10.12066/j.issn.1007-2861.2197

• Research Articles • Previous Articles    

Distant supervision for relation extraction via attention CNNs

XING Yixue1, ZHU Yonghua1(), GAO Haiyan2, ZHOU Jin1, ZHANG Ke1   

  1. 1. Shanghai Film Academy, Shanghai University, Shanghai 200072, China
    2. School of Life Sciences, Shanghai University, Shanghai 200444, China
  • Received:2019-08-20 Online:2021-10-31 Published:2021-10-22
  • Contact: ZHU Yonghua E-mail:zyh@shu.edu.cn

Abstract:

The process of relation extraction is a significant step in several information extraction systems designed to mine structured facts from text. However, two problems surface when traditional distant supervision methods are employed to conduct the entity relation extraction task. First, the distant supervision heuristic aligns the text in the corpus using existing knowledge marked with entities and relations, after which the alignment results are treated as annotation data; this leads to inevitable labeling errors. Second, current statistical methods rely extensively on natural language processing tools to extract features, and the noise accumulating during the entire process significantly affects the extraction results. In this study, an end-to-end, attention mechanism-based convolutional neural network (CNN) is proposed. First, the attention mechanism is added to the input layer for automatic detection of more subtle clues and learning of parts of sentences that are relevant to relation extraction. Second, the sentence is encoded based on the position feature and word feature, a piecewise CNN (PCNN) is used to extract sentence features and classify relationships, and finally a max-margin loss function with a higher efficiency is used on the network. The accuracy of this method when used on the New York Times dataset is 2.0% higher than that of the classical PCNN+MIL model, and 1.0% higher than that of the classical APCNN+D model. The experimental results therefore demonstrate excellent accuracy of the proposed model when compared with that of other baselinemodels.

Key words: entity relation extraction, attention mechanism, deep learning, distant supervision

CLC Number: