Distant Supervised Relation Extraction on Pre-train Model with Improved Multi-label Attention Mechanism

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

Relation extraction serves as the cornerstone for numerous natural language processing tasks. Supervised methods necessitate manual data labeling, incurring significant costs, while unsupervised approaches often suffer from low precision. Consequently, distant supervised relation extraction has emerged as a hot spot. However, the underlying assumption of distant supervision relation extraction introduces considerable noise into data, markedly impairing performances. Therefore, this paper aims to mitigate the impact of noisy sentences and enhance overall performance. Herein, we proposed a modification to the conventional approach by employing a model integrating a piecewise convolutional neural network (PCNN) with a sentence-level attention mechanism as the baseline. This model comprises two components: a sentence encoder and an attention layer. For the sentence encoder, we substitute the PCNN with a pre-trained GPT model, leveraging its superior ability to capture sentence features. Additionally, we enhance the attention layer by utilizing all possible relations as queries to compute attention weights for sentences. Subsequently, we aggregate these weighted representations to obtain a comprehensive feature representation for various relations. Experimental results confirm the superior performance of the proposed model, which combines pre-trained Transformer-combined encoders with the refined sentence-level attention mechanism. Specially, employing Transformer as a sentence encoder yields significant precision improvements, particularly at high recall levels. Meanwhile, the enhanced multi-label sentence-level attention mechanism enhances precision, particularly in scenarios with low recall.

Original languageEnglish
Title of host publicationKnowledge Science, Engineering and Management - 17th International Conference, KSEM 2024, Proceedings
EditorsCungeng Cao, Huajun Chen, Liang Zhao, Junaid Arshad, Yonghao Wang, Taufiq Asyhari
PublisherSpringer Science and Business Media Deutschland GmbH
Pages310-321
Number of pages12
ISBN (Print)9789819754915
DOIs
StatePublished - 2024
Event17th International Conference on Knowledge Science, Engineering and Management, KSEM 2024 - Birmingham, United Kingdom
Duration: 16 Aug 202418 Aug 2024

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume14884 LNAI
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference17th International Conference on Knowledge Science, Engineering and Management, KSEM 2024
Country/TerritoryUnited Kingdom
CityBirmingham
Period16/08/2418/08/24

Keywords

  • Attention mechanism
  • Distant Supervision
  • Knowledge Graph
  • Relation Extraction
  • Transformer

Fingerprint

Dive into the research topics of 'Distant Supervised Relation Extraction on Pre-train Model with Improved Multi-label Attention Mechanism'. Together they form a unique fingerprint.

Cite this