TY - GEN
T1 - Distant Supervised Relation Extraction on Pre-train Model with Improved Multi-label Attention Mechanism
AU - Zhao, Qiming
AU - Yin, Chuantao
AU - Fan, Xin
AU - Chen, Hui
AU - Chai, Yanmei
AU - Ouyang, Yuanxin
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2024.
PY - 2024
Y1 - 2024
N2 - Relation extraction serves as the cornerstone for numerous natural language processing tasks. Supervised methods necessitate manual data labeling, incurring significant costs, while unsupervised approaches often suffer from low precision. Consequently, distant supervised relation extraction has emerged as a hot spot. However, the underlying assumption of distant supervision relation extraction introduces considerable noise into data, markedly impairing performances. Therefore, this paper aims to mitigate the impact of noisy sentences and enhance overall performance. Herein, we proposed a modification to the conventional approach by employing a model integrating a piecewise convolutional neural network (PCNN) with a sentence-level attention mechanism as the baseline. This model comprises two components: a sentence encoder and an attention layer. For the sentence encoder, we substitute the PCNN with a pre-trained GPT model, leveraging its superior ability to capture sentence features. Additionally, we enhance the attention layer by utilizing all possible relations as queries to compute attention weights for sentences. Subsequently, we aggregate these weighted representations to obtain a comprehensive feature representation for various relations. Experimental results confirm the superior performance of the proposed model, which combines pre-trained Transformer-combined encoders with the refined sentence-level attention mechanism. Specially, employing Transformer as a sentence encoder yields significant precision improvements, particularly at high recall levels. Meanwhile, the enhanced multi-label sentence-level attention mechanism enhances precision, particularly in scenarios with low recall.
AB - Relation extraction serves as the cornerstone for numerous natural language processing tasks. Supervised methods necessitate manual data labeling, incurring significant costs, while unsupervised approaches often suffer from low precision. Consequently, distant supervised relation extraction has emerged as a hot spot. However, the underlying assumption of distant supervision relation extraction introduces considerable noise into data, markedly impairing performances. Therefore, this paper aims to mitigate the impact of noisy sentences and enhance overall performance. Herein, we proposed a modification to the conventional approach by employing a model integrating a piecewise convolutional neural network (PCNN) with a sentence-level attention mechanism as the baseline. This model comprises two components: a sentence encoder and an attention layer. For the sentence encoder, we substitute the PCNN with a pre-trained GPT model, leveraging its superior ability to capture sentence features. Additionally, we enhance the attention layer by utilizing all possible relations as queries to compute attention weights for sentences. Subsequently, we aggregate these weighted representations to obtain a comprehensive feature representation for various relations. Experimental results confirm the superior performance of the proposed model, which combines pre-trained Transformer-combined encoders with the refined sentence-level attention mechanism. Specially, employing Transformer as a sentence encoder yields significant precision improvements, particularly at high recall levels. Meanwhile, the enhanced multi-label sentence-level attention mechanism enhances precision, particularly in scenarios with low recall.
KW - Attention mechanism
KW - Distant Supervision
KW - Knowledge Graph
KW - Relation Extraction
KW - Transformer
UR - https://www.scopus.com/pages/publications/85200760515
U2 - 10.1007/978-981-97-5492-2_24
DO - 10.1007/978-981-97-5492-2_24
M3 - 会议稿件
AN - SCOPUS:85200760515
SN - 9789819754915
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 310
EP - 321
BT - Knowledge Science, Engineering and Management - 17th International Conference, KSEM 2024, Proceedings
A2 - Cao, Cungeng
A2 - Chen, Huajun
A2 - Zhao, Liang
A2 - Arshad, Junaid
A2 - Wang, Yonghao
A2 - Asyhari, Taufiq
PB - Springer Science and Business Media Deutschland GmbH
T2 - 17th International Conference on Knowledge Science, Engineering and Management, KSEM 2024
Y2 - 16 August 2024 through 18 August 2024
ER -