TY - GEN
T1 - Enhancing Dual-Encoders with Question and Answer Cross-Embeddings for Answer Retrieval
AU - Wang, Yanmeng
AU - Bai, Jun
AU - Wang, Ye
AU - Zhang, Jianfei
AU - Rong, Wenge
AU - Ji, Zongcheng
AU - Wang, Shaojun
AU - Xiao, Jing
N1 - Publisher Copyright:
© 2021 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - Dual-Encoders is a promising mechanism for answer retrieval in question answering (QA) systems. Currently most conventional DualEncoders learn the semantic representations of questions and answers merely through matching score. Researchers proposed to introduce the QA interaction features in scoring function but at the cost of low efficiency in inference stage. To keep independent encoding of questions and answers during inference stage, variational auto-encoder is further introduced to reconstruct answers (questions) from question (answer) embeddings as an auxiliary task to enhance QA interaction in representation learning in training stage. However, the needs of text generation and answer retrieval are different, which leads to hardness in training. In this work, we propose a framework to enhance the Dual-Encoders model with question answer cross-embeddings and a novel Geometry Alignment Mechanism (GAM) to align the geometry of embeddings from Dual-Encoders with that from Cross-Encoders. Extensive experimental results show that our framework significantly improves Dual-Encoders model and outperforms the state-of-the-art method on multiple answer retrieval datasets.
AB - Dual-Encoders is a promising mechanism for answer retrieval in question answering (QA) systems. Currently most conventional DualEncoders learn the semantic representations of questions and answers merely through matching score. Researchers proposed to introduce the QA interaction features in scoring function but at the cost of low efficiency in inference stage. To keep independent encoding of questions and answers during inference stage, variational auto-encoder is further introduced to reconstruct answers (questions) from question (answer) embeddings as an auxiliary task to enhance QA interaction in representation learning in training stage. However, the needs of text generation and answer retrieval are different, which leads to hardness in training. In this work, we propose a framework to enhance the Dual-Encoders model with question answer cross-embeddings and a novel Geometry Alignment Mechanism (GAM) to align the geometry of embeddings from Dual-Encoders with that from Cross-Encoders. Extensive experimental results show that our framework significantly improves Dual-Encoders model and outperforms the state-of-the-art method on multiple answer retrieval datasets.
UR - https://www.scopus.com/pages/publications/85129203047
U2 - 10.18653/v1/2021.findings-emnlp.198
DO - 10.18653/v1/2021.findings-emnlp.198
M3 - 会议稿件
AN - SCOPUS:85129203047
T3 - Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
SP - 2306
EP - 2315
BT - Findings of the Association for Computational Linguistics, Findings of ACL
A2 - Moens, Marie-Francine
A2 - Huang, Xuanjing
A2 - Specia, Lucia
A2 - Yih, Scott Wen-Tau
PB - Association for Computational Linguistics (ACL)
T2 - 2021 Findings of the Association for Computational Linguistics, Findings of ACL: EMNLP 2021
Y2 - 7 November 2021 through 11 November 2021
ER -