Logo Logo
Help
Contact
Switch Language to German

Yue, Zhenrui; Zeng, Huimin; Kratzwald, Bernhard; Feuerriegel, Stefan ORCID logoORCID: https://orcid.org/0000-0001-7856-8729 and Wang, Dong (2022): QA Domain Adaptation using Hidden Space Augmentation and Self-Supervised Contrastive Adaptation. EMNLP 2022, Abu Dhabi, 07.12.2022-11.12.2022. arXiv.

Full text not available from 'Open Access LMU'.

Abstract

Question answering (QA) has recently shown impressive results for answering questions from customized domains. Yet, a common challenge is to adapt QA models to an unseen target domain. In this paper, we propose a novel self-supervised framework called QADA for QA domain adaptation. QADA introduces a novel data augmentation pipeline used to augment training QA samples. Different from existing methods, we enrich the samples via hidden space augmentation. For questions, we introduce multi-hop synonyms and sample augmented token embeddings with Dirichlet distributions. For contexts, we develop an augmentation method which learns to drop context spans via a custom attentive sampling strategy. Additionally, contrastive learning is integrated in the proposed self-supervised adaptation framework QADA. Unlike existing approaches, we generate pseudo labels and propose to train the model via a novel attention-based contrastive adaptation method. The attention weights are used to build informative features for discrepancy estimation that helps the QA model separate answers and generalize across source and target domains. To the best of our knowledge, our work is the first to leverage hidden space augmentation and attention-based contrastive adaptation for self-supervised domain adaptation in QA. Our evaluation shows that QADA achieves considerable improvements on multiple target datasets over state-of-the-art baselines in QA domain adaptation.

Actions (login required)

View Item View Item