Enhancing dialogue generation via multi-level contrastive learning
Most of the existing works for dialogue generation are data-driven models trained directly on
corpora crawled from websites. They mainly focus on improving the model architecture to
produce better responses but pay little attention to considering the quality of the training data
contrastively. In this paper, we propose a multi-level contrastive learning paradigm to model
the fine-grained quality of the responses with respect to the query. A Rank-aware Calibration
(RC) network is designed to construct the multi-level contrastive optimization objectives …
corpora crawled from websites. They mainly focus on improving the model architecture to
produce better responses but pay little attention to considering the quality of the training data
contrastively. In this paper, we propose a multi-level contrastive learning paradigm to model
the fine-grained quality of the responses with respect to the query. A Rank-aware Calibration
(RC) network is designed to construct the multi-level contrastive optimization objectives …
Most of the existing works for dialogue generation are data-driven models trained directly on corpora crawled from websites. They mainly focus on improving the model architecture to produce better responses but pay little attention to considering the quality of the training data contrastively. In this paper, we propose a multi-level contrastive learning paradigm to model the fine-grained quality of the responses with respect to the query. A Rank-aware Calibration (RC) network is designed to construct the multi-level contrastive optimization objectives. Since these objectives are calculated based on the sentence level, which may erroneously encourage/suppress the generation of uninformative/informative words. To tackle this incidental issue, on one hand, we design an exquisite token-level strategy for estimating the instance loss more accurately. On the other hand, we build a Knowledge Inference (KI) component to capture the keyword knowledge from the reference during training and exploit such information to encourage the generation of informative words. We evaluate the proposed model on a carefully annotated dialogue dataset and the results suggest that our model can generate more relevant and diverse responses compared to the baseline models.
arxiv.org
顯示最佳搜尋結果。 查看所有結果