Augesc: Dialogue augmentation with large language models for emotional support conversation
Published in Findings of the Association for Computational Linguistics: ACL 2023, 2023
Crowdsourced dialogue corpora are usually limited in scale and topic coverage due to the expensive cost of data curation. This would hinder the generalization of downstream dialogue models to open-domain topics. In this work, we leverage large language models for dialogue augmentation in the task of emotional support conversation (ESC). By treating dialogue augmentation as a dialogue completion task, we prompt a fine-tuned language model to complete full dialogues from available dialogue posts of various topics, which are then postprocessed based on heuristics. Applying this approach, we construct AugESC, an augmented dataset for the ESC task, which largely extends the scale and topic coverage of the crowdsourced ESConv corpus. Through comprehensive human evaluation, we demonstrate that our approach is superior to strong baselines of dialogue augmentation and that AugESC has comparable dialogue quality to the crowdsourced corpus. We also conduct human interactive evaluation and prove that post-training on AugESC improves downstream dialogue models’ generalization ability to open-domain topics. These results suggest the utility of AugESC and highlight the potential of large language models in improving data-scarce dialogue generation tasks.
Citation:
@inproceedings{zheng-etal-2023-augesc,
title = "{A}ug{ESC}: Dialogue Augmentation with Large Language Models for Emotional Support Conversation",
author = "Zheng, Chujie and
Sabour, Sahand and
Wen, Jiaxin and
Zhang, Zheng and
Huang, Minlie",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.findings-acl.99",
doi = "10.18653/v1/2023.findings-acl.99",
pages = "1552--1568"
}