1 AIT Asian Institute of Technology

Contrastive learning with template’s prompt in multiple aspect dialogue summarization

AuthorWorachot Nakduk
Call NumberAIT Thesis no.DSAI-23-11
Subject(s)Natural language processing (Computer science)

NoteA thesis submitted in partial fulfillment of the requirements for the degree of Master of Engineering in Data Science and Artificial Intelligence
PublisherAsian Institute of Technology
AbstractDialogue summarizing compresses the essential information from a conversation into a concise paragraph, enabling individuals to efficiently understand the main ideas without needing to look into the surrounding context. Recently, pre-trained language models have the potential to be helpful for dialogue summaries. Due to the engagement of sev eral participants, topic deviations, frequent references to previous statements, diverse interaction messages, and specific vocabulary, these variables contribute to increased complexity for summarizers. It can generate misinformation and generate misleading content that only covers partial facts of a conversation. Furthermore, a single discourse has the ability to incorporate multiple topics without a clearly defined boundary between them. The realistic importance of the dialogue summarization model can be better char acterized by assessing numerous aspects, such as the speaker’s goal or the specific topic under discussion. However, current summarization systems produce generic summaries that lack personalization and fail to align with customer preferences and expectations. To address this limitation, create customized different aspects of the produced summaries by engaging with the summarization process through textual submission in a format of descriptive prompts. In this paper, we present Contrastive Learning with a Topic Length Template Prompt for Dialogue Summarization. We select a topic either from the goal summary or attractive topics. These topics serve as a control signal, guided by a template’s prompt. We utilize contrastive learning methods to enhance the diversity of prompts in the template by using synonym replacement and random topics. These methods enable us to generate both positive and negative topics, thereby increasing the quantity of meaningful information available for training. Additionally, we use special tokens to highlight words and prompts to focus on important keywords. Experimentally, we show that our model can increase the ROUGE score in the DialogSum testing dataset. Our models are available at https://github.com/worachot-n/topic-length.git
Year2023
TypeThesis
SchoolSchool of Engineering and Technology
DepartmentDepartment of Information and Communications Technologies (DICT)
Academic Program/FoSData Science and Artificial Intelligence (DSAI)
Chairperson(s)Chaklam Silpasuwanchai
Examination Committee(s)Mongkol Ekpanyapong;Dailey, Matthew N.
Scholarship Donor(s)His Majesty the King’s Scholarships (Thailand)
DegreeThesis (M. Eng.) - Asian Institute of Technology, 2023


Usage Metrics
View Detail0
Read PDF0
Download PDF0