Publication
Title
BART for knowledge grounded conversations
Author
Abstract
Transformers have shown incredible capabilities for conversation modeling, however, they store factual knowledge in their learned parameters, which is costly to update with new knowledge since it requires retraining. Models trained before the Coronavirus pandemic do not know about COVID-19. In this paper, we investigate how a BART model can be adapted to a knowledge grounded conversational setup. We introduce the notion of ๐‘˜๐‘’๐‘ฆ and ๐‘ž๐‘ข๐‘’๐‘Ÿ๐‘ฆ tokens to retrieve knowledge stored in an external database, that can easily be updated with new knowledge. As factual knowledge can hardly be reduced to a single sentence or vector, we allow the model to retrieve multiple sentences from the memory. Our analysis shows perplexity decreases with the number of passages retrieved from memory. Second, our analysis shows a shared encoder for knowledge retrieval, and conversation understanding reduces the model size and perform as well as a specialized module.
Language
English
Source (book)
Proceedings of KDD Workshop on Conversational Systems Towards Mainstream Adoption (KDD Converseโ€™20). ACM, New York, NY, USA
Publication
2020
Volume/pages
2666 (2020) , p. 1-6
Full text (open access)
UAntwerpen
Faculty/Department
Research group
Publication type
Subject
Affiliation
Publications with a UAntwerp address
External links
VABB-SHW
Record
Identifier
Creation 10.09.2020
Last edited 17.06.2024
To cite this reference