Encode-Attend-Refine Decode: Enriching Contextual Representations for Natural Language Generation
Date19th Apr 2022
Time02:00 PM
Venue meet.google.com/qhf-zisu-bdw
PAST EVENT
Details
Sequence-to-Sequence (encode-attend-decode) models have become ubiquitous to a wide variety of Natural Language Generation (NLG) tasks. Although the generation quality across tasks have improved significantly, there are still several limitations. For instance, i) task-agnostic problems like repeating-phrases, hallucinations, and so on, still persist, ii) task-specific characteristics, like conciseness in abstractive summarization, answerability for question generation, better alignment in machine translation, etc., are not explicitly modeled or rewarded in the vanilla sequence-to-sequence models. In this thesis, we integrate the refine module as (encode-attend-refine-decode) in a vanilla sequence-to-sequence model. The refine module improves upon the context representations learnt by the attend module, i) to mitigate task-agnostic: repeating phrase problem, and ii) to enrich task-specific characteristics in the model.
Speakers
Preksha Nema
Computer Science and Engineering