-
EACL 20232023Missing information is a common issue of dialogue summarization where some information in the reference summaries is not covered in the generated summaries. To address this issue, we propose to utilize natural language inference (NLI) models to improve coverage while avoiding introducing factual inconsistencies. Specifically, we use NLI to compute fine-grained training signals to encourage the model to
-
ACL Findings 20232023The Open-Domain Question Answering (ODQA) task involves retrieving and subsequently generating answers from fine-grained relevant passages within a database. Current systems leverage Pre-trained Language Models (PLMs) to model the relationship between questions and passages. However, the diversity in surface form expressions can hinder the model’s ability to capture accurate correlations, especially within
-
ACL 20232023Large pre-trained language models (PLMs) have been shown to retain implicit knowledge within their parameters. To enhance this implicit knowledge, we propose Knowledge Injection into Language Models (KILM), a novel approach that injects entity-related knowledge into encoder-decoder PLMs, via a generative knowledge infilling objective through continued pre-training. This is done without architectural modifications
-
ACL 20232023Automatic melody-to-lyric generation is a task in which song lyrics are generated to go with a given melody. It is of significant practical interest and more challenging than unconstrained lyric generation as the music imposes additional constraints onto the lyrics. The training data is limited as most songs are copyrighted, resulting in models that underfit the complicated cross-modal relationship between
-
ACL Findings 20232023Deep Neural Networks (DNNs) are becoming integral components of real world services relied upon by millions of users. Unfortunately, architects of these systems can find it difficult to ensure reliable performance as irrelevant details like random initialization can unexpectedly change the outputs of a trained system with potentially disastrous consequences. We formulate the model stability problem by studying
Related content
-
June 15, 2021Alexa Fund company unlocks voice-based computing for people who have trouble using their voices.
-
June 11, 2021Proteno model dramatically increases the efficiency of the first step in text-to-speech conversion.
-
June 10, 2021Recasting different natural-language tasks in the same form dramatically improves few-shot multitask learning.
-
June 04, 2021Topics range from the predictable, such as speech recognition and noise cancellation, to singing separation and automatic video dubbing.
-
June 03, 2021Rastrow discussed the continued challenges and expanded role of speech recognition, and some of the interesting research and themes that emerged from ICASSP 2021.
-
June 03, 2021Amazon Scholar Heng Ji says that deep learning could benefit from the addition of a little linguistic intuition.