-
NAACL 20222022The massive amount of trainable parameters in the pre-trained language models (PLMs) makes them hard to be deployed to multiple downstream tasks. To address this issue, parameter-efficient transfer learning methods have been proposed to tune only a few parameters during fine-tuning while freezing the rest. This paper looks at existing methods along this line through the kernel lens. Motivated by the connection
-
NAACL 20222022Providing conversation models with background knowledge has been shown to make open-domain dialogues more informative and engaging. Existing models treat knowledge selection as a sentence ranking or classification problem where each sentence is handled individually, ignoring the internal semantic connection among sentences in background document. In this work, we propose to automatically convert the background
-
NAACL 20222022With the growing popularity of deep-learning models, model understanding becomes more important. Much effort has been devoted to demystify deep neural networks for better interpretability. Some feature attribution methods have shown promising results in computer vision, especially the gradient-based methods where effectively smoothing the gradients with reference data is key to a robust and faithful result
-
NAACL 20222022Self-learning paradigms in large-scale conversational AI agents tend to leverage user feedback in bridging between what they say and what they mean. However, such learning, particularly in Markov-based query rewriting systems have far from addressed the impact of these models on future training where successive feedback is inevitably contingent on the rewrite itself, especially in a continually updating
-
NAACL 20222022Unsupervised word alignments offer a lightweight and interpretable method to transfer labels from high- to low-resource languages, as long as semantically related words have the same label across languages. But such an assumption is often not true in industrial NLP pipelines, where multilingual annotation guidelines are complex and deviate from semantic consistency due to various factors (such as annotation
Related content
-
August 01, 2022McKeown awarded IEEE Innovation in Societal Infrastructure Award and named a member of the American Philosophical Society.
-
July 28, 2022Donato Crisostomi talks about how his mother helped spark a love of knowledge that led him to two science internships at Amazon.
-
July 22, 2022New EMNLP workshop will feature talks, papers, posters, and a competition built around the 50-plus-language, million-utterance MASSIVE dataset.
-
July 15, 2022New method optimizes the twin demands of retrieving relevant content and filtering out bad content.
-
July 14, 2022To become the interface for the Internet of things, conversational agents will need to learn on their own. Alexa has already started down that path.
-
July 13, 2022Four MIT professors are the recipients of the inaugural call for research projects.