-
NAACL 20222022In production SLU systems, new training data becomes available with time so that ML models need to be updated on a regular basis. Specifically, releasing new features adds new classes of data while the old data remains constant. However, retraining the full model each time from scratch is computationally expensive. To address this problem, we propose to consider production releases from the curriculum learning
-
Improving distantly supervised document-level relation extraction through natural language inferenceNAACL 2022 Workshop on Deep Learning for Low-Resource NLP2022The distant supervision (DS) paradigm has been widely used for relation extraction (RE) to alleviate the need for expensive annotations. However, it suffers from noisy labels, which leads to worse performance than models trained on human-annotated data, even when trained using hundreds of times more data. We present a systematic study on the use of natural language inference (NLI) to improve distantly supervised
-
MIT Sloan Sports Analytics Conference 20222022Sports broadcasters are increasingly sharing statistical insights throughout the game to tell a richer story for the audience. Thanks to abundant data and advanced statistics, broadcasters can quickly tell stories and make comparisons between teams and players to keep viewers engaged. To keep up with the fast-paced nature of many games, broadcasters rely on template-generated narratives to speak about in-game
-
NeurIPS 2022 Workshop on Trustworthy and Socially Responsible Machine Learning (TSRML)2022We study the problem of differentially private (DP) fine-tuning of large pre-trained models — a recent privacy-preserving approach suitable for solving downstream tasks with sensitive data. Existing work has demonstrated that high accuracy is possible under strong privacy constraint, yet requires significant computational overhead or modifications to the network architecture. We propose differentially private
-
ACL 20222022The goal of meta-learning is to learn to adapt to a new task with only a few labeled examples. Inspired by the recent progress in large language models, we propose in-context tuning (ICT), which recasts task adaptation and prediction as a simple sequence prediction problem: to form the input sequence, we concatenate the task instruction, labeled in-context examples, and the target input to predict; to metatrain
Related content
-
November 19, 2020AI models exceed human performance on public data sets; modified training and testing could help ensure that they aren’t exploiting short cuts.
-
November 16, 2020Amazon Scholar Julia Hirschberg on why speech understanding and natural-language understanding are intertwined.
-
November 11, 2020With a new machine learning system, Alexa can infer that an initial question implies a subsequent request.
-
November 10, 2020Alexa senior applied scientist provides career advice to graduate students considering a research role in industry.
-
November 9, 2020Watch a recording of the EMNLP 2020 session featuring a discussion with Amazon scholars and academics on the state of conversational AI.
-
November 6, 2020Work aims to improve accuracy of models both on- and off-device.