Customer-obsessed science
Research areas
-
December 5, 20256 min readA multiagent architecture separates data perception, tool knowledge, execution history, and code generation, enabling ML automation that works with messy, real-world inputs.
-
-
-
November 20, 20254 min read
-
October 20, 20254 min read
Featured news
-
CVPR 2021 Workshop on Learning from Unlabled Videos2021In this paper, we explore learning end-to-end deep neural trackers without tracking annotations. This is important as large-scale training data is essential for deep neural trackers, while tracking annotations are expensive to acquire. We first hallucinate videos from images with bounding box annotations using motion transformations along with simulated video effects to create a diverse tracking dataset
-
Interspeech 20212021To improve customer privacy, commercial speech applications are reducing human transcription of customer data. This has a negative impact on language model training due to a smaller amount of in-domain transcripts. Prior work demonstrated that training on automated transcripts alone provides modest gains due to reinforcement of recognition errors. We consider a new condition, where a model trained on historical
-
ICLR 2021 Workshop on Synthetic Data Generation2021Generalization is a central problem in machine learning, especially when data is limited. Using prior information to enforce constraints is the principled way of encouraging generalization. In this work, we propose to leverage the prior information embedded in pretrained language models (LM) to improve generalization for intent classification and slot labeling tasks with limited training data. Specifically
-
Interspeech 20212021With the expanding role of voice-controlled devices, bootstrapping spoken language understanding models from little labeled data becomes essential. Semi-supervised learning is a common technique to improve model performance when labeled data is scarce. In a real-world production system, the labeled data and the online test data often may come from different distributions. In this work, we use semi-supervised
-
ACL-IJCNLP 2021 Workshop on Meta-Learning and its Applications to NLP2021Meta-learning has recently been proposed to learn models and algorithms that can generalize from a handful of examples. However, applications to structured prediction and textual tasks pose challenges for meta-learning algorithms. In this paper, we apply two metalearning algorithms, Prototypical Networks and Reptile, to few-shot Named Entity recognition (NER), including a method for incorporating language
Collaborations
View allWhether you're a faculty member or student, there are number of ways you can engage with Amazon.
View all