-
PRML 20232023Deep neural networks are a powerful tool for a wide range of applications, including natural language processing (NLP) and computer vision (CV). However, training these networks can be a challenging task, as it requires careful selection of hyperparameters such as learning rates and scheduling strategies. Despite significant advances in designing dynamic (and adaptive) learning rate schedulers, choosing
-
ASRU 20232023We propose a neural language modeling system based on low-rank adaptation (LoRA) for speech recognition output rescoring. Although pretrained language models (LMs) like BERT have shown superior performance in second-pass rescoring, the high computational cost of scaling up the pretraining stage and adapting the pretrained models to specific domains limit their practical use in rescoring. Here we present
-
ACL 20232023Parameter-efficient tuning (PET) methods fit pre-trained language models (PLMs) to downstream tasks by either computing a small compressed update for a subset of model parameters, or appending and fine-tuning a small number of new model parameters to the pretrained network. Hand-designed PET architectures from the literature perform well in practice, but have the potential to be improved via automated neural
-
KDD 2023 International Workshop on Multimodal Learning2023E-commerce platforms enable brands to connect with relevant online shoppers. While major brands are easily identifiable by shoppers, smaller and emerging brands often lean on advertising campaigns in e-commerce platforms to reach a wide audience. For such advertising campaigns, brands need to come up with a leading ad creative which may be shown together with their listed products. Designing such creatives
-
IJCNLP-AACL 20232023Conventional text style transfer approaches focus on sentence-level style transfer without considering contextual information, and the style is described with attributes (e.g., formality). When applying style transfer in conversations such as task-oriented dialogues, existing approaches suffer from these limitations as context can play an important role and the style attributes are often difficult to define
Related content
-
May 6, 2020Leveraging semantic content improves performance of acoustic-only model for detecting device-directed speech.
-
May 4, 2020Alexa scientist Ariya Rastrow on the blurring boundaries between acoustic processing and language understanding.
-
April 30, 2020Letting a machine learning system label its own examples improves performance.
-
April 27, 2020An end-to-end deep-learning-based solution circumvents the “permutation problem”.
-
April 23, 2020Generating natural sounding, human-like speech has been a goal of scientists for decades.
-
April 21, 2020Bezos’s Shareholder Letter has become a must read, along the lines of Warren Buffet’s letter to Berkshire Hathaway shareholders, or the Bill & Melinda Gates Annual Letter.