Customer-obsessed science
Research areas
-
May 15, 20265 min readA new scaling law that relates particular architectural choices to loss helps identify models that improve throughput by up to 47% with no loss of accuracy.
-
May 14, 202616 min read
-
-
April 15, 20268 min read
Featured news
-
PRML 20232023Deep neural networks are a powerful tool for a wide range of applications, including natural language processing (NLP) and computer vision (CV). However, training these networks can be a challenging task, as it requires careful selection of hyperparameters such as learning rates and scheduling strategies. Despite significant advances in designing dynamic (and adaptive) learning rate schedulers, choosing
-
ECML PKDD 2023 International Workshop on Machine Learning for Irregular Time Series2023Mixup is a domain-agnostic approach for data augmentation, originally proposed for training Deep Neural Networks (DNNs) for image classification. It obtains additional data for training by sampling from linear interpolations of model inputs and their labels. While proven to be effective for computer vision (CV) and natural language processing (NLP) tasks, it remains unknown if mixup can bring performance
-
ACM BuildSys 2023 Workshop on RLEM2023Building operations account for a significant portion of global emissions, contributing approximately 28% of global greenhouse gas emissions, according to the International Energy Agency. With the anticipated increase in cooling demand due to rising global temperatures, the optimization of rooftop units (RTUs) in buildings becomes crucial for reducing energy consumption and associated emissions. We focus
-
ASRU 20232023We propose a neural language modeling system based on low-rank adaptation (LoRA) for speech recognition output rescoring. Although pretrained language models (LMs) like BERT have shown superior performance in second-pass rescoring, the high computational cost of scaling up the pretraining stage and adapting the pretrained models to specific domains limit their practical use in rescoring. Here we present
-
CVPR 20232023Prompt learning is an efficient approach to adapt transformers by inserting learnable set of parameters into the input and intermediate representations of a pre-trained model. In this work, we present Expressive Prompts with Residuals (EXPRES) which modifies the prompt learning paradigm specifically for effective adaptation of vision transformers (ViT). Our method constructs downstream representations via
Collaborations
View allWhether you're a faculty member or student, there are number of ways you can engage with Amazon.
View all