Customer-obsessed science


Research areas
-
August 8, 2025A new philosophy for developing LLM architectures reduces energy requirements, speeds up runtime, and preserves pretrained-model performance.
Featured news
-
ISTFA 20252025As integrated circuits continue to scale down to ever-smaller dimensions and increased complexity, traditional failure isolation & analysis (FA/FI), and sample preparation techniques face significant limitations. The increasing density of modern semiconductor devices, now advancing beyond 3 nm technology nodes, presents unprecedented challenges in precise layer-by-layer analysis. Over the past decade, Xenon
-
EMNLP 20252025The Multimodal Large Language Models (MLLMs) are continually pre-trained on a mixture of image-text caption data and interleaved document data, while the high-quality data filtering towards image-text interleaved document data is under-explored. We propose to train an efficient MLLM as a Unified Mulitmodal Data Quality Classifier to Filter both high-quality image-text caption and interleaved data (UniFilter
-
CIKM 20252025Current methods for evaluating large language models (LLMs) typically focus on high-level tasks such as text generation, without targeting a particular AI application. This approach is not sufficient for evaluating LLMs for Responsible AI dimensions like fairness, since protected attributes that are highly relevant in one application may be less relevant in another. In this work, we construct a dataset
-
IEEE CDC 20252025Online advertising is typically implemented via real-time bidding, and advertising campaigns are then defined as extremely high-dimensional optimization problems. To solve these problems in light of large scale and significant uncertainties, the optimization problems are modularized in a manner that makes feedback control a critical component of the solution. The control problem, however, is challenging
-
IEEE 2025 Workshop on Automatic Speech Recognition and Understanding2025Speech Recognition has seen a dramatic shift towards adopting Large Language Models (LLMs). This shift is partly driven by good scalability properties demonstrated by LLMs, ability to leverage large amounts of labelled, unlabelled speech and text data, streaming capabilities with autoregressive framework and multi-tasking with instruction following characteristics of LLMs. However, simple next-token prediction
Academia
View allWhether you're a faculty member or student, there are number of ways you can engage with Amazon.
View all