Customer-obsessed science
Research areas
-
March 20, 202615 min readSimplifying and clarifying the assembly code for core operations enabled automated optimization and verification.
-
March 19, 202611 min read
-
February 17, 20263 min read
-
-
January 13, 20267 min read
Featured news
-
2026Despite multilingual pretraining, large language models often struggle with non-English tasks, particularly in language control–the ability to respond in the intended language. We identify and characterize two key failure modes: the multilingual transfer bottleneck (correct language, incorrect task response) and the language consistency bottleneck (correct task response, wrong language). To systematically
-
ECIR 20262026Personalized experiences in multimodal assistants rely on accurate user understanding, yet large-scale training for personalization remains limited by privacy constraints and data sparsity. We introduce a framework for generating Comprehensive Synthetic Personas (CSPs) and personalized synthetic training data through taxonomy-guided knowledge enrichment, in-context learning, and Chain-of-Thought (CoT) knowledge
-
CHI 20262026As AI agents become collaborative partners in complex tasks, understanding how agent personality affects human-AI interaction becomes critical. While recent work explores personality customization in language models, little is known about how personality affects AI coding agents. We conducted the first exploratory study investigating: if OCEAN (Openness, Conscientiousness, Extraversion, Agreeableness, Neuroticism
-
2026Agentic vision–language models are increasingly trained to 'think with images' by calling image operations. However, we show that high final-answer accuracy often hides unfaithful visual reasoning: models may invoke tools on irrelevant regions or ignore tool outputs entirely, yet still guess the correct answer. In this work, we first propose a faithfulness evaluation protocol that measures whether intermediate
-
2026Generating long, coherent text remains a challenge for large language models (LLMs), as they lack hierarchical planning and structured organization in discourse generation. We introduce Structural Alignment, a novel method that aligns LLMs with human-like discourse structures to enhance long-form text generation. By integrating linguistically grounded discourse frameworks into reinforcement learning, our
Collaborations
View allWhether you're a faculty member or student, there are number of ways you can engage with Amazon.
View all