Customer-obsessed science
Research areas
-
November 20, 20254 min readA new evaluation pipeline called FiSCo uncovers hidden biases and offers an assessment framework that evolves alongside language models.
-
October 20, 20254 min read
-
October 14, 20257 min read
-
October 2, 20253 min read
-
Featured news
-
EACL 20232023When upgrading neural models to a newer version, new errors that were not encountered in the legacy version can be introduced, known as regression errors. This inconsistent behavior during model upgrade often outweighs the benefits of accuracy gain and hinders the adoption of new models. To mitigate regression errors from model upgrade, distillation and ensemble have proven to be viable solutions without
-
ICLR 20232023Question answering over knowledge bases (KBs) aims to answer natural language questions with factual information such as entities and relations in KBs. Previous methods either generate logical forms that can be executed over KBs to obtain final answers or predict answers directly. Empirical results show that the former often produces more accurate answers, but it suffers from non-execution issues due to
-
IWSDS 20232023Effective evaluation methods remain a significant challenge for research on open-domain conversational dialogue systems. Explicit satisfaction ratings can be elicited from users, but users often do not provide ratings when asked, and those they give can be highly subjective. Post-hoc ratings by experts are an alternative, but these can be both expensive and complex to collect. Here, we explore the creation
-
LAK 20232023A/B testing at scale provides opportunities for learning analytics researchers to learn from large sample sizes. Deploying and running live intervention experiments with such large samples, however, raises infrastructural challenges. This paper discusses some of those challenges, and reports on two possible implementations that address those challenges in a workforce learning context at a large technology
-
ICSE 20232023Manual code reviews and static code analyzers are the traditional mechanisms to verify if source code complies with coding policies. However, they are hard to scale. We formulate code compliance assessment as a machine learning (ML) problem, to take as input a natural language policy and code, and generate a prediction on the code’s compliance, non-compliance, or irrelevance. Our intention for ML-based
Collaborations
View allWhether you're a faculty member or student, there are number of ways you can engage with Amazon.
View all