-
Instruction following is a key capability for LLMs. However, recent studies have shown that LLMs often struggle with instructions containing multiple constraints (e.g. a request to create a social media post “in a funny tone” with “no hashtag”). Despite this, most evaluations focus solely on synthetic data. To address this, we introduce RealInstruct, the first benchmark designed to evaluate LLMs’ ability
-
2024Entity matching is the task of linking records from different sources that refer to the same real-world entity. Past work has primarily treated entity linking as a standard supervised learning problem. However, supervised entity matching models often do not generalize well to new data, and collecting exhaustive labeled training data is often cost prohibitive. Further, recent efforts have adopted LLMs for
-
2024Training with mixed data distributions is a common and important part of creating multi-task and instruction-following models. The diversity of the data distributions and cost of joint training makes the optimization procedure extremely challenging. Data mixing methods partially address this problem, albeit having a suboptimal performance across data sources and require multiple expensive training runs.
-
2024We propose a constraint learning schema for fine-tuning Large Language Models (LLMs) with attribute control. Given a training corpus and control criteria formulated as a sequencelevel constraint on model outputs, our method fine-tunes the LLM on the training corpus while enhancing constraint satisfaction with minimal impact on its utility and generation quality. Specifically, our approach regularizes the
-
Findings of EMNLP 20242024Language Models for text classification often produce overconfident predictions for both indistribution and out-of-distribution samples, i.e. the model’s output probabilities do not match their accuracy. Prior work showed that simple post-hoc approaches are effective for mitigating this issue, but are not robust in noisy settings, e.g., when the distribution shift is caused by spelling mistakes. In this
Related content
-
July 11, 2022The SCOT science team used lessons from the past — and improved existing tools — to contend with “a peak that lasted two years”.
-
July 08, 2022Industry track chair and Amazon principal research scientist Rashmi Gangadharaiah on trends in industry papers and the challenges of building practical dialogue systems.
-
July 08, 2022New model sets new standard in accuracy while enabling 60-fold speedups.
-
July 07, 2022The breadth and originality of Amazon’s natural-language-processing research are on display at the annual meeting of the North American chapter of the Association for Computational Linguistics.
-
June 29, 2022President’s visit part of a mission to preserve the Icelandic language in the digital age.
-
June 28, 2022Amazon’s TabTransformer model is now available through SageMaker JumpStart and the official release of the Keras open-source library.