-
2024Convolution augmented Transformer architectures have dominated the field of automatic speech recognition by showing better WER results when the models are trained on relatively smaller training data. In this work, we revisit the necessity of convolution modules in the ASR encoder architecture, given that the inductive bias brought by the convolution modules may only boost performance in a low training data
-
2024Large Language Models (LLMs) frequently hallucinate, impeding their reliability in mission-critical situations. One approach to address this issue is to provide citations to relevant sources alongside generated content, enhancing the verifiability of generations. However, citing passages accurately in answers remains a substantial challenge. This paper proposes a weakly-supervised fine-tuning method leveraging
-
SIGIR Forum2024Consumers on a shopping mission often leverage both product search and information seeking systems, such as web search engines and Question Answering (QA) systems, in an iterative process to improve their understanding of available products and reach a purchase decision. While product search is useful for shoppers to find the actual products meeting their requirements in the catalog, information seeking
-
Instruction fine-tuning (IFT) elicits instruction following capabilities and steers the behavior of large language models (LLMs) via supervised learning. However, existing models trained on open-source IFT datasets only have the ability to follow instructions from users, and often fail to follow complex role and rules specified by developers, a.k.a. system prompts. The ability to follow these roles and
-
2024In this paper, we investigate which questions are challenging for retrieval-based Question Answering (QA). We (i) propose retrieval complexity (RC), a novel metric conditioned on the completeness of retrieved documents, which measures the difficulty of answering questions, and (ii) propose an unsupervised pipeline to measure RC given an arbitrary retrieval system. Our proposed pipeline measures RC more
Related content
-
March 31, 2021Throughout the pandemic, the Alexa team has continued to invent on behalf of our customers.
-
March 26, 2021In the future, says Amazon Scholar Emine Yilmaz, users will interact with computers to identify just the information they need, rather than scrolling through long lists of results.
-
March 24, 2021Human-evaluation studies validate metrics, and experiments show evidence of bias in popular language models.
-
March 19, 2021A model that uses both local and global context improves on the state of the art by 6% and 11% on two benchmark datasets.
-
March 16, 2021Amanda Cullen, a PhD candidate in informatics at the University of California, Irvine, wanted to do work that had an impact outside of academia — she found an ideal opportunity at Twitch.
-
March 11, 2021Watch a recording of the presentation and Q&A roundtable featuring Amazon scientists and scholars.