-
2024Semi-supervised dialogue summarization (SSDS) leverages model-generated summaries to reduce reliance on human-labeled data and improve the performance of summarization models. While addressing label noise, previous works on semi-supervised learning primarily focus on natural language understanding tasks, assuming each sample has a unique label. However, these methods are not directly applicable to SSDS,
-
2024Scaling up model and data size has been quite successful for the evolution of LLMs. However, the scaling law for the diffusion based text-to-image (T2I) models is not fully explored. It is also unclear how to efficiently scale the model for better performance at reduced cost. The different training settings and expensive training cost make a fair model comparison extremely difficult. In this work, we empirically
-
2024While the recommendation system (RS) has advanced significantly through deep learning, current RS approaches usually train and finetune models on task-specific datasets, limiting their generalizability to new recommendation tasks and their ability to leverage external knowledge due to model scale and data size constraints. Thus, we designed an LLM-powered autonomous recommender agent, RecMind, which is
-
2024Despite being widely spoken, dialectal variants of languages are frequently considered low in resources due to lack of writing standards and orthographic inconsistencies. As a result, training natural language understanding (NLU) systems relies primarily on standard language resources leading to biased and inequitable NLU technology that underserves dialectal speakers. In this paper, we propose to address
-
Transactions of the Association for Computational Linguistics (TACL)2024Answering factual questions from heterogenous sources, such as graphs and text, is a key capacity of intelligent systems. Current approaches either (i) perform question answering over text and structured sources as separate pipelines followed by a merge step or (ii) provide an early integration giving up the strengths of particular information sources. To solve this problem, we present "HumanIQ", a method
Related content
-
July 15, 2021Hirschberg explains why mastering empathetic speech is critical for successful dialogue systems.
-
July 15, 2021The paper, which received honorable mention at EACL, presents guidelines for better analysis and construction of datasets.
-
July 12, 2021New method uses cross-attention and multitask training to improve the accuracy and training efficiency of video moment retrieval.
-
July 09, 2021The conference’s mission is to bring together stakeholders working toward improving the truthfulness and trustworthiness of online communications.
-
July 08, 2021Amazon Visiting Academic Barbara Poblete helps to build safer, more-diverse online communities — and to aid disaster response.
-
July 02, 2021Giving a neural generation model “control knobs” enables modulation of the content of generated language.