-
INLG 20232023In this work, we investigate Data Augmentation methods to improve the performance of state-of-the-art models for four different downstream tasks. Specifically, we propose Generative Adversarial Network using Language Models (GAN-LM) approach that combines a deep generative model with a pre-trained language model to produce diverse augmentations. We compare the GAN-LM to various conventional methods in non-contextual
-
ICCV 20232023Text-to-video retrieval systems have recently made significant progress by utilizing pre-trained models trained on large-scale image-text pairs. However, most of the latest methods primarily focus on the video modality while disregarding the audio signal for this task. Nevertheless, a recent advancement by ECLIPSE has improved long-range text-to-video retrieval by developing an audiovisual video representation
-
Graph-aware language model pre-training on a large graph corpus can help multiple graph applicationsKDD 20232023Model pre-training on large text corpora has been demonstrated effective for various downstream applications in the NLP domain. In the graph mining domain, a similar analogy can be drawn for pre-training graph models on large graphs in the hope of benefiting downstream graph applications, which has also been explored by several recent studies. However, no existing study has ever investigated the pre-training
-
ICLR 20232023Predicting the responses of a cell under perturbations may bring important benefits to drug discovery and personalized therapeutics. In this work, we propose a novel graph variational Bayesian causal inference framework to predict a cell’s gene expressions under counterfactual perturbations (perturbations that this cell did not factually receive), leveraging information representing biological knowledge
-
KDD 20232023Pre-trained language models (PLMs) such as BERT, RoBERTa, and DeBERTa have achieved state-of-the-art performance on various downstream tasks. The enormous sizes of PLMs hinder their deployment in resource-constrained scenarios, e.g., on edge and mobile devices. To address this issue, many model compression approaches have been proposed to reduce the number of model parameters. This paper focuses on compressing
Related content
-
August 19, 2021How Minghui He turned her Amazon internship into a full-time research scientist role.
-
July 30, 2021“Anytime query” approach adapts to the available resources.
-
July 12, 2021New method uses cross-attention and multitask training to improve the accuracy and training efficiency of video moment retrieval.
-
July 08, 2021Amazon Visiting Academic Barbara Poblete helps to build safer, more-diverse online communities — and to aid disaster response.
-
June 23, 2021A new method based on Transformers and trained with self-supervised learning achieves state-of-the-art performance.
-
June 16, 2021Relative to human evaluation of question-answering models, the new method has an error rate of only 7%.