-
2024In this paper, we introduce AdaSelection, an adaptive sub-sampling method to identify the most informative sub-samples within each minibatch to speed up the training of large-scale deep learning models without sacrificing model performance. Our method is able to flexibly combines an arbitrary number of baseline sub-sampling methods incorporating the method-level importance and intra-method sample-level
-
2024Pre-trained language models, trained on largescale corpora, demonstrate strong generalizability across various NLP tasks. Finetuning these models for specific tasks typically involves updating all parameters, which is resource-intensive. Parameter-efficient finetuning (PEFT) methods, such as the popular LoRA family, introduce low-rank matrices to learn only a few parameters efficiently. However, during
-
2024Code generation models are not robust to small perturbations, which often lead to incorrect generations and significantly degrade the performance of these models. Although improving the robustness of code generation models is crucial to enhancing user experience in real-world applications, existing research efforts do not address this issue. To fill this gap, we propose CodeFort, a framework to improve
-
2024Hierarchical Text Classification (HTC) is a sub-class of multi-label classification. It is challenging because the hierarchy typically has a large number of diverse topics. Existing methods for HTC fall within two categories, local methods (a classifier for each level, node, or parent) or global methods (a single classifier for everything). Local methods are computationally expensive, whereas global methods
-
2024Warning: this paper contains content that may be inappropriate or offensive. As generative models become available for public use in various applications, testing and analyzing vulnerabilities of these models has become a priority. In this work, we propose an automatic red teaming framework that evaluates a given black-box model and exposes its vulnerabilities against unsafe and inappropriate content generation
Related content
-
April 1, 2021Why conditional demographic disparity matters for developers using SageMaker Clarify.
-
March 30, 2021Learn how Bill Smart wants to simplify the ways that robots and people work together — and why waiting on a date one night changed his career path.
-
March 29, 2021Amazon distinguished scientist and conference general chair Alex Smola on what makes MLSys unique — both thematically and culturally.
-
March 23, 2021Politecnico di Milano professor Stefano Ceri is working to integrate genomic datasets into a single accessible system with the support of an Amazon Machine Learning Research Award.
-
March 16, 2021Amanda Cullen, a PhD candidate in informatics at the University of California, Irvine, wanted to do work that had an impact outside of academia — she found an ideal opportunity at Twitch.
-
March 10, 2021Exploring and analyzing possible techniques to make ML algorithms capable of learning fairer models by utilizing empirical risk minimization theory.