-
2024Posterior sampling in contextual bandits with a Gaussian prior can be implemented exactly or approximately using the Laplace approximation. The Gaussian prior is computationally efficient but it cannot describe complex distributions. In this work, we propose approximate posterior sampling algorithms for contextual bandits with a diffusion model prior. The key idea is to sample from a chain of approximate
-
2024In the domain of code generation, self-debugging is crucial. It allows LLMs to refine their generated code based on execution feedback. This is particularly important because generating correct solutions in one attempt proves challenging for complex tasks. Prior works on self-debugging mostly focus on prompting methods by providing LLMs with few-shot examples, which work poorly on small open-sourced LLMs
-
RecSys 2024 Workshop on Context-Aware Recommender Systems2024Sequential recommendation systems often struggle to make predictions or take action when dealing with cold-start items that have limited amount of interactions. In this work, we propose SimRec – a new approach to mitigate the cold-start problem in sequential recommendation systems. SimRec addresses this challenge by leveraging the inherent similarity among items, incorporating item similarities into the
-
MLTEC 20242024The increasing popularity of wireless sensing applications has led to a growing demand for large datasets of realistic wireless data. However, collecting such wireless data is often time-consuming and expensive. To address this challenge, we propose a synthetic data generation pipeline using human mesh generated from videos that can generate data at scale. The pipeline first generates a 3D mesh of the human
-
2024Fine-tuning large language models (LLMs) has achieved remarkable performance across various natural language processing tasks, yet it demands more and more memory as model sizes keep growing. To address this issue, the recently proposed Memory-efficient Zeroth-order (MeZO) methods attempt to fine-tune LLMs using only forward passes, thereby avoiding the need for a backpropagation graph. However, significant
Related content
-
July 13, 2022Allowing separate tasks to converge on their own schedules and using knowledge distillation to maintain performance improves accuracy.
-
July 12, 2022Fun visual essays explain key concepts of machine learning.
-
July 07, 2022Walid’s 2010 paper on distributed caching algorithms for content distribution networks cited for its “significant impact on the research community”.
-
July 06, 2022Expanded program aimed at engineering undergraduate and graduate students builds off the success of inaugural program.
-
June 28, 2022Amazon’s TabTransformer model is now available through SageMaker JumpStart and the official release of the Keras open-source library.
-
June 24, 2022Technique that mixes public and private training data can meet differential-privacy criteria while cutting error increase by 60%-70%.