Search Results for author: Akash Srivastava

Found 50 papers, 23 papers with code

Sculpting Subspaces: Constrained Full Fine-Tuning in LLMs for Continual Learning

1 code implementation9 Apr 2025 Nikhil Shivakumar Nayak, KrishnaTeja Killamsetty, Ligong Han, Abhishek Bhandwaldar, Prateek Chanda, Kai Xu, Hao Wang, Aldo Pareja, Oleg Silkin, Mustafa Eyceoz, Akash Srivastava

Continual learning in large language models (LLMs) is prone to catastrophic forgetting, where adapting to new tasks significantly degrades performance on previously learned ones.

Continual Learning Decoder +1

SQuat: Subspace-orthogonal KV Cache Quantization

no code implementations31 Mar 2025 Hao Wang, Ligong Han, Kai Xu, Akash Srivastava

The key-value (KV) cache accelerates LLMs decoding by storing KV tensors from previously generated tokens.

Quantization

Activation-Informed Merging of Large Language Models

1 code implementation4 Feb 2025 Amin Heyrani Nobari, Kaveh Alimohammadi, Ali ArjomandBigdeli, Akash Srivastava, Faez Ahmed, Navid Azizan

Model merging, a method that combines the parameters and embeddings of multiple fine-tuned large language models (LLMs), offers a promising approach to enhance model performance across various tasks while maintaining computational efficiency.

Computational Efficiency Continual Learning +1

A Probabilistic Inference Approach to Inference-Time Scaling of LLMs using Particle-Based Monte Carlo Methods

1 code implementation3 Feb 2025 Isha Puri, Shivchander Sudalairaj, Guangxuan Xu, Kai Xu, Akash Srivastava

In this paper, we instead cast inference-time scaling as a probabilistic inference task and leverage sampling-based techniques to explore the typical set of the state distribution of a state-space model with an approximate likelihood, rather than optimize for its mode directly.

Math Mathematical Reasoning

Unveiling the Secret Recipe: A Guide For Supervised Fine-Tuning Small LLMs

no code implementations17 Dec 2024 Aldo Pareja, Nikhil Shivakumar Nayak, Hao Wang, KrishnaTeja Killamsetty, Shivchander Sudalairaj, Wenlong Zhao, Seungwook Han, Abhishek Bhandwaldar, Guangxuan Xu, Kai Xu, Ligong Han, Luke Inglis, Akash Srivastava

The rise of large language models (LLMs) has created a significant disparity: industrial research labs with their computational resources, expert teams, and advanced infrastructures, can effectively fine-tune LLMs, while individual developers and small organizations face barriers due to limited resources.

MMLU

Dr. SoW: Density Ratio of Strong-over-weak LLMs for Reducing the Cost of Human Annotation in Preference Tuning

no code implementations4 Nov 2024 Guangxuan Xu, Kai Xu, Shivchander Sudalairaj, Hao Wang, Akash Srivastava

In this paper, we introduce Dr. SoW (Density Ratio of Strong over Weak) a cost-effective method that eliminates the reliance for human annotation by leveraging off-the-shelf LLMs for preference data annotation.

Urban context and delivery performance: Modelling service time for cargo bikes and vans across diverse urban environments

no code implementations27 Aug 2024 Maxwell Schrader, Navish Kumar, Esben Sørig, Soonmyeong Yoon, Akash Srivastava, Kai Xu, Maria Astefanoaei, Nicolas Collignon

We introduce two datasets that allow for in-depth analysis and modelling of service times of cargo bikes and use existing datasets to reason about differences in delivery performance across vehicle types.

LInK: Learning Joint Representations of Design and Performance Spaces through Contrastive Learning for Mechanism Synthesis

1 code implementation31 May 2024 Amin Heyrani Nobari, Akash Srivastava, Dan Gutfreund, Kai Xu, Faez Ahmed

Moreover, we introduce a significantly more challenging benchmark, named LINK ABC, which involves synthesizing linkages that trace the trajectories of English capital alphabets, an inverse design benchmark task that existing methods struggle with due to large nonlinearities and tiny feasible space.

Contrastive Learning

Differentially Private Synthetic Data Generation for Relational Databases

1 code implementation29 May 2024 Kaveh Alimohammadi, Hao Wang, Ojas Gulati, Akash Srivastava, Navid Azizan

Existing differentially private (DP) synthetic data generation mechanisms typically assume a single-source table.

Synthetic Data Generation

Value Augmented Sampling for Language Model Alignment and Personalization

1 code implementation10 May 2024 Seungwook Han, Idan Shenfeld, Akash Srivastava, Yoon Kim, Pulkit Agrawal

Aligning Large Language Models (LLMs) to cater to different human preferences, learning new skills, and unlearning harmful behavior is an important problem.

Language Modeling Language Modelling +1

LAB: Large-Scale Alignment for ChatBots

1 code implementation2 Mar 2024 Shivchander Sudalairaj, Abhishek Bhandwaldar, Aldo Pareja, Kai Xu, David D. Cox, Akash Srivastava

This work introduces LAB (Large-scale Alignment for chatBots), a novel methodology designed to overcome the scalability challenges in the instruction-tuning phase of large language model (LLM) training.

Instruction Following Language Modeling +3

Curiosity-driven Red-teaming for Large Language Models

1 code implementation29 Feb 2024 Zhang-Wei Hong, Idan Shenfeld, Tsun-Hsuan Wang, Yung-Sung Chuang, Aldo Pareja, James Glass, Akash Srivastava, Pulkit Agrawal

To probe when an LLM generates unwanted content, the current paradigm is to recruit a \textit{red team} of human testers to design input prompts (i. e., test cases) that elicit undesirable responses from LLMs.

Red Teaming Reinforcement Learning (RL)

Learning to Deliver: a Foundation Model for the Montreal Capacitated Vehicle Routing Problem

no code implementations28 Feb 2024 Samuel J. K. Chin, Matthias Winkenbach, Akash Srivastava

In this paper, we present the Foundation Model for the Montreal Capacitated Vehicle Routing Problem (FM-MCVRP), a novel Deep Learning (DL) model that approximates high-quality solutions to a variant of the Capacitated Vehicle Routing Problem (CVRP) that characterizes many real-world applications.

Language Modelling Large Language Model

Private Synthetic Data Meets Ensemble Learning

no code implementations15 Oct 2023 Haoyuan Sun, Navid Azizan, Akash Srivastava, Hao Wang

When machine learning models are trained on synthetic data and then deployed on real data, there is often a performance drop due to the distribution shift between synthetic and real data.

Diversity Ensemble Learning

Constraining Generative Models for Engineering Design with Negative Data

1 code implementation27 Jun 2023 Lyle Regenwetter, Giorgio Giannone, Akash Srivastava, Dan Gutfreund, Faez Ahmed

Our negative-data generative model (NDGM) formulation easily outperforms classic models, generating 1/6 as many constraint-violating samples using 1/8 as much data in certain problems.

Diversity valid

A Probabilistic Framework for Modular Continual Learning

1 code implementation11 Jun 2023 Lazar Valkov, Akash Srivastava, Swarat Chaudhuri, Charles Sutton

Comparing to a wide range of approaches, we show that PICLE is the first modular CL algorithm to achieve perceptual, few-shot and latent transfer while scaling well to large search spaces, outperforming previous state-of-the-art modular CL approaches on long problem sequences.

Continual Learning

Improving Tuning-Free Real Image Editing with Proximal Guidance

1 code implementation8 Jun 2023 Ligong Han, Song Wen, Qi Chen, Zhixing Zhang, Kunpeng Song, Mengwei Ren, Ruijiang Gao, Anastasis Stathopoulos, Xiaoxiao He, Yuxiao Chen, Di Liu, Qilong Zhangli, Jindong Jiang, Zhaoyang Xia, Akash Srivastava, Dimitris Metaxas

Null-text inversion (NTI) optimizes null embeddings to align the reconstruction and inversion trajectories with larger CFG scales, enabling real image editing with cross-attention control.

Estimating the Density Ratio between Distributions with High Discrepancy using Multinomial Logistic Regression

no code implementations1 May 2023 Akash Srivastava, Seungwook Han, Kai Xu, Benjamin Rhodes, Michael U. Gutmann

We show that if these auxiliary densities are constructed such that they overlap with $p$ and $q$, then a multi-class logistic regression allows for estimating $\log p/q$ on the domain of any of the $K+2$ distributions and resolves the distribution shift problems of the current state-of-the-art methods.

Binary Classification Density Ratio Estimation +4

Multi-Symmetry Ensembles: Improving Diversity and Generalization via Opposing Symmetries

1 code implementation4 Mar 2023 Charlotte Loh, Seungwook Han, Shivchander Sudalairaj, Rumen Dangovski, Kai Xu, Florian Wenzel, Marin Soljacic, Akash Srivastava

In this work, we present Multi-Symmetry Ensembles (MSE), a framework for constructing diverse ensembles by capturing the multiplicity of hypotheses along symmetry axes, which explore the hypothesis space beyond stochastic perturbations of model weights and hyperparameters.

Diversity Representation Learning +1

Beyond Statistical Similarity: Rethinking Metrics for Deep Generative Models in Engineering Design

no code implementations6 Feb 2023 Lyle Regenwetter, Akash Srivastava, Dan Gutfreund, Faez Ahmed

This paper doubles as a review and a practical guide to evaluation metrics for deep generative models (DGMs) in engineering design.

Drug Discovery Learning Theory +1

On the Importance of Calibration in Semi-supervised Learning

no code implementations10 Oct 2022 Charlotte Loh, Rumen Dangovski, Shivchander Sudalairaj, Seungwook Han, Ligong Han, Leonid Karlinsky, Marin Soljacic, Akash Srivastava

State-of-the-art (SOTA) semi-supervised learning (SSL) methods have been highly successful in leveraging a mix of labeled and unlabeled data by combining techniques of consistency regularization and pseudo-labeling.

LINKS: A dataset of a hundred million planar linkage mechanisms for data-driven kinematic design

1 code implementation30 Aug 2022 Amin Heyrani Nobari, Akash Srivastava, Dan Gutfreund, Faez Ahmed

LINKS is made up of various components including 100 million mechanisms, the simulation data for each mechanism, normalized paths generated by each mechanism, a curated set of paths, the code used to generate the data and simulate mechanisms, and a live web demo for interactive design of linkage mechanisms.

Retrieval

A Bayesian-Symbolic Approach to Reasoning and Learning in Intuitive Physics

no code implementations NeurIPS 2021 Kai Xu, Akash Srivastava, Dan Gutfreund, Felix Sosa, Tomer Ullman, Josh Tenenbaum, Charles Sutton

In this paper, we propose a Bayesian-symbolic framework (BSP) for physical reasoning and learning that is close to human-level sample-efficiency and accuracy.

Bayesian Inference Bilevel Optimization +3

Targeted Neural Dynamical Modeling

3 code implementations NeurIPS 2021 Cole Hurwitz, Akash Srivastava, Kai Xu, Justin Jude, Matthew G. Perich, Lee E. Miller, Matthias H. Hennig

These approaches, however, are limited in their ability to capture the underlying neural dynamics (e. g. linear) and in their ability to relate the learned dynamics back to the observed behaviour (e. g. no time lag).

Decoder

Equivariant Contrastive Learning

2 code implementations28 Oct 2021 Rumen Dangovski, Li Jing, Charlotte Loh, Seungwook Han, Akash Srivastava, Brian Cheung, Pulkit Agrawal, Marin Soljačić

In state-of-the-art self-supervised learning (SSL) pre-training produces semantically good representations by encouraging them to be invariant under meaningful transformations prescribed from human knowledge.

Contrastive Learning Self-Supervised Learning

Equivariant Self-Supervised Learning: Encouraging Equivariance in Representations

no code implementations ICLR 2022 Rumen Dangovski, Li Jing, Charlotte Loh, Seungwook Han, Akash Srivastava, Brian Cheung, Pulkit Agrawal, Marin Soljacic

In state-of-the-art self-supervised learning (SSL) pre-training produces semantically good representations by encouraging them to be invariant under meaningful transformations prescribed from human knowledge.

Self-Supervised Learning

Scaling Densities For Improved Density Ratio Estimation

no code implementations29 Sep 2021 Akash Srivastava, Seungwook Han, Benjamin Rhodes, Kai Xu, Michael U. Gutmann

As such, estimating density ratios accurately using only samples from $p$ and $q$ is of high significance and has led to a flurry of recent work in this direction.

Binary Classification Density Ratio Estimation

A Bayesian-Symbolic Approach to Learning and Reasoning for Intuitive Physics

no code implementations1 Jan 2021 Kai Xu, Akash Srivastava, Dan Gutfreund, Felix Sosa, Tomer Ullman, Joshua B. Tenenbaum, Charles Sutton

As such, learning the laws is then reduced to symbolic regression and Bayesian inference methods are used to obtain the distribution of unobserved properties.

Bayesian Inference Common Sense Reasoning +2

Sequential Transfer Machine Learning in Networks: Measuring the Impact of Data and Neural Net Similarity on Transferability

no code implementations29 Mar 2020 Robin Hirt, Akash Srivastava, Carlos Berg, Niklas Kühl

As the number of data sets in business networks grows and not every neural net transfer is successful, indicators are needed for its impact on the target performance-its transferability.

CZ-GEM: A FRAMEWORK FOR DISENTANGLED REPRESENTATION LEARNING

no code implementations ICLR 2020 Akash Srivastava, Yamini Bansal, Yukun Ding, Bernhard Egger, Prasanna Sattigeri, Josh Tenenbaum, David D. Cox, Dan Gutfreund

In this work, we tackle a slightly more intricate scenario where the observations are generated from a conditional distribution of some known control variate and some latent noise variate.

Disentanglement

BreGMN: scaled-Bregman Generative Modeling Networks

no code implementations1 Jun 2019 Akash Srivastava, Kristjan Greenewald, Farzaneh Mirzazadeh

Well-definedness of f-divergences, however, requires the distributions of the data and model to overlap completely in every time step of training.

Generative Ratio Matching Networks

no code implementations ICLR 2020 Akash Srivastava, Kai Xu, Michael U. Gutmann, Charles Sutton

In this work, we take their insight of using kernels as fixed adversaries further and present a novel method for training deep generative models that does not involve saddlepoint optimization.

Variational Inference In Pachinko Allocation Machines

no code implementations21 Apr 2018 Akash Srivastava, Charles Sutton

The Pachinko Allocation Machine (PAM) is a deep topic model that allows representing rich correlation structures among topics by a directed acyclic graph over topics.

Variational Inference

HOUDINI: Lifelong Learning as Program Synthesis

2 code implementations NeurIPS 2018 Lazar Valkov, Dipak Chaudhari, Akash Srivastava, Charles Sutton, Swarat Chaudhuri

We present a neurosymbolic framework for the lifelong learning of algorithmic tasks that mix perception and procedural reasoning.

Lifelong learning Program Synthesis +1

VEEGAN: Reducing Mode Collapse in GANs using Implicit Variational Learning

1 code implementation NeurIPS 2017 Akash Srivastava, Lazar Valkov, Chris Russell, Michael U. Gutmann, Charles Sutton

Deep generative models provide powerful tools for distributions over complicated manifolds, such as those of natural images.

Autoencoding Variational Inference For Topic Models

6 code implementations4 Mar 2017 Akash Srivastava, Charles Sutton

A promising approach to address this problem is autoencoding variational Bayes (AEVB), but it has proven diffi- cult to apply to topic models in practice.

Topic Models Variational Inference

Clustering with a Reject Option: Interactive Clustering as Bayesian Prior Elicitation

no code implementations19 Jun 2016 Akash Srivastava, James Zou, Ryan P. Adams, Charles Sutton

A good clustering can help a data analyst to explore and understand a data set, but what constitutes a good clustering may depend on domain-specific and application-specific criteria.

Clustering

Clustering with a Reject Option: Interactive Clustering as Bayesian Prior Elicitation

no code implementations22 Feb 2016 Akash Srivastava, James Zou, Charles Sutton

A good clustering can help a data analyst to explore and understand a data set, but what constitutes a good clustering may depend on domain-specific and application-specific criteria.

Clustering Computational Efficiency

Cannot find the paper you are looking for? You can Submit a new open access paper.