Search Results for author: Ashok Vardhan Makkuva

Found 11 papers, 4 papers with code

Attention with Markov: A Framework for Principled Analysis of Transformers via Markov Chains

1 code implementation6 Feb 2024 Ashok Vardhan Makkuva, Marco Bondaschi, Adway Girish, Alliot Nagle, Martin Jaggi, Hyeji Kim, Michael Gastpar

Inspired by the Markovianity of natural languages, we model the data as a Markovian source and utilize this framework to systematically study the interplay between the data-distributional properties, the transformer architecture, the learnt distribution, and the final model performance.

Machine Learning-Aided Efficient Decoding of Reed-Muller Subcodes

no code implementations16 Jan 2023 Mohammad Vahid Jamali, Xiyang Liu, Ashok Vardhan Makkuva, Hessam Mahdavifar, Sewoong Oh, Pramod Viswanath

Next, we derive the soft-decision based version of our algorithm, called soft-subRPA, that not only improves upon the performance of subRPA but also enables a differentiable decoding algorithm.

CRISP: Curriculum based Sequential Neural Decoders for Polar Code Family

1 code implementation1 Oct 2022 S Ashwin Hebbar, Viraj Nadkarni, Ashok Vardhan Makkuva, Suma Bhat, Sewoong Oh, Pramod Viswanath

We design a principled curriculum, guided by information-theoretic insights, to train CRISP and show that it outperforms the successive-cancellation (SC) decoder and attains near-optimal reliability performance on the Polar(32, 16) and Polar(64, 22) codes.

KO codes: Inventing Nonlinear Encoding and Decoding for Reliable Wireless Communication via Deep-learning

1 code implementation29 Aug 2021 Ashok Vardhan Makkuva, Xiyang Liu, Mohammad Vahid Jamali, Hessam Mahdavifar, Sewoong Oh, Pramod Viswanath

In this paper, we construct KO codes, a computationaly efficient family of deep-learning driven (encoder, decoder) pairs that outperform the state-of-the-art reliability performance on the standardized AWGN channel.

Benchmarking

Reed-Muller Subcodes: Machine Learning-Aided Design of Efficient Soft Recursive Decoding

no code implementations2 Feb 2021 Mohammad Vahid Jamali, Xiyang Liu, Ashok Vardhan Makkuva, Hessam Mahdavifar, Sewoong Oh, Pramod Viswanath

To lower the complexity of our decoding algorithm, referred to as subRPA in this paper, we investigate different ways for pruning the projections.

Information Theory Information Theory

Towards Principled Objectives for Contrastive Disentanglement

no code implementations25 Sep 2019 Anwesa Choudhuri, Ashok Vardhan Makkuva, Ranvir Rana, Sewoong Oh, Girish Chowdhary, Alexander Schwing

%In fact, contrastive disentanglement and unsupervised recovery are often combined in that we seek additional variations that exhibit salient factors/properties.

Disentanglement

Optimal transport mapping via input convex neural networks

2 code implementations ICML 2020 Ashok Vardhan Makkuva, Amirhossein Taghvaei, Sewoong Oh, Jason D. Lee

Building upon recent advances in the field of input convex neural networks, we propose a new framework where the gradient of one convex function represents the optimal transport mapping.

Learning in Gated Neural Networks

no code implementations6 Jun 2019 Ashok Vardhan Makkuva, Sewoong Oh, Sreeram Kannan, Pramod Viswanath

Gating is a key feature in modern neural networks including LSTMs, GRUs and sparsely-gated deep neural networks.

Learning One-hidden-layer Neural Networks under General Input Distributions

no code implementations9 Oct 2018 Weihao Gao, Ashok Vardhan Makkuva, Sewoong Oh, Pramod Viswanath

Significant advances have been made recently on training neural networks, where the main challenge is in solving an optimization problem with abundant critical points.

Breaking the gridlock in Mixture-of-Experts: Consistent and Efficient Algorithms

no code implementations21 Feb 2018 Ashok Vardhan Makkuva, Sewoong Oh, Sreeram Kannan, Pramod Viswanath

Once the experts are known, the recovery of gating parameters still requires an EM algorithm; however, we show that the EM algorithm for this simplified problem, unlike the joint EM algorithm, converges to the true parameters.

Ensemble Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.