no code implementations • ICML 2020 • Benjamin Coleman, Anshumali Shrivastava, Richard Baraniuk
We present the first sublinear memory sketch that can be queried to find the nearest neighbors in a dataset.
1 code implementation • 9 Mar 2025 • Alexander Scarlatos, Naiming Liu, Jaewook Lee, Richard Baraniuk, Andrew Lan
Specifically, we generate a set of candidate tutor utterances and score them using (1) an LLM-based student model to predict the chance of correct student responses and (2) a pedagogical rubric evaluated by GPT-4o.
no code implementations • 29 Aug 2024 • Sina AlEMohammad, Ahmed Imtiaz Humayun, Shruti Agarwal, John Collomosse, Richard Baraniuk
Unfortunately, training new generative models with synthetic data from current or past generation models creates an autophagous (self-consuming) loop that degrades the quality and/or diversity of the synthetic data in what has been termed model autophagy disorder (MAD) and model collapse.
Ranked #1 on
Image Generation
on ImageNet 64x64
no code implementations • 9 Aug 2024 • Randall Balestriero, Ahmed Imtiaz Humayun, Richard Baraniuk
In this paper, we overview one promising avenue of progress at the mathematical foundation of deep learning: the connection between deep networks and function approximation by affine splines (continuous piecewise linear functions in multiple dimensions).
1 code implementation • 1 Jul 2024 • Naiming Liu, Shashank Sonkar, MyCo Le, Richard Baraniuk
We propose the Malgorithm Identification task, where LLMs are assessed based on their ability to identify corresponding malgorithm given an incorrect answer choice.
no code implementations • 19 Jun 2024 • Naiming Liu, Zichao Wang, Richard Baraniuk
Despite rapid advancements in large language models (LLMs), QG remains a challenging problem due to its complicated process, open-ended nature, and the diverse settings in which question generation occurs.
1 code implementation • 14 Jun 2024 • Omer Ronen, Ahmed Imtiaz Humayun, Randall Balestriero, Richard Baraniuk, Bin Yu
We develop Scalable Latent Exploration Score (ScaLES) to mitigate over-exploration in Latent Space Optimization (LSO), a popular method for solving black-box discrete optimization problems.
1 code implementation • 23 Feb 2024 • Ahmed Imtiaz Humayun, Randall Balestriero, Richard Baraniuk
Grokking, or delayed generalization, is a phenomenon where generalization in a deep neural network (DNN) occurs long after achieving near zero training error.
no code implementations • 7 Dec 2023 • Micah Goldblum, Anima Anandkumar, Richard Baraniuk, Tom Goldstein, Kyunghyun Cho, Zachary C Lipton, Melanie Mitchell, Preetum Nakkiran, Max Welling, Andrew Gordon Wilson
The goal of this series is to chronicle opinions and issues in the field of machine learning as they stand today and as they change over time.
no code implementations • 19 Oct 2023 • Ahmed Imtiaz Humayun, Randall Balestriero, Richard Baraniuk
First, we present a novel statistic that encompasses the local complexity (LC) of the DN based on the concentration of linear regions inside arbitrary dimensional neighborhoods around data points.
1 code implementation • 7 Jul 2023 • Zichao Wang, Richard Baraniuk
We study the new problem of automatic question generation (QG) from multi-modal sources containing images and texts, significantly expanding the scope of most of the existing work that focuses exclusively on QG from only textual sources.
1 code implementation • CVPR 2023 • Ahmed Imtiaz Humayun, Randall Balestriero, Guha Balakrishnan, Richard Baraniuk
In this paper, we go one step further by developing the first provably exact method for computing the geometry of a DN's mapping - including its decision boundary - over a specified region of the data space.
no code implementations • 2 Feb 2023 • Fernando Gama, Nicolas Zilberstein, Martin Sevilla, Richard Baraniuk, Santiago Segarra
Thus, the crux of particle filters lies in designing sampling distributions that are both easy to sample from and lead to accurate estimators.
2 code implementations • 23 Aug 2022 • Zichao Wang, Weili Nie, Zhuoran Qiao, Chaowei Xiao, Richard Baraniuk, Anima Anandkumar
On various tasks ranging from simple design criteria to a challenging real-world scenario for designing lead compounds that bind to the SARS-CoV-2 main protease, we demonstrate our approach extrapolates well beyond the retrieval database, and achieves better performance and wider applicability than previous methods.
1 code implementation • 19 May 2022 • Nigel Fernandez, Aritra Ghosh, Naiming Liu, Zichao Wang, Benoît Choffin, Richard Baraniuk, Andrew Lan
Our approach, in-context BERT fine-tuning, produces a single shared scoring model for all items with a carefully-designed input structure to provide contextual information on each item.
1 code implementation • CVPR 2022 • Gowthami Somepalli, Liam Fowl, Arpit Bansal, Ping Yeh-Chiang, Yehuda Dar, Richard Baraniuk, Micah Goldblum, Tom Goldstein
We also use decision boundary methods to visualize double descent phenomena.
1 code implementation • 4 Mar 2022 • Ahmed Imtiaz Humayun, Randall Balestriero, Anastasios Kyrillidis, Richard Baraniuk
We propose to remedy such a scenario by introducing a maximal radius constraint $r$ on the clusters formed by the centroids, i. e., samples from the same cluster should not be more than $2r$ apart in terms of $\ell_2$ distance.
1 code implementation • CVPR 2022 • Ahmed Imtiaz Humayun, Randall Balestriero, Richard Baraniuk
We present Polarity Sampling, a theoretically justified plug-and-play method for controlling the generation quality and diversity of pre-trained deep generative networks DGNs).
Ranked #1 on
Image Generation
on LSUN Car 512 x 384
no code implementations • 16 Feb 2022 • Romain Cosentino, Randall Balestriero, Yanis Bahroun, Anirvan Sengupta, Richard Baraniuk, Behnaam Aazhang
This enables (i) the reduction of intrinsic nuisances associated with the data, reducing the complexity of the clustering task and increasing performances and producing state-of-the-art results, (ii) clustering in the input space of the data, leading to a fully interpretable clustering algorithm, and (iii) the benefit of convergence guarantees.
no code implementations • 25 Oct 2021 • Hossein Babaei, Sina AlEMohammad, Richard Baraniuk
Covariate balancing methods increase the similarity between the distributions of the two groups' covariates.
1 code implementation • ICLR 2022 • Ahmed Imtiaz Humayun, Randall Balestriero, Richard Baraniuk
Deep Generative Networks (DGNs) are extensively employed in Generative Adversarial Networks (GANs), Variational Autoencoders (VAEs), and their variants to approximate the data manifold and distribution.
Ranked #4 on
Image Generation
on FFHQ 1024 x 1024
no code implementations • 29 Sep 2021 • Aditya Desai, Shashank Sonkar, Anshumali Shrivastava, Richard Baraniuk
Grounded on this framework, we show that many algorithms ranging across different domains are, in fact, searching for continuous stable coloring solutions of an underlying graph corresponding to the domain.
no code implementations • 29 Sep 2021 • Zichao Wang, Weili Nie, Zhenwei Dai, Richard Baraniuk
Many existing approaches either require extensive training/fine-tuning of the LM for each single attribute under control or are slow to generate text.
1 code implementation • 18 Aug 2021 • Vishwanath Saragadam, Akshat Dave, Ashok Veeraraghavan, Richard Baraniuk
We introduce DeepIR, a new thermal image processing framework that combines physically accurate sensor modeling with deep network-based image representation.
no code implementations • AKBC 2021 • Vinay K. Chaudhri, Matthew Boggess, Han Lin Aung, Debshila Basu Mallick, Andrew C Waters, Richard Baraniuk
Ontology graphs are graphs in which the nodes are generic classes and edges have labels that specify the relationships between the classes.
no code implementations • 7 Jun 2021 • Lorenzo Luzi, Yehuda Dar, Richard Baraniuk
We show that overparameterization can improve generalization performance and accelerate the training process.
no code implementations • 25 Apr 2021 • Mengxue Zhang, Zichao Wang, Richard Baraniuk, Andrew Lan
Feedback on student answers and even during intermediate steps in their solutions to open-ended questions is an important element in math education.
no code implementations • 1 Apr 2021 • Randall Balestriero, Richard Baraniuk
Jacobian-vector products (JVPs) form the backbone of many recent developments in Deep Networks (DNs), with applications including faster constrained optimization, regularization with generalization guarantees, and adversarial example sensitivity assessments.
1 code implementation • 7 Jan 2021 • Haoran You, Randall Balestriero, Zhihan Lu, Yutong Kou, Huihong Shi, Shunyao Zhang, Shang Wu, Yingyan Lin, Richard Baraniuk
In this paper, we study the importance of pruning in Deep Networks (DNs) and the yin & yang relationship between (1) pruning highly overparametrized DNs that have been trained from random initialization and (2) training small DNs that have been "cleverly" initialized.
1 code implementation • 28 Dec 2020 • Vishwanath Saragadam, Michael DeZeeuw, Richard Baraniuk, Ashok Veeraraghavan, Aswin Sankaranarayanan
Hence, a scene-adaptive spatial sampling of an hyperspectral scene, guided by its super-pixel segmented image, is capable of obtaining high-quality reconstructions.
no code implementations • 16 Dec 2020 • Romain Cosentino, Randall Balestriero, Yanis Bahroun, Anirvan Sengupta, Richard Baraniuk, Behnaam Aazhang
We design an interpretable clustering algorithm aware of the nonlinear structure of image manifolds.
2 code implementations • 9 Dec 2020 • Sina AlEMohammad, Randall Balestriero, Zichao Wang, Richard Baraniuk
Kernels derived from deep neural networks (DNNs) in the infinite-width regime provide not only high performance in a range of machine learning tasks but also new theoretical insights into DNN training dynamics and generalization.
no code implementations • NeurIPS 2020 • Randall Balestriero, Sebastien Paris, Richard Baraniuk
Deep Generative Networks (DGNs) with probabilistic modeling of their output and latent space are currently trained via Variational Autoencoders (VAEs).
no code implementations • 20 Sep 2020 • Romain Cosentino, Randall Balestriero, Richard Baraniuk, Behnaam Aazhang
Our regularizations leverage recent advances in the group of transformation learning to enable AEs to better approximate the data manifold without explicitly defining the group underlying the manifold.
no code implementations • ICLR 2021 • Sina Al-E-Mohammad, Zichao Wang, Randall Balestriero, Richard Baraniuk
The study of deep neural networks (DNNs) in the infinite-width limit, via the so-called neural tangent kernel (NTK) approach, has provided new insights into the dynamics of learning, generalization, and the impact of initialization.
no code implementations • 27 May 2020 • Zichao Wang, Yi Gu, Andrew Lan, Richard Baraniuk
We propose VarFA, a variational inference factor analysis framework that extends existing factor analysis models for educational data mining to efficiently output uncertainty estimation in the model's estimated factors.
1 code implementation • 26 Feb 2020 • Randall Balestriero, Sebastien Paris, Richard Baraniuk
We also derive the output probability density mapped onto the generated manifold in terms of the latent space density, which enables the computation of key statistics such as its Shannon entropy.
no code implementations • 25 Sep 2019 • Lorenzo Luzi, Randall Balestriero, Richard Baraniuk
We define a goodness of fit measure for generative networks which captures how well the network can generate the training data, which is necessary to learn the true data distribution.
no code implementations • 10 Jul 2019 • Yue Wang, Jianghao Shen, Ting-Kuei Hu, Pengfei Xu, Tan Nguyen, Richard Baraniuk, Zhangyang Wang, Yingyan Lin
State-of-the-art convolutional neural networks (CNNs) yield record-breaking predictive performance, yet at the cost of high-energy-consumption inference, that prohibits their widely deployments in resource-constrained Internet of Things (IoT) applications.
1 code implementation • NeurIPS 2019 • Randall Balestriero, Romain Cosentino, Behnaam Aazhang, Richard Baraniuk
The subdivision process constrains the affine maps on the (exponentially many) power diagram regions to greatly reduce their complexity.
no code implementations • ICLR 2019 • Zichao Wang, Randall Balestriero, Richard Baraniuk
Second, we show that the affine parameter of an RNN corresponds to an input-specific template, from which we can interpret an RNN as performing a simple template matching (matched filtering) given the input.
no code implementations • NIPS Workshop CDNNRIA 2018 • Yue Wang, Tan Nguyen, Yang Zhao, Zhangyang Wang, Yingyan Lin, Richard Baraniuk
The prohibitive energy cost of running high-performance Convolutional Neural Networks (CNNs) has been limiting their deployment on resource-constrained platforms including mobile and wearable devices.
no code implementations • ICML 2018 • Randall Balestriero, Romain Cosentino, Herve Glotin, Richard Baraniuk
We propose to tackle the problem of end-to-end learning for raw waveform signals by introducing learnable continuous time-frequency atoms.
no code implementations • 17 May 2018 • Randall Balestriero, Richard Baraniuk
For instance, conditioned on the input signal, the output of a MASO DN can be written as a simple affine transformation of the input.
no code implementations • 27 Feb 2018 • Randall Balestriero, Herve Glotin, Richard Baraniuk
Deep Neural Networks (DNNs) provide state-of-the-art solutions in several difficult machine perceptual tasks.
no code implementations • 25 Dec 2017 • Romain Cosentino, Randall Balestriero, Richard Baraniuk, Ankit Patel
In this work, we derive a generic overcomplete frame thresholding scheme based on risk minimization.
no code implementations • 25 Oct 2017 • Randall Balestriero, Richard Baraniuk
Deep Neural Networks (DNNs) are universal function approximators providing state-of- the-art solutions on wide range of applications.
2 code implementations • 1 Sep 2015 • M. Salman Asif, Ali Ayremlou, Aswin Sankaranarayanan, Ashok Veeraraghavan, Richard Baraniuk
FlatCam is a thin form-factor lensless camera that consists of a coded mask placed on top of a bare, conventional sensor array.
2 code implementations • 16 Jan 2015 • Tom Goldstein, Christoph Studer, Richard Baraniuk
This is a user manual for the software package FASTA.
Mathematical Software Numerical Analysis Numerical Analysis
no code implementations • 1 Dec 2014 • Ali Ayremlou, Thomas Goldstein, Ashok Veeraraghavan, Richard Baraniuk
Sparse approximations using highly over-complete dictionaries is a state-of-the-art tool for many imaging applications including denoising, super-resolution, compressive sensing, light-field analysis, and object recognition.
4 code implementations • 13 Nov 2014 • Tom Goldstein, Christoph Studer, Richard Baraniuk
Non-differentiable and constrained optimization play a key role in machine learning, signal and image processing, communications, and beyond.
Numerical Analysis G.1.6
no code implementations • NeurIPS 2013 • Divyanshu Vats, Richard Baraniuk
We consider the problem of accurately estimating a high-dimensional sparse vector using a small number of linear measurements that are contaminated by noise.
1 code implementation • 14 Nov 2013 • Tom Goldstein, Lina Xu, Kevin F. Kelly, Richard Baraniuk
Compressed sensing enables the reconstruction of high-resolution signals from under-sampled data.
1 code implementation • 2 May 2013 • Tom Goldstein, Min Li, Xiaoming Yuan, Ernie Esser, Richard Baraniuk
The Primal-Dual hybrid gradient (PDHG) method is a powerful optimization scheme that breaks complex problems into simple sub-steps.
Numerical Analysis 65K15 G.1.6
no code implementations • NeurIPS 2011 • Andrew E. Waters, Aswin C. Sankaranarayanan, Richard Baraniuk
We consider the problem of recovering a matrix $\mathbf{M}$ that is the sum of a low-rank matrix $\mathbf{L}$ and a sparse matrix $\mathbf{S}$ from a small set of linear measurements of the form $\mathbf{y} = \mathcal{A}(\mathbf{M}) = \mathcal{A}({\bf L}+{\bf S})$.
no code implementations • NeurIPS 2008 • Volkan Cevher, Marco F. Duarte, Chinmay Hegde, Richard Baraniuk
Compressive Sensing (CS) combines sampling and compression into a single sub-Nyquist linear measurement process for sparse and compressible signals.
no code implementations • NeurIPS 2007 • Chinmay Hegde, Michael Wakin, Richard Baraniuk
First, we show that with a small number $M$ of {\em random projections} of sample points in $\reals^N$ belonging to an unknown $K$-dimensional Euclidean manifold, the intrinsic dimension (ID) of the sample set can be estimated to high accuracy.