no code implementations • 13 Mar 2024 • Ziyuan Lin, Deanna Needell
By doing so, our model can learn both linear and nonlinear similarity information and automatically generate the most appropriate kernel.
no code implementations • 11 Mar 2024 • Kedar Karhadkar, Erin George, Michael Murray, Guido Montúfar, Deanna Needell
The problem of benign overfitting asks whether it is possible for a model to perfectly fit noisy training data and still generalize well.
no code implementations • 2 Mar 2024 • Halyun Jeong, Deanna Needell, Elizaveta Rebrova
We propose SGD-exp, a stochastic gradient descent approach for linear and ReLU regressions under Massart noise (adversarial semi-random corruption model) for the fully streaming setting.
no code implementations • 16 Dec 2023 • Yuchen Li, Laura Balzano, Deanna Needell, Hanbaek Lyu
Block majorization-minimization (BMM) is a simple iterative algorithm for nonconvex optimization that sequentially minimizes a majorizing surrogate of the objective function in each block coordinate while the other block coordinates are held fixed.
no code implementations • 17 Nov 2023 • James Chapman, Yotam Yaniv, Deanna Needell
Non-negative matrix factorization (NMF) is an important technique for obtaining low dimensional representations of datasets.
1 code implementation • 8 Jul 2023 • Joyce Chew, Edward De Brouwer, Smita Krishnaswamy, Deanna Needell, Michael Perlmutter
We introduce a class of manifold neural networks (MNNs) that we call Manifold Filter-Combine Networks (MFCNs), that aims to further our understanding of MNNs, analogous to how the aggregate-combine framework helps with the understanding of graph neural networks (GNNs).
no code implementations • 7 Jun 2023 • Rachel Grotheer, Shuang Li, Anna Ma, Deanna Needell, Jing Qin
Sparse signal recovery is one of the most fundamental problems in various applications, including medical imaging and remote sensing.
no code implementations • 23 May 2023 • Erin George, Joyce Chew, Deanna Needell
To evaluate this method, we perform a series of common tests and demonstrate that measures of bias in the word embeddings are reduced in exchange for minor reduction in the semantic quality of the embeddings.
1 code implementation • 6 May 2023 • HanQin Cai, Zehan Chao, Longxiu Huang, Deanna Needell
We study the tensor robust principal component analysis (TRPCA) problem, a tensorial extension of matrix robust principal component analysis (RPCA), that aims to split the given tensor into an underlying low-rank component and a sparse outlier component.
no code implementations • 20 Apr 2023 • Halyun Jeong, Deanna Needell
The Kaczmarz method (KZ) and its variants, which are types of stochastic gradient descent (SGD) methods, have been extensively studied due to their simplicity and efficiency in solving linear equation systems.
no code implementations • 28 Feb 2023 • Tyler Will, Runyu Zhang, Eli Sadovnik, Mengdi Gao, Joshua Vendrow, Jamie Haddock, Denali Molitor, Deanna Needell
We introduce a new method based on nonnegative matrix factorization, Neural NMF, for detecting latent hierarchical structure in data.
no code implementations • 24 Feb 2023 • Longxiu Huang, Xia Li, Deanna Needell
Additionally, the efficiency of the proposed methods for solving convex problems is shown in simulations with the presence of adversaries.
no code implementations • 20 Feb 2023 • Halyun Jeong, Deanna Needell, Jing Qin
In particular, federated learning (FL) provides such a solution to learn a shared model while keeping training data at local clients.
no code implementations • 23 Dec 2022 • Joyce Chew, Deanna Needell, Michael Perlmutter
Moreover, in this work, the authors provide a numerical scheme for implementing such neural networks when the manifold is unknown and one only has access to finitely many sample points.
no code implementations • 19 Dec 2022 • Michael R. Lindstrom, Xiaofu Ding, Feng Liu, Anand Somayajula, Deanna Needell
Nonnegative matrix factorization can be used to automatically detect topics within a corpus in an unsupervised fashion.
no code implementations • 1 Dec 2022 • Zehan Chao, Denali Molitor, Deanna Needell, Mason A. Porter
We then infer a ``media-bias chart'' of (bias, quality) coordinates for the media outlets by integrating the (bias, quality) measurements of the tweets of the media outlets.
no code implementations • 10 Nov 2022 • Jocelyn T. Chi, Deanna Needell
We present convergence guarantees for the sketched predictions on new data within a fixed number of iterations.
1 code implementation • 28 Aug 2022 • Elena Sizikova, Joshua Vendrow, Xu Cao, Rachel Grotheer, Jamie Haddock, Lara Kassab, Alona Kryshchenko, Thomas Merkh, R. W. M. A. Madushani, Kenny Moise, Annie Ulichney, Huy V. Vo, Chuntian Wang, Megan Coffee, Kathryn Leonard, Deanna Needell
Automatic infectious disease classification from images can facilitate needed medical diagnoses.
1 code implementation • 20 Aug 2022 • HanQin Cai, Longxiu Huang, Pengyu Li, Deanna Needell
While uniform sampling has been widely studied in the matrix completion literature, CUR sampling approximates a low-rank matrix via row and column samples.
no code implementations • 17 Aug 2022 • Joyce Chew, Matthew Hirn, Smita Krishnaswamy, Deanna Needell, Michael Perlmutter, Holly Steach, Siddharth Viswanath, Hau-Tieng Wu
Our proposed framework includes previous work on geometric scattering as special cases but also applies to more general settings such as directed graphs, signed graphs, and manifolds with boundary.
no code implementations • 17 Jul 2022 • Shuang Li, William J. Swartworth, Martin Takáč, Deanna Needell, Robert M. Gower
We take a step further and develop a method for solving the interpolation equations that uses the local second-order approximation of the model.
1 code implementation • 21 Jun 2022 • Joyce Chew, Holly R. Steach, Siddharth Viswanath, Hau-Tieng Wu, Matthew Hirn, Deanna Needell, Smita Krishnaswamy, Michael Perlmutter
The manifold scattering transform is a deep feature extractor for data defined on a Riemannian manifold.
no code implementations • 28 Feb 2022 • Xia Li, Longxiu Huang, Deanna Needell
Developing large-scale distributed methods that are robust to the presence of adversarial or corrupted workers is an important part of making such methods practical for real-world problems.
no code implementations • 28 Feb 2022 • Jamie Haddock, Lara Kassab, Sixian Li, Alona Kryshchenko, Rachel Grotheer, Elena Sizikova, Chuntian Wang, Thomas Merkh, RWMA Madushani, Miju Ahn, Deanna Needell, Kathryn Leonard
We propose new semi-supervised nonnegative matrix factorization (SSNMF) models for document classification and provide motivation for these models as maximum likelihood estimators.
no code implementations • 31 Jan 2022 • Pengyu Li, Christine Tseng, Yaxuan Zheng, Joyce A. Chew, Longxiu Huang, Benjamin Jarman, Deanna Needell
Classification and topic modeling are popular techniques in machine learning that extract information from large-scale datasets.
1 code implementation • 7 Oct 2021 • Andrew Sack, Wenzhao Jiang, Michael Perlmutter, Palina Salanevich, Deanna Needell
We propose a method for noise reduction, the task of producing a clean audio signal from a recording corrupted by additive noise.
1 code implementation • 30 Sep 2021 • Joshua Vendrow, Jamie Haddock, Deanna Needell
Hierarchical NTF (HNTF) methods have been proposed, however these methods do not naturally generalize their matrix-based counterparts.
no code implementations • 23 Aug 2021 • HanQin Cai, Zehan Chao, Longxiu Huang, Deanna Needell
We study the problem of tensor robust principal component analysis (TRPCA), which aims to separate an underlying low-multilinear-rank tensor and a sparse outlier tensor from their sum.
no code implementations • 28 Apr 2021 • Ryan Budahazy, Lu Cheng, Yihuan Huang, Andrew Johnson, Pengyu Li, Joshua Vendrow, Zhoutong Wu, Denali Molitor, Elizaveta Rebrova, Deanna Needell
The California Innocence Project (CIP), a clinical law school program aiming to free wrongfully convicted prisoners, evaluates thousands of mails containing new requests for assistance and corresponding case files.
1 code implementation • 19 Mar 2021 • HanQin Cai, Keaton Hamm, Longxiu Huang, Deanna Needell
Low rank tensor approximation is a fundamental tool in modern machine learning and data science.
no code implementations • 5 Jan 2021 • HanQin Cai, Keaton Hamm, Longxiu Huang, Deanna Needell
Additionally, we consider hybrid randomized and deterministic sampling methods which produce a compact CUR decomposition of a given matrix, and apply this to video sequences to produce canonical frames thereof.
no code implementations • 1 Jan 2021 • Joshua Vendrow, Jamie Haddock, Deanna Needell
We propose a new hierarchical nonnegative CANDECOMP/PARAFAC (CP) decomposition (hierarchical NCPD) model and a training method, Neural NCPD, for performing hierarchical topic modeling on multi-modal tensor data.
no code implementations • 10 Nov 2020 • Hanbaek Lyu, Georg Menz, Deanna Needell, Christopher Strohmeier
Online nonnegative matrix factorization (ONMF) is a matrix factorization technique in the online setting where data are acquired in a streaming fashion and the matrix factors are updated each time.
1 code implementation • 22 Oct 2020 • Joshua Vendrow, Jamie Haddock, Elizaveta Rebrova, Deanna Needell
Fully unsupervised topic models have found fantastic success in document clustering and classification.
1 code implementation • 15 Oct 2020 • Jamie Haddock, Lara Kassab, Sixian Li, Alona Kryshchenko, Rachel Grotheer, Elena Sizikova, Chuntian Wang, Thomas Merkh, R. W. M. A. Madushani, Miju Ahn, Deanna Needell, Kathryn Leonard
We propose several new models for semi-supervised nonnegative matrix factorization (SSNMF) and provide motivation for SSNMF models as maximum likelihood estimators given specific distributions of uncertainty.
Ranked #13 on Text Classification on 20NEWS
1 code implementation • 4 Oct 2020 • Lara Kassab, Alona Kryshchenko, Hanbaek Lyu, Denali Molitor, Deanna Needell, Elizaveta Rebrova, Jiahong Yuan
Further, we propose quantitative ways to measure the topic length and demonstrate the ability of S-NCPD (as well as its online variant) to discover short and long-lasting temporal topics in a controlled manner in semi-synthetic and real-world data including news headlines.
1 code implementation • 16 Sep 2020 • Hanbaek Lyu, Christopher Strohmeier, Deanna Needell
We prove that our algorithm converges almost surely to the set of stationary points of the objective function under the hypothesis that the sequence of data tensors is generated by an underlying Markov chain.
no code implementations • EMNLP (NLP-COVID19) 2020 • Rachel Grotheer, Yihuan Huang, Pengyu Li, Elizaveta Rebrova, Deanna Needell, Longxiu Huang, Alona Kryshchenko, Xia Li, Kyung Ha, Oleksandr Kryshchenko
A dataset of COVID-19-related scientific literature is compiled, combining the articles from several online libraries and selecting those with open access and full text available.
no code implementations • 24 Aug 2020 • Joshua Vendrow, Jamie Haddock, Deanna Needell, Lorraine Johnson
We first analyze the general performance of the model and then identify the most important features for predicting participant answers to GROC.
no code implementations • 30 Jul 2020 • Deanna Needell, Aaron A. Nelson, Rayan Saab, Palina Salanevich, Olov Schavemaker
We provide a (corrected) rigorous proof that the Igelnik and Pao construction is a universal approximator for continuous functions on compact domains, with approximation error decaying asymptotically like $O(1/\sqrt{n})$ for the number $n$ of network nodes.
2 code implementations • 20 Apr 2020 • Hanbaek Lyu, Christopher Strohmeier, Georg Menz, Deanna Needell
One of the main sources of difficulty is that a very limited amount of daily COVID-19 case data is available, and with few exceptions, the majority of countries are currently in the "exponential spread stage," and thus there is scarce information available which would enable one to predict the phase transition between spread and containment.
no code implementations • 2 Jan 2020 • Miju Ahn, Nicole Eikmeier, Jamie Haddock, Lara Kassab, Alona Kryshchenko, Kathryn Leonard, Deanna Needell, R. W. M. A. Madushani, Elena Sizikova, Chuntian Wang
There is currently an unprecedented demand for large-scale temporal data analysis due to the explosive growth of data.
2 code implementations • 1 Dec 2019 • Yuchen Guo, Nicholas Hanoian, Zhexiao Lin, Nicholas Liskij, Hanbaek Lyu, Deanna Needell, Jiahao Qu, Henry Sojico, Yuliang Wang, Zhe Xiong, Zhenhong Zou
We propose a novel model for a topic-aware chatbot by combining the traditional Recurrent Neural Network (RNN) encoder-decoder model with a topic attention layer based on Nonnegative Matrix Factorization (NMF).
1 code implementation • 5 Nov 2019 • Hanbaek Lyu, Deanna Needell, Laura Balzano
As the main application, by combining online non-negative matrix factorization and a recent MCMC algorithm for sampling motifs from networks, we propose a novel framework of Network Dictionary Learning, which extracts ``network dictionary patches' from a given network in an online manner that encodes main features of the network.
2 code implementations • 9 Sep 2019 • Robert Gower, Denali Molitor, Jacob Moorman, Deanna Needell
We present new adaptive sampling rules for the sketch-and-project method for solving linear systems.
Numerical Analysis Numerical Analysis 15A06, 15B52, 65F10, 68W20, 65N75, 65Y20, 68Q25, 68W40, 90C20
no code implementations • 22 Aug 2019 • Rachel Grotheer, Shuang Li, Anna Ma, Deanna Needell, Jing Qin
In this paper, we utilize the same tensor version of the Restricted Isometry Property (RIP) to extend these results for tensors with low CANDECOMP/PARAFAC (CP) rank.
no code implementations • 26 Jul 2019 • Denali Molitor, Deanna Needell, Rachel Ward
Gradient descent is a simple and widely used optimization method for machine learning.
no code implementations • 31 May 2019 • Jesus A. De Loera, Jamie Haddock, Anna Ma, Deanna Needell
Machine learning algorithms typically rely on optimization subroutines and are well-known to provide very effective outcomes for many types of problems.
no code implementations • 17 Apr 2019 • Christian Parkinson, Kevin Huynh, Deanna Needell
Matrix completion is a classical problem in data science wherein one attempts to reconstruct a low-rank matrix while only observing some subset of the entries.
no code implementations • 9 Sep 2018 • Denali Molitor, Deanna Needell
Building on a recently designed simple framework for classification using binary data, we demonstrate that one can improve classification accuracy of this approach through iterative applications whose output serves as input to the next application.
1 code implementation • 13 Aug 2018 • Deanna Needell, Sam Nelson, Yingqi Shi
We provide examples to illustrate the computation of the invariant and show that the enhancement is proper.
Geometric Topology Quantum Algebra 57M27, 57M25
no code implementations • 23 Jul 2018 • Denali Molitor, Deanna Needell
In classification problems, especially those that categorize data into a large number of classes, the classes often naturally follow a hierarchical structure.
no code implementations • 31 May 2018 • Saiprasad Ravishankar, Anna Ma, Deanna Needell
Sparsity-based models and techniques have been exploited in many signal processing and imaging applications.
no code implementations • 1 Feb 2018 • Saiprasad Ravishankar, Anna Ma, Deanna Needell
Alternating minimization algorithms have been particularly popular in dictionary or transform learning.
no code implementations • 29 Jan 2018 • Denali Molitor, Deanna Needell
We propose adjusting the standard nuclear norm minimization strategy for matrix completion to account for such structural differences between observed and unobserved entries by regularizing the values of the unobserved entries.
no code implementations • 6 Jul 2017 • Deanna Needell, Rayan Saab, Tina Woolf
Binary, or one-bit, representations of data arise naturally in many applications, and are appealing in both hardware implementations and algorithm design.
1 code implementation • 9 Apr 2017 • WonHyuk Choi, Deanna Needell, Sam Nelson
In this paper, we build on the biquasiles and dual graph diagrams introduced in arXiv:1610. 06969.
Geometric Topology Quantum Algebra 57M25, 57M27
no code implementations • 12 Jan 2017 • Deanna Needell, Tina Woolf
Asynchronous algorithms are often studied to solve optimization problems where the cost function takes the form $\sum_{i=1}^M f_i(x)$, with a common assumption that each $f_i$ is sparse; that is, each $f_i$ acts only on a small number of components of $x\in\mathbb{R}^n$.
1 code implementation • 19 Jun 2015 • Mark A. Davenport, Andrew K. Massimino, Deanna Needell, Tina Woolf
Suppose that we wish to estimate a vector $\mathbf{x} \in \mathbb{C}^n$ from a small number of noisy linear measurements of the form $\mathbf{y} = \mathbf{A x} + \mathbf{z}$, where $\mathbf{z}$ represents measurement noise.
Information Theory Information Theory
no code implementations • 8 May 2014 • Anna Ma, Arjuna Flenner, Deanna Needell, Allon G. Percus
We propose a method to improve image clustering using sparse text and the wisdom of the crowds.
no code implementations • 22 Apr 2014 • Ran Zhao, Deanna Needell, Christopher Johansen, Jerry L. Grenard
In this paper, we compare and analyze clustering methods with missing data in health behavior research.
no code implementations • NeurIPS 2014 • Deanna Needell, Nathan Srebro, Rachel Ward
Furthermore, we show how reweighting the sampling distribution (i. e. importance sampling) is necessary in order to further improve convergence, and obtain a linear dependence in the average smoothness, dominating previous results.
no code implementations • 11 Oct 2012 • Deanna Needell, Rachel Ward
Consider the problem of reconstructing a multidimensional signal from an underdetermined set of measurements, as in the setting of compressed sensing.
no code implementations • 19 Aug 2012 • Deanna Needell, Joel A. Tropp
The block Kaczmarz method is an iterative scheme for solving overdetermined least-squares problems.
Numerical Analysis Numerical Analysis 65F10, 65F20, 68W20, 41A65
no code implementations • 25 Aug 2010 • Yonina C. Eldar, Deanna Needell
The Kaczmarz method is an algorithm for finding the solution to an overdetermined consistent system of linear equations Ax=b by iteratively projecting onto the solution spaces.
Numerical Analysis