Search Results for author: Ping Li

Found 139 papers, 7 papers with code

Optimal Estimator for Unlabeled Linear Regression

no code implementations ICML 2020 Hang Zhang, Ping Li

Unlabeled linear regression, or ``linear regression with an unknown permutation'', has attracted increasing attentions due to its applications in linkage record and de-anonymization.

A Predicate-Function-Argument Annotation of Natural Language for Open-Domain Information eXpression

no code implementations EMNLP 2020 Mingming Sun, Wenyue Hua, Zoey Liu, Xin Wang, Kangjie Zheng, Ping Li

Based on the same platform of OIX, the OIE strategies are reusable, and people can select a set of strategies to assemble their algorithm for a specific task so that the adaptability may be significantly increased.

Open Information Extraction

Toward Faster and Simpler Matrix Normalization via Rank-1 Update

no code implementations ECCV 2020 Tan Yu, Yunfeng Cai, Ping Li

To boost the efficiency in the GPU platform, recent methods rely on Newton-Schulz (NS) iteration to approximate the matrix square-root.

Discriminative Similarity for Data Clustering

no code implementations17 Sep 2021 Yingzhen Yang, Ping Li

Similarity-based clustering methods separate data into clusters according to the pairwise similarity between the data, and the pairwise similarity is crucial for their performance.

Toward Communication Efficient Adaptive Gradient Method

no code implementations10 Sep 2021 Xiangyi Chen, Xiaoyun Li, Ping Li

While adaptive gradient methods have been proven effective for training neural nets, the study of adaptive gradient methods in federated learning is scarce.

Distributed Optimization Federated Learning

C-MinHash: Practically Reducing Two Permutations to Just One

no code implementations10 Sep 2021 Xiaoyun Li, Ping Li

That is, one single permutation is used for both the initial pre-processing step to break the structures in the data and the circulant hashing step to generate $K$ hashes.

Extreme Bandits using Robust Statistics

no code implementations9 Sep 2021 Sujay Bhatt, Ping Li, Gennady Samorodnitsky

We consider a multi-armed bandit problem motivated by situations where only the extreme values, as opposed to expected values in the classical bandit setting, are of interest.

C-MinHash: Rigorously Reducing $K$ Permutations to Two

no code implementations7 Sep 2021 Xiaoyun Li, Ping Li

Unlike classical MinHash, these $K$ hashes are obviously correlated, but we are able to provide rigorous proofs that we still obtain an unbiased estimate of the Jaccard similarity and the theoretical variance is uniformly smaller than that of the classical MinHash with $K$ independent permutations.

On the Convergence of Decentralized Adaptive Gradient Methods

no code implementations7 Sep 2021 Xiangyi Chen, Belhal Karimi, Weijie Zhao, Ping Li

Adaptive gradient methods including Adam, AdaGrad, and their variants have been very successful for training deep learning models, such as neural networks.

Distributed Computing Distributed Optimization

Provable Data Clustering via Innovation Search

no code implementations16 Aug 2021 Weiwei Li, Mostafa Rahmani, Ping Li

It is shown that in contrast to most of the existing methods which require the subspaces to be sufficiently incoherent with each other, Innovation Pursuit only requires the innovative components of the subspaces to be sufficiently incoherent with each other.

Non-Local Feature Aggregation on Graphs via Latent Fixed Data Structures

no code implementations16 Aug 2021 Mostafa Rahmani, Rasoul Shafipour, Ping Li

The proposed approach is used to design several novel global feature aggregation methods based on the choice of the LFDS.

S$^2$-MLPv2: Improved Spatial-Shift MLP Architecture for Vision

no code implementations2 Aug 2021 Tan Yu, Xu Li, Yunfeng Cai, Mingming Sun, Ping Li

More recently, using smaller patches with a pyramid structure, Vision Permutator (ViP) and Global Filter Network (GFNet) achieve better performance than S$^2$-MLP.

Rethinking Token-Mixing MLP for MLP-based Vision Backbone

no code implementations28 Jun 2021 Tan Yu, Xu Li, Yunfeng Cai, Mingming Sun, Ping Li

By introducing the inductive bias from the image processing, convolution neural network (CNN) has achieved excellent performance in numerous computer vision tasks and has been established as \emph{de facto} backbone.

A Systematic Collection of Medical Image Datasets for Deep Learning

1 code implementation24 Jun 2021 Johann Li, Guangming Zhu, Cong Hua, Mingtao Feng, BasheerBennamoun, Ping Li, Xiaoyuan Lu, Juan Song, Peiyi Shen, Xu Xu, Lin Mei, Liang Zhang, Syed Afaq Ali Shah, Mohammed Bennamoun

Thus, as comprehensive as possible, this paper provides a collection of medical image datasets with their associated challenges for deep learning research.

Closed-Form, Provable, and Robust PCA via Leverage Statistics and Innovation Search

no code implementations23 Jun 2021 Mostafa Rahmani, Ping Li

In the application of Innovation Search for outlier detection, the directions of innovation were utilized to measure the innovation of the data points.

Outlier Detection

Patchwise Generative ConvNet: Training Energy-Based Models From a Single Natural Image for Internal Learning

no code implementations CVPR 2021 Zilong Zheng, Jianwen Xie, Ping Li

Exploiting internal statistics of a single natural image has long been recognized as a significant research paradigm where the goal is to learn the distribution of patches within the image without relying on external training data.

Image Generation Super-Resolution

Learning Deep Latent Variable Models by Short-Run MCMC Inference With Optimal Transport Correction

no code implementations CVPR 2021 Dongsheng An, Jianwen Xie, Ping Li

Learning latent variable models with deep top-down architectures typically requires inferring the latent variables for each training example based on the posterior distribution of these latent variables.

Latent Variable Models

S$^2$-MLP: Spatial-Shift MLP Architecture for Vision

no code implementations14 Jun 2021 Tan Yu, Xu Li, Yunfeng Cai, Mingming Sun, Ping Li

We discover that the token-mixing MLP is a variant of the depthwise convolution with a global reception field and spatial-specific configuration.

Cross-lingual Cross-modal Pretraining for Multimodal Retrieval

no code implementations NAACL 2021 Hongliang Fei, Tan Yu, Ping Li

Recent pretrained vision-language models have achieved impressive performance on cross-modal retrieval tasks in English.

Cross-Modal Retrieval Machine Translation

High-quality Low-dose CT Reconstruction Using Convolutional Neural Networks with Spatial and Channel Squeeze and Excitation

no code implementations1 Apr 2021 Jingfeng Lu, Shuo Wang, Ping Li, Dong Ye

Low-dose computed tomography (CT) allows the reduction of radiation risk in clinical applications at the expense of image quality, which deteriorates the diagnosis accuracy of radiologists.

Computed Tomography (CT) Image Reconstruction

Quantization Algorithms for Random Fourier Features

no code implementations25 Feb 2021 Xiaoyun Li, Ping Li

Closely related to RP, the method of random Fourier features (RFF) has also become popular, for approximating the Gaussian kernel.

Dimensionality Reduction Quantization

Simulation on the Transparency of Electrons and Ion Back Flow for a Time Projection Chamber based on Staggered Multiple THGEMs

no code implementations16 Feb 2021 Mengzhi Wu, Qian Liu, Ping Li, Shi Chen, Binlong Wang, Wenhan Shen, Shiping Chen, Yangheng Zheng, Yigang Xie, Jin Li

The IBF and the transparent rate of electrons are two essential indicators of TPC, which affect the energy resolution and counting rate respectively.

Instrumentation and Detectors High Energy Physics - Experiment

Learning Energy-Based Generative Models via Coarse-to-Fine Expanding and Sampling

no code implementations ICLR 2021 Yang Zhao, Jianwen Xie, Ping Li

Energy-based models (EBMs) for generative modeling parametrize a single net and can be directly trained by maximum likelihood estimation.

Unsupervised Image-To-Image Translation

Convergent Adaptive Gradient Methods in Decentralized Optimization

no code implementations1 Jan 2021 Xiangyi Chen, Belhal Karimi, Weijie Zhao, Ping Li

Specifically, we propose a general algorithmic framework that can convert existing adaptive gradient methods to their decentralized counterparts.

Distributed Optimization

MISSO: Minimization by Incremental Stochastic Surrogate Optimization for Large Scale Nonconvex and Nonsmooth Problems

no code implementations1 Jan 2021 Belhal Karimi, Hoi To Wai, Eric Moulines, Ping Li

Many constrained, nonconvex and nonsmooth optimization problems can be tackled using the majorization-minimization (MM) method which alternates between constructing a surrogate function which upper bounds the objective function, and then minimizing this surrogate.

Latent Variable Models Variational Inference

Cross-Probe BERT for Efficient and Effective Cross-Modal Search

no code implementations1 Jan 2021 Tan Yu, Hongliang Fei, Ping Li

Inspired by the great success of BERT in NLP tasks, many text-vision BERT models emerged recently.

Text-Image Retrieval

Learning Energy-Based Model with Variational Auto-Encoder as Amortized Sampler

no code implementations29 Dec 2020 Jianwen Xie, Zilong Zheng, Ping Li

In this paper, we propose to learn a variational auto-encoder (VAE) to initialize the finite-step MCMC, such as Langevin dynamics that is derived from the energy function, for efficient amortized sampling of the EBM.

Towards Better Generalization of Adaptive Gradient Methods

no code implementations NeurIPS 2020 Yingxue Zhou, Belhal Karimi, Jinxing Yu, Zhiqiang Xu, Ping Li

Adaptive gradient methods such as AdaGrad, RMSprop and Adam have been optimizers of choice for deep learning due to their fast training speed.

Ratio Trace Formulation of Wasserstein Discriminant Analysis

no code implementations NeurIPS 2020 Hexuan Liu, Yunfeng Cai, You-Lin Chen, Ping Li

We reformulate the Wasserstein Discriminant Analysis (WDA) as a ratio trace problem and present an eigensolver-based algorithm to compute the discriminative subspace of WDA.

RANet: Region Attention Network for Semantic Segmentation

1 code implementation NeurIPS 2020 Dingguo Shen, Yuanfeng Ji, Ping Li, Yi Wang, Di Lin

In contrast to the previous methods, RANet configures the information pathways between the pixels in different regions, enabling the region interaction to exchange the regional context for enhancing all of the pixels in the image.

Semantic Segmentation

Optimal Prediction of the Number of Unseen Species with Multiplicity

no code implementations NeurIPS 2020 Yi Hao, Ping Li

Based on a sample of size $n$, we consider estimating the number of symbols that appear at least $\mu$ times in an independent sample of size $a \cdot n$, where $a$ is a given parameter.

Identification of Matrix Joint Block Diagonalization

no code implementations2 Nov 2020 Yunfeng Cai, Ping Li

This paper considers the identification problem for BJBDP, that is, under what conditions and by what means, we can identify the diagonalizer $A$ and the block diagonal structure of $\Sigma_i$, especially when there is noise in $C_i$'s.

Tensor Completion via Tensor Networks with a Tucker Wrapper

no code implementations29 Oct 2020 Yunfeng Cai, Ping Li

In this paper, we propose to solve LRTC via tensor networks with a Tucker wrapper.

Tensor Networks Video Inpainting

Exploring global diverse attention via pairwise temporal relation for video summarization

no code implementations23 Sep 2020 Ping Li, Qinghao Ye, Luming Zhang, Li Yuan, Xianghua Xu, Ling Shao

In this paper, we propose an efficient convolutional neural network architecture for video SUMmarization via Global Diverse Attention called SUM-GDA, which adapts attention mechanism in a global perspective to consider pairwise temporal relations of video frames.

Video Summarization

A Framework of Randomized Selection Based Certified Defenses Against Data Poisoning Attacks

no code implementations18 Sep 2020 Ruoxin Chen, Jie Li, Chentao Wu, Bin Sheng, Ping Li

Random selection based defenses can achieve certified robustness by averaging the classifiers' predictions on the sub-datasets sampled from the training set.

Data Poisoning

Understanding and Detecting Convergence for Stochastic Gradient Descent with Momentum

no code implementations27 Aug 2020 Jerry Chee, Ping Li

We construct a statistical diagnostic test for convergence to the stationary phase using the inner product between successive gradients and demonstrate that the proposed diagnostic works well.

Stochastic Optimization

FedSKETCH: Communication-Efficient and Private Federated Learning via Sketching

no code implementations11 Aug 2020 Farzin Haddadpour, Belhal Karimi, Ping Li, Xiaoyun Li

Communication complexity and privacy are the two key challenges in Federated Learning where the goal is to perform a distributed learning through a large volume of devices.

Federated Learning

MeDaS: An open-source platform as service to help break the walls between medicine and informatics

no code implementations12 Jul 2020 Liang Zhang, Johann Li, Ping Li, Xiaoyuan Lu, Peiyi Shen, Guangming Zhu, Syed Afaq Shah, Mohammed Bennarmoun, Kun Qian, Björn W. Schuller

To the best of our knowledge, MeDaS is the first open-source platform proving a collaborative and interactive service for researchers from a medical background easily using DL related toolkits, and at the same time for scientists or engineers from information sciences to understand the medical knowledge side.

Learning Interpretable Relationships between Entities, Relations and Concepts via Bayesian Structure Learning on Open Domain Facts

no code implementations ACL 2020 Jingyuan Zhang, Mingming Sun, Yue Feng, Ping Li

Compared to the state-of-the-art methods, the learned network structures help improving the identification of concepts for entities based on the relations of entities on both datasets.

Cooperative Rate-Splitting for Secrecy Sum-Rate Enhancement in Multi-antenna Broadcast Channels

no code implementations3 Jun 2020 Ping Li, Ming Chen, Yijie Mao, Zhaohui Yang, Bruno Clerckx, Mohammad Shikh-Bahaei

In this paper, we employ Cooperative Rate-Splitting (CRS) technique to enhance the Secrecy Sum Rate (SSR) for the Multiple Input Single Output (MISO) Broadcast Channel (BC), consisting of two legitimate users and one eavesdropper, with perfect Channel State Information (CSI) available at all nodes.

Estimate the Implicit Likelihoods of GANs with Application to Anomaly Detection

1 code implementation20 Apr 2020 Shaogang Ren, Dingcheng Li, Zhixin Zhou, Ping Li

The thriving of deep models and generative models provides approaches to model high dimensional distributions.

Anomaly Detection

Randomized Kernel Multi-view Discriminant Analysis

no code implementations2 Apr 2020 Xiaoyun Li, Jie Gui, Ping Li

In this paper, we propose the kernel version of multi-view discriminant analysis, called kernel multi-view discriminant analysis (KMvDA).

Object Recognition

Distributed Primal-Dual Optimization for Online Multi-Task Learning

no code implementations2 Apr 2020 Peng Yang, Ping Li

Conventional online multi-task learning algorithms suffer from two critical limitations: 1) Heavy communication caused by delivering high velocity of sequential data to a central machine; 2) Expensive runtime complexity for building task relatedness.

Multi-Task Learning

An Inverse-free Truncated Rayleigh-Ritz Method for Sparse Generalized Eigenvalue Problem

no code implementations24 Mar 2020 Yunfeng Cai, Ping Li

Particularly, a new truncation strategy is proposed, which is able to find the support set of the leading eigenvector effectively.

Solving the Robust Matrix Completion Problem via a System of Nonlinear Equations

no code implementations24 Mar 2020 Yunfeng Cai, Ping Li

We consider the problem of robust matrix completion, which aims to recover a low rank matrix $L_*$ and a sparse matrix $S_*$ from incomplete observations of their sum $M=L_*+S_*\in\mathbb{R}^{m\times n}$.

Matrix Completion

Meta-CoTGAN: A Meta Cooperative Training Paradigm for Improving Adversarial Text Generation

no code implementations12 Mar 2020 Haiyan Yin, Dingcheng Li, Xu Li, Ping Li

To this end, we introduce a cooperative training paradigm, where a language model is cooperatively trained with the generator and we utilize the language model to efficiently shape the data distribution of the generator against mode collapse.

Adversarial Text Language Modelling +2

Distributed Hierarchical GPU Parameter Server for Massive Scale Deep Learning Ads Systems

no code implementations12 Mar 2020 Weijie Zhao, Deping Xie, Ronglai Jia, Yulei Qian, Ruiquan Ding, Mingming Sun, Ping Li

For example, a sponsored online advertising system can contain more than $10^{11}$ sparse features, making the neural network a massive model with around 10 TB parameters.

Selective Convolutional Network: An Efficient Object Detector with Ignoring Background

no code implementations4 Feb 2020 Hefei Ling, Yangyang Qin, Li Zhang, Yuxuan Shi, Ping Li

It is well known that attention mechanisms can effectively improve the performance of many CNNs including object detectors.

Structure-Feature based Graph Self-adaptive Pooling

no code implementations30 Jan 2020 Liang Zhang, Xudong Wang, Hongsheng Li, Guangming Zhu, Peiyi Shen, Ping Li, Xiaoyuan Lu, Syed Afaq Ali Shah, Mohammed Bennamoun

To solve these problems mentioned above, we propose a novel graph self-adaptive pooling method with the following objectives: (1) to construct a reasonable pooled graph topology, structure and feature information of the graph are considered simultaneously, which provide additional veracity and objectivity in node selection; and (2) to make the pooled nodes contain sufficiently effective graph information, node feature information is aggregated before discarding the unimportant nodes; thus, the selected nodes contain information from neighbor nodes, which can enhance the use of features of the unselected nodes.

Graph Classification

Outlier Detection and Data Clustering via Innovation Search

no code implementations30 Dec 2019 Mostafa Rahmani, Ping Li

In this paper, we present a new discovery that the directions of innovation can be used to design a provable and strong robust (to outlier) PCA method.

Outlier Detection

Towards Practical Alternating Least-Squares for CCA

no code implementations NeurIPS 2019 Zhiqiang Xu, Ping Li

To promote the practical use of ALS for CCA, we propose truly alternating least-squares.

Random Projections with Asymmetric Quantization

no code implementations NeurIPS 2019 Xiaoyun Li, Ping Li

The method of random projection has been a popular tool for data compression, similarity search, and machine learning.

Quantization

Generalization Error Analysis of Quantized Compressive Learning

no code implementations NeurIPS 2019 Xiaoyun Li, Ping Li

In this paper, we consider the learning problem where the projected data is further compressed by scalar quantization, which is called quantized compressive learning.

Quantization

A Fourier Analytical Approach to Estimation of Smooth Functions in Gaussian Shift Model

no code implementations5 Nov 2019 Fan Zhou, Ping Li

Let $\mathbf{x}_j = \mathbf{\theta} + \mathbf{\epsilon}_j$, $j=1,\dots, n$ be i. i. d.

On Efficient Retrieval of Top Similarity Vectors

no code implementations IJCNLP 2019 Shulong Tan, Zhixin Zhou, Zhaozhuo Xu, Ping Li

Retrieval of relevant vectors produced by representation learning critically influences the efficiency in natural language processing (NLP) tasks.

Representation Learning

Reinforced Product Metadata Selection for Helpfulness Assessment of Customer Reviews

no code implementations IJCNLP 2019 Miao Fan, Chao Feng, Mingming Sun, Ping Li

Given a product, a selector (agent) learns from both the keys in the product metadata and one of its reviews to take an action that selects the correct value, and a successive predictor (network) makes the free-text review attend to this value to obtain better neural representations for helpfulness assessment.

Graph Analysis and Graph Pooling in the Spatial Domain

no code implementations3 Oct 2019 Mostafa Rahmani, Ping Li

The proposed approach leverages a spatial representation of the graph which makes the neural network aware of the differences between the nodes and also their locations in the graph.

Graph Embedding

The Benefits of Diversity: Permutation Recovery in Unlabeled Sensing from Multiple Measurement Vectors

no code implementations5 Sep 2019 Hang Zhang, Martin Slawski, Ping Li

For the case in which both the signal and permutation are unknown, the problem is reformulated as a bi-convex optimization problem with an auxiliary variable, which can be solved by the Alternating Direction Method of Multipliers (ADMM).

Multi-Spectral Visual Odometry without Explicit Stereo Matching

no code implementations23 Aug 2019 Weichen Dai, Yu Zhang, Donglei Sun, Naira Hovakimyan, Ping Li

Moreover, the proposed method can also provide a metric 3D reconstruction in semi-dense density with multi-spectral information, which is not available from existing multi-spectral methods.

3D Reconstruction Stereo Matching +2

On Convergence of Distributed Approximate Newton Methods: Globalization, Sharper Bounds and Beyond

no code implementations6 Aug 2019 Xiao-Tong Yuan, Ping Li

We first introduce a simple variant of DANE equipped with backtracking line search, for which global asymptotic convergence and sharper local non-asymptotic convergence rate guarantees can be proved for both quadratic and non-quadratic strongly convex functions.

A Two-Stage Approach to Multivariate Linear Regression with Sparsely Mismatched Data

no code implementations16 Jul 2019 Martin Slawski, Emanuel Ben-David, Ping Li

A tacit assumption in linear regression is that (response, predictor)-pairs correspond to identical observational units.

Integration of Knowledge Graph Embedding Into Topic Modeling with Hierarchical Dirichlet Process

no code implementations NAACL 2019 Dingcheng Li, Siamak Zamani, Jingyuan Zhang, Ping Li

Leveraging domain knowledge is an effective strategy for enhancing the quality of inferred low-dimensional representations of documents by topic models.

Document Classification General Classification +3

DEEP GEOMETRICAL GRAPH CLASSIFICATION

no code implementations ICLR 2019 Mostafa Rahmani, Ping Li

In the second step, the GNN is applied to the point-cloud representation of the graph provided by the embedding method.

Classification General Classification +3

Logician: A Unified End-to-End Neural Approach for Open-Domain Information Extraction

no code implementations29 Apr 2019 Mingming Sun, Xu Li, Xin Wang, Miao Fan, Yue Feng, Ping Li

In this paper, we consider the problem of open information extraction (OIE) for extracting entity and relation level intermediate structures from sentences in open-domain.

Global Optimization Open Information Extraction +1

Cycle-SUM: Cycle-consistent Adversarial LSTM Networks for Unsupervised Video Summarization

no code implementations17 Apr 2019 Li Yuan, Francis EH Tay, Ping Li, Li Zhou, Jiashi Feng

The evaluator defines a learnable information preserving metric between original video and summary video and "supervises" the selector to identify the most informative frames to form the summary video.

Unsupervised Video Summarization

An Optimistic Acceleration of AMSGrad for Nonconvex Optimization

no code implementations ICLR 2020 Jun-Kun Wang, Xiaoyun Li, Belhal Karimi, Ping Li

We propose a new variant of AMSGrad, a popular adaptive gradient based optimization algorithm widely used for training deep neural networks.

RGB-D SLAM in Dynamic Environments Using Point Correlations

no code implementations8 Nov 2018 Weichen Dai, Yu Zhang, Ping Li, Zheng Fang, Sebastian Scherer

This method utilizes the correlation between map points to separate points that are part of the static scene and points that are part of different moving objects into different groups.

Motion Estimation Simultaneous Localization and Mapping

Logician and Orator: Learning from the Duality between Language and Knowledge in Open Domain

no code implementations EMNLP 2018 Mingming Sun, Xu Li, Ping Li

We propose the task of Open-Domain Information Narration (OIN) as the reverse task of Open Information Extraction (OIE), to implement the dual structure between language and knowledge in the open domain.

Open Information Extraction Relation Classification

Several Tunable GMM Kernels

no code implementations8 May 2018 Ping Li

In this study, we propose a series of "tunable GMM kernels" which are simple and perform largely comparably to tree methods on the same datasets.

General Classification Multi-class Classification

Sign-Full Random Projections

no code implementations26 Apr 2018 Ping Li

At high similarity ($\rho\rightarrow1$), the asymptotic variance of recommended estimator is only $\frac{4}{3\pi} \approx 0. 4$ of the estimator for sign-sign projections.

Simple strategies for recovering inner products from coarsely quantized random projections

no code implementations NeurIPS 2017 Ping Li, Martin Slawski

Random projections have been increasingly adopted for a diverse set of tasks in machine learning involving dimensionality reduction.

Dimensionality Reduction Quantization

Partial Hard Thresholding: Towards A Principled Analysis of Support Recovery

no code implementations NeurIPS 2017 Jie Shen, Ping Li

In machine learning and compressed sensing, it is of central importance to understand when a tractable algorithm recovers the support of a sparse signal from its compressed measurements.

On the Iteration Complexity of Support Recovery via Hard Thresholding Pursuit

no code implementations ICML 2017 Jie Shen, Ping Li

Recovering the support of a sparse signal from its compressed samples has been one of the most important problems in high dimensional statistics.

Tunable GMM Kernels

no code implementations9 Jan 2017 Ping Li

The linearized GMM kernel was extensively compared in with linearized radial basis function (RBF) kernel.

General Classification

Generalized Intersection Kernel

no code implementations29 Dec 2016 Ping Li

Following the very recent line of work on the ``generalized min-max'' (GMM) kernel, this study proposes the ``generalized intersection'' (GInt) kernel and the related ``normalized generalized min-max'' (NGMM) kernel.

Exact Recovery of Hard Thresholding Pursuit

no code implementations NeurIPS 2016 Xiaotong Yuan, Ping Li, Tong Zhang

In this paper, we bridge this gap by showing, for the first time, that exact recovery of the global sparse minimizer is possible for HTP-style methods under restricted strong condition number bounding conditions.

Quantized Random Projections and Non-Linear Estimation of Cosine Similarity

no code implementations NeurIPS 2016 Ping Li, Michael Mitzenmacher, Martin Slawski

Random projections constitute a simple, yet effective technique for dimensionality reduction with applications in learning and search problems.

Dimensionality Reduction Quantization

Learning Additive Exponential Family Graphical Models via \ell_{2,1}-norm Regularized M-Estimation

no code implementations NeurIPS 2016 Xiaotong Yuan, Ping Li, Tong Zhang, Qingshan Liu, Guangcan Liu

We investigate a subclass of exponential family graphical models of which the sufficient statistics are defined by arbitrary additive forms.

Relational Multi-Manifold Co-Clustering

no code implementations16 Nov 2016 Ping Li, Jiajun Bu, Chun Chen, Zhanying He, Deng Cai

In this study, we focus on improving the co-clustering performance via manifold ensemble learning, which is able to maximally approximate the intrinsic manifolds of both the sample and feature spaces.

Ensemble Learning

Constrained Low-Rank Learning Using Least Squares-Based Regularization

no code implementations15 Nov 2016 Ping Li, Jun Yu, Meng Wang, Luming Zhang, Deng Cai, Xuelong. Li

To achieve this goal, we cast the problem into a constrained rank minimization framework by adopting the least squares regularization.

General Classification Image Categorization +2

Theory of the GMM Kernel

no code implementations1 Aug 2016 Ping Li, Cun-Hui Zhang

We prove the theoretical limit of GMM and the consistency result, assuming that the data follow an elliptical distribution, which is a very general family of distributions and includes the multivariate $t$-distribution as a special case.

Nystrom Method for Approximating the GMM Kernel

no code implementations12 Jul 2016 Ping Li

In order to use the GMM kernel for large-scale datasets, the prior work resorted to the (generalized) consistent weighted sampling (GCWS) to convert the GMM kernel to linear kernel.

Engineering Deep Representations for Modeling Aesthetic Perception

no code implementations25 May 2016 Yanxiang Chen, Yuxing Hu, Luming Zhang, Ping Li, Chao Zhang

To remedy these problems, we develop a deep architecture to learn aesthetically-relevant visual attributes from Flickr1, which are localized by multiple textual attributes in a weakly-supervised setting.

A Tight Bound of Hard Thresholding

no code implementations5 May 2016 Jie Shen, Ping Li

This paper is concerned with the hard thresholding operator which sets all but the $k$ largest absolute elements of a vector to zero.

Methods for Sparse and Low-Rank Recovery under Simplex Constraints

no code implementations2 May 2016 Ping Li, Syama Sundar Rangapuram, Martin Slawski

The de-facto standard approach of promoting sparsity by means of $\ell_1$-regularization becomes ineffective in the presence of simplex constraints, i. e.,~the target is known to have non-negative entries summing up to a given constant.

Density Estimation Portfolio Optimization +1

A Comparison Study of Nonlinear Kernels

no code implementations21 Mar 2016 Ping Li

In this paper, we compare 5 different nonlinear kernels: min-max, RBF, fRBF (folded RBF), acos, and acos-$\chi^2$, on a wide range of publicly available datasets.

b-bit Marginal Regression

no code implementations NeurIPS 2015 Martin Slawski, Ping Li

We consider the problem of sparse signal recovery from $m$ linear measurements quantized to $b$ bits.

Quantization

Sign Stable Random Projections for Large-Scale Learning

no code implementations27 Apr 2015 Ping Li

When $\alpha =2$, it is known that the corresponding nonlinear kernel is the arc-cosine kernel.

General Classification

Regularization-free estimation in trace regression with symmetric positive semidefinite matrices

no code implementations NeurIPS 2015 Martin Slawski, Ping Li, Matthias Hein

Over the past few years, trace regression models have received considerable attention in the context of matrix completion, quantum state tomography, and compressed sensing.

Matrix Completion Quantum State Tomography

Efficient Online Minimization for Low-Rank Subspace Clustering

no code implementations28 Mar 2015 Jie Shen, Ping Li, Huan Xu

Low-rank representation~(LRR) has been a significant method for segmenting data that are generated from a union of subspaces.

One Scan 1-Bit Compressed Sensing

no code implementations8 Mar 2015 Ping Li

Based on $\alpha$-stable random projections with small $\alpha$, we develop a simple algorithm for compressed sensing (sparse signal recovery) by utilizing only the signs (i. e., 1-bit) of the measurements.

Min-Max Kernels

no code implementations5 Mar 2015 Ping Li

Via an extensive empirical study, we show that this 0-bit scheme does not lose essential information.

General Classification

Object Proposal with Kernelized Partial Ranking

no code implementations5 Feb 2015 Jing Wang, Jie Shen, Ping Li

In order to determine a small set of proposals with a high recall, a common scheme is extracting multiple features followed by a ranking algorithm which however, incurs two major challenges: {\bf 1)} The ranking model often imposes pairwise constraints between each proposal, rendering the problem away from an efficient training/testing phase; {\bf 2)} Linear kernels are utilized due to the computational and memory bottleneck of training a kernelized model.

Online Optimization for Max-Norm Regularization

no code implementations NeurIPS 2014 Jie Shen, Huan Xu, Ping Li

The key technique in our algorithm is to reformulate the max-norm into a matrix factorization form, consisting of a basis component and a coefficients one.

Matrix Completion

Asymmetric Minwise Hashing

1 code implementation14 Nov 2014 Anshumali Shrivastava, Ping Li

Minwise hashing (Minhash) is a widely popular indexing scheme in practice.

Improved Asymmetric Locality Sensitive Hashing (ALSH) for Maximum Inner Product Search (MIPS)

no code implementations20 Oct 2014 Anshumali Shrivastava, Ping Li

In the prior work, the authors use asymmetric transformations which convert the problem of approximate MIPS into the problem of approximate near neighbor search which can be efficiently solved using hashing.

Recovery of Sparse Signals Using Multiple Orthogonal Least Squares

no code implementations9 Oct 2014 Jian Wang, Ping Li

We study the problem of recovering sparse signals from compressed linear measurements.

Compressed Sensing with Very Sparse Gaussian Random Projections

no code implementations11 Aug 2014 Ping Li, Cun-Hui Zhang

We have developed two estimators: (i) the {\em tie estimator}, and (ii) the {\em absolute minimum estimator}.

In Defense of MinHash Over SimHash

no code implementations16 Jul 2014 Anshumali Shrivastava, Ping Li

To provide a common basis for comparison, we evaluate retrieval results in terms of $\mathcal{S}$ for both MinHash and SimHash.

Improved Densification of One Permutation Hashing

1 code implementation18 Jun 2014 Anshumali Shrivastava, Ping Li

The existing work on densification of one permutation hashing reduces the query processing cost of the $(K, L)$-parameterized Locality Sensitive Hashing (LSH) algorithm with minwise hashing, from $O(dKL)$ to merely $O(d + KL)$, where $d$ is the number of nonzeros of the data vector, $K$ is the number of hashes in each hash table, and $L$ is the number of hash tables.

Online Optimization for Large-Scale Max-Norm Regularization

no code implementations12 Jun 2014 Jie Shen, Huan Xu, Ping Li

Max-norm regularizer has been extensively studied in the last decade as it promotes an effective low-rank estimation for the underlying data.

Matrix Completion

Asymmetric LSH (ALSH) for Sublinear Time Maximum Inner Product Search (MIPS)

no code implementations NeurIPS 2014 Anshumali Shrivastava, Ping Li

Our proposal is based on an interesting mathematical phenomenon in which inner products, after independent asymmetric transformations, can be converted into the problem of approximate near neighbor search.

CoRE Kernels

no code implementations24 Apr 2014 Ping Li

The term "CoRE kernel" stands for correlation-resemblance kernel.

General Classification

Graph Kernels via Functional Embedding

no code implementations21 Apr 2014 Anshumali Shrivastava, Ping Li

We propose a representation of graph as a functional object derived from the power iteration of the underlying adjacency matrix.

General Classification Graph Classification

Advancing Matrix Completion by Modeling Extra Structures beyond Low-Rankness

no code implementations17 Apr 2014 Guangcan Liu, Ping Li

To better handle non-uniform data, in this paper we propose a method termed Low-Rank Factor Decomposition (LRFD), which imposes an additional restriction that the data points must be represented as linear combinations of the bases in a dictionary constructed or learnt in advance.

Low-Rank Matrix Completion

A New Space for Comparing Graphs

no code implementations17 Apr 2014 Anshumali Shrivastava, Ping Li

We show that the proposed matrix representation encodes the spectrum of the underlying adjacency matrix and it also contains information about the counts of small sub-structures present in the graph such as triangles and small paths.

General Classification

Recovery of Coherent Data via Low-Rank Dictionary Pursuit

no code implementations NeurIPS 2014 Guangcan Liu, Ping Li

More precisely, we mathematically prove that if the dictionary itself is low-rank then LRR is immune to the coherence parameter which increases with the underlying cluster number.

Coding for Random Projections and Approximate Near Neighbor Search

no code implementations31 Mar 2014 Ping Li, Michael Mitzenmacher, Anshumali Shrivastava

This technical note compares two coding (quantization) schemes for random projections in the context of sub-linear time approximate near neighbor search.

Quantization

Multi-label ensemble based on variable pairwise constraint projection

no code implementations8 Mar 2014 Ping Li, Hong Li, Min Wu

For the boosting-like strategy, we employ both the variable pairwise constraints and the bootstrap steps to diversify the base classifiers.

Classification General Classification +1

Sparse Recovery with Very Sparse Compressed Counting

no code implementations31 Dec 2013 Ping Li, Cun-Hui Zhang, Tong Zhang

In this paper, we adopt very sparse Compressed Counting for nonnegative signal recovery.

Adaptive Stochastic Alternating Direction Method of Multipliers

no code implementations16 Dec 2013 Peilin Zhao, Jinwei Yang, Tong Zhang, Ping Li

The Alternating Direction Method of Multipliers (ADMM) has been studied for years.

Beyond Pairwise: Provably Fast Algorithms for Approximate k-Way Similarity Search

no code implementations NeurIPS 2013 Anshumali Shrivastava, Ping Li

We go beyond the notion of pairwise similarity and look into search problems with $k$-way similarity functions.

Sign Cauchy Projections and Chi-Square Kernel

no code implementations NeurIPS 2013 Ping Li, Gennady Samorodnitsk, John Hopcroft

The method of Cauchy random projections is popular for computing the $l_1$ distance in high dimension.

Gradient Hard Thresholding Pursuit for Sparsity-Constrained Optimization

no code implementations22 Nov 2013 Xiao-Tong Yuan, Ping Li, Tong Zhang

Numerical evidences show that our method is superior to the state-of-the-art greedy selection methods in sparse logistic regression and sparse precision matrix estimation tasks.

Compressive Sensing

Learning Pairwise Graphical Models with Nonlinear Sufficient Statistics

no code implementations21 Nov 2013 Xiao-Tong Yuan, Ping Li, Tong Zhang

We investigate a generic problem of learning pairwise exponential family graphical models with pairwise sufficient statistics defined by a global mapping function, e. g., Mercer kernels.

Compressed Counting Meets Compressed Sensing

no code implementations3 Oct 2013 Ping Li, Cun-Hui Zhang, Tong Zhang

In particular, when p->0 the required number of measurements is essentially M=K\log N, where K is the number of nonzero coordinates of the signal.

Coding for Random Projections

no code implementations9 Aug 2013 Ping Li, Michael Mitzenmacher, Anshumali Shrivastava

The method of random projections has become very popular for large-scale applications in statistical learning, information retrieval, bio-informatics and other applications.

Information Retrieval Quantization

Sign Stable Projections, Sign Cauchy Projections and Chi-Square Kernels

no code implementations5 Aug 2013 Ping Li, Gennady Samorodnitsky, John Hopcroft

The method of stable random projections is popular for efficiently computing the Lp distances in high dimension (where 0<p<=2), using small space.

Entropy Estimations Using Correlated Symmetric Stable Random Projections

no code implementations NeurIPS 2012 Ping Li, Cun-Hui Zhang

Methods for efficiently estimating the Shannon entropy of data streams have important applications in learning, data mining, and network anomaly detections (e. g., the DDoS attacks).

One Permutation Hashing

no code implementations NeurIPS 2012 Ping Li, Art Owen, Cun-Hui Zhang

While minwise hashing is promising for large-scale learning in massive binary data, the preprocessing cost is prohibitive as it requires applying (e. g.,) $k=500$ permutations on the data.

Hashing Algorithms for Large-Scale Learning

no code implementations NeurIPS 2011 Ping Li, Anshumali Shrivastava, Joshua L. Moore, Arnd C. König

Minwise hashing is a standard technique in the context of search for efficiently computing set similarities.

b-Bit Minwise Hashing for Large-Scale Linear SVM

1 code implementation23 May 2011 Ping Li, Joshua Moore, Christian Konig

Interestingly, our proof for the positive definiteness of the b-bit minwise hashing kernel naturally suggests a simple strategy to integrate b-bit hashing with linear SVM.

One sketch for all: Theory and Application of Conditional Random Sampling

no code implementations NeurIPS 2008 Ping Li, Kenneth W. Church, Trevor J. Hastie

Conditional Random Sampling (CRS) was originally proposed for efficiently computing pairwise ($l_2$, $l_1$) distances, in static, large-scale, and sparse data sets such as text and Web data.

Cannot find the paper you are looking for? You can Submit a new open access paper.