no code implementations • 16 Aug 2023 • Xinmeng Huang, Ping Li, Xiaoyun Li
The existing approaches either cannot accommodate arbitrary data heterogeneity or partial participation, or require stringent conditions on compression.
no code implementations • 10 Aug 2023 • Guyu Jiang, Xiaoyun Li, Rongrong Jing, Ruoqi Zhao, Xingliang Ni, Guodong Cao, Ning Hu
Click-through rate (CTR) prediction is a crucial task in the context of an online on-demand food delivery (OFD) platform for precisely estimating the probability of a user clicking on food items.
no code implementations • 13 Jun 2023 • Xiaoyun Li, Ping Li
Minwise hashing (MinHash) is a standard algorithm widely used in the industry, for large-scale search and learning applications with the binary (0/1) Jaccard similarity.
no code implementations • 22 May 2023 • Ping Li, Xiaoyun Li
Among the presented algorithms, iDP-SignRP is remarkably effective under the setting of ``individual differential privacy'' (iDP), based on sign random projections (SignRP).
no code implementations • 26 Apr 2023 • Xinyi Zheng, Weijie Zhao, Xiaoyun Li, Ping Li
To retrieve personalized campaigns and creatives while protecting user privacy, digital advertising is shifting from member-based identity to cohort-based identity.
no code implementations • 7 Feb 2023 • Ping Li, Xiaoyun Li
We show that the estimation variance is essentially: $(s-1)A + \frac{D-k}{D-1}\frac{1}{k}\left[ (1-\rho^2)^2 -2A\right]$, where $A\geq 0$ is a function of the data ($u, v$).
no code implementations • 25 Nov 2022 • Xiaoyun Li, Ping Li
Moreover, we develop a new analysis of the EF under partial client participation, which is an important scenario in FL.
no code implementations • 26 Jun 2022 • Chenglin Fan, Ping Li, Xiaoyun Li
When designing clustering algorithms, the choice of initial centers is crucial for the quality of the learned clusters.
no code implementations • ICLR 2022 • Xiaoyun Li, Belhal Karimi, Ping Li
We study COMP-AMS, a distributed optimization framework based on gradient averaging and adaptive AMSGrad algorithm.
no code implementations • 5 Jan 2022 • Weijie Zhao, Xuewu Jiao, Mingqing Hu, Xiaoyun Li, Xiangyu Zhang, Ping Li
In this paper, we propose a hardware-aware training workflow that couples the hardware topology into the algorithm design.
no code implementations • 18 Nov 2021 • Xiaoyun Li, Ping Li
Note that C-MinHash is different from the well-known work on "One Permutation Hashing (OPH)" published in NIPS'12.
no code implementations • 1 Oct 2021 • Belhal Karimi, Ping Li, Xiaoyun Li
In the emerging paradigm of Federated Learning (FL), large amount of clients such as mobile devices are used to train possibly high-dimensional models on their respective data.
no code implementations • 29 Sep 2021 • Chenglin Fan, Ping Li, Xiaoyun Li
Our method, named the HST initialization, can also be easily extended to the setting of differential privacy (DP) to generate private initial centers.
no code implementations • 29 Sep 2021 • Xiaoyun Li, Ping Li
We show the locality-sensitivity of SignRFF, and propose a new measure, called ranking efficiency, to theoretically compare different Locality-Sensitive Hashing (LSH) methods with practical implications.
no code implementations • 29 Sep 2021 • Xiaoyun Li, Ping Li
Minwise hashing (MinHash) is an important and practical algorithm for generating random hashes to approximate the Jaccard (resemblance) similarity in massive binary (0/1) data.
no code implementations • 10 Sep 2021 • Xiangyi Chen, Xiaoyun Li, Ping Li
While adaptive gradient methods have been proven effective for training neural nets, the study of adaptive gradient methods in federated learning is scarce.
no code implementations • 10 Sep 2021 • Xiaoyun Li, Ping Li
That is, one single permutation is used for both the initial pre-processing step to break the structures in the data and the circulant hashing step to generate $K$ hashes.
no code implementations • 7 Sep 2021 • Xiaoyun Li, Ping Li
Unlike classical MinHash, these $K$ hashes are obviously correlated, but we are able to provide rigorous proofs that we still obtain an unbiased estimate of the Jaccard similarity and the theoretical variance is uniformly smaller than that of the classical MinHash with $K$ independent permutations.
no code implementations • 25 Feb 2021 • Xiaoyun Li, Ping Li
Closely related to RP, the method of random Fourier features (RFF) has also become popular, for approximating the Gaussian kernel.
no code implementations • 11 Aug 2020 • Farzin Haddadpour, Belhal Karimi, Ping Li, Xiaoyun Li
Communication complexity and privacy are the two key challenges in Federated Learning where the goal is to perform a distributed learning through a large volume of devices.
no code implementations • 2 Apr 2020 • Xiaoyun Li, Jie Gui, Ping Li
In this paper, we propose the kernel version of multi-view discriminant analysis, called kernel multi-view discriminant analysis (KMvDA).
no code implementations • 2 Apr 2020 • Xiaoyun Li, Chengxi Wu, Ping Li
Feature selection is an important tool to deal with high dimensional data.
no code implementations • NeurIPS 2019 • Xiaoyun Li, Ping Li
The method of random projection has been a popular tool for data compression, similarity search, and machine learning.
no code implementations • NeurIPS 2019 • Ping Li, Xiaoyun Li, Cun-Hui Zhang
Jaccard similarity is widely used as a distance measure in many machine learning and search applications.
no code implementations • NeurIPS 2019 • Xiaoyun Li, Ping Li
In this paper, we consider the learning problem where the projected data is further compressed by scalar quantization, which is called quantized compressive learning.
no code implementations • 25 Sep 2019 • Jun-Kun Wang, Xiaoyun Li, Ping Li
Perhaps the only methods that enjoy convergence guarantees are the ones that sample the perturbed points uniformly from a unit sphere or from a multivariate Gaussian distribution with an isotropic covariance.
no code implementations • ICLR 2019 • Jun-Kun Wang, Xiaoyun Li, Ping Li
We consider new variants of optimization algorithms.
no code implementations • ICLR 2020 • Jun-Kun Wang, Xiaoyun Li, Belhal Karimi, Ping Li
We propose a new variant of AMSGrad, a popular adaptive gradient based optimization algorithm widely used for training deep neural networks.