A framework for learned sparse sketches

1 Jan 2021  ·  Simin Liu, Tianrui Liu, Ali Vakilian, Yulin Wan, David Woodruff ·

Sketching is a compression technique that can be applied to many problems to solve them quickly and approximately. The matrices used to project data to smaller dimensions are called sketches. In this work, we consider the problem of optimizing sketches to obtain low approximation error over a data distribution. We introduce a general framework for learning and applying input-independent sparse sketches. The sketch optimization procedure has two stages: one for optimizing the placement of the sketch's non-zero entries and another for optimizing their values. Next, we provide a way to apply learned sketches that has worst-case guarantees for approximation error. We instantiate this framework with three sketching applications: least-squares regression, low-rank approximation (LRA), and k-means clustering. Our experiments demonstrate that our approach substantially decreases approximation error compared to classical and naively learned sketches. Finally, we investigate the theoretical aspects of our approach. For regression and LRA, we show that our method obtains state-of-the art accuracy for fixed time complexity. For LRA, we prove that it is strictly better to include the first optimization stage. For k-means, we derive a more straightforward means of retaining approximation guarantees.

PDF Abstract
No code implementations yet. Submit your code now

Datasets


  Add Datasets introduced or used in this paper

Results from the Paper


  Submit results from this paper to get state-of-the-art GitHub badges and help the community compare results to other papers.

Methods


No methods listed for this paper. Add relevant methods here