no code implementations • 23 May 2023 • WeiYe Zhao, Rui Chen, Yifan Sun, Ruixuan Liu, Tianhao Wei, Changliu Liu
Due to the diversity of algorithms and tasks, it remains difficult to compare existing safe RL algorithms.
no code implementations • 29 Nov 2022 • Christoph Breunig, Ruixuan Liu, Zhengfei Yu
We study a double robust Bayesian inference procedure on the average treatment effect (ATE) under unconfoundedness.
no code implementations • 18 Apr 2022 • Ruixuan Liu, Yanlin Wang, Yang Cao, Lingjuan Lyu, Weike Pan, Yun Chen, Hong Chen
Collecting and training over sensitive personal data raise severe privacy concerns in personalized recommendation systems, and federated learning can potentially alleviate the problem by training models over decentralized user data. However, a theoretically private solution in both the training and serving stages of federated recommendation is essential but still lacking. Furthermore, naively applying differential privacy (DP) to the two stages in federated recommendation would fail to achieve a satisfactory trade-off between privacy and utility due to the high-dimensional characteristics of model gradients and hidden representations. In this work, we propose a federated news recommendation method for achieving a better utility in model training and online serving under a DP guarantee. We first clarify the DP definition over behavior data for each round in the life-circle of federated recommendation systems. Next, we propose a privacy-preserving online serving mechanism under this definition based on the idea of decomposing user embeddings with public basic vectors and perturbing the lower-dimensional combination coefficients.
no code implementations • 16 Feb 2022 • Ruixuan Liu, Fangzhao Wu, Chuhan Wu, Yanlin Wang, Lingjuan Lyu, Hong Chen, Xing Xie
In this way, all the clients can participate in the model learning in FL, and the final model can be big and powerful enough.
1 code implementation • EMNLP 2021 • Jingwei Yi, Fangzhao Wu, Chuhan Wu, Ruixuan Liu, Guangzhong Sun, Xing Xie
However, the computation and communication cost of directly learning many existing news recommendation models in a federated way are unacceptable for user clients.
no code implementations • 16 Aug 2021 • Ruixuan Liu, Changliu Liu
Predicting human intention is critical to facilitating safe and efficient human-robot collaboration (HRC).
1 code implementation • 17 Sep 2020 • Ruixuan Liu, Yang Cao, Hong Chen, Ruoyang Guo, Masatoshi Yoshikawa
In this work, by leveraging the \textit{privacy amplification} effect in the recently proposed shuffle model of differential privacy, we achieve the best of two worlds, i. e., accuracy in the curator model and strong privacy without relying on any trusted party.
no code implementations • 24 Mar 2020 • Ruixuan Liu, Yang Cao, Masatoshi Yoshikawa, Hong Chen
To prevent privacy leakages from gradients that are calculated on users' sensitive data, local differential privacy (LDP) has been considered as a privacy guarantee in federated SGD recently.