no code implementations • NeurIPS 2023 • Rachael Hwee Ling Sim, Yehong Zhang, Trong Nghia Hoang, Xinyi Xu, Bryan Kian Hsiang Low, Patrick Jaillet
Finally, the mediator rewards each party with different posterior samples of the model parameters.
no code implementations • 22 Feb 2024 • Han Zhang, Lin Gui, Yu Lei, Yuanzhao Zhai, Yehong Zhang, Yulan He, Hui Wang, Yue Yu, Kam-Fai Wong, Bin Liang, Ruifeng Xu
Reinforcement Learning from Human Feedback (RLHF) is commonly utilized to improve the alignment of Large Language Models (LLMs) with human preferences.
1 code implementation • 1 Jan 2024 • Jinglong Luo, Yehong Zhang, Zhuo Zhang, JiaQi Zhang, Xin Mu, Hui Wang, Yue Yu, Zenglin Xu
However, the application of SMPC in Privacy-Preserving Inference (PPI) for large language models, particularly those based on the Transformer architecture, often leads to considerable slowdowns or declines in performance.
no code implementations • 19 Dec 2023 • Xin Mu, Yu Wang, Zhengan Huang, Junzuo Lai, Yehong Zhang, Hui Wang, Yue Yu
In the rapidly growing digital economy, protecting intellectual property (IP) associated with digital products has become increasingly important.
no code implementations • 4 Aug 2023 • Xin Mu, Yu Wang, Yehong Zhang, JiaQi Zhang, Hui Wang, Yang Xiang, Yue Yu
Understanding the life cycle of the machine learning (ML) model is an intriguing area of research (e. g., understanding where the model comes from, how it is trained, and how it is used).
no code implementations • 1 Aug 2023 • Mohit Rajpal, Lac Gia Tran, Yehong Zhang, Bryan Kian Hsiang Low
Derivative-free approaches such as Bayesian Optimization mitigate the dependency on the quality of gradient feedback, but are known to scale poorly in the high-dimension setting of complex decision making models.
no code implementations • 26 Jun 2023 • Jinglong Luo, Yehong Zhang, JiaQi Zhang, Shuang Qin, Hui Wang, Yue Yu, Zenglin Xu
In contrast to existing studies that protect the data privacy of GPR via homomorphic encryption, differential privacy, or federated learning, our proposed method is more practical and can be used to preserve the data privacy of both the model inputs and outputs for various data-sharing scenarios (e. g., horizontally/vertically-partitioned data).
4 code implementations • 26 Apr 2021 • Wei Zeng, Xiaozhe Ren, Teng Su, Hui Wang, Yi Liao, Zhiwei Wang, Xin Jiang, ZhenZhang Yang, Kaisheng Wang, Xiaoda Zhang, Chen Li, Ziyan Gong, Yifan Yao, Xinjing Huang, Jun Wang, Jianfeng Yu, Qi Guo, Yue Yu, Yan Zhang, Jin Wang, Hengtao Tao, Dasen Yan, Zexuan Yi, Fang Peng, Fangqing Jiang, Han Zhang, Lingfeng Deng, Yehong Zhang, Zhe Lin, Chao Zhang, Shaojie Zhang, Mingyue Guo, Shanzhi Gu, Gaojun Fan, YaoWei Wang, Xuefeng Jin, Qun Liu, Yonghong Tian
To enhance the generalization ability of PanGu-$\alpha$, we collect 1. 1TB high-quality Chinese data from a wide range of domains to pretrain the model.
Ranked #1 on Reading Comprehension (One-Shot) on DuReader
Cloze (multi-choices) (Few-Shot) Cloze (multi-choices) (One-Shot) +19
no code implementations • 1 Jan 2021 • Mohit Rajpal, Yehong Zhang, Bryan Kian Hsiang Low
Pruning is an approach to alleviate overparameterization of deep neural networks (DNN) by zeroing out or pruning DNN elements with little to no efficacy at a given task.
no code implementations • ICML 2020 • Rachael Hwee Ling Sim, Yehong Zhang, Mun Choon Chan, Bryan Kian Hsiang Low
This paper proposes to value a party's reward based on Shapley value and information gain on model parameters given its data.
no code implementations • 5 Dec 2019 • Tong Teng, Jie Chen, Yehong Zhang, Kian Hsiang Low
To achieve this, we represent the probabilistic kernel as an additional variational variable in a variational inference (VI) framework for SGPR models where its posterior belief is learned together with that of the other variational variables (i. e., inducing variables and kernel hyperparameters).
no code implementations • 17 Jun 2019 • Yehong Zhang, Zhongxiang Dai, Kian Hsiang Low
This paper presents novel mixed-type Bayesian optimization (BO) algorithms to accelerate the optimization of a target objective function by exploiting correlated auxiliary information of binary type that can be more cheaply obtained, such as in policy search for reinforcement learning and hyperparameter tuning of machine learning models with early stopping.
1 code implementation • 21 Nov 2015 • Yehong Zhang, Trong Nghia Hoang, Kian Hsiang Low, Mohan Kankanhalli
This paper addresses the problem of active learning of a multi-output Gaussian process (MOGP) model representing multiple types of coexisting correlated environmental phenomena.