no code implementations • 19 Jul 2024 • Ka-Ho Chow, Sihao Hu, Tiansheng Huang, Ling Liu
Second, we incorporate a perceptibility optimization to preserve the visual quality of the protected facial images.
no code implementations • 2 Jul 2024 • Yuxuan Zhu, Michael Mandulak, Kerui Wu, George Slota, Yuseok Jeon, Ka-Ho Chow, Lei Yu
Meanwhile, graph reduction techniques, including coarsening and sparsification, which have long been employed to improve the scalability of large graph computational tasks, have recently emerged as effective methods for accelerating GNN training on large-scale graphs.
1 code implementation • 15 Apr 2024 • Nawrin Tabassum, Ka-Ho Chow, Xuyu Wang, Wenbin Zhang, Yanzhao Wu
Second, we propose three early-stopping techniques to effectively reduce the computational costs of these privacy attacks.
1 code implementation • 5 Apr 2024 • Selim Furkan Tekin, Fatih Ilhan, Tiansheng Huang, Sihao Hu, Ka-Ho Chow, Margaret L. Loper, Ling Liu
This paper presents FusionShot, a focal diversity optimized few-shot ensemble learning approach for boosting the robustness and generalization performance of pre-trained few-shot models.
no code implementations • 6 Feb 2024 • Lei Yu, Meng Han, Yiming Li, Changting Lin, Yao Zhang, Mingyang Zhang, Yan Liu, Haiqin Weng, Yuseok Jeon, Ka-Ho Chow, Stacy Patterson
Vertical Federated Learning (VFL) is a federated learning paradigm where multiple participants, who share the same set of samples but hold different features, jointly train machine learning models.
no code implementations • 2 Jan 2024 • Ka-Ho Chow, Wenqi Wei, Lei Yu
This paper proposes Imperio, which harnesses the language understanding capabilities of NLP models to enrich backdoor attacks.
1 code implementation • 17 Nov 2023 • Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, Ling Liu
This paper presents a novel deep ensemble pruning approach, which can efficiently identify smaller deep ensembles and provide higher ensemble accuracy than the entire deep ensemble of a large number of member networks.
1 code implementation • 3 Oct 2023 • Yanzhao Wu, Ka-Ho Chow, Wenqi Wei, Ling Liu
We show that this two-tier heterogeneity driven ensemble construction method can compose an ensemble team that promotes high ensemble diversity and low negative correlation among member models of the ensemble, strengthening ensemble robustness against both negative examples and adversarial attacks.
1 code implementation • 10 May 2023 • Wenqi Wei, Ling Liu, Jingya Zhou, Ka-Ho Chow, Yanzhao Wu
Next, we present a gradient leakage resilient approach to securing distributed SGD in federated learning, with differential privacy controlled noise as the tool.
1 code implementation • CVPR 2023 • Ka-Ho Chow, Ling Liu, Wenqi Wei, Fatih Ilhan, Yanzhao Wu
Based on the insights, we introduce a three-tier forensic framework to identify and expel Trojaned gradients and reclaim the performance over the course of FL.
1 code implementation • 15 Jan 2023 • Fatih Ilhan, Ka-Ho Chow, Sihao Hu, Tiansheng Huang, Selim Tekin, Wenqi Wei, Yanzhao Wu, Myungjin Lee, Ramana Kompella, Hugo Latapie, Gaowen Liu, Ling Liu
Instead of having every sample go through all DNN layers during prediction, EENet learns an early exit scheduler, which can intelligently terminate the inference earlier for certain predictions, which the model has high confidence of early exit.
1 code implementation • CVPR 2021 • Yanzhao Wu, Ling Liu, Zhongwei Xie, Ka-Ho Chow, Wenqi Wei
Our new metrics significantly improve the intrinsic correlation between high ensemble diversity and high ensemble accuracy.
1 code implementation • 20 Oct 2020 • Yanzhao Wu, Ling Liu, Zhongwei Xie, Juhyun Bae, Ka-Ho Chow, Wenqi Wei
Ensemble learning is gaining renewed interests in recent years.
1 code implementation • 11 Jul 2020 • Ka-Ho Chow, Ling Liu, Mehmet Emre Gursoy, Stacey Truex, Wenqi Wei, Yanzhao Wu
We demonstrate that the proposed framework can serve as a methodical benchmark for analyzing adversarial behaviors and risks in real-time object detection systems.
no code implementations • 5 Jun 2020 • Stacey Truex, Ling Liu, Ka-Ho Chow, Mehmet Emre Gursoy, Wenqi Wei
However, in federated learning model parameter updates are collected iteratively from each participant and consist of high dimensional, continuous values with high precision (10s of digits after the decimal point), making existing LDP protocols inapplicable.
2 code implementations • 22 Apr 2020 • Wenqi Wei, Ling Liu, Margaret Loper, Ka-Ho Chow, Mehmet Emre Gursoy, Stacey Truex, Yanzhao Wu
FL offers default client privacy by allowing clients to keep their sensitive data on local devices and to only share local training parameter updates with the federated server.
2 code implementations • 9 Apr 2020 • Ka-Ho Chow, Ling Liu, Mehmet Emre Gursoy, Stacey Truex, Wenqi Wei, Yanzhao Wu
The rapid growth of real-time huge data capturing has pushed the deep learning and data analytic computing to the edge systems.
no code implementations • 1 Oct 2019 • Wenqi Wei, Ling Liu, Margaret Loper, Ka-Ho Chow, Emre Gursoy, Stacey Truex, Yanzhao Wu
Deep neural network (DNN) has demonstrated its success in multiple domains.
no code implementations • 29 Aug 2019 • Ling Liu, Wenqi Wei, Ka-Ho Chow, Margaret Loper, Emre Gursoy, Stacey Truex, Yanzhao Wu
In this paper we first give an overview of the concept of ensemble diversity and examine the three types of ensemble diversity in the context of DNN classifiers.
no code implementations • 21 Aug 2019 • Ka-Ho Chow, Wenqi Wei, Yanzhao Wu, Ling Liu
Deep neural networks (DNNs) have demonstrated impressive performance on many challenging machine learning tasks.
1 code implementation • 18 Aug 2019 • Yanzhao Wu, Ling Liu, Juhyun Bae, Ka-Ho Chow, Arun Iyengar, Calton Pu, Wenqi Wei, Lei Yu, Qi Zhang
Learning Rate (LR) is an important hyper-parameter to tune for effective training of deep neural networks (DNNs).
no code implementations • 18 Jan 2019 • Yifeng Zhang, Ka-Ho Chow, S. -H. Gary Chan
In this paper, we develop a Depth-Adaptive Long Short-Term Memory (DA-LSTM) architecture, which can dynamically adjust the structure depending on information distribution without prior knowledge.
no code implementations • 20 Nov 2018 • Ka-Ho Chow, Anish Hiranandani, Yifeng Zhang, S. -H. Gary Chan
Representation learning of pedestrian trajectories transforms variable-length timestamp-coordinate tuples of a trajectory into a fixed-length vector representation that summarizes spatiotemporal characteristics.