no code implementations • 7 Feb 2024 • Yuji Roh, Qingyun Liu, Huan Gui, Zhe Yuan, Yujin Tang, Steven Euijong Whang, Liang Liu, Shuchao Bi, Lichan Hong, Ed H. Chi, Zhe Zhao
By combining two complementing models, LEVI effectively suppresses problematic features in both the fine-tuning data and pre-trained model and preserves useful features for new tasks.
no code implementations • 5 Feb 2023 • Yuji Roh, Kangwook Lee, Steven Euijong Whang, Changho Suh
First, we analytically show that existing in-processing fair algorithms have fundamental limits in accuracy and group fairness.
no code implementations • 13 Dec 2021 • Steven Euijong Whang, Yuji Roh, Hwanjun Song, Jae-Gil Lee
In this survey, we study the research landscape for data collection and data quality primarily for deep learning applications.
no code implementations • NeurIPS 2021 • Yuji Roh, Kangwook Lee, Steven Euijong Whang, Changho Suh
In this work, we propose a sample selection-based algorithm for fair and robust training.
no code implementations • 15 Jan 2021 • Steven Euijong Whang, Ki Hyun Tae, Yuji Roh, Geon Heo
Second, responsible AI must be broadly supported, preferably in all steps of machine learning.
1 code implementation • ICLR 2021 • Yuji Roh, Kangwook Lee, Steven Euijong Whang, Changho Suh
We address this problem via the lens of bilevel optimization.
no code implementations • 7 Apr 2020 • Geon Heo, Yuji Roh, Seonghyeon Hwang, Dayun Lee, Steven Euijong Whang
We propose Inspector Gadget, an image labeling system that combines crowdsourcing, data augmentation, and data programming to produce weak labels at scale for image classification.
1 code implementation • ICML 2020 • Yuji Roh, Kangwook Lee, Steven Euijong Whang, Changho Suh
Trustworthy AI is a critical issue in machine learning where, in addition to training a model that is accurate, one must consider both fair and robust training in the presence of data bias and poisoning.
no code implementations • 25 Sep 2019 • Yuji Roh, Kangwook Lee, Gyeong Jo Hwang, Steven Euijong Whang, Changho Suh
We consider the problem of fair and robust model training in the presence of data poisoning.
no code implementations • 22 Apr 2019 • Ki Hyun Tae, Yuji Roh, Young Hun Oh, Hyunsu Kim, Steven Euijong Whang
As machine learning is used in sensitive applications, it becomes imperative that the trained model is accurate, fair, and robust to attacks.
no code implementations • 8 Nov 2018 • Yuji Roh, Geon Heo, Steven Euijong Whang
Interestingly, recent research in data collection comes not only from the machine learning, natural language, and computer vision communities, but also from the data management community due to the importance of handling large amounts of data.