Search Results for author: Wonje Jeung

Found 6 papers, 3 papers with code

Representation Bending for Large Language Model Safety

1 code implementation2 Apr 2025 Ashkan Yousefpour, Taeheon Kim, Ryan S. Kwon, Seungbeen Lee, Wonje Jeung, Seungju Han, Alvin Wan, Harrison Ngan, Youngjae Yu, Jonghyun Choi

Large Language Models (LLMs) have emerged as powerful tools, but their inherent safety risks - ranging from harmful content generation to broader societal harms - pose significant challenges.

Language Modeling Language Modelling +2

Adversarial Sample-Based Approach for Tighter Privacy Auditing in Final Model-Only Scenarios

no code implementations2 Dec 2024 Sangyeon Yoon, Wonje Jeung, Albert No

Auditing Differentially Private Stochastic Gradient Descent (DP-SGD) in the final model setting is challenging and often results in empirical lower bounds that are significantly looser than theoretical privacy guarantees.

Large Language Models Still Exhibit Bias in Long Text

no code implementations23 Oct 2024 Wonje Jeung, Dongjae Jeon, Ashkan Yousefpour, Jonghyun Choi

Existing fairness benchmarks for large language models (LLMs) primarily focus on simple tasks, such as multiple-choice questions, overlooking biases that may arise in more complex scenarios like long-text generation.

Fairness Multiple-choice +1

ReALFRED: An Embodied Instruction Following Benchmark in Photo-Realistic Environments

1 code implementation26 Jul 2024 Taewoong Kim, Cheolhong Min, Byeonghwi Kim, Jinyeon Kim, Wonje Jeung, Jonghyun Choi

To bridge the gap between these learning environments and deploying (i. e., real) environments, we propose the ReALFRED benchmark that employs real-world scenes, objects, and room layouts to learn agents to complete household tasks by understanding free-form language instructions and interacting with objects in large, multi-room and 3D-captured scenes.

Instruction Following

An Information Theoretic Evaluation Metric For Strong Unlearning

no code implementations28 May 2024 Dongjae Jeon, Wonje Jeung, Taeheon Kim, Albert No, Jonghyun Choi

Machine unlearning (MU) aims to remove the influence of specific data from trained models, addressing privacy concerns and ensuring compliance with regulations such as the "right to be forgotten."

CoLA Machine Unlearning +1

Learning Equi-angular Representations for Online Continual Learning

1 code implementation CVPR 2024 Minhyuk Seo, Hyunseo Koh, Wonje Jeung, Minjae Lee, San Kim, Hankook Lee, Sungjun Cho, Sungik Choi, Hyunwoo Kim, Jonghyun Choi

Online continual learning suffers from an underfitted solution due to insufficient training for prompt model update (e. g., single-epoch training).

Continual Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.