Search Results for author: Yujian Liu

Found 15 papers, 11 papers with code

Defending LLM Watermarking Against Spoofing Attacks with Contrastive Representation Learning

1 code implementation9 Apr 2025 Li An, Yujian Liu, Yepeng Liu, Yang Zhang, Yuheng Bu, Shiyu Chang

We identify two core challenges that make defending against spoofing difficult: (1) the need for watermarks to be both sensitive to semantic-distorting changes and insensitive to semantic-preserving edits, and (2) the contradiction between the need to detect global semantic shifts and the local, auto-regressive nature of most watermarking schemes.

Representation Learning

ThinkPrune: Pruning Long Chain-of-Thought of LLMs via Reinforcement Learning

1 code implementation2 Apr 2025 Bairu Hou, Yang Zhang, Jiabao Ji, Yujian Liu, Kaizhi Qian, Jacob Andreas, Shiyu Chang

To fill this gap, ThinkPrune offers a simple solution that continuously trains the long-thinking LLMs via reinforcement learning (RL) with an added token limit, beyond which any unfinished thoughts and answers will be discarded, resulting in a zero reward.

Reinforcement Learning (RL)

Fictitious Synthetic Data Can Improve LLM Factuality via Prerequisite Learning

1 code implementation25 Oct 2024 Yujian Liu, Shiyu Chang, Tommi Jaakkola, Yang Zhang

Recent studies have identified one aggravating factor of LLM hallucinations as the knowledge inconsistency between pre-training and fine-tuning, where unfamiliar fine-tuning data mislead the LLM to fabricate plausible but wrong outputs.

Revisiting Who's Harry Potter: Towards Targeted Unlearning from a Causal Intervention Perspective

1 code implementation24 Jul 2024 Yujian Liu, Yang Zhang, Tommi Jaakkola, Shiyu Chang

This paper investigates Who's Harry Potter (WHP), a pioneering yet insufficiently understood method for LLM unlearning.

Reversing the Forget-Retain Objectives: An Efficient LLM Unlearning Framework from Logit Difference

1 code implementation12 Jun 2024 Jiabao Ji, Yujian Liu, Yang Zhang, Gaowen Liu, Ramana Rao Kompella, Sijia Liu, Shiyu Chang

To achieve both goals, a mainstream class of LLM unlearning methods introduces an optimization framework with a combination of two objectives - maximizing the prediction loss on the forget documents while minimizing that on the retain documents, which suffers from two challenges, degenerated output and catastrophic forgetting.

Towards Efficient Information Fusion: Concentric Dual Fusion Attention Based Multiple Instance Learning for Whole Slide Images

no code implementations21 Mar 2024 Yujian Liu, Ruoxuan Wu, Xinjie Shen, Zihuang Lu, Lingyu Liang, Haiyu Zhou, Shipu Xu, Shaoai Cai, Shidang Xu

In the realm of digital pathology, multi-magnification Multiple Instance Learning (multi-mag MIL) has proven effective in leveraging the hierarchical structure of Whole Slide Images (WSIs) to reduce information loss and redundant data.

Multiple Instance Learning whole slide images

Augment before You Try: Knowledge-Enhanced Table Question Answering via Table Expansion

1 code implementation28 Jan 2024 Yujian Liu, Jiabao Ji, Tong Yu, Ryan Rossi, Sungchul Kim, Handong Zhao, Ritwik Sinha, Yang Zhang, Shiyu Chang

Table question answering is a popular task that assesses a model's ability to understand and interact with structured data.

Question Answering

Correcting Diffusion Generation through Resampling

1 code implementation CVPR 2024 Yujian Liu, Yang Zhang, Tommi Jaakkola, Shiyu Chang

Despite diffusion models' superior capabilities in modeling complex distributions, there are still non-trivial distributional discrepancies between generated and ground-truth images, which has resulted in several notable problems in image generation, including missing object errors in text-to-image generation and low image quality.

Object Text to Image Generation +1

Decomposing Uncertainty for Large Language Models through Input Clarification Ensembling

1 code implementation15 Nov 2023 Bairu Hou, Yujian Liu, Kaizhi Qian, Jacob Andreas, Shiyu Chang, Yang Zhang

We show that, when aleatoric uncertainty arises from ambiguity or under-specification in LLM inputs, this approach makes it possible to factor an (unclarified) LLM's predictions into separate aleatoric and epistemic terms, using a decomposition similar to the one employed by Bayesian neural networks.

Uncertainty Quantification

All Things Considered: Detecting Partisan Events from News Media with Cross-Article Comparison

no code implementations28 Oct 2023 Yujian Liu, Xinliang Frederick Zhang, Kaijian Zou, Ruihong Huang, Nick Beauchamp, Lu Wang

Public opinion is shaped by the information news media provide, and that information in turn may be shaped by the ideological preferences of media outlets.

All

Harnessing the Spatial-Temporal Attention of Diffusion Models for High-Fidelity Text-to-Image Synthesis

1 code implementation ICCV 2023 Qiucheng Wu, Yujian Liu, Handong Zhao, Trung Bui, Zhe Lin, Yang Zhang, Shiyu Chang

We then impose spatial attention control by combining the attention over the entire text description and that over the local description of the particular object in the corresponding pixel region of that object.

Denoising Image Generation

Uncovering the Disentanglement Capability in Text-to-Image Diffusion Models

1 code implementation CVPR 2023 Qiucheng Wu, Yujian Liu, Handong Zhao, Ajinkya Kale, Trung Bui, Tong Yu, Zhe Lin, Yang Zhang, Shiyu Chang

Based on this finding, we further propose a simple, light-weight image editing algorithm where the mixing weights of the two text embeddings are optimized for style matching and content preservation.

Denoising Disentanglement

Cannot find the paper you are looking for? You can Submit a new open access paper.