Search Results for author: Haoqin Tu

Found 17 papers, 13 papers with code

AttnGCG: Enhancing Jailbreaking Attacks on LLMs with Attention Manipulation

1 code implementation11 Oct 2024 Zijun Wang, Haoqin Tu, Jieru Mei, Bingchen Zhao, Yisen Wang, Cihang Xie

This paper studies the vulnerabilities of transformer-based Large Language Models (LLMs) to jailbreaking attacks, focusing specifically on the optimization-based Greedy Coordinate Gradient (GCG) strategy.

Safety Alignment

VHELM: A Holistic Evaluation of Vision Language Models

1 code implementation9 Oct 2024 Tony Lee, Haoqin Tu, Chi Heem Wong, Wenhao Zheng, Yiyang Zhou, Yifan Mai, Josselin Somerville Roberts, Michihiro Yasunaga, Huaxiu Yao, Cihang Xie, Percy Liang

Current benchmarks for assessing vision-language models (VLMs) often focus on their perception or problem-solving capabilities and neglect other critical aspects such as fairness, multilinguality, or toxicity.

Fairness

A Preliminary Study of o1 in Medicine: Are We Closer to an AI Doctor?

no code implementations23 Sep 2024 Yunfei Xie, Juncheng Wu, Haoqin Tu, Siwei Yang, Bingchen Zhao, Yongshuo Zong, Qiao Jin, Cihang Xie, Yuyin Zhou

Large language models (LLMs) have exhibited remarkable capabilities across various domains and tasks, pushing the boundaries of our knowledge in learning and cognition.

Hallucination MedQA +1

MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?

1 code implementation5 Jul 2024 Zhaorun Chen, Yichao Du, Zichen Wen, Yiyang Zhou, Chenhang Cui, Zhenzhen Weng, Haoqin Tu, Chaoqi Wang, Zhengwei Tong, Qinglan Huang, Canyu Chen, Qinghao Ye, Zhihong Zhu, Yuqing Zhang, Jiawei Zhou, Zhuokai Zhao, Rafael Rafailov, Chelsea Finn, Huaxiu Yao

Compared with open-source VLMs, smaller-sized scoring models can provide better feedback regarding text-image alignment and image quality, while VLMs provide more accurate feedback regarding safety and generation bias due to their stronger reasoning capabilities.

Hallucination Text-to-Image Generation

Autoregressive Pretraining with Mamba in Vision

1 code implementation11 Jun 2024 Sucheng Ren, Xianhang Li, Haoqin Tu, Feng Wang, Fangxun Shu, Lei Zhang, Jieru Mei, Linjie Yang, Peng Wang, Heng Wang, Alan Yuille, Cihang Xie

The vision community has started to build with the recently developed state space model, Mamba, as the new backbone for a range of tasks.

Mamba

How Far Are We From AGI: Are LLMs All We Need?

1 code implementation16 May 2024 Tao Feng, Chuanyang Jin, Jingyu Liu, Kunlun Zhu, Haoqin Tu, Zirui Cheng, GuanYu Lin, Jiaxuan You

The evolution of artificial intelligence (AI) has profoundly impacted human society, driving significant advancements in multiple sectors.

Tuning LayerNorm in Attention: Towards Efficient Multi-Modal LLM Finetuning

no code implementations18 Dec 2023 Bingchen Zhao, Haoqin Tu, Chen Wei, Jieru Mei, Cihang Xie

This paper introduces an efficient strategy to transform Large Language Models (LLMs) into Multi-Modal Large Language Models (MLLMs).

Domain Adaptation

How Many Unicorns Are in This Image? A Safety Evaluation Benchmark for Vision LLMs

1 code implementation27 Nov 2023 Haoqin Tu, Chenhang Cui, Zijun Wang, Yiyang Zhou, Bingchen Zhao, Junlin Han, Wangchunshu Zhou, Huaxiu Yao, Cihang Xie

Different from prior studies, we shift our focus from evaluating standard performance to introducing a comprehensive safety evaluation suite, covering both out-of-distribution (OOD) generalization and adversarial robustness.

Adversarial Robustness Visual Question Answering (VQA) +1

Sight Beyond Text: Multi-Modal Training Enhances LLMs in Truthfulness and Ethics

1 code implementation13 Sep 2023 Haoqin Tu, Bingchen Zhao, Chen Wei, Cihang Xie

Multi-modal large language models (MLLMs) are trained based on large language models (LLM), with an enhanced capability to comprehend multi-modal inputs and generate textual responses.

Ethics TruthfulQA

ZeroGen: Zero-shot Multimodal Controllable Text Generation with Multiple Oracles

1 code implementation29 Jun 2023 Haoqin Tu, Bowen Yang, Xianfeng Zhao

Automatically generating textual content with desired attributes is an ambitious task that people have pursued long.

News Generation Sentence

ReSee: Responding through Seeing Fine-grained Visual Knowledge in Open-domain Dialogue

1 code implementation23 May 2023 Haoqin Tu, Yitong Li, Fei Mi, Zhongliang Yang

To demonstrate the superiority and universality of the provided visual knowledge, we propose a simple but effective framework ReSee to add visual representation into vanilla dialogue models by modality concatenations.

An Overview on Controllable Text Generation via Variational Auto-Encoders

1 code implementation15 Nov 2022 Haoqin Tu, Yitong Li

Recent advances in neural-based generative modeling have reignited the hopes of having computer systems capable of conversing with humans and able to understand natural language.

Text Generation

PCAE: A Framework of Plug-in Conditional Auto-Encoder for Controllable Text Generation

1 code implementation7 Oct 2022 Haoqin Tu, Zhongliang Yang, Jinshuai Yang, Siyu Zhang, Yongfeng Huang

Visualization of the local latent prior well confirms the primary devotion in hidden space of the proposed model.

Text Generation

AdaVAE: Exploring Adaptive GPT-2s in Variational Auto-Encoders for Language Modeling

1 code implementation12 May 2022 Haoqin Tu, Zhongliang Yang, Jinshuai Yang, Yongfeng Huang

Variational Auto-Encoder (VAE) has become the de-facto learning paradigm in achieving representation learning and generation for natural language at the same time.

Conditional Text Generation Decoder +3

Cannot find the paper you are looking for? You can Submit a new open access paper.