Search Results for author: Xiaoyu Liang

Found 14 papers, 6 papers with code

Content-Rich AIGC Video Quality Assessment via Intricate Text Alignment and Motion-Aware Consistency

1 code implementation6 Feb 2025 Shangkun Sun, Xiaoyu Liang, Bowen Qu, Wei Gao

These models substantially mitigate flickering artifacts prevalent in prior models, enable longer and complex text prompts and generate longer videos with intricate, diverse motion patterns.

Video Generation Video Quality Assessment +1

Dynamic Token Reduction during Generation for Vision Language Models

no code implementations24 Jan 2025 Xiaoyu Liang, Chaofeng Guan, Jiaying Lu, Huiyao Chen, Huan Wang, Haoji Hu

Vision-Language Models (VLMs) have achieved notable success in multimodal tasks but face practical limitations due to the quadratic complexity of decoder attention mechanisms and autoregressive generation.

Decoder Token Reduction

IE-Bench: Advancing the Measurement of Text-Driven Image Editing for Human Perception Alignment

no code implementations17 Jan 2025 Shangkun Sun, Bowen Qu, Xiaoyu Liang, Songlin Fan, Wei Gao

In this work, we introduce the Text-driven Image Editing Benchmark suite (IE-Bench) to enhance the assessment of text-driven edited images.

Image Generation

Enhancing Facial Consistency in Conditional Video Generation via Facial Landmark Transformation

no code implementations12 Dec 2024 Lianrui Mu, Xingze Zhou, Wenjie Zheng, Jiangnan Ye, Xiaoyu Liang, Yuchen Yang, Jianhong Bai, Jiedong Zhuang, Haoji Hu

Existing methods often fail to maintain facial feature consistency due to mismatches between the facial landmarks extracted from source videos and the target facial features in the reference image.

Video Generation

Mitigating Hallucination in Visual-Language Models via Re-Balancing Contrastive Decoding

no code implementations10 Sep 2024 Xiaoyu Liang, Jiayuan Yu, Lianrui Mu, Jiedong Zhuang, Jiaqi Hu, Yuchen Yang, Jiangnan Ye, Lu Lu, Jian Chen, Haoji Hu

Concurrently, the visual branch focuses on the selection of significant tokens, refining the attention mechanism to highlight the primary subject.

Hallucination Image Captioning +2

VE-Bench: Subjective-Aligned Benchmark Suite for Text-Driven Video Editing Quality Assessment

2 code implementations21 Aug 2024 Shangkun Sun, Xiaoyu Liang, Songlin Fan, Wenxu Gao, Wei Gao

To the best of our knowledge, VE-Bench introduces the first quality assessment dataset for video editing and an effective subjective-aligned quantitative metric for this domain.

Video Alignment Video Editing +2

FALIP: Visual Prompt as Foveal Attention Boosts CLIP Zero-Shot Performance

no code implementations8 Jul 2024 Jiedong Zhuang, Jiaqi Hu, Lianrui Mu, Rui Hu, Xiaoyu Liang, Jiangnan Ye, Haoji Hu

CLIP has achieved impressive zero-shot performance after pre-training on a large-scale dataset consisting of paired image-text data.

image-classification Image Classification

NTIRE 2024 Quality Assessment of AI-Generated Content Challenge

no code implementations25 Apr 2024 Xiaohong Liu, Xiongkuo Min, Guangtao Zhai, Chunyi Li, Tengchuan Kou, Wei Sun, HaoNing Wu, Yixuan Gao, Yuqin Cao, ZiCheng Zhang, Xiele Wu, Radu Timofte, Fei Peng, Huiyuan Fu, Anlong Ming, Chuanming Wang, Huadong Ma, Shuai He, Zifei Dou, Shu Chen, Huacong Zhang, Haiyi Xie, Chengwei Wang, Baoying Chen, Jishen Zeng, Jianquan Yang, Weigang Wang, Xi Fang, Xiaoxin Lv, Jun Yan, Tianwu Zhi, Yabin Zhang, Yaohui Li, Yang Li, Jingwen Xu, Jianzhao Liu, Yiting Liao, Junlin Li, Zihao Yu, Yiting Lu, Xin Li, Hossein Motamednia, S. Farhad Hosseini-Benvidi, Fengbin Guan, Ahmad Mahmoudi-Aznaveh, Azadeh Mansouri, Ganzorig Gankhuyag, Kihwan Yoon, Yifang Xu, Haotian Fan, Fangyuan Kong, Shiling Zhao, Weifeng Dong, Haibing Yin, Li Zhu, Zhiling Wang, Bingchen Huang, Avinab Saha, Sandeep Mishra, Shashank Gupta, Rajesh Sureddi, Oindrila Saha, Luigi Celona, Simone Bianco, Paolo Napoletano, Raimondo Schettini, Junfeng Yang, Jing Fu, Wei zhang, Wenzhi Cao, Limei Liu, Han Peng, Weijun Yuan, Zhan Li, Yihang Cheng, Yifan Deng, Haohui Li, Bowen Qu, Yao Li, Shuqing Luo, Shunzhou Wang, Wei Gao, Zihao Lu, Marcos V. Conde, Xinrui Wang, Zhibo Chen, Ruling Liao, Yan Ye, Qiulin Wang, Bing Li, Zhaokun Zhou, Miao Geng, Rui Chen, Xin Tao, Xiaoyu Liang, Shangkun Sun, Xingyuan Ma, Jiaze Li, Mengduo Yang, Haoran Xu, Jie zhou, Shiding Zhu, Bohan Yu, Pengfei Chen, Xinrui Xu, Jiabin Shen, Zhichao Duan, Erfan Asadi, Jiahe Liu, Qi Yan, Youran Qu, Xiaohui Zeng, Lele Wang, Renjie Liao

A total of 196 participants have registered in the video track.

Image Quality Assessment Image Restoration +2

Exploring AIGC Video Quality: A Focus on Visual Harmony, Video-Text Consistency and Domain Distribution Gap

1 code implementation21 Apr 2024 Bowen Qu, Xiaoyu Liang, Shangkun Sun, Wei Gao

The recent advancements in Text-to-Video Artificial Intelligence Generated Content (AIGC) have been remarkable.

Common Sense Reasoning

Towards the Desirable Decision Boundary by Moderate-Margin Adversarial Training

no code implementations16 Jul 2022 Xiaoyu Liang, Yaguan Qian, Jianchang Huang, Xiang Ling, Bin Wang, Chunming Wu, Wassim Swaileh

Adversarial training, as one of the most effective defense methods against adversarial attacks, tends to learn an inclusive decision boundary to increase the robustness of deep learning models.

Cannot find the paper you are looking for? You can Submit a new open access paper.