1 code implementation • 17 Apr 2025 • Xin Li, Kun Yuan, Bingchen Li, Fengbin Guan, Yizhen Shao, Zihao Yu, Xijun Wang, Yiting Lu, Wei Luo, Suhang Yao, Ming Sun, Chao Zhou, Zhibo Chen, Radu Timofte, Yabin Zhang, Ao-Xiang Zhang, Tianwu Zhi, Jianzhao Liu, Yang Li, Jingwen Xu, Yiting Liao, Yushen Zuo, Mingyang Wu, Renjie Li, Shengyun Zhong, Zhengzhong Tu, Yufan Liu, Xiangguang Chen, Zuowei Cao, Minhao Tang, Shan Liu, Kexin Zhang, Jingfen Xie, Yan Wang, Kai Chen, Shijie Zhao, Yunchen Zhang, Xiangkai Xu, Hong Gao, Ji Shi, Yiming Bao, Xiugang Dong, Xiangsheng Zhou, Yaofeng Tu, Ying Liang, Yiwen Wang, Xinning Chai, Yuxuan Zhang, Zhengxue Cheng, Yingsheng Qin, Yucai Yang, Rong Xie, Li Song, Wei Sun, Kang Fu, Linhan Cao, Dandan Zhu, Kaiwei Zhang, Yucheng Zhu, ZiCheng Zhang, Menghan Hu, Xiongkuo Min, Guangtao Zhai, Zhi Jin, Jiawei Wu, Wei Wang, Wenjian Zhang, Yuhai Lan, Gaoxiong Yi, Hengyuan Na, Wang Luo, Di wu, MingYin Bai, Jiawang Du, Zilong Lu, Zhenyu Jiang, Hui Zeng, Ziguan Cui, Zongliang Gan, Guijin Tang, Xinglin Xie, Kehuan Song, Xiaoqiang Lu, Licheng Jiao, Fang Liu, Xu Liu, Puhua Chen, Ha Thu Nguyen, Katrien De Moor, Seyed Ali Amirshahi, Mohamed-Chaker Larabi, Qi Tang, Linfeng He, Zhiyong Gao, Zixuan Gao, Guohua Zhang, Zhiye Huang, Yi Deng, Qingmiao Jiang, Lu Chen, Yi Yang, Xi Liao, Nourine Mohammed Nadir, YuXuan Jiang, Qiang Zhu, Siyue Teng, Fan Zhang, Shuyuan Zhu, Bing Zeng, David Bull, Meiqin Liu, Chao Yao, Yao Zhao
This paper presents a review for the NTIRE 2025 Challenge on Short-form UGC Video Quality Assessment and Enhancement.
no code implementations • 15 Jun 2024 • Ying Fu, Yu Li, ShaoDi You, Boxin Shi, Linwei Chen, Yunhao Zou, Zichun Wang, Yichen Li, Yuze Han, Yingkai Zhang, Jianan Wang, Qinglin Liu, Wei Yu, Xiaoqian Lv, Jianing Li, Shengping Zhang, Xiangyang Ji, Yuanpei Chen, Yuhan Zhang, Weihang Peng, Liwen Zhang, Zhe Xu, Dingyong Gou, Cong Li, Senyan Xu, Yunkang Zhang, Siyuan Jiang, Xiaoqiang Lu, Licheng Jiao, Fang Liu, Xu Liu, Lingling Li, Wenping Ma, Shuyuan Yang, Haiyang Xie, Jian Zhao, Shihua Huang, Peng Cheng, Xi Shen, Zheng Wang, Shuai An, Caizhi Zhu, Xuelong Li, Tao Zhang, Liang Li, Yu Liu, Chenggang Yan, Gengchen Zhang, Linyan Jiang, Bingyi Song, Zhuoyu An, Haibo Lei, Qing Luo, Jie Song, YuAn Liu, Haoyuan Zhang, Lingfeng Wang, Wei Chen, Aling Luo, Cheng Li, Jun Cao, Shu Chen, Zifei Dou, Xinyu Liu, Jing Zhang, Kexin Zhang, Yuting Yang, Xuejian Gou, Qinliang Wang, Yang Liu, Shizhan Zhao, Yanzhao Zhang, Libo Yan, Yuwei Guo, Guoxin Li, Qiong Gao, Chenyue Che, Long Sun, Xiang Chen, Hao Li, Jinshan Pan, Chuanlong Xie, Hongming Chen, Mingrui Li, Tianchen Deng, Jingwei Huang, Yufeng Li, Fei Wan, Bingxin Xu, Jian Cheng, Hongzhe Liu, Cheng Xu, Yuxiang Zou, Weiguo Pan, Songyin Dai, Sen Jia, Junpei Zhang, Puhua Chen, Qihang Li
The intersection of physics-based vision and deep learning presents an exciting frontier for advancing computer vision technologies.
no code implementations • 10 May 2021 • Bin Zhao, Haopeng Li, Xiaoqiang Lu, Xuelong Li
Then, the videos are summarized by exploiting both the local and global dependencies among shots.
no code implementations • 9 Mar 2021 • Yuan Yuan, Hailong Ning, Xiaoqiang Lu
In this paper, a novel VAP method is proposed to generate visual attention map via bio-inspired representation learning.
no code implementations • IEEE Transactions on Image Processing 2021 • Hao Sun, Xiangtao Zheng, Xiaoqiang Lu
To explore the spatial information for HSI classification, pixels with its adjacent pixels are usually directly cropped from hyperspectral data to form HSI cubes in CNN-based methods.
no code implementations • 29 May 2019 • Xuelong. Li, Aihong Yuan, Xiaoqiang Lu
To make full use of these information, this paper attempt to exploit the text guided attention and semantic-guided attention (SA) to find the more correlated spatial information and reduce the semantic gap between vision and language.
no code implementations • 28 Apr 2019 • Bin Zhao, Xuelong. Li, Xiaoqiang Lu
Compared to traditional RNNs, H-RNN is more suitable to video summarization, since it can exploit long temporal dependency among frames, meanwhile, the computation operations are significantly lessened.
no code implementations • 24 Apr 2019 • Bin Zhao, Xuelong. Li, Xiaoqiang Lu, Zhigang Wang
To address this problem, we make the first attempt to view weather recognition as a multi-label classification task, i. e., assigning an image more than one labels according to the displayed weather conditions.
no code implementations • 24 Apr 2019 • Xuelong. Li, Bin Zhao, Xiaoqiang Lu
Besides, the property-weights are learned for edited videos and raw videos, respectively.
no code implementations • 21 Apr 2019 • Aihong Yuan, Xuelong. Li, Xiaoqiang Lu
In this paper, we propose a model with 3-gated model which fuses the global and local image features together for the task of image caption generation.
no code implementations • 20 Apr 2019 • Xuelong. Li, Aihong Yuan, Xiaoqiang Lu
And in the testing step, when an image is imported to our multi-modal GRU model, a sentence which describes the image content is generated.
no code implementations • CVPR 2018 • Bin Zhao, Xuelong. Li, Xiaoqiang Lu
Although video summarization has achieved great success in recent years, few approaches have realized the influence of video structure on the summarization results.
2 code implementations • 21 Dec 2017 • Xiaoqiang Lu, Binqiang Wang, Xiangtao Zheng, Xuelong. Li
Finally, a comprehensive review is presented on the proposed data set to fully advance the task of remote sensing caption.
no code implementations • ICCV 2017 • Xuelong. Li, Di Hu, Xiaoqiang Lu
Image is usually taken for expressing some kinds of emotions or purposes, such as love, celebrating Christmas.
4 code implementations • 1 Mar 2017 • Gong Cheng, Junwei Han, Xiaoqiang Lu
During the past years, significant efforts have been made to develop various datasets or present a variety of approaches for scene classification from remote sensing images.
no code implementations • CVPR 2016 • Di Hu, Xuelong. Li, Xiaoqiang Lu
Recently, audiovisual speech recognition based the MRBM has attracted much attention, and the MRBM shows its effectiveness in learning the joint representation across audiovisual modalities.