no code implementations • 29 Aug 2024 • Mian Zou, Baosheng Yu, Yibing Zhan, Siwei Lyu, Kede Ma
In this paper, we delve deeper into semantics-oriented multitask learning for DeepFake detection, leveraging the relationships among face semantics via joint embedding.
no code implementations • 18 Jul 2024 • Peibei Cao, Haoyu Chen, Jingzhe Ma, Yu-Chieh Yuan, Zhiyong Xie, Xin Xie, Haiqing Bai, Kede Ma
High dynamic range (HDR) capture and display have seen significant growth in popularity driven by the advancements in technology and increasing consumer demand for superior image quality.
1 code implementation • 14 Jul 2024 • Jiaqi He, Zhihua Wang, Leon Wang, Tsein-I Liu, Yuming Fang, Qilin Sun, Kede Ma
Contemporary color difference (CD) measures for photographic images typically operate by comparing co-located pixels, patches in a ``perceptually uniform'' color space, or features in a learned latent space.
1 code implementation • 13 Jul 2024 • Wei Shang, Dongwei Ren, Wanying Zhang, Yuming Fang, WangMeng Zuo, Kede Ma
Arbitrary-scale video super-resolution (AVSR) aims to enhance the resolution of video frames, potentially at various scaling factors, which presents several challenges regarding spatial detail reproduction, temporal consistency, and computational complexity.
no code implementations • 14 May 2024 • Mian Zou, Baosheng Yu, Yibing Zhan, Siwei Lyu, Kede Ma
In recent years, deep learning has greatly streamlined the process of generating realistic fake face images.
no code implementations • 18 Apr 2024 • Ronghuan Wu, Wanchao Su, Kede Ma, Jing Liao
To generate cartoon-style and smooth motion, we first define B\'{e}zier curves over keypoints of the clipart image as a form of motion regularization.
1 code implementation • 10 Apr 2024 • Kehua Feng, Keyan Ding, Kede Ma, Zhihua Wang, Qiang Zhang, Huajun Chen
The past years have witnessed a proliferation of large language models (LLMs).
1 code implementation • CVPR 2024 • Kanglong Fan, Wen Wen, Mu Li, Yifan Peng, Kede Ma
Panoramic videos have the advantage of providing an immersive and interactive viewing experience.
1 code implementation • 16 Mar 2024 • Tianhe Wu, Kede Ma, Jie Liang, Yujiu Yang, Lei Zhang
While Multimodal Large Language Models (MLLMs) have experienced significant advancement in visual understanding and reasoning, their potential to serve as powerful, flexible, interpretable, and text-driven models for Image Quality Assessment (IQA) remains largely unexplored.
no code implementations • 11 Mar 2024 • Weixia Zhang, Dingquan Li, Guangtao Zhai, Xiaokang Yang, Kede Ma
Contemporary no-reference image quality assessment (NR-IQA) models can effectively quantify the perceived image quality, with high correlations between model predictions and human perceptual scores on fixed test sets.
1 code implementation • CVPR 2024 • Wen Wen, Mu Li, Yabin Zhang, Yiting Liao, Junlin Li, Li Zhang, Kede Ma
Blind video quality assessment (BVQA) plays a pivotal role in evaluating and improving the viewing experience of end-users across a wide range of video-based platforms and services.
1 code implementation • 17 Feb 2024 • Dingquan Li, Kede Ma, Jing Wang, Ge Li
The content-dependent hierarchical prior is constructed at the encoder side, which enables coarse-to-fine super resolution of the point cloud geometry at the decoder side.
1 code implementation • CVPR 2024 • Peibei Cao, Rafal K. Mantiuk, Kede Ma
Existing quality models are mostly designed for low dynamic range (LDR) images, and do not align well with human perception of HDR image quality.
1 code implementation • 26 Jul 2023 • Wei Sun, Wen Wen, Xiongkuo Min, Long Lan, Guangtao Zhai, Kede Ma
By minimalistic, we restrict our family of BVQA models to build only upon basic blocks: a video preprocessor (for aggressive spatiotemporal downsampling), a spatial quality analyzer, an optional temporal quality analyzer, and a quality regressor, all with the simplest possible instantiations.
no code implementations • 4 May 2023 • Mu Li, Kanglong Fan, Kede Ma
Predicting human scanpaths when exploring panoramic videos is a challenging task due to the spherical geometry and the multimodality of the input, and the inherent uncertainty and diversity of the output.
no code implementations • 27 Apr 2023 • Ronghuan Wu, Wanchao Su, Kede Ma, Jing Liao
More importantly, we demonstrate the flexibility of IconShop with multiple novel icon synthesis tasks, including icon editing, icon interpolation, icon semantic combination, and icon design auto-suggestion.
2 code implementations • CVPR 2023 • Wei Shang, Dongwei Ren, Yi Yang, Hongzhi Zhang, Kede Ma, WangMeng Zuo
Moreover, on the seemingly implausible x16 interpolation task, our method outperforms existing methods by more than 1. 5 dB in terms of PSNR.
1 code implementation • CVPR 2023 • Haoyu Chen, Zhihua Wang, Yang Yang, Qilin Sun, Kede Ma
Most well-established and widely used color difference (CD) metrics are handcrafted and subject-calibrated against uniformly colored patches, which do not generalize well to photographic images characterized by natural scene complexities.
1 code implementation • CVPR 2023 • Weixia Zhang, Guangtao Zhai, Ying WEI, Xiaokang Yang, Kede Ma
We aim at advancing blind image quality assessment (BIQA), which predicts the human perception of image quality without any reference information.
no code implementations • 5 Oct 2022 • Haoyu Chen, Linqi Song, Zhenxing Qian, Xinpeng Zhang, Kede Ma
As an instantiation, we adopt a SinGAN, a pyramid of generative adversarial networks (GANs), to learn the patch distribution of one cover image.
1 code implementation • 3 Oct 2022 • Weixia Zhang, Dingquan Li, Xiongkuo Min, Guangtao Zhai, Guodong Guo, Xiaokang Yang, Kede Ma
No-reference image quality assessment (NR-IQA) aims to quantify how humans perceive visual distortions of digital images without access to their undistorted references.
no code implementations • 29 Jul 2022 • Peibei Cao, Dingquan Li, Kede Ma
Learning-based image quality assessment (IQA) has made remarkable progress in the past decade, but nearly all consider the two key components -- model and data -- in isolation.
no code implementations • 18 Jun 2022 • Peibei Cao, Chenyang Le, Yuming Fang, Kede Ma
In Stage two, the input HDR image is self-calibrated to compute the final LDR image.
1 code implementation • 13 Jun 2022 • Wen Wen, Mu Li, Yiru Yao, Xiangjie Sui, Yabin Zhang, Long Lan, Yuming Fang, Kede Ma
Investigating how people perceive virtual reality (VR) videos in the wild (i. e., those captured by everyday users) is a crucial and challenging task in VR-related applications due to complex authentic distortions localized in space and time.
1 code implementation • 26 May 2022 • Zhihua Wang, Keshuo Xu, Yang Yang, Jianlei Dong, Shuhang Gu, Lihao Xu, Yuming Fang, Kede Ma
Measuring perceptual color differences (CDs) is of great importance in modern smartphone photography.
2 code implementations • 20 Jan 2022 • Xiangyang Zhu, Kede Ma, Wufeng Xue
Predicting cardiac indices has long been a focal point in the medical imaging community.
1 code implementation • 25 Dec 2021 • Mu Li, Kede Ma, Jinxing Li, David Zhang
We first describe parametric pseudocylindrical representation as a generalization of common pseudocylindrical map projections.
no code implementations • 19 Oct 2021 • Kede Ma, Yuming Fang
This tutorial provides the audience with the basic theories, methodologies, and current progresses of image quality assessment (IQA).
no code implementations • 16 Oct 2021 • Keyan Ding, Yi Liu, Xueyi Zou, Shiqi Wang, Kede Ma
The latest advances in full-reference image quality assessment (IQA) involve unifying structure and texture similarity based on deep representations.
Ranked #34 on Video Quality Assessment on MSU SR-QA Dataset
no code implementations • 1 Sep 2021 • Chenyang Le, Jiebin Yan, Yuming Fang, Kede Ma
We describe a deep high-dynamic-range (HDR) image tone mapping operator that is computationally efficient and perceptually optimized.
2 code implementations • 28 Jul 2021 • Weixia Zhang, Kede Ma, Guangtao Zhai, Xiaokang Yang
In this paper, we present a simple yet effective continual learning method for blind image quality assessment (BIQA) with improved quality prediction accuracy, plasticity-stability trade-off, and task-order/-length robustness.
1 code implementation • 26 Jun 2021 • Zhihua Wang, Dingquan Li, Kede Ma
Ensemble methods are generally regarded to be better than a single model if the base learners are deemed to be "accurate" and "diverse."
no code implementations • CVPR 2021 • Zhihua Wang, Haotao Wang, Tianlong Chen, Zhangyang Wang, Kede Ma
Recently, the group maximum differentiation competition (gMAD) has been used to improve blind image quality assessment (BIQA) models, with the help of full-reference metrics.
1 code implementation • 27 Feb 2021 • Jiebin Yan, Yu Zhong, Yuming Fang, Zhangyang Wang, Kede Ma
A natural question then arises: Does the superior performance on the closed (and frequently re-used) test sets transfer to the open visual world with unconstrained variations?
1 code implementation • 19 Feb 2021 • Weixia Zhang, Dingquan Li, Chao Ma, Guangtao Zhai, Xiaokang Yang, Kede Ma
In this paper, we formulate continual learning for BIQA, where a model learns continually from a stream of IQA datasets, building on what was learned from previously seen data.
no code implementations • 1 Jan 2021 • Haotao Wang, Tianlong Chen, Zhangyang Wang, Kede Ma
Image segmentation lays the foundation for many high-stakes vision applications such as autonomous driving and medical image analysis.
1 code implementation • 28 May 2020 • Weixia Zhang, Kede Ma, Guangtao Zhai, Xiaokang Yang
Nevertheless, due to the distributional shift between images simulated in the laboratory and captured in the wild, models trained on databases with synthetic distortions remain particularly weak at handling realistic distortions (and vice versa).
2 code implementations • 21 May 2020 • Xiangjie Sui, Kede Ma, Yiru Yao, Yuming Fang
We first carry out a psychophysical experiment to investigate the interplay among the VR viewing conditions, the user viewing behaviors, and the perceived quality of 360{\deg} images.
1 code implementation • 4 May 2020 • Keyan Ding, Kede Ma, Shiqi Wang, Eero P. Simoncelli
The performance of objective image quality assessment (IQA) models has been evaluated primarily by comparing model predictions to human quality judgments.
2 code implementations • 16 Apr 2020 • Keyan Ding, Kede Ma, Shiqi Wang, Eero P. Simoncelli
Objective measures of image quality generally operate by comparing pixels of a "degraded" image to those of the original.
Ranked #31 on Video Quality Assessment on MSU SR-QA Dataset
no code implementations • 8 Mar 2020 • Zhihua Wang, Kede Ma
We then seek pairs of images by comparing the baseline model with a set of full-reference IQA methods in gMAD.
1 code implementation • ICLR 2020 • Haotao Wang, Tianlong Chen, Zhangyang Wang, Kede Ma
On the other hand, the trained classifiers have traditionally been evaluated on small and fixed sets of test images, which are deemed to be extremely sparsely distributed in the space of all natural images.
1 code implementation • 5 Jul 2019 • Weixia Zhang, Kede Ma, Jia Yan, Dexiang Deng, Zhou Wang
We propose a deep bilinear model for blind image quality assessment (BIQA) that handles both synthetic and authentic distortions.
Ranked #2 on Video Quality Assessment on MSU NR VQA Database
1 code implementation • 3 Jul 2019 • Keyan Ding, Kede Ma, Shiqi Wang
The goal of research in automatic image popularity assessment (IPA) is to develop computational models that can accurately predict the potential of a social image to go viral on the Internet.
1 code implementation • 1 Jul 2019 • Weixia Zhang, Kede Ma, Guangtao Zhai, Xiaokang Yang
Computational models for blind image quality assessment (BIQA) are typically trained in well-controlled laboratory environments with limited generalizability to realistically distorted images.
2 code implementations • 24 Jun 2019 • Mu Li, Kede Ma, Jane You, David Zhang, WangMeng Zuo
For the former, we directly apply a CCN to the binarized representation of an image to compute the Bernoulli distribution of each code for entropy estimation.
no code implementations • 13 Apr 2019 • Kede Ma, Wentao Liu, Tongliang Liu, Zhou Wang, DaCheng Tao
One of the biggest challenges in learning BIQA models is the conflict between the gigantic image space (which is in the dimension of the number of image pixels) and the extremely limited reliable ground truth data for training.
no code implementations • 5 Dec 2016 • Kede Ma, Huan Fu, Tongliang Liu, Zhou Wang, DaCheng Tao
The human visual system excels at detecting local blur of visual images, but the underlying mechanism is not well understood.
no code implementations • CVPR 2016 • Kede Ma, Qingbo Wu, Zhou Wang, Zhengfang Duanmu, Hongwei Yong, Hongliang Li, Lei Zhang
We first build a database that contains 4, 744 source natural images, together with 94, 880 distorted images created from them.