no code implementations • 12 Mar 2025 • Yuhang Liu, Dong Gong, Erdun Gao, Zhen Zhang, Biwei Huang, Mingming Gong, Anton Van Den Hengel, Javen Qinfeng Shi
The remarkable achievements of large language models (LLMs) have led many to conclude that they exhibit a form of intelligence.
no code implementations • 13 Jan 2025 • Xinyu Zhang, Zicheng Duan, Dong Gong, Lingqiao Liu
We then design a motion consistency loss to maintain similar feature correlation patterns in the generated video, using the gradient of this loss in the latent space to guide the generation process for precise motion control.
no code implementations • 1 Dec 2024 • Haodong Lu, Chongyang Zhao, Jason Xue, Lina Yao, Kristen Moore, Dong Gong
We investigate whether the pre-trained knowledge of vision-language models (VLMs), such as CLIP, can be retained or even enhanced during continual learning (CL) while absorbing knowledge from a data stream.
no code implementations • 1 Dec 2024 • Xin Xie, Dong Gong
Text-to-image diffusion model alignment is critical for improving the alignment between the generated images and human preferences.
no code implementations • 1 Dec 2024 • Chongyang Zhao, Dong Gong
By formulating CL as a sequence prediction task, meta-continual learning (MCL) enables to meta-learn an efficient continual learner based on the recent advanced sequence models, e. g., Transformers.
1 code implementation • 16 Oct 2024 • Zixin Wang, Dong Gong, Sen Wang, Zi Huang, Yadan Luo
To address these questions, we propose Token Condensation as Adaptation (TCA), a training-free adaptation method for CLIP by pruning class-irrelevant visual tokens while merging class-ambiguous tokens.
no code implementations • 1 Oct 2024 • Saurav Jha, Shiqi Yang, Masato Ishii, Mengjie Zhao, Christian Simon, Muhammad Jehanzeb Mirza, Dong Gong, Lina Yao, Shusuke Takahashi, Yuki Mitsufuji
Personalized text-to-image diffusion models have grown popular for their ability to efficiently acquire a new concept from user-defined text descriptions and a few images.
1 code implementation • 30 Sep 2024 • Cheng Zhang, Dong Gong, Jiumei He, Yu Zhu, Jinqiu Sun, Yanning Zhang
Inspired by the success of deep generative models and fine-tuning techniques, we proposed a universal image restoration framework based on multiple low-rank adapters (LoRA) from multi-domain transfer learning.
2 code implementations • 20 May 2024 • Zidu Yin, Zhen Zhang, Dong Gong, Stefano V. Albrecht, Javen Q. Shi
Building on this observation, we introduce the highway graph to model state transitions.
1 code implementation • 28 Mar 2024 • Saurav Jha, Dong Gong, Lina Yao
Cooperating with the diverse range of existing prompting methods, CLAP can surpass the predominant deterministic finetuning approaches for CL with CLIP.
no code implementations • 27 Mar 2024 • Huiyi Wang, Haodong Lu, Lina Yao, Dong Gong
Continual learning (CL) aims to continually accumulate knowledge from a non-stationary data stream without catastrophic forgetting of learned knowledge, requiring a balance between stability and adaptability.
no code implementations • 23 Mar 2024 • Yuhang Liu, Zhen Zhang, Dong Gong, Mingming Gong, Biwei Huang, Anton Van Den Hengel, Kun Zhang, Javen Qinfeng Shi
This work establishes a {sufficient} and {necessary} condition characterizing the types of distribution shifts for identifiability in the context of latent additive noise models.
1 code implementation • 12 Mar 2024 • De Cheng, Yanling Ji, Dong Gong, Yan Li, Nannan Wang, Junwei Han, Dingwen Zhang
It considers the characteristics of the image restoration task with multiple degenerations in continual learning, and the knowledge for different degenerations can be shared and accumulated in the unified network structure.
1 code implementation • 12 Mar 2024 • Mark D. McDonnell, Dong Gong, Ehsan Abbasnejad, Anton Van Den Hengel
We show here that the combination of a large language model and an image generation model can similarly provide useful premonitions as to how a continual learning challenge might develop over time.
no code implementations • 19 Feb 2024 • Jialei Xu, Wei Yin, Dong Gong, Junjun Jiang, Xianming Liu
We suggest building virtual pinhole cameras to resolve the distortion problem of fisheye cameras and unify the processing for the two types of 360$^\circ$ cameras.
no code implementations • 9 Feb 2024 • Yuhang Liu, Zhen Zhang, Dong Gong, Biwei Huang, Mingming Gong, Anton Van Den Hengel, Kun Zhang, Javen Qinfeng Shi
Multimodal contrastive representation learning methods have proven successful across a range of domains, partly due to their ability to generate meaningful shared representations of complex phenomena.
1 code implementation • 5 Feb 2024 • Haodong Lu, Dong Gong, Shuo Wang, Jason Xue, Lina Yao, Kristen Moore
To tackle these issues, we propose PrototypicAl Learning with a Mixture of prototypes (PALM) which models each class with multiple prototypes to capture the sample diversities, and learns more faithful and compact samples embeddings to enhance OOD detection.
Out-of-Distribution Detection
Out of Distribution (OOD) Detection
+1
no code implementations • 24 Oct 2023 • Yuhang Liu, Zhen Zhang, Dong Gong, Mingming Gong, Biwei Huang, Anton Van Den Hengel, Kun Zhang, Javen Qinfeng Shi
However, this progress rests on the assumption that the causal relationships among latent causal variables adhere strictly to linear Gaussian models.
no code implementations • 17 Oct 2023 • Xueyang Kang, Fengze Han, Abdur R. Fayjie, Patrick Vandewalle, Kourosh Khoshelham, Dong Gong
Most existing methods for depth estimation from a focal stack of images employ convolutional neural networks (CNNs) using 2D or 3D convolutions over a fixed set of images.
no code implementations • 30 Sep 2023 • Hailan Ma, Zhenhong Sun, Daoyi Dong, Dong Gong
QST aims to recover the density matrix or the properties of the quantum state from the measured frequencies.
2 code implementations • NeurIPS 2023 • Mark D. McDonnell, Dong Gong, Amin Parveneh, Ehsan Abbasnejad, Anton Van Den Hengel
In this paper, we propose a concise and effective approach for CL with pre-trained models.
1 code implementation • CVPR 2023 • Rui Li, Dong Gong, Wei Yin, Hao Chen, Yu Zhu, Kaixuan Wang, Xiaozhi Chen, Jinqiu Sun, Yanning Zhang
To let the geometric perception learned from multi-view cues in static areas propagate to the monocular representation in dynamic areas and let monocular cues enhance the representation of multi-view cost volume, we propose a cross-cue fusion (CCF) module, which includes the cross-cue attention (CCA) to encode the spatially non-local relative intra-relations from each source to enhance the representation of the other.
1 code implementation • 5 Mar 2023 • Junyan Wang, Zhenhong Sun, Yichen Qian, Dong Gong, Xiuyu Sun, Ming Lin, Maurice Pagnucco, Yang song
In this work, we propose to automatically design efficient 3D CNN architectures via a novel training-free neural architecture search approach tailored for 3D CNNs considering the model complexity.
Ranked #84 on
Action Recognition
on Something-Something V2
1 code implementation • 1 Sep 2022 • Saurav Jha, Dong Gong, Xuesong Wang, Richard E. Turner, Lina Yao
We shed light on their potential to bring several recent advances in other deep learning domains under one umbrella.
no code implementations • 30 Aug 2022 • Yuhang Liu, Zhen Zhang, Dong Gong, Mingming Gong, Biwei Huang, Anton Van Den Hengel, Kun Zhang, Javen Qinfeng Shi
Within this new paradigm, we present an intricate causal generative model by introducing latent noises across domains, along with a latent content variable and a latent style variable to achieve more nuanced rendering of observational data.
1 code implementation • 30 Aug 2022 • Zhen Zhang, Ignavier Ng, Dong Gong, Yuhang Liu, Ehsan M Abbasnejad, Mingming Gong, Kun Zhang, Javen Qinfeng Shi
Recovering underlying Directed Acyclic Graph (DAG) structures from observational data is highly challenging due to the combinatorial nature of the DAG-constrained optimization problem.
no code implementations • 30 Aug 2022 • Yuhang Liu, Zhen Zhang, Dong Gong, Mingming Gong, Biwei Huang, Anton Van Den Hengel, Kun Zhang, Javen Qinfeng Shi
The task of causal representation learning aims to uncover latent higher-level causal representations that affect lower-level observations.
no code implementations • 25 May 2022 • Eduardo Pérez-Pellitero, Sibi Catley-Chandar, Richard Shaw, Aleš Leonardis, Radu Timofte, Zexin Zhang, Cen Liu, Yunbo Peng, Yue Lin, Gaocheng Yu, Jin Zhang, Zhe Ma, Hongbin Wang, Xiangyu Chen, Xintao Wang, Haiwei Wu, Lin Liu, Chao Dong, Jiantao Zhou, Qingsen Yan, Song Zhang, Weiye Chen, Yuhang Liu, Zhen Zhang, Yanning Zhang, Javen Qinfeng Shi, Dong Gong, Dan Zhu, Mengdi Sun, Guannan Chen, Yang Hu, Haowei Li, Baozhu Zou, Zhen Liu, Wenjie Lin, Ting Jiang, Chengzhi Jiang, Xinpeng Li, Mingyan Han, Haoqiang Fan, Jian Sun, Shuaicheng Liu, Juan Marín-Vega, Michael Sloth, Peter Schneider-Kamp, Richard Röttger, Chunyang Li, Long Bao, Gang He, Ziyao Xu, Li Xu, Gen Zhan, Ming Sun, Xing Wen, Junlin Li, Shuang Feng, Fei Lei, Rui Liu, Junxiang Ruan, Tianhong Dai, Wei Li, Zhan Lu, Hengyan Liu, Peian Huang, Guangyu Ren, Yonglin Luo, Chang Liu, Qiang Tu, Fangya Li, Ruipeng Gang, Chenghua Li, Jinjing Li, Sai Ma, Chenming Liu, Yizhen Cao, Steven Tel, Barthelemy Heyrman, Dominique Ginhac, Chul Lee, Gahyeon Kim, Seonghyun Park, An Gia Vien, Truong Thanh Nhat Mai, Howoon Yoon, Tu Vo, Alexander Holston, Sheir Zaheer, Chan Y. Park
The challenge is composed of two tracks with an emphasis on fidelity and complexity constraints: In Track 1, participants are asked to optimize objective fidelity scores while imposing a low-complexity constraint (i. e. solutions can not exceed a given number of operations).
no code implementations • CVPR 2022 • Dong Gong, Qingsen Yan, Yuhang Liu, Anton Van Den Hengel, Javen Qinfeng Shi
This minimizes the interference between parameters for different tasks.
Ranked #5 on
Continual Learning
on Tiny-ImageNet (10tasks)
no code implementations • ICCV 2021 • Dong Gong, Frederic Z. Zhang, Javen Qinfeng Shi, Anton Van Den Hengel
This motivates us to propose a memory-augmented dynamic neural relational inference method, which maintains two associative memory pools: one for the interactive relations and the other for the individual entities.
no code implementations • 23 Apr 2020 • Qingsen Yan, Bo wang, Dong Gong, Chuan Luo, Wei Zhao, Jianhu Shen, Qinfeng Shi, Shuo Jin, Liang Zhang, Zheng You
Inspired by the observation that the boundary of the infected lung can be enhanced by adjusting the global intensity, in the proposed deep CNN, we introduce a feature variation block which adaptively adjusts the global properties of the features for segmenting COVID-19 infection.
no code implementations • 13 Jan 2020 • Xin-Yu Zhang, Dong Gong, Jiewei Cao, Chunhua Shen
Due to the lack of supervision in the target domain, it is crucial to identify the underlying similarity-and-dissimilarity relationships among the unlabelled samples in the target domain.
no code implementations • 9 Jan 2020 • Hai-Ming Xu, Lingqiao Liu, Dong Gong
Our insight is that the prediction target in SemSL can be modeled as the latent factor in the predictor for the SlfSL target.
no code implementations • 8 Jan 2020 • Dong Gong, Wei Sun, Qinfeng Shi, Anton Van Den Hengel, Yanning Zhang
Most learning-based super-resolution (SR) methods aim to recover high-resolution (HR) image from a given low-resolution (LR) image via learning on LR-HR image pairs.
no code implementations • ECCV 2020 • Tong He, Dong Gong, Zhi Tian, Chunhua Shen
3D point cloud semantic and instance segmentation is crucial and fundamental for 3D scene understanding.
Ranked #29 on
3D Instance Segmentation
on ScanNet(v2)
1 code implementation • 13 Sep 2019 • Xin-Yu Zhang, Rufeng Zhang, Jiewei Cao, Dong Gong, Mingyu You, Chunhua Shen
Finally, we aggregate the global appearance and part features to improve the feature performance further.
1 code implementation • 29 Jul 2019 • Tong Shen, Dong Gong, Wei zhang, Chunhua Shen, Tao Mei
To tackle the unsupervised domain adaptation problem, we explore the possibilities to generate high-quality labels as proxy labels to supervise the training on target data.
no code implementations • 14 May 2019 • Yinglong Wang, Dong Gong, Jie Yang, Qinfeng Shi, Anton Van Den Hengel, Dehua Xie, Bing Zeng
Removing rain effects from an image is of importance for various applications such as autonomous driving, drone piloting, and photo editing.
5 code implementations • CVPR 2019 • Qingsen Yan, Dong Gong, Qinfeng Shi, Anton Van Den Hengel, Chunhua Shen, Ian Reid, Yanning Zhang
Ghosting artifacts caused by moving objects or misalignments is a key challenge in high dynamic range (HDR) imaging for dynamic scenes.
5 code implementations • ICCV 2019 • Dong Gong, Lingqiao Liu, Vuong Le, Budhaditya Saha, Moussa Reda Mansour, Svetha Venkatesh, Anton Van Den Hengel
At the test stage, the learned memory will be fixed, and the reconstruction is obtained from a few selected memory records of the normal data.
1 code implementation • CVPR 2019 • Tong He, Chunhua Shen, Zhi Tian, Dong Gong, Changming Sun, Youliang Yan
To tackle this dilemma, we propose a knowledge distillation method tailored for semantic segmentation to improve the performance of the compact FCNs with large overall stride.
no code implementations • CVPR 2019 • Jie Li, Yu Liu, Dong Gong, Qinfeng Shi, Xia Yuan, Chunxia Zhao, Ian Reid
RGB images differentiate from depth images as they carry more details about the color and texture information, which can be utilized as a vital complementary to depth for boosting the performance of 3D semantic scene completion (SSC).
Ranked #20 on
3D Semantic Scene Completion
on NYUv2
no code implementations • CVPR 2019 • Yuhang Liu, Wenyong Dong, Lei Zhang, Dong Gong, Qinfeng Shi
Then, we incorporate such a prior into inferring the joint posterior over network weights and the variance in the hierarchical prior, with which both the network training and the dropout rate estimation can be cast into a joint optimization problem.
no code implementations • 23 Oct 2018 • Gerard Snaauw, Dong Gong, Gabriel Maicas, Anton Van Den Hengel, Wiro J. Niessen, Johan Verjans, Gustavo Carneiro
In this paper, we propose a learning method to train diagnosis models, where our approach is designed to work with relatively small datasets.
no code implementations • 12 Oct 2018 • Dong Gong, Mingkui Tan, Qinfeng Shi, Anton Van Den Hengel, Yanning Zhang
Compared to existing methods, MPTV is less sensitive to the choice of the trade-off parameter between data fitting and regularization.
no code implementations • ECCV 2018 • Yuhang Liu, Wenyong Dong, Dong Gong, Lei Zhang, Qinfeng Shi
Existing sparsity-based priors are usually rooted in modeling the response of images to some specific filters (e. g., image gradients), which are insufficient to capture the complicated image structures.
1 code implementation • ECCV 2018 • Jie Yang, Dong Gong, Lingqiao Liu, Qinfeng Shi
Reflections often obstruct the desired scene when taking photos through glass panels.
1 code implementation • 10 Apr 2018 • Dong Gong, Zhen Zhang, Qinfeng Shi, Anton Van Den Hengel, Chunhua Shen, Yanning Zhang
Extensive experiments on synthetic benchmarks and challenging real-world images demonstrate that the proposed deep optimization method is effective and robust to produce favorable results as well as practical for real-world image deblurring applications.
no code implementations • ICCV 2017 • Dong Gong, Mingkui Tan, Yanning Zhang, Anton Van Den Hengel, Qinfeng Shi
Rather than attempt to identify outliers to the model a priori, we instead propose to sequentially identify inliers, and gradually incorporate them into the estimation process.
no code implementations • CVPR 2017 • Dong Gong, Jie Yang, Lingqiao Liu, Yanning Zhang, Ian Reid, Chunhua Shen, Anton Van Den Hengel, Qinfeng Shi
The critical observation underpinning our approach is thus that learning the motion flow instead allows the model to focus on the cause of the blur, irrespective of the image content.
no code implementations • CVPR 2016 • Dong Gong, Mingkui Tan, Yanning Zhang, Anton Van Den Hengel, Qinfeng Shi
We show here that a subset of the image gradients are adequate to estimate the blur kernel robustly, no matter the gradient image is sparse or not.