no code implementations • 17 Oct 2024 • Xuezhi Xiang, Xi Wang, Lei Zhang, Denis Ombati, Himaloy Himu, XianTong Zhen
Scene flow estimation aims to generate the 3D motion field of points between two consecutive frames of point clouds, which has wide applications in various fields.
no code implementations • 14 Oct 2024 • Xuezhi Xiang, Yibo Ning, Lei Zhang, Denis Ombati, Himaloy Himu, XianTong Zhen
In this paper, we propose a remote-sensing image semantic segmentation network named LKASeg, which combines Large Kernel Attention(LSKA) and Full-Scale Skip Connections(FSC).
no code implementations • 26 Sep 2024 • Xuezhi Xiang, Yao Wang, Lei Zhang, Denis Ombati, Himaloy Himu, XianTong Zhen
Self-supervised monocular depth estimation has emerged as a promising approach since it does not rely on labeled training data.
no code implementations • 26 Sep 2024 • Xuezhi Xiang, Zhushan Ma, Lei Zhang, Denis Ombati, Himaloy Himu, XianTong Zhen
Using attention mechanism to capture global and local features is crucial to solve the challenge of high similarity between classes in vehicle Re-ID tasks.
no code implementations • 19 Aug 2024 • Zhi Qiao, Xuhui Liu, Xiaopeng Wang, Runkun Liu, XianTong Zhen, Pei Dong, Zhen Qian
Intraoperative CT imaging serves as a crucial resource for surgical guidance; however, it may not always be readily accessible or practical to implement.
no code implementations • 19 Aug 2024 • Zhi Qiao, Hanqiang Ouyang, Dongheng Chu, Huishu Yuan, XianTong Zhen, Pei Dong, Zhen Qian
For surgical planning and intra-operation imaging, CT reconstruction using X-ray images can potentially be an important alternative when CT imaging is not available or not feasible.
no code implementations • 19 Aug 2024 • Zhi Qiao, Linbin Han, XianTong Zhen, Jia-Hong Gao, Zhen Qian
In light of the inherent entailment relations between images and text, hyperbolic point vector embeddings, leveraging the hierarchical modeling advantages of hyperbolic space, have been utilized for visual semantic representation learning.
no code implementations • 18 Jul 2024 • Xuhui Liu, Zhi Qiao, Runkun Liu, Hong Li, Juan Zhang, XianTong Zhen, Zhen Qian, Baochang Zhang
Computed tomography (CT) is widely utilized in clinical settings because it delivers detailed 3D images of the human body.
1 code implementation • 24 Jun 2024 • Yuchen Yang, Yingdong Shi, Cheems Wang, XianTong Zhen, Yuxuan Shi, Jun Xu
Fine-tuning pretrained large models to downstream tasks is an important problem, which however suffers from huge memory overhead due to large-scale parameters.
1 code implementation • NeurIPS 2023 • Jiayi Shen, XianTong Zhen, QI, Wang, Marcel Worring
This paper focuses on the data-insufficiency problem in multi-task learning within an episodic training setup.
1 code implementation • ICCV 2023 • Mengmeng Jing, XianTong Zhen, Jingjing Li, Cees Snoek
To alleviate this problem, data augmentation coupled with consistency regularization are commonly adopted to make the model less sensitive to domain-specific attributes.
1 code implementation • ICCV 2023 • Guanghui Li, Mingqi Gao, Heng Liu, XianTong Zhen, Feng Zheng
Referring video object segmentation (RVOS), as a supervised learning task, relies on sufficient annotated data for a given scene.
Referring Video Object Segmentation Semantic Segmentation +1
no code implementations • 2 Sep 2023 • Tom van Sonsbeek, XianTong Zhen, Marcel Worring
We show the use of this embedding on two tasks namely disease classification of X-ray reports and image classification.
no code implementations • ICCV 2023 • Baoshuo Kan, Teng Wang, Wenpeng Lu, XianTong Zhen, Weili Guan, Feng Zheng
Pre-trained vision-language models, e. g., CLIP, working with manually designed prompts have demonstrated great capacity of transfer learning.
1 code implementation • 8 Jul 2023 • Liqi Xue, Tianyi Xu, Yongbao Song, Yan Liu, Lei Zhang, XianTong Zhen, Jun Xu
But the majority of media images on the internet remain in 8-bit standard dynamic range (SDR) format.
no code implementations • 8 Jul 2023 • Sameer Ambekar, Zehao Xiao, Jiayi Shen, XianTong Zhen, Cees G. M. Snoek
We formulate the generalization at test time as a variational inference problem, by modeling pseudo labels as distributions, to consider the uncertainty during generalization and alleviate the misleading signal of inaccurate pseudo labels.
no code implementations • 8 Jun 2023 • Yingjun Du, Jiayi Shen, XianTong Zhen, Cees G. M. Snoek
By learning to retain and recall the learning process of past training tasks, EMO nudges parameter updates in the right direction, even when the gradients provided by a limited number of examples are uninformative.
no code implementations • 17 May 2023 • Hao Zheng, Jinbao Wang, XianTong Zhen, Hong Chen, Jingkuan Song, Feng Zheng
Recently, Transformers have emerged as the go-to architecture for both vision and language modeling tasks, but their computational efficiency is limited by the length of the input sequence.
1 code implementation • 17 May 2023 • Wenfang Sun, Yingjun Du, XianTong Zhen, Fan Wang, Ling Wang, Cees G. M. Snoek
To account for the uncertainty caused by the limited training tasks, we propose a variational MetaModulation where the modulation parameters are treated as latent variables.
no code implementations • CVPR 2023 • Yingjun Du, Jiayi Shen, XianTong Zhen, Cees G. M. Snoek
Modern image classifiers perform well on populated classes, while degrading considerably on tail classes with only a few instances.
1 code implementation • CVPR 2023 • Sicheng Gao, Xuhui Liu, Bohan Zeng, Sheng Xu, Yanjing Li, Xiaoyan Luo, Jianzhuang Liu, XianTong Zhen, Baochang Zhang
IDM integrates an implicit neural representation and a denoising diffusion model in a unified end-to-end framework, where the implicit neural representation is adopted in the decoding process to learn continuous-resolution representation.
Ranked #1 on Image Super-Resolution on CelebA-HQ 128x128
no code implementations • 20 Mar 2023 • Yuxuan Shi, Lingxiao Yang, Wangpeng An, XianTong Zhen, Liuqing Wang
The channel attention mechanism is a useful technique widely employed in deep convolutional neural networks to boost the performance for image processing tasks, eg, image classification and image super-resolution.
1 code implementation • 28 Feb 2023 • Ivona Najdenkoska, XianTong Zhen, Marcel Worring
Existing methods are trying to communicate visual concepts as prompts to frozen language models, but rely on hand-engineered task induction to reduce the hypothesis space.
1 code implementation • 22 Feb 2023 • Zehao Xiao, XianTong Zhen, Shengcai Liao, Cees G. M. Snoek
In this paper, we propose energy-based sample adaptation at test time for domain generalization.
1 code implementation • 17 Feb 2023 • Zhi Zhang, Helen Yannakoudakis, XianTong Zhen, Ekaterina Shutova
The task of multimodal referring expression comprehension (REC), aiming at localizing an image region described by a natural language expression, has recently received increasing attention within the research comminity.
1 code implementation • 19 Oct 2022 • Mengmeng Jing, XianTong Zhen, Jingjing Li, Cees G. M. Snoek
Our model perturbation provides a new probabilistic way for domain adaptation which enables efficient adaptation to target domains while maximally preserving knowledge in source models.
no code implementations • 13 Oct 2022 • Tom van Sonsbeek, XianTong Zhen, Dwarikanath Mahapatra, Marcel Worring
This shows how two-stage learning of labels from coarse to fine-grained, in particular with object level annotations, is an effective method for more optimal annotation usage.
1 code implementation • 10 Oct 2022 • Jiayi Shen, Zehao Xiao, XianTong Zhen, Cees G. M. Snoek, Marcel Worring
To generalize to such test data, it is crucial for individual tasks to leverage knowledge from related tasks.
1 code implementation • 12 Apr 2022 • Mohammad Mahdi Derakhshani, Ivona Najdenkoska, Tom van Sonsbeek, XianTong Zhen, Dwarikanath Mahapatra, Marcel Worring, Cees G. M. Snoek
Task and class incremental learning of diseases address the issue of classifying new samples without re-training the models from scratch, while cross-domain incremental learning addresses the issue of dealing with datasets originating from different institutions while retaining the previously obtained knowledge.
1 code implementation • 5 Mar 2022 • Jin Liang, Yuchen Yang, Anran Zhang, Jun Xu, Hui Li, XianTong Zhen
For image exposure enhancement, the tasks of Single-Exposure Correction (SEC) and Multi-Exposure Fusion (MEF) are widely studied in the image processing community.
1 code implementation • ICLR 2022 • Zehao Xiao, XianTong Zhen, Ling Shao, Cees G. M. Snoek
We leverage a meta-learning paradigm to learn our model to acquire the ability of adaptation with single samples at training time so as to further adapt itself to each single test sample at test time.
Ranked #1 on Domain Adaptation on PACS
no code implementations • 26 Dec 2021 • Mohammad Mahdi Derakhshani, XianTong Zhen, Ling Shao, Cees G. M. Snoek
Kernel continual learning by \citet{derakhshani2021kernel} has recently emerged as a strong continual learner due to its non-parametric ability to tackle task interference and catastrophic forgetting.
1 code implementation • ICLR 2022 • Yingjun Du, XianTong Zhen, Ling Shao, Cees G. M. Snoek
To explore and exploit the importance of different semantic levels, we further propose to learn the weights associated with the prototype at each level in a data-driven way, which enables the model to adaptively choose the most generalizable features.
no code implementations • NeurIPS 2021 • Ze Wang, Zichen Miao, XianTong Zhen, Qiang Qiu
In contrast to sparse Gaussian processes, we define a set of dense inducing variables to be of a much larger size than the support set in each task, which collects prior knowledge from experienced tasks.
no code implementations • 10 Nov 2021 • Jiayi Shen, XianTong Zhen, Marcel Worring, Ling Shao
Our multi-task neural processes methodologically expand the scope of vanilla neural processes and provide a new way of exploring task relatedness in function spaces for multi-task learning.
1 code implementation • NeurIPS 2021 • Jiayi Shen, XianTong Zhen, Marcel Worring, Ling Shao
Multi-task learning aims to explore task relatedness to improve individual tasks, which is of particular significance in the challenging scenario that only limited data is available for each task.
no code implementations • ICCV 2021 • Hongjun Chen, Jinbao Wang, Hong Cai Chen, XianTong Zhen, Feng Zheng, Rongrong Ji, Ling Shao
Annotation burden has become one of the biggest barriers to semantic segmentation.
Weakly supervised Semantic Segmentation Weakly-Supervised Semantic Segmentation
no code implementations • 15 Jul 2021 • Ivona Najdenkoska, XianTong Zhen, Marcel Worring, Ling Shao
The topics are inferred in a conditional variational inference framework, with each topic governing the generation of a sentence in the report.
1 code implementation • 12 Jul 2021 • Mohammad Mahdi Derakhshani, XianTong Zhen, Ling Shao, Cees G. M. Snoek
We further introduce variational random features to learn a data-driven kernel for each task.
no code implementations • ACL 2021 • Yingjun Du, Nithin Holla, XianTong Zhen, Cees G. M. Snoek, Ekaterina Shutova
A critical challenge faced by supervised word sense disambiguation (WSD) is the lack of large annotated datasets with sufficient coverage of words in their diversity of senses.
1 code implementation • 14 May 2021 • Haoliang Sun, Xiankai Lu, Haochen Wang, Yilong Yin, XianTong Zhen, Cees G. M. Snoek, Ling Shao
We define a global latent variable to represent the prototype of each object category, which we model as a probabilistic distribution.
1 code implementation • 9 May 2021 • Zehao Xiao, Jiayi Shen, XianTong Zhen, Ling Shao, Cees G. M. Snoek
Domain generalization is challenging due to the domain shift and the uncertainty caused by the inaccessibility of target domain data.
1 code implementation • 8 May 2021 • Yingjun Du, Haoliang Sun, XianTong Zhen, Jun Xu, Yilong Yin, Ling Shao, Cees G. M. Snoek
Specifically, we propose learning variational random features in a data-driven manner to obtain task-specific kernels by leveraging the shared knowledge provided by related tasks in a meta-learning setting.
no code implementations • 19 Mar 2021 • Tom van Sonsbeek, XianTong Zhen, Marcel Worring, Ling Shao
It is challenging to incorporate this information into disease classification due to the high reliance on clinician input in EHRs, limiting the possibility for automated diagnosis.
no code implementations • 1 Jan 2021 • Zehao Xiao, Jiayi Shen, XianTong Zhen, Ling Shao, Cees G. M. Snoek
In the probabilistic modeling framework, we introduce a domain-invariant principle to explore invariance across domains in a unified way.
no code implementations • ICLR 2021 • Yingjun Du, XianTong Zhen, Ling Shao, Cees G. M. Snoek
Batch normalization plays a crucial role when training deep neural networks.
no code implementations • 1 Jan 2021 • Jiayi Shen, XianTong Zhen, Marcel Worring, Ling Shao
Multi-task learning aims to improve the overall performance of a set of tasks by leveraging their relatedness.
no code implementations • 23 Dec 2020 • Haoliang Sun, XianTong Zhen, Chris Bailey, Parham Rasoulinejad, Yilong Yin, Shuo Li
The Cobb angle that quantitatively evaluates the spinal curvature plays an important role in the scoliosis diagnosis and treatment.
1 code implementation • NeurIPS 2020 • XianTong Zhen, Yingjun Du, Huan Xiong, Qiang Qiu, Cees G. M. Snoek, Ling Shao
The variational semantic memory accrues and stores semantic information for the probabilistic inference of class prototypes in a hierarchical Bayesian framework.
2 code implementations • ICML 2020 • Xiantong Zhen, Haoliang Sun, Ying-Jun Du, Jun Xu, Yilong Yin, Ling Shao, Cees Snoek
We propose meta variational random features (MetaVRF) to learn adaptive kernels for the base-learner, which is developed in a latent variable model by treating the random feature basis as the latent variable.
no code implementations • 21 Oct 2015 • Xiantong Zhen, Shuo Li
Direct methods have recently emerged as an effective and efficient tool in automated medical image analysis and become a trend to solve diverse challenging tasks in clinical practise.
no code implementations • CVPR 2015 • Xiantong Zhen, Zhijie Wang, Mengyang Yu, Shuo Li
In this paper, we propose a novel supervised descriptor learning (SDL) algorithm to establish a discriminative and compact feature representation for multi-output regression.