no code implementations • 2 Dec 2024 • Wenbo Zhang, Junyu Chen, Christopher Kanan
Due to their large sizes, volumetric scans and whole-slide pathology images (WSIs) are often processed by extracting embeddings from local regions and then an aggregator makes predictions from this set.
no code implementations • 24 Oct 2024 • Kexuan Xin, Qingyun Wang, Junyu Chen, Pengfei Yu, Huimin Zhao, Heng Ji
In the rapidly evolving field of metabolic engineering, the quest for efficient and precise gene target identification for metabolite production enhancement presents significant challenges.
no code implementations • 14 Oct 2024 • Enze Xie, Junsong Chen, Junyu Chen, Han Cai, Haotian Tang, Yujun Lin, Zhekai Zhang, Muyang Li, Ligeng Zhu, Yao Lu, Song Han
We introduce Sana, a text-to-image framework that can efficiently generate images up to 4096$\times$4096 resolution.
1 code implementation • 14 Oct 2024 • Junyu Chen, Han Cai, Junsong Chen, Enze Xie, Shang Yang, Haotian Tang, Muyang Li, Yao Lu, Song Han
With these designs, we improve the autoencoder's spatial compression ratio up to 128 while maintaining the reconstruction quality.
1 code implementation • 14 Oct 2024 • Haotian Tang, Yecheng Wu, Shang Yang, Enze Xie, Junsong Chen, Junyu Chen, Zhuoyang Zhang, Han Cai, Yao Lu, Song Han
To address these challenges, we present the hybrid tokenizer, which decomposes the continuous latents from the autoencoder into two components: discrete tokens representing the big picture and continuous tokens representing the residual components that cannot be represented by the discrete tokens.
1 code implementation • 20 Sep 2024 • Junyu Chen, Yihao Liu, Shuwen Wei, Aaron Carass, Yong Du
Affine registration plays a crucial role in PET/CT imaging, where aligning PET with CT images is challenging due to their respective functional and anatomical representations.
1 code implementation • 6 Sep 2024 • Yecheng Wu, Zhuoyang Zhang, Junyu Chen, Haotian Tang, Dacheng Li, Yunhao Fang, Ligeng Zhu, Enze Xie, Hongxu Yin, Li Yi, Song Han, Yao Lu
VILA-U is a Unified foundation model that integrates Video, Image, Language understanding and generation.
no code implementations • 14 Aug 2024 • Junyu Chen, Long Shi, Badong Chen
This approach enables each enhancement to focus on different GSFs, thereby achieving diverse feature representation in the enhanced structure.
no code implementations • 9 Aug 2024 • Jian Lu, Shikhar Srivastava, Junyu Chen, Robik Shrestha, Manoj Acharya, Kushal Kafle, Christopher Kanan
With the advent of multi-modal large language models (MLLMs), datasets used for visual question answering (VQA) and referring expression comprehension have seen a resurgence.
1 code implementation • 14 Jul 2024 • Yihao Liu, Junyu Chen, Lianrui Zuo, Aaron Carass, Jerry L. Prince
VFA uses neural networks to extract multi-resolution feature maps from the fixed and moving images and then retrieves pixel-level correspondences based on feature similarity.
no code implementations • 8 Mar 2024 • Junyu Chen, Yihao Liu, Shuwen Wei, Zhangxing Bian, Aaron Carass, Yong Du
Here, we propose a novel framework to concurrently estimate both the epistemic and aleatoric segmentation uncertainties for image registration.
no code implementations • 31 Jan 2024 • Zhangxing Bian, Ahmed Alshareef, Shuwen Wei, Junyu Chen, Yuli Wang, Jonghye Woo, Dzung L. Pham, Jiachen Zhuo, Aaron Carass, Jerry L. Prince
This is a factor that has been overlooked in prior research on tMRI post-processing.
no code implementations • CVPR 2024 • Zifan Wang, Junyu Chen, Ziqing Chen, Pengwei Xie, Rui Chen, Li Yi
This paper presents GenH2R a framework for learning generalizable vision-based human-to-robot (H2R) handover skills.
no code implementations • 1 Jan 2024 • Zifan Wang, Junyu Chen, Ziqing Chen, Pengwei Xie, Rui Chen, Li Yi
We further introduce a distillation-friendly demonstration generation method that automatically generates a million high-quality demonstrations suitable for learning.
1 code implementation • 22 Dec 2023 • Junyu Chen, Binh T. Nguyen, Shang Hui Koh, Yong Sheng Soh
The relaxation can be viewed as the Lagrangian dual of the GW distance augmented with constraints that relate to the linear and quadratic terms of transportation plans.
no code implementations • 13 Dec 2023 • Zifan Wang, Zhuorui Ye, Haoran Wu, Junyu Chen, Li Yi
To tackle this challenging problem, we properly model the synergetic relationship between future forecasting and semantic scene completion through a novel network named SCSFNet.
1 code implementation • journal 2023 • Junyu Chen, Qianqian Xu, Zhiyong Yang, Ke Ma, Xiaochun Cao, Qingming Huang
For the motif-based node representation learning process, we propose a Motif Coarsening strategy for incorporating motif structure into the graph representation learning process.
2 code implementations • 15 Sep 2023 • Junyu Chen, Susmitha Vekkot, Pancham Shukla
Music source separation (MSS) aims to extract 'vocals', 'drums', 'bass' and 'other' tracks from a piece of mixed music.
Ranked #10 on Music Source Separation on MUSDB18-HQ
no code implementations • 25 Aug 2023 • Md Yousuf Harun, Jhair Gallardo, Junyu Chen, Christopher Kanan
Compared to uniform balanced sampling, GRASP achieves the same performance with 40% fewer updates.
no code implementations • 5 Aug 2023 • Zhangxing Bian, Shuwen Wei, Yihao Liu, Junyu Chen, Jiachen Zhuo, Fangxu Xing, Jonghye Woo, Aaron Carass, Jerry L. Prince
We introduce a novel "momenta, shooting, and correction" framework for Lagrangian motion estimation in the presence of repetitive patterns and large motion.
no code implementations • 28 Jul 2023 • Junyu Chen, Yihao Liu, Shuwen Wei, Zhangxing Bian, Shalini Subramanian, Aaron Carass, Jerry L. Prince, Yong Du
Deep learning technologies have dramatically reshaped the field of medical image registration over the past decade.
no code implementations • 8 May 2023 • Junyu Chen, Jie An, Hanjia Lyu, Christopher Kanan, Jiebo Luo
Assessing the artness of AI-generated images continues to be a challenge within the realm of image generation.
no code implementations • 28 Mar 2023 • Jingyang Lin, Junyu Chen, Hanjia Lyu, Igor Khodak, Divya Chhabra, Colby L Day Richardson, Irina Prelipcean, Andrew M Dylag, Jiebo Luo
In this work, we first analyze the correlations between three adverse neonatal outcomes and then formulate the diagnosis of multiple neonatal outcomes as a multi-task learning (MTL) problem.
no code implementations • 24 Mar 2023 • Junyu Chen, Norihiro Yoshida, Hiroaki Takada
And in the development of machine learning systems, the most widely used are publicly available datasets.
no code implementations • 10 Mar 2023 • Junyu Chen, Yihao Liu, Yufan He, Yong Du
In the past, optimization-based registration models have used spatially-varying regularization to account for deformation variations in different image regions.
no code implementations • 10 Mar 2023 • Junyu Chen, Yihao Liu, Yufan He, Yong Du
Transformers have recently shown promise for medical image applications, leading to an increasing interest in developing such models for medical image registration.
no code implementations • 8 Feb 2023 • Gary Y. Li, Junyu Chen, Se-In Jang, Kuang Gong, Quanzheng Li
Inspired by the recent success of Vision Transformers and advances in multi-modal image analysis, we propose a novel segmentation model, debuted, Cross-Modal Swin Transformer (SwinCross), with cross-modal attention (CMA) module to incorporate cross-modal feature extraction at multiple resolutions. To validate the effectiveness of the proposed method, we performed experiments on the HECKTOR 2021 challenge dataset and compared it with the nnU-Net (the backbone of the top-5 methods in HECKTOR 2021) and other state-of-the-art transformer-based methods such as UNETR, and Swin UNETR.
no code implementations • ICCV 2023 • Yunze Liu, Junyu Chen, Zekai Zhang, Jingwei Huang, Li Yi
With such frames, we can factorize geometry and motion to facilitate a feature-space geometric reconstruction for more effective 4D learning.
no code implementations • 21 Dec 2022 • Ye Li, Junyu Chen, Se-In Jang, Kuang Gong, Quanzheng Li
Inspired by the recent success of Transformers for Natural Language Processing and vision Transformer for Computer Vision, many researchers in the medical imaging community have flocked to Transformer-based networks for various main stream medical tasks such as classification, segmentation, and estimation.
1 code implementation • 12 Dec 2022 • Yihao Liu, Junyu Chen, Shuwen Wei, Aaron Carass, Jerry Prince
For digital transformations, |J| is commonly approximated using a central difference, but this strategy can yield positive |J|'s for transformations that are clearly not diffeomorphic -- even at the voxel resolution level.
1 code implementation • 23 Nov 2022 • Junyu Chen, Jie An, Hanjia Lyu, Christopher Kanan, Jiebo Luo
Visual-textual sentiment analysis aims to predict sentiment with the input of a pair of image and text, which poses a challenge in learning effective features for diverse input images.
1 code implementation • Conference 2022 • Junyu Chen, Qianqian Xu, Zhiyong Yang, Ke Ma, Xiaochun Cao, Qingming Huang
To attack this problem, we propose a recursive meta-learning model with the user's behavior sequence prediction as a separate training task.
2 code implementations • Proceedings of the 30th ACM International Conference on Multimedia 2022 • Junyu Chen, Qianqian Xu, Zhiyong Yang, Xiaochun Cao, Qingming Huang
We develop a multi-class AUC optimization work to deal with the class imbalance problem.
1 code implementation • 7 Sep 2022 • Se-In Jang, Tinsu Pan, Ye Li, Pedram Heidari, Junyu Chen, Quanzheng Li, Kuang Gong
In this work, we proposed an efficient spatial and channel-wise encoder-decoder transformer, Spach Transformer, that can leverage spatial and channel information based on local and global MSAs.
no code implementations • 2 Jun 2022 • Jun Li, Junyu Chen, Yucheng Tang, Ce Wang, Bennett A. Landman, S. Kevin Zhou
Transformer, the latest technological advance of deep learning, has gained prevalence in natural language processing or computer vision.
no code implementations • 15 Mar 2022 • Ye Li, Jianan Cui, Junyu Chen, Guodong Zeng, Scott Wollenweber, Floris Jansen, Se-In Jang, Kyungsang Kim, Kuang Gong, Quanzheng Li
Our hypothesis is that by explicitly providing the local relative noise level of the input image to a deep convolutional neural network (DCNN), the DCNN can outperform itself trained on image appearance only.
1 code implementation • 19 Nov 2021 • Junyu Chen, Eric C. Frey, Yufan He, William P. Segars, Ye Li, Yong Du
Recently Vision Transformer architectures have been proposed to address the shortcomings of ConvNets and have produced state-of-the-art performances in many medical imaging applications.
Ranked #1 on Medical Image Registration on OASIS
no code implementations • 29 Sep 2021 • Junyu Chen, Evren Asma, Chung Chan
In this study, we present Targeted Gradient Descent (TGD), a novel fine-tuning method that can extend a pre-trained network to a new task without revisiting data from the previous task while preserving the knowledge acquired from previous training.
1 code implementation • 17 Apr 2021 • Junyu Chen, Ye Li, Licia P. Luna, Hyun Woo Chung, Steven P. Rowe, Yong Du, Lilja B. Solnes, Eric C. Frey
The results demonstrated that the proposed method provides fast and robust lesion and bone segmentation for QBSPECT/CT.
1 code implementation • 13 Apr 2021 • Junyu Chen, Yufan He, Eric C. Frey, Ye Li, Yong Du
However, the performances of ConvNets are still limited by lacking the understanding of long-range spatial relations in an image.
Ranked #4 on Medical Image Registration on OASIS
1 code implementation • MIDL 2019 • Junyu Chen, Eric C. Frey
For the majority of the learning-based segmentation methods, a large quantity of high-quality training data is required.
1 code implementation • 6 Dec 2019 • Junyu Chen, Ye Li, Yong Du, Eric C. Frey
In this work, we present a novel image registration method for creating highly anatomically detailed anthropomorphic phantoms from a single digital phantom.
1 code implementation • 5 Jul 2019 • Junyu Chen, Eric C. Frey
Pixel intensity is a widely used feature for clustering and segmentation algorithms, the resulting segmentation using only intensity values might suffer from noises and lack of spatial context information.