1 code implementation • 28 May 2024 • Mihir Chauhan, Mohammad Abuzar Shaikh, Bina Ramamurthy, Mingchen Gao, Siwei Lyu, Sargur Srihari
We present SSL-HV: Self-Supervised Learning approaches applied to the task of Handwriting Verification.
no code implementations • 27 Mar 2024 • Dana Moukheiber, Saurabh Mahindre, Lama Moukheiber, Mira Moukheiber, Mingchen Gao
In this study, we propose a framework to achieve accurate diagnostic outcomes and ensure fairness across intersectional groups in high-dimensional chest X- ray multi-label classification.
no code implementations • 5 Feb 2024 • Yan Shen, Zhanghexuan Ji, Chunwei Ma, Mingchen Gao
We prove that with the introduction of a pre-trained source-only domain discriminator, the empirical estimation error of $\gH$-divergence related adversarial loss is reduced from the source domain side.
no code implementations • 1 Feb 2023 • Zhanghexuan Ji, Dazhou Guo, Puyang Wang, Ke Yan, Le Lu, Minfeng Xu, Jingren Zhou, Qifeng Wang, Jia Ge, Mingchen Gao, Xianghua Ye, Dakai Jin
Deep learning empowers the mainstream medical image segmentation methods.
no code implementations • ICCV 2023 • Zhanghexuan Ji, Dazhou Guo, Puyang Wang, Ke Yan, Le Lu, Minfeng Xu, Qifeng Wang, Jia Ge, Mingchen Gao, Xianghua Ye, Dakai Jin
In this work, we propose a new architectural CSS learning framework to learn a single deep segmentation model for segmenting a total of 143 whole-body organs.
no code implementations • CVPR 2023 • Zhenyi Wang, Li Shen, Donglin Zhan, Qiuling Suo, Yanjun Zhu, Tiehang Duan, Mingchen Gao
To make them trustworthy and robust to corruptions deployed in safety-critical scenarios, we propose a meta-learning framework of self-adaptive data augmentation to tackle the corruption robustness in CL.
1 code implementation • 3 Sep 2022 • Zhenyi Wang, Li Shen, Le Fang, Qiuling Suo, Donglin Zhan, Tiehang Duan, Mingchen Gao
Two key challenges arise in this more realistic setting: (i) how to use unlabeled data in the presence of a large amount of unlabeled out-of-distribution (OOD) data; and (ii) how to prevent catastrophic forgetting on previously learned task distributions due to the task distribution shift.
no code implementations • 28 Jul 2022 • Chunwei Ma, Zhanghexuan Ji, Ziyun Huang, Yan Shen, Mingchen Gao, Jinhui Xu
Exemplar-free Class-incremental Learning (CIL) is a challenging problem because rehearsing data from previous phases is strictly prohibited, causing catastrophic forgetting of Deep Neural Networks (DNNs).
1 code implementation • 15 Jul 2022 • Zhenyi Wang, Li Shen, Le Fang, Qiuling Suo, Tiehang Duan, Mingchen Gao
To address these problems, for the first time, we propose a principled memory evolution framework to dynamically evolve the memory data distribution by making the memory buffer gradually harder to be memorized with distributionally robust optimization (DRO).
no code implementations • 5 May 2022 • Yan Shen, Zhanghexuan Ji, Chunwei Ma, Mingchen Gao
Based on our particle filter inference algorithm, a semi-supervised learn-ing algorithm is utilized for learning tracking network on intermittent labeled frames by variational inference.
no code implementations • 4 May 2022 • Yan Shen, Fan Yang, Mingchen Gao, Wen Dong
Traditional machine learning approaches capture complex system dynamics either with dynamic Bayesian networks and state space models, which is hard to scale because it is non-trivial to prescribe the dynamics with a sparse graph or a system of differential equations; or a deep neural networks, where the distributed representation of the learned dynamics is hard to interpret.
1 code implementation • 5 Feb 2022 • Chunwei Ma, Ziyun Huang, Mingchen Gao, Jinhui Xu
One observation is that the widely embraced ProtoNet model is essentially a Voronoi Diagram (VD) in the feature space.
1 code implementation • CVPR 2022 • Zhenyi Wang, Li Shen, Tiehang Duan, Donglin Zhan, Le Fang, Mingchen Gao
We propose a domain shift detection technique to capture latent domain change and equip the meta optimizer with it to work in this setting.
1 code implementation • 16 Oct 2021 • Yan Shen, Jian Du, Han Zhao, Benyu Zhang, Zhanghexuan Ji, Mingchen Gao
Federated adversary domain adaptation is a unique distributed minimax training task due to the prevalence of label imbalance among clients, with each client only seeing a subset of the classes of labels required to train a global model.
no code implementations • ICLR 2022 • Chunwei Ma, Ziyun Huang, Mingchen Gao, Jinhui Xu
One observation is that the widely embraced ProtoNet model is essentially a Dirichlet Tessellation (Voronoi Diagram) in the feature space.
1 code implementation • ICCV 2021 • Zhenyi Wang, Tiehang Duan, Le Fang, Qiuling Suo, Mingchen Gao
In this paper, we explore a more practical and challenging setting where task distribution changes over time with domain shift.
1 code implementation • 4 Sep 2021 • Zhanghexuan Ji, Mohammad Abuzar Shaikh, Dana Moukheiber, Sargur Srihari, Yifan Peng, Mingchen Gao
Self-supervised learning provides an opportunity to explore unlabeled chest X-rays and their associated free-text reports accumulated in clinical routine without manual supervision.
1 code implementation • 4 Sep 2021 • Mohammad Abuzar Hashemi, Zhanghexuan Li, Mihir Chauhan, Yan Shen, Abhishek Satbhai, Mir Basheer Ali, Mingchen Gao, Sargur Srihari
Pre-training visual and textual representations from large-scale image-text pairs is becoming a standard approach for many downstream vision-language tasks.
no code implementations • 1 Sep 2021 • Yan Shen, Zhanghexuan Ji, Mingchen Gao
Many segmentation tasks for biomedical images can be modeled as the minimization of an energy function and solved by a class of max-flow and min-cut optimization algorithms.
1 code implementation • 25 Jun 2021 • Chunwei Ma, Ziyun Huang, Jiayi Xian, Mingchen Gao, Jinhui Xu
Deep Neural Networks (DNNs), despite their tremendous success in recent years, could still cast doubts on their predictions due to the intrinsic uncertainty associated with their learning process.
1 code implementation • 5 May 2021 • Soumyya Kanti Datta, Mohammad Abuzar Shaikh, Sargur N. Srihari, Mingchen Gao
Soft-Attention mechanism enables a neural network toachieve this goal.
Ranked #1 on Lesion Classification on ISIC 2017
no code implementations • 1 Jan 2021 • Yan Shen, Jian Du, Chunwei Ma, Mingchen Gao, Benyu Zhang
Our introduced SGLD oracle would lower generalization errors in local node's parameter learning and provide local node DP protections.
no code implementations • 5 Nov 2019 • Zhanghexuan Ji, Yan Shen, Chunwei Ma, Mingchen Gao
In this paper, we use only two kinds of weak labels, i. e., scribbles on whole tumor and healthy brain tissue, and global labels for the presence of each substructure, to train a deep learning model to segment all the sub-regions.
1 code implementation • 20 Sep 2019 • Chunwei Ma, Zhanghexuan Ji, Mingchen Gao
Three-dimensional medical image segmentation is one of the most important problems in medical image analysis and plays a key role in downstream diagnosis and treatment.
Ranked #2 on Cardiovascular MR Segmentaiton on HVSMR 2016
no code implementations • 15 Apr 2019 • Yan Shen, Mingchen Gao
We design a brain tumor segmentation algorithm that is robust to the absence of any modality.
no code implementations • 17 Aug 2018 • Yan Shen, Mingchen Gao
We present and evaluate a new deep neural network architecture for automatic thoracic disease detection on chest X-rays.
no code implementations • 19 Jan 2017 • Mingchen Gao, Ziyue Xu, Le Lu, Adam P. Harrison, Ronald M. Summers, Daniel J. Mollura
Accurately predicting and detecting interstitial lung disease (ILD) patterns given any computed tomography (CT) slice without any pre-processing prerequisites, such as manually delineated regions of interest (ROIs), is a clinically desirable, yet challenging goal.
no code implementations • 21 Sep 2016 • Mario Buty, Ziyue Xu, Mingchen Gao, Ulas Bagci, Aaron Wu, Daniel J. Mollura
Both sets of features were combined to estimate the nodule malignancy using a random forest classifier.
no code implementations • 10 Feb 2016 • Hoo-chang Shin, Holger R. Roth, Mingchen Gao, Le Lu, Ziyue Xu, Isabella Nogues, Jianhua Yao, Daniel Mollura, Ronald M. Summers
Another effective method is transfer learning, i. e., fine-tuning CNN models pre-trained from natural image dataset to medical image tasks.