Search Results for author: Shizhong Han

Found 19 papers, 4 papers with code

FutureDepth: Learning to Predict the Future Improves Video Depth Estimation

no code implementations19 Mar 2024 Rajeev Yasarla, Manish Kumar Singh, Hong Cai, Yunxiao Shi, Jisoo Jeong, Yinhao Zhu, Shizhong Han, Risheek Garrepalli, Fatih Porikli

In this paper, we propose a novel video depth estimation approach, FutureDepth, which enables the model to implicitly leverage multi-frame and motion cues to improve depth estimation by making it learn to predict the future at training.

Future prediction Monocular Depth Estimation

OpenShape: Scaling Up 3D Shape Representation Towards Open-World Understanding

1 code implementation NeurIPS 2023 Minghua Liu, Ruoxi Shi, Kaiming Kuang, Yinhao Zhu, Xuanlin Li, Shizhong Han, Hong Cai, Fatih Porikli, Hao Su

Due to their alignment with CLIP embeddings, our learned shape representations can also be integrated with off-the-shelf CLIP-based models for various applications, such as point cloud captioning and point cloud-conditioned image generation.

3D Classification 3D Shape Representation +4

Validating uncertainty in medical image translation

1 code implementation11 Feb 2020 Jacob C. Reinhold, Yufan He, Shizhong Han, Yunqiang Chen, Dashan Gao, Junghoon Lee, Jerry L. Prince, Aaron Carass

Medical images are increasingly used as input to deep neural networks to produce quantitative values that aid researchers and clinicians.

Translation

Finding novelty with uncertainty

2 code implementations11 Feb 2020 Jacob C. Reinhold, Yufan He, Shizhong Han, Yunqiang Chen, Dashan Gao, Junghoon Lee, Jerry L. Prince, Aaron Carass

Medical images are often used to detect and characterize pathology and disease; however, automatically identifying and segmenting pathology in medical images is challenging because the appearance of pathology across diseases varies widely.

Segmentation

Outlier Guided Optimization of Abdominal Segmentation

no code implementations10 Feb 2020 Yuchen Xu, Olivia Tang, Yucheng Tang, Ho Hin Lee, Yunqiang Chen, Dashan Gao, Shizhong Han, Riqiang Gao, Michael R. Savona, Richard G. Abramson, Yuankai Huo, Bennett A. Landman

We built on a pre-trained 3D U-Net model for abdominal multi-organ segmentation and augmented the dataset either with outlier data (e. g., exemplars for which the baseline algorithm failed) or inliers (e. g., exemplars for which the baseline algorithm worked).

Active Learning Computed Tomography (CT) +2

Contrast Phase Classification with a Generative Adversarial Network

no code implementations14 Nov 2019 Yucheng Tang, Ho Hin Lee, Yuchen Xu, Olivia Tang, Yunqiang Chen, Dashan Gao, Shizhong Han, Riqiang Gao, Camilo Bermudez, Michael R. Savona, Richard G. Abramson, Yuankai Huo, Bennett A. Landman

Dynamic contrast enhanced computed tomography (CT) is an imaging technique that provides critical information on the relationship of vascular structure and dynamics in the context of underlying anatomy.

Anatomy Classification +4

Semi-Supervised Multi-Organ Segmentation through Quality Assurance Supervision

no code implementations12 Nov 2019 Ho Hin Lee, Yucheng Tang, Olivia Tang, Yuchen Xu, Yunqiang Chen, Dashan Gao, Shizhong Han, Riqiang Gao, Michael R. Savona, Richard G. Abramson, Yuankai Huo, Bennett A. Landman

The contributions of the proposed method are threefold: We show that (1) the QA scores can be used as a loss function to perform semi-supervised learning for unlabeled data, (2) the well trained discriminator is learnt by QA score rather than traditional true/false, and (3) the performance of multi-organ segmentation on unlabeled datasets can be fine-tuned with more robust and higher accuracy than the original baseline method.

Image Segmentation Medical Image Segmentation +3

Feature-level and Model-level Audiovisual Fusion for Emotion Recognition in the Wild

no code implementations6 Jun 2019 Jie Cai, Zibo Meng, Ahmed Shehab Khan, Zhiyuan Li, James O'Reilly, Shizhong Han, Ping Liu, Min Chen, Yan Tong

In this paper, we proposed two strategies to fuse information extracted from different modalities, i. e., audio and visual.

Emotion Recognition

Identity-Free Facial Expression Recognition using conditional Generative Adversarial Network

no code implementations19 Mar 2019 Jie Cai, Zibo Meng, Ahmed Shehab Khan, Zhiyuan Li, James O'Reilly, Shizhong Han, Yan Tong

A novel Identity-Free conditional Generative Adversarial Network (IF-GAN) was proposed for Facial Expression Recognition (FER) to explicitly reduce high inter-subject variations caused by identity-related facial attributes, e. g., age, race, and gender.

Facial Expression Recognition Facial Expression Recognition (FER) +1

Optimizing Filter Size in Convolutional Neural Networks for Facial Action Unit Recognition

no code implementations CVPR 2018 Shizhong Han, Zibo Meng, Zhiyuan Li, James O'Reilly, Jie Cai, Xiao-Feng Wang, Yan Tong

Most recently, Convolutional Neural Networks (CNNs) have shown promise for facial AU recognition, where predefined and fixed convolution filter sizes are employed.

Facial Action Unit Detection

Incremental Boosting Convolutional Neural Network for Facial Action Unit Recognition

no code implementations NeurIPS 2016 Shizhong Han, Zibo Meng, Ahmed Shehab Khan, Yan Tong

Experimental results on four benchmark AU databases have demonstrated that the IB-CNN yields significant improvement over the traditional CNN and the boosting CNN without incremental learning, as well as outperforming the state-of-the-art CNN-based methods in AU recognition.

Facial Action Unit Detection Incremental Learning

Improving Speech Related Facial Action Unit Recognition by Audiovisual Information Fusion

no code implementations29 Jun 2017 Zibo Meng, Shizhong Han, Ping Liu, Yan Tong

Instead of solely improving visual observations, this paper presents a novel audiovisual fusion framework, which makes the best use of visual and acoustic cues in recognizing speech-related facial AUs.

Facial Action Unit Detection

Listen to Your Face: Inferring Facial Action Units from Audio Channel

no code implementations23 Jun 2017 Zibo Meng, Shizhong Han, Yan Tong

Different from all prior work that utilized visual observations for facial AU recognition, this paper presents a novel approach that recognizes speech-related AUs exclusively from audio signals based on the fact that facial activities are highly correlated with voice during speech.

Facial Expression Recognition via a Boosted Deep Belief Network

no code implementations CVPR 2014 Ping Liu, Shizhong Han, Zibo Meng, Yan Tong

A training process for facial expression recognition is usually performed sequentially in three individual stages: feature learning, feature selection, and classifier construction.

Facial Expression Recognition Facial Expression Recognition (FER) +1

Cannot find the paper you are looking for? You can Submit a new open access paper.