1 code implementation • 8 Mar 2024 • Weibin Liao, Yinghao Zhu, Xinyuan Wang, Chengwei Pan, Yasha Wang, Liantao Ma
This highlights the potential of Mamba in facilitating model lightweighting.
no code implementations • 30 Jan 2024 • Weibin Liao, Yinghao Zhu, Zixiang Wang, Xu Chu, Yasha Wang, Liantao Ma
PAI no longer introduces any imputed data but constructs a learnable prompt to model the implicit preferences of the downstream model for missing values, resulting in a significant performance improvement for all EHR analysis models.
1 code implementation • 25 Jan 2024 • Yinghao Zhu, Zixiang Wang, Junyi Gao, Yuning Tong, Jingkun An, Weibin Liao, Ewen M. Harrison, Liantao Ma, Chengwei Pan
The inherent complexity of structured longitudinal Electronic Health Records (EHR) data poses a significant challenge when integrated with Large Language Models (LLMs), which are traditionally tailored for natural language processing.
no code implementations • 6 Oct 2023 • Weibin Liao, Xuhong LI, Qingzhong Wang, Yanwu Xu, Zhaozheng Yin, Haoyi Xiong
While pre-training on object detection tasks, such as Common Objects in Contexts (COCO) [1], could significantly boost the performance of cell segmentation, it still consumes on massive fine-annotated cell images [2] with bounding boxes, masks, and cell types for every cell in every image, to fine-tune the pre-trained model.
no code implementations • 3 Oct 2023 • Weibin Liao, Haoyi Xiong, Qingzhong Wang, Yan Mo, Xuhong LI, Yi Liu, Zeyu Chen, Siyu Huang, Dejing Dou
In this work, we study a novel self-supervised pre-training pipeline, namely Multi-task Self-super-vised Continual Learning (MUSCLE), for multiple medical imaging tasks, such as classification and segmentation, using X-ray images collected from multiple body parts, including heads, lungs, and bones.