no code implementations • 10 Mar 2025 • Yanlong Wang, Jian Xu, Tiantian Gao, Hongkang Zhang, Shao-Lun Huang, Danny Dongning Sun, Xiao-Ping Zhang
Despite the growing attention to time series forecasting in recent years, many studies have proposed various solutions to address the challenges encountered in time series prediction, aiming to improve forecasting performance.
no code implementations • 10 Mar 2025 • Yanlong Wang, Jian Xu, Shao-Lun Huang, Danny Dongning Sun, Xiao-Ping Zhang
We introduce a novel deep learning model that utilizes a Gaussian mixture distribution to capture the complex, time-varying nature of asset return distributions in the Chinese stock market.
no code implementations • 6 Feb 2025 • Qingyue Zhang, Haohao Fu, Guanbo Huang, Yaoyuan Liang, Chang Chu, Tianren Peng, Yanru Wu, Qi Li, Yang Li, Shao-Lun Huang
Additionally, we develop an architecture-agnostic and data-efficient algorithm OTQMS to implement our theoretical results for training deep multi-source transfer learning models.
no code implementations • 3 Nov 2024 • Yanlong Wang, Jian Xu, Fei Ma, Shao-Lun Huang, Danny Dongning Sun, Xiao-Ping Zhang
Time series forecasting remains a critical challenge across various domains, often complicated by high-dimensional data and long-term dependencies.
no code implementations • 24 Jun 2024 • Xiao Liang, Xinyu Hu, Simiao Zuo, Yeyun Gong, Qiang Lou, Yi Liu, Shao-Lun Huang, Jian Jiao
On average, TRAIT improves LLM performance by 8% in the advertisement domain and 7. 5% in the math domain.
no code implementations • International Conference on Acoustics, Speech, and Signal Processing 2024 • Yuqing Li, Haoming Huang, Jian Xu, Shao-Lun Huang
The presence of noisy correspondence within cross-modal matching has significantly undermined the performance of existing matching methods.
1 code implementation • 26 Oct 2023 • Xiao Liang, Tao Shi, Yaoyuan Liang, Te Tao, Shao-Lun Huang
In this paper, we propose DiffusionVG, a novel framework with diffusion models that formulates video grounding as a conditional generation task, where the target span is generated from Gaussian noise inputs and interatively refined in the reverse diffusion process.
1 code implementation • 25 Oct 2023 • Tao Shi, Xiao Liang, Yaoyuan Liang, Xinyi Tong, Shao-Lun Huang
To address these challenges, we propose an efficient and model-agnostic SCL framework named Supervised Sample-Label Contrastive Learning with Soft-HGR Maximal Correlation (SSLCL), which eliminates the need for a large batch size and can be seamlessly integrated with existing ERC models without introducing any model-specific assumptions.
4 code implementations • 20 Jun 2023 • Jian Xu, Xinyi Tong, Shao-Lun Huang
Data heterogeneity is one of the most challenging issues in federated learning, which motivates a variety of approaches to learn personalized models for participating clients.
no code implementations • 11 Mar 2023 • Jian Xu, Meiling Yang, Wenbo Ding, Shao-Lun Huang
The label distribution skew induced data heterogeniety has been shown to be a significant obstacle that limits the model performance in federated learning, which is particularly developed for collaborative model training over decentralized data sources while preserving user privacy.
no code implementations • 20 Dec 2022 • Yajie Bao, Yang Li, Shao-Lun Huang, Lin Zhang, Lizhong Zheng, Amir Zamir, Leonidas Guibas
Task transfer learning is a popular technique in image processing applications that uses pre-trained models to reduce the supervision cost of related tasks.
1 code implementation • 24 Oct 2022 • Xili Dai, Mingyang Li, Pengyuan Zhai, Shengbang Tong, Xingjian Gao, Shao-Lun Huang, Zhihui Zhu, Chong You, Yi Ma
We show that such models have equally strong empirical performance on CIFAR-10, CIFAR-100, and ImageNet datasets when compared to conventional neural networks.
no code implementations • 12 Jul 2022 • Yang Tan, Enming Zhang, Yang Li, Shao-Lun Huang, Xiao-Ping Zhang
We propose two novel transferability metrics F-OTCE (Fast Optimal Transport based Conditional Entropy) and JC-OTCE (Joint Correspondence OTCE) to evaluate how much the source model (task) can benefit the learning of the target task and to learn more transferable representations for cross-domain cross-task transfer learning.
no code implementations • 11 Jul 2022 • Jingge Wang, Liyan Xie, Yao Xie, Shao-Lun Huang, Yang Li
Domain generalization aims at learning a universal model that performs well on unseen target domains, incorporating knowledge from multiple source domains.
no code implementations • 12 Jun 2022 • Runpeng Yu, Hong Zhu, Kaican Li, Lanqing Hong, Rui Zhang, Nanyang Ye, Shao-Lun Huang, Xiuqiang He
Due to the poor generalization performance of traditional empirical risk minimization (ERM) in the case of distributional shift, Out-of-Distribution (OoD) generalization algorithms receive increasing attention.
no code implementations • NeurIPS 2021 • Xinyi Tong, Xiangxiang Xu, Shao-Lun Huang, Lizhong Zheng
Current transfer learning algorithm designs mainly focus on the similarities between source and target tasks, while the impacts of the sample sizes of these tasks are often not sufficiently addressed.
no code implementations • 30 Sep 2021 • Yang Tan, Yang Li, Shao-Lun Huang
Recent analytical transferability metrics are mainly designed for image classification problem, and currently there is no specific investigation for the transferability estimation of semantic segmentation task, which is an essential problem in autonomous driving, medical image analysis, etc.
1 code implementation • ICLR 2022 • Zifeng Wang, Shao-Lun Huang, Ercan E. Kuruoglu, Jimeng Sun, Xi Chen, Yefeng Zheng
Then, we build an IIW-based information bottleneck on the trade-off between accuracy and information complexity of NNs, namely PIB.
no code implementations • NeurIPS Workshop DLDE 2021 • Feng Zhao, Xiang Chen, Jun Wang, Zuoqiang Shi, Shao-Lun Huang
Traditionally, we provide technical parameters for ODE solvers, such as the order, the stepsize and the local error threshold.
no code implementations • 14 Sep 2021 • Xiangxiang Xu, Shao-Lun Huang
Specifically, we consider the distributed hypothesis testing (DHT) problem where two distributed nodes are constrained to transmit a constant number of bits to a central decoder.
3 code implementations • 13 Sep 2021 • Jian Xu, Shao-Lun Huang, Linqi Song, Tian Lan
To this end, previous work either makes use of auxiliary data at parameter server to verify the received gradients (e. g., by computing validation error rate) or leverages statistic-based methods (e. g. median and Krum) to identify and remove malicious gradients from Byzantine clients.
no code implementations • 24 Aug 2021 • Fei Ma, Xiangxiang Xu, Shao-Lun Huang, Lin Zhang
Moreover, we develop a generalized form of the softmax function to effectively implement maximum likelihood estimation in an end-to-end manner.
no code implementations • 30 Jul 2021 • Guangfeng Yan, Shao-Lun Huang, Tian Lan, Linqi Song
Gradient quantization is an emerging technique in reducing communication costs in distributed learning.
no code implementations • 19 Jun 2021 • Yang Tan, Yang Li, Shao-Lun Huang
Transferability estimation is an essential problem in transfer learning to predict how good the performance is when transferring a source model (or source task) to a target task.
3 code implementations • CVPR 2021 • Yang Tan, Yang Li, Shao-Lun Huang
Specifically, we use optimal transport to estimate domain difference and the optimal coupling between source and target distributions, which is then used to derive the conditional entropy of the target task (task difference).
1 code implementation • 27 Feb 2021 • Zifeng Wang, Yifan Yang, Rui Wen, Xi Chen, Shao-Lun Huang, Yefeng Zheng
Current deep learning based disease diagnosis systems usually fall short in catastrophic forgetting, i. e., directly fine-tuning the disease diagnosis model on new tasks usually leads to abrupt decay of performance on previous tasks.
no code implementations • 1 Jan 2021 • Guangfeng Yan, Shao-Lun Huang, Tian Lan, Linqi Song
This paper addresses this issue by proposing a novel dynamic quantized SGD (DQSGD) framework, which enables us to optimize the quantization strategy for each gradient descent step by exploring the trade-off between communication cost and modeling error.
no code implementations • 17 Dec 2020 • Zelong Yang, Yan Wang, Piji Li, Shaobin Lin, Shuming Shi, Shao-Lun Huang, Wei Bi
The multiplayer online battle arena (MOBA) games have become increasingly popular in recent years.
no code implementations • COLING 2022 • Zifeng Wang, Rui Wen, Xi Chen, Shao-Lun Huang, Ningyu Zhang, Yefeng Zheng
Distant supervision (DS) is a strong way to expand the datasets for enhancing relation extraction (RE) models but often suffers from high label noise.
1 code implementation • NeurIPS 2020 • Zifeng Wang, Xi Chen, Rui Wen, Shao-Lun Huang, Ercan E. Kuruoglu, Yefeng Zheng
Counterfactual learning for dealing with missing-not-at-random data (MNAR) is an intriguing topic in the recommendation literature since MNAR data are ubiquitous in modern recommender systems.
no code implementations • 6 Sep 2020 • Zifeng Wang, Rui Wen, Xi Chen, Shilei Cao, Shao-Lun Huang, Buyue Qian, Yefeng Zheng
We propose a Healthcare Graph Convolutional Network (HealGCN) to offer disease self-diagnosis service for online users based on Electronic Healthcare Records (EHRs).
no code implementations • 14 Aug 2020 • Zelong Yang, Zhufeng Pan, Yan Wang, Deng Cai, Xiaojiang Liu, Shuming Shi, Shao-Lun Huang
With the rapid prevalence and explosive development of MOBA esports (Multiplayer Online Battle Arena electronic sports), much research effort has been devoted to automatically predicting game results (win predictions).
no code implementations • 25 Jan 2020 • Zifeng Wang, Xi Chen, Rui Wen, Shao-Lun Huang
Observed events in recommendation are consequence of the decisions made by a policy, thus they are usually selectively labeled, namely the data are Missing Not At Random (MNAR), which often causes large bias to the estimate of true outcomes risk.
1 code implementation • 3 Dec 2019 • Zifeng Wang, Hong Zhu, Zhenhua Dong, Xiuqiang He, Shao-Lun Huang
In the time of Big Data, training complex models on large-scale data sets is challenging, making it appealing to reduce data volume for saving computation resources by subsampling.
no code implementations • 20 Nov 2019 • Shao-Lun Huang, Anuran Makur, Gregory W. Wornell, Lizhong Zheng
We consider the problem of identifying universal low-dimensional features from high-dimensional data for inference tasks in settings involving learning.
no code implementations • 8 Oct 2019 • Shao-Lun Huang, Xiangxiang Xu, Lizhong Zheng
In this paper, we propose an information-theoretic approach to design the functional representations to extract the hidden common structure shared by a set of random variables.
no code implementations • 16 May 2019 • Shao-Lun Huang, Xiangxiang Xu, Lizhong Zheng, Gregory W. Wornell
It is commonly believed that the hidden layers of deep neural networks (DNNs) attempt to extract informative features for learning tasks.
1 code implementation • ICLR 2019 • Yajie Bao, Yang Li, Shao-Lun Huang, Lin Zhang, Amir R. Zamir, Leonidas J. Guibas
An important question in task transfer learning is to determine task transferability, i. e. given a common input domain, estimating to what extent representations learned from a source task can help in learning a target task.
no code implementations • 22 Nov 2018 • Lichen Wang, Jiaxiang Wu, Shao-Lun Huang, Lizhong Zheng, Xiangxiang Xu, Lin Zhang, Junzhou Huang
We further generalize the framework to handle more than two modalities and missing modalities.