no code implementations • 20 Nov 2024 • Yifei Zhang, Tianxu Jiang, Bo Pan, Jingyu Wang, Guangji Bai, Liang Zhao
A Visual Explanation Distribution Consistency loss further reinforces visual coherence by aligning the generated visual explanations with dataset-level patterns, enabling the model to effectively learn from incomplete multimodal supervision.
no code implementations • 18 Oct 2024 • Guangji Bai, Yijiang Li, Zilinghan Li, Liang Zhao, Kibaek Kim
Large Language Models (LLMs) achieve state-of-the-art performance but are challenging to deploy due to their high computational and storage demands.
1 code implementation • 25 May 2024 • Zekun Cai, Guangji Bai, Renhe Jiang, Xuan Song, Liang Zhao
Temporal Domain Generalization (TDG) addresses the challenge of training predictive models under temporally varying data distributions.
no code implementations • 25 May 2024 • Qilong Zhao, Shiyu Wang, Guangji Bai, Bo Pan, Zhaohui Qin, Liang Zhao
This is due to the long-lasting challenge of jointly identifying key latent variables, their causal relations, and their correlation with properties of interest, as well as how to leverage their discoveries toward causally controlled data generation.
2 code implementations • 23 Apr 2024 • Xiongxiao Xu, Canyu Chen, Yueqing Liang, Baixiang Huang, Guangji Bai, Liang Zhao, Kai Shu
To meet the objectives, we propose a multi-scale hybrid Mamba-Transformer experts model State Space Transformer (SST).
2 code implementations • 28 Feb 2024 • Guangji Bai, Yijiang Li, Chen Ling, Kibaek Kim, Liang Zhao
The transformative impact of large language models (LLMs) like LLaMA and GPT on natural language processing is countered by their prohibitive computational demands.
1 code implementation • 15 Feb 2024 • Chen Ling, Xujiang Zhao, Xuchao Zhang, Wei Cheng, Yanchi Liu, Yiyou Sun, Mika Oishi, Takao Osaki, Katsushi Matsuda, Jie Ji, Guangji Bai, Liang Zhao, Haifeng Chen
Existing works have been devoted to quantifying the uncertainty in LLM's response, but they often overlook the complex nature of LLMs and the uniqueness of in-context learning.
1 code implementation • 1 Jan 2024 • Guangji Bai, Zheng Chai, Chen Ling, Shiyu Wang, Jiaying Lu, Nan Zhang, Tingwei Shi, Ziyang Yu, Mengdan Zhu, Yifei Zhang, Xinyuan Song, Carl Yang, Yue Cheng, Liang Zhao
We categorize methods based on their optimization focus: computational, memory, energy, financial, and network resources and their applicability across various stages of an LLM's lifecycle, including architecture design, pretraining, finetuning, and system design.
no code implementations • 19 Dec 2023 • Junxiang Wang, Guangji Bai, Wei Cheng, Zhengzhang Chen, Liang Zhao, Haifeng Chen
In order to tackle these challenges simultaneously, in this paper, we introduce PrOmpt-based domaiN Discrimination (POND), the first framework to utilize prompts for time series domain adaptation.
1 code implementation • 12 Oct 2023 • Yifei Zhang, Siyi Gu, Bo Pan, Guangji Bai, Meikang Qiu, Xiaofeng Yang, Liang Zhao
However, in many real-world situations, it is usually desired to prompt the model with visual attention without model retraining.
no code implementations • 12 Oct 2023 • Yifei Zhang, Siyi Gu, James Song, Bo Pan, Guangji Bai, Liang Zhao
Our proposed benchmarks facilitate a fair evaluation and comparison of visual explanation methods.
1 code implementation • 6 Oct 2023 • Guangji Bai, Qilong Zhao, Xiaoyang Jiang, Yifei Zhang, Liang Zhao
Continual Learning is a burgeoning domain in next-generation AI, focusing on training neural networks over a sequence of tasks akin to human learning.
no code implementations • 25 Aug 2023 • Guangji Bai, Ziyang Yu, Zheng Chai, Yue Cheng, Liang Zhao
It utilizes an offline memory to cache historical information (e. g., node embedding) as an affordable approximation of the exact value and achieves high concurrency.
no code implementations • 19 May 2023 • Shiyu Wang, Guangji Bai, Qingyang Zhu, Zhaohui Qin, Liang Zhao
As a result, domain generalization graph transformation that predicts graphs not available in the training data is under-explored, with multiple key challenges to be addressed including (1) the extreme space complexity when training on all input-output mode combinations, (2) difference of graph topologies between the input and the output modes, and (3) how to generalize the model to (unseen) target domains that are not in the training data.
no code implementations • 4 Feb 2023 • Tanmoy Chowdhury, Chen Ling, Xuchao Zhang, Xujiang Zhao, Guangji Bai, Jian Pei, Haifeng Chen, Liang Zhao
Knowledge-enhanced neural machine reasoning has garnered significant attention as a cutting-edge yet challenging research area with numerous practical applications.
1 code implementation • 26 Dec 2022 • Guangji Bai, Chen Ling, Yuyang Gao, Liang Zhao
Specifically, we innovatively propose to store the part of the image most important to the tasks in episodic memory by saliency map extraction and memory encoding.
1 code implementation • 3 Oct 2022 • Dazhou Yu, Guangji Bai, Yun Li, Liang Zhao
Spatial domain generalization is a spatial extension of domain generalization, which can generalize to unseen spatial domains in continuous 2D space.
1 code implementation • 3 Jul 2022 • Guangji Bai, Liang Zhao
Specifically, we propose to model the task relation as the similarity between task input gradients, with a theoretical analysis of their equivalency.
1 code implementation • 27 Jun 2022 • Yuyang Gao, Tong Steven Sun, Guangji Bai, Siyi Gu, Sungsoo Ray Hong, Liang Zhao
Despite the fast progress of explanation techniques in modern Deep Neural Networks (DNNs) where the main focus is handling "how to generate the explanations", advanced research questions that examine the quality of the explanation itself (e. g., "whether the explanations are accurate") and improve the explanation quality (e. g., "how to adjust the model to generate more accurate explanations when explanations are inaccurate") are still relatively under-explored.
no code implementations • 31 May 2022 • Zheng Chai, Guangji Bai, Liang Zhao, Yue Cheng
Traditional sampling-based methods accelerate GNN training by dropping edges and nodes, which impairs the graph integrity and model performance.
1 code implementation • 21 May 2022 • Guangji Bai, Chen Ling, Liang Zhao
Temporal domain generalization is a promising yet extremely challenging area where the goal is to learn models under temporally changing data distributions and generalize to unseen data distributions following the trends of the change.
no code implementations • 22 Feb 2021 • Johnny Torres, Guangji Bai, Junxiang Wang, Liang Zhao, Carmen Vaca, Cristina Abad
Multi-task learning is a framework that enforces different learning tasks to share their knowledge to improve their generalization performance.