no code implementations • COLING 2022 • Jiayi Liu, Wei Wei, Zhixuan Chu, Xing Gao, Ji Zhang, Tan Yan, Yulin kang
Although the Conditional Variational Auto-Encoder (CVAE) model can generate more diversified responses than the traditional Seq2Seq model, the responses often have low relevance with the input words or are illogical with the question.
1 code implementation • 20 Dec 2024 • Zhenjie Xu, Wenqing Chen, Yi Tang, Xuanying Li, Cheng Hu, Zhixuan Chu, Kui Ren, Zibin Zheng, Zhichao Lu
Our experiments conducted on two datasets and two models demonstrate that MOMA reduces bias scores by up to 87. 7%, with only a marginal performance degradation of up to 6. 8% in the BBQ dataset.
no code implementations • 30 Oct 2024 • Yanchu Guan, Dong Wang, Yan Wang, Haiqing Wang, Renen Sun, Chenyi Zhuang, Jinjie Gu, Zhixuan Chu
In this paper, we propose an Explainable Behavior Cloning LLM Agent (EBC-LLMAgent), a novel approach that combines large language models (LLMs) with behavior cloning by learning demonstrations to create intelligent and explainable agents for autonomous mobile app interaction.
1 code implementation • 21 Oct 2024 • Shiyu Wang, Jiawei Li, Xiaoming Shi, Zhou Ye, Baichuan Mo, Wenze Lin, Shengtong Ju, Zhixuan Chu, Ming Jin
Specifically, we introduce a general-purpose TSPM that processes multi-scale time series using (1) multi-resolution time imaging (MRTI), (2) time image decomposition (TID), (3) multi-scale mixing (MCM), and (4) multi-resolution mixing (MRM) to extract comprehensive temporal patterns.
no code implementations • 6 Oct 2024 • Siqiao Xue, Tingting Chen, Fan Zhou, Qingyang Dai, Zhixuan Chu, Hongyuan Mei
Our benchmark aims to evaluate the abilities of multimodal large language models (MLLMs) in answering questions that require advanced financial knowledge and sophisticated reasoning.
no code implementations • 29 Jul 2024 • Zhixuan Chu, Hui Ding, Guang Zeng, Shiyu Wang, Yiming Li
Although the widespread use of AI systems in today's world is growing, many current AI systems are found vulnerable due to hidden bias and missing information, especially in the most commonly used forecasting system.
no code implementations • 29 Jul 2024 • Shiyu Wang, Zhixuan Chu, Yinbo Sun, Yu Liu, Yuliang Guo, Yang Chen, HuiYang Jian, Lintao Ma, Xingyu Lu, Jun Zhou
Despite recent advances with transformer-based forecasting models, challenges remain due to the non-stationary, nonlinear characteristics of workload time series and the long-term dependencies.
no code implementations • 24 Jun 2024 • Yichen Sun, Zhixuan Chu, Zhan Qin, Kui Ren
To address this problem, we introduce a novel diffusion-based framework to significantly enhance the alignment of generated images with their corresponding descriptions, addressing the inconsistency between visual output and textual input.
no code implementations • 6 Jun 2024 • Lei Liu, Xiaoyan Yang, Junchi Lei, Yue Shen, Jian Wang, Peng Wei, Zhixuan Chu, Zhan Qin, Kui Ren
With the advent of Large Language Models (LLMs), medical artificial intelligence (AI) has experienced substantial technological progress and paradigm shifts, highlighting the potential of LLMs to streamline healthcare delivery and improve patient outcomes.
no code implementations • 24 May 2024 • Zhibo Wang, Peng Kuang, Zhixuan Chu, Jingyi Wang, Kui Ren
To answer the questions, we revisit biased distributions in existing benchmarks and real-world datasets, and propose a fine-grained framework for analyzing dataset bias by disentangling it into the magnitude and prevalence of bias.
no code implementations • 7 May 2024 • Zhixuan Chu, Yan Wang, Longfei Li, Zhibo Wang, Zhan Qin, Kui Ren
Large Language Models (LLMs) have shown impressive performance in natural language tasks, but their outputs can exhibit undesirable attributes or biases.
1 code implementation • 7 May 2024 • Zhixuan Chu, Lei Zhang, Yichen Sun, Siqiao Xue, Zhibo Wang, Zhan Qin, Kui Ren
Leveraging the state-of-the-art keyframe extraction techniques and multimodal large language models, SoraDetector first evaluates the consistency between extracted video content summary and textual prompts, then constructs static and dynamic knowledge graphs (KGs) from frames to detect hallucination both in single frames and across frames.
no code implementations • 16 Feb 2024 • Guangya Wan, Yuqi Wu, Mengxuan Hu, Zhixuan Chu, Sheng Li
Causal discovery (CD) and Large Language Models (LLMs) represent two emerging fields of study with significant implications for artificial intelligence.
1 code implementation • 11 Feb 2024 • Fangyu Ding, Haiyang Wang, Zhixuan Chu, Tianming Li, Zhaoping Hu, Junchi Yan
Many recent endeavors of GIL focus on extracting the invariant subgraph from the input graph for prediction as a regularization strategy to improve the generalization performance of graph learning.
no code implementations • 6 Feb 2024 • Zhixuan Chu, Yan Wang, Feng Zhu, Lu Yu, Longfei Li, Jinjie Gu
The advent of large language models (LLMs) such as ChatGPT, PaLM, and GPT-4 has catalyzed remarkable advances in natural language processing, demonstrating human-like language fluency and reasoning capacities.
no code implementations • 16 Jan 2024 • Zhixuan Chu, Yan Wang, Qing Cui, Longfei Li, Wenqing Chen, Zhan Qin, Kui Ren
As personalized recommendation systems become vital in the age of information overload, traditional methods relying solely on historical user interactions often fail to fully capture the multifaceted nature of human interests.
no code implementations • 20 Dec 2023 • Zhixuan Chu, Mengxuan Hu, Qing Cui, Longfei Li, Sheng Li
To address this, we propose a Task-Driven Causal Feature Distillation model (TDCFD) to transform original feature values into causal feature attributions for the specific risk prediction task.
no code implementations • 4 Dec 2023 • Yanchu Guan, Dong Wang, Zhixuan Chu, Shiyu Wang, Feiyue Ni, Ruihua Song, Longfei Li, Jinjie Gu, Chenyi Zhuang
This paper proposes a novel LLM-based virtual assistant that can automatically perform multi-step operations within mobile apps based on high-level user requests.
no code implementations • 7 Oct 2023 • Zhixuan Chu, Huaiyu Guo, Xinyuan Zhou, Yijia Wang, Fei Yu, Hong Chen, Wanqing Xu, Xin Lu, Qing Cui, Longfei Li, Jun Zhou, Sheng Li
Large language models (LLMs) show promise for natural language tasks but struggle when applied directly to complex domains like finance.
2 code implementations • 3 Oct 2023 • Ming Jin, Shiyu Wang, Lintao Ma, Zhixuan Chu, James Y. Zhang, Xiaoming Shi, Pin-Yu Chen, Yuxuan Liang, Yuan-Fang Li, Shirui Pan, Qingsong Wen
We begin by reprogramming the input time series with text prototypes before feeding it into the frozen LLM to align the two modalities.
Ranked #1 on Time Series Forecasting on ETTh1 (48)
no code implementations • 23 Sep 2023 • Zhichao Chen, Leilei Ding, Zhixuan Chu, Yucheng Qi, Jianmin Huang, Hao Wang
Time-Series Forecasting based on Cumulative Data (TSFCD) is a crucial problem in decision-making across various industrial scenarios.
no code implementations • 6 Sep 2023 • Yan Wang, Zhixuan Chu, Tao Zhou, Caigao Jiang, Hongyan Hao, Minjie Zhu, Xindong Cai, Qing Cui, Longfei Li, james Y zhang, Siqiao Xue, Jun Zhou
Asynchronous time series, also known as temporal event sequences, are the basis of many applications throughout different industries.
no code implementations • 23 Aug 2023 • Ronghang Zhu, Dongliang Guo, Daiqing Qi, Zhixuan Chu, Xiang Yu, Sheng Li
Inspired by the concepts in trustworthy AI, we proposed the first trustworthy representation learning across domains framework which includes four concepts, i. e, robustness, privacy, fairness, and explainability, to give a comprehensive literature review on this research direction.
no code implementations • 21 Aug 2023 • Yan Wang, Zhixuan Chu, Xin Ouyang, Simeng Wang, Hongyan Hao, Yue Shen, Jinjie Gu, Siqiao Xue, james Y zhang, Qing Cui, Longfei Li, Jun Zhou, Sheng Li
In this paper, we propose a novel approach that leverages large language models (LLMs) to construct personalized reasoning graphs.
no code implementations • 21 Aug 2023 • Zhixuan Chu, Hongyan Hao, Xin Ouyang, Simeng Wang, Yan Wang, Yue Shen, Jinjie Gu, Qing Cui, Longfei Li, Siqiao Xue, james Y zhang, Sheng Li
In this paper, we propose RecSysLLM, a novel pre-trained recommendation model based on LLMs.
no code implementations • 29 Jul 2023 • Hongyan Hao, Zhixuan Chu, Shiyi Zhu, Gangwei Jiang, Yan Wang, Caigao Jiang, James Zhang, Wei Jiang, Siqiao Xue, Jun Zhou
In order to surmount this challenge and effectively integrate new sample distribution, we propose a density-based sample selection strategy that utilizes kernel density estimation to calculate sample density as a reference to compute sample weight, and employs weight sampling to construct a new memory set.
1 code implementation • 16 Jul 2023 • Siqiao Xue, Xiaoming Shi, Zhixuan Chu, Yan Wang, Hongyan Hao, Fan Zhou, Caigao Jiang, Chen Pan, James Y. Zhang, Qingsong Wen, Jun Zhou, Hongyuan Mei
In this paper, we present EasyTPP, the first central repository of research assets (e. g., data, models, evaluation programs, documentations) in the area of event sequence modeling.
no code implementations • 16 May 2023 • Yunyi Zhou, Zhixuan Chu, Yijia Ruan, Ge Jin, Yuchen Huang, Sheng Li
However, the choice of model highly relies on the characteristics of the input time series and the fixed distribution that the model is based on.
no code implementations • 3 Mar 2023 • Zhixuan Chu, Ruopeng Li, Stephen Rathbun, Sheng Li
We propose a Continual Causal Effect Representation Learning method for estimating causal effects with observational data, which are incrementally available from non-stationary data distributions.
1 code implementation • 25 Feb 2023 • Dongliang Guo, Zhixuan Chu, Sheng Li
To our best knowledge, FairAC is the first method that jointly addresses the graph attribution completion and graph unfairness problems.
no code implementations • 2 Feb 2023 • Zhixuan Chu, Jianmin Huang, Ruopeng Li, Wei Chu, Sheng Li
Causal inference has numerous real-world applications in many domains, such as health care, marketing, political science, and online advertising.
no code implementations • 3 Jan 2023 • Zhixuan Chu, Sheng Li
A further understanding of cause and effect within observational data is critical across many domains, such as economics, health care, public policy, web mining, online advertising, and marketing campaigns.
no code implementations • 20 Sep 2022 • Jiayi Liu, Wei Wei, Zhixuan Chu, Xing Gao, Ji Zhang, Tan Yan, Yulin kang
Although the Conditional Variational AutoEncoder (CVAE) model can generate more diversified responses than the traditional Seq2Seq model, the responses often have low relevance with the input words or are illogical with the question.
no code implementations • 22 Aug 2022 • Zhixuan Chu, Hui Ding, Guang Zeng, Yuchen Huang, Tan Yan, Yulin kang, Sheng Li
In this paper, we provide an in-depth analysis of the underlying parse tree-like structure involved in the effect prediction task and we further establish a Hierarchical Capsule Prediction Network (HapNet) for predicting the effects of marketing campaigns.
no code implementations • 10 Mar 2022 • Zhixuan Chu, Stephen L. Rathbun, Sheng Li
In our paper, the basket trial is employed as an intuitive example to present this new causal inference setting.
no code implementations • 22 Feb 2022 • Zhixuan Chu, Stephen Rathbun, Sheng Li
In this paper, we reveal the weaknesses of these strategies, i. e., they lead to the loss of predictive information when enforcing the domain invariance; and the treatment effect estimation performance is unstable, which heavily relies on the characteristics of the domain distributions and the choice of domain divergence metrics.
no code implementations • 29 Sep 2021 • Zhixuan Chu, Tan Yan, Yue Wu, Yi Xu, Cheng Zhang, Yulin kang
Time series forecasting has historically been a key area of academic research and industrial applications.
no code implementations • 5 Jun 2021 • Zhixuan Chu, Stephen L. Rathbun, Sheng Li
The foremost challenge in treatment effect estimation is how to capture hidden confounders.
no code implementations • 1 Jan 2021 • Zhixuan Chu, Stephen Rathbun, Sheng Li
We propose a Continual Causal Effect Representation Learning method for estimating causal effect with observational data, which are incrementally available from non-stationary data distributions.
no code implementations • 15 Sep 2020 • Zhixuan Chu, Stephen L. Rathbun, Sheng Li
The dramatically growing availability of observational data is being witnessed in various domains of science and technology, which facilitates the study of causal inference.
1 code implementation • 5 Feb 2020 • Liuyi Yao, Zhixuan Chu, Sheng Li, Yaliang Li, Jing Gao, Aidong Zhang
Embraced with the rapidly developed machine learning area, various causal effect estimation methods for observational data have sprung up.