no code implementations • 11 Dec 2024 • Panlong Wu, Kangshuo Li, Junbao Nan, Fangxin Wang
In this paper, we propose a novel privacy-preserving Federated In-Context LLM Agent Learning (FICAL) algorithm, which to our best knowledge for the first work unleashes the power of in-context learning to train diverse LLM agents through FL.
no code implementations • 11 Dec 2024 • Jinke Ren, Yaping Sun, Hongyang Du, Weiwen Yuan, Chongjie Wang, Xianda Wang, Yingbin Zhou, Ziwei Zhu, Fangxin Wang, Shuguang Cui
This system features two LLM-based AI agents at both the transmitter and receiver, serving as "brains" to enable powerful information understanding and content regeneration capabilities, respectively.
no code implementations • 5 Nov 2024 • Xunze Liu, Yifei Sun, Zhaorui Wang, Lizhao You, Haoyuan Pan, Fangxin Wang, Shuguang Cui
To solve this problem, this paper presents a receiver-centric generative semantic communication system, where each transmission is initialized by the receiver.
1 code implementation • 12 Oct 2024 • Fangxin Wang, Kay Liu, Sourav Medya, Philip S. Yu
Graph self-training is a semi-supervised learning method that iteratively selects a set of unlabeled data to retrain the underlying graph neural network (GNN) model and improve its prediction performance.
no code implementations • 4 Oct 2024 • Jiahao Zheng, Jinke Ren, Peng Xu, Zhihao Yuan, Jie Xu, Fangxin Wang, Gui Gui, Shuguang Cui
Semantic communication is a promising technology to improve communication efficiency by transmitting only the semantic information of the source data.
no code implementations • 25 Sep 2024 • Boyan Li, YongTing Chen, Dayou Zhang, Fangxin Wang
Volumetric video streaming offers immersive 3D experiences but faces significant challenges due to high bandwidth requirements and latency issues in transmitting detailed content in real time.
no code implementations • 10 May 2024 • Rongyu Zhang, Yun Chen, Chenrui Wu, Fangxin Wang, Bo Li
Federated learning (FL) offers a privacy-centric distributed learning framework, enabling model training on individual clients and central aggregation without necessitating data exchange.
no code implementations • 11 Mar 2024 • Fangxin Wang, Yuqing Liu, Kay Liu, Yibo Wang, Sourav Medya, Philip S. Yu
Therefore, identifying, quantifying, and utilizing uncertainty are essential to enhance the performance of the model for the downstream tasks as well as the reliability of the GNN predictions.
no code implementations • 14 Feb 2024 • Chen Wang, Fangxin Wang, Ruocheng Guo, Yueqing Liang, Kay Liu, Philip S. Yu
Recognizing the critical role of confidence in aligning training objectives with evaluation metrics, we propose CPFT, a versatile framework that enhances recommendation confidence by integrating Conformal Prediction (CP)-based losses with CE loss during fine-tuning.
no code implementations • 4 Feb 2024 • Duo Wu, Xianda Wang, Yaqi Qiao, Zhi Wang, Junchen Jiang, Shuguang Cui, Fangxin Wang
Motivated by the recent success of large language models (LLMs), this work studies the LLM adaptation for networking to explore a more sustainable design philosophy.
no code implementations • 5 Jan 2024 • Panlong Wu, Qi Liu, Yanjie Dong, Fangxin Wang
In the first step, we optimize the seller's pricing decision and propose an Iterative Model Pricing (IMP) algorithm that optimizes the prices of large models iteratively by reasoning customers' future rental decisions, which is able to achieve a near-optimal pricing solution.
1 code implementation • 29 Dec 2023 • Kay Liu, Hengrui Zhang, Ziqing Hu, Fangxin Wang, Philip S. Yu
To bridge this gap, we introduce GODM, a novel data augmentation for mitigating class imbalance in supervised Graph Outlier detection via latent Diffusion Models.
no code implementations • 26 Dec 2023 • Panlong Wu, Kangshuo Li, Ting Wang, Fangxin Wang
In this paper, we propose a novel two-stage federated learning algorithm called FedMS.
no code implementations • 19 Aug 2023 • Duo Wu, Dayou Zhang, Miao Zhang, Ruoyu Zhang, Fangxin Wang, Shuguang Cui
The high-accuracy and resource-intensive deep neural networks (DNNs) have been widely adopted by live video analytics (VA), where camera videos are streamed over the network to resource-rich edge/cloud servers for DNN inference.
no code implementations • 6 Apr 2023 • Chenrui Wu, Zexi Li, Fangxin Wang, Chao Wu
It includes a noise-resilient local solver and a robust global aggregator.
no code implementations • 27 Mar 2023 • Rongyu Zhang, Xiaowei Chi, Guiliang Liu, Wenyi Zhang, Yuan Du, Fangxin Wang
Multimodal learning has seen great success mining data features from multiple modalities with remarkable model performance improvement.
no code implementations • 7 Mar 2023 • Kaiyuan Hu, Yili Jin, Haowen Yang, Junhua Liu, Fangxin Wang
Recent years have witnessed a rapid development of immersive multimedia which bridges the gap between the real world and virtual space.