1 code implementation • 9 Aug 2024 • Mingze Gong, Shuoyao Wang, Suzhi Bi, Yuan Wu, LiPing Qian
In particular, in the FE stage, we learn the representation ability of semantic information by end-to-end training the encoder and decoder in an analog manner.
1 code implementation • 8 Aug 2024 • Yupeng Chang, Yi Chang, Yuan Wu
Large language models (LLMs) have exhibited remarkable proficiency across a diverse array of natural language processing (NLP) tasks.
1 code implementation • 17 May 2024 • Tingyu Xia, Bowen Yu, Yuan Wu, Yi Chang, Chang Zhou
In this paper, we initiate our discussion by demonstrating how Large Language Models (LLMs), when tasked with responding to queries, display a more even probability distribution in their answers if they are more adept, as opposed to their less skilled counterparts.
1 code implementation • 5 May 2024 • Xu Wang, Cheng Li, Yi Chang, Jindong Wang, Yuan Wu
The results are revealing: NegativePrompt markedly enhances the performance of LLMs, evidenced by relative improvements of 12. 89% in Instruction Induction tasks and 46. 25% in BIG-Bench tasks.
1 code implementation • 5 May 2024 • Juntao Hu, Yuan Wu
We propose a novel data augmentation method termed You Only Need hAlf (YONA), which simplifies the augmentation process.
1 code implementation • 24 Apr 2024 • Yahan Li, Yuan Wu
Unsupervised domain adaptation (UDA) aims to transfer knowledge from a labeled source domain to an unlabeled target domain.
no code implementations • 1 Mar 2024 • Yuan Wu
Subsequently, we propose a margin discrepancy-based adversarial training (MDAT) approach for MDTC, in accordance with our theoretical analysis.
no code implementations • 7 Feb 2024 • Xi Chen, Yang Cai, Yuan Wu, Bo Xiong, Taesung Park
Recently, MBConv blocks, initially designed for efficiency in resource-limited settings and later adapted for cutting-edge image classification performances, have demonstrated significant potential in image classification tasks.
1 code implementation • 27 Jan 2024 • Yue Zhou, Chenlu Guo, Xu Wang, Yi Chang, Yuan Wu
Leveraging large models, these data augmentation techniques have outperformed traditional approaches.
no code implementations • CVPR 2024 • Yanlu Cai, Weizhong Zhang, Yuan Wu, Cheng Jin
A natural solution is to artificially synthesize some samples i. e. 2D-3D pose pairs under massive new camera settings.
no code implementations • 18 Dec 2023 • Juntao Hu, Yuan Wu
The most successful multi-domain text classification (MDTC) approaches employ the shared-private paradigm to facilitate the enhancement of domain-invariant features through domain-specific attributes.
2 code implementations • 6 Nov 2023 • Li Ping Qian, Yi Zhang, Sikai Lyu, Huijie Zhu, Yuan Wu, Xuemin Sherman Shen, Xiaoniu Yang
Particularly, at the transmitter side, a high-precision image semantic segmentation algorithm is proposed to extract the semantic information of the image to achieve significant compression of the image data.
no code implementations • 21 Oct 2023 • Peichun Li, Hanwen Zhang, Yuan Wu, LiPing Qian, Rong Yu, Dusit Niyato, Xuemin Shen
Distributed Artificial Intelligence (AI) model training over mobile edge networks encounters significant challenges due to the data and resource heterogeneity of edge devices.
no code implementations • 18 Aug 2023 • Yapeng Zhao, Qingqing Wu, Guangji Chen, Wen Chen, Ruiqi Liu, Ming-Min Zhao, Yuan Wu, Shaodan Ma
Moreover, the results indicate that the DT assisted MEC system can precisely achieve the balance between local computing and task offloading since real-time system status can be obtained with the help of DT.
no code implementations • 9 Aug 2023 • Xumin Huang, Yuan Wu, Jiawen Kang, Jiangtian Nie, Weifeng Zhong, Dong In Kim, Shengli Xie
A single-leader multi-follower Stackelberg game is formulated between the MSP and users while each user optimizes an offloading probability to minimize the weighted sum of time, energy consumption and monetary cost.
no code implementations • 14 Jul 2023 • Xumin Huang, Peichun Li, Hongyang Du, Jiawen Kang, Dusit Niyato, Dong In Kim, Yuan Wu
Artificial intelligence generated content (AIGC) has emerged as a promising technology to improve the efficiency, quality, diversity and flexibility of the content creation process by adopting a variety of generative AI models.
1 code implementation • 6 Jul 2023 • Yupeng Chang, Xu Wang, Jindong Wang, Yuan Wu, Linyi Yang, Kaijie Zhu, Hao Chen, Xiaoyuan Yi, Cunxiang Wang, Yidong Wang, Wei Ye, Yue Zhang, Yi Chang, Philip S. Yu, Qiang Yang, Xing Xie
Large language models (LLMs) are gaining increasing popularity in both academia and industry, owing to their unprecedented performance in various applications.
no code implementations • 8 Jan 2023 • Peichun Li, Guoliang Cheng, Xumin Huang, Jiawen Kang, Rong Yu, Yuan Wu, Miao Pan
We propose a cost-adjustable FL framework, named AnycostFL, that enables diverse edge devices to efficiently perform local updates under a wide range of efficiency constraints.
no code implementations • 18 Nov 2022 • Huawei Hou, Suzhi Bi, Lili Zheng, Xiaohui Lin, Yuan Wu, Zhi Quan
In this paper, we propose a Domain-Agnostic and Sample-Efficient wireless indoor crowd Counting (DASECount) framework that suffices to attain robust cross-domain detection accuracy given very limited data samples in new domains.
no code implementations • 20 Sep 2022 • Jiadong Yu, Yang Li, Xiaolan Liu, Bo Sun, Yuan Wu, Danny H. K. Tsang
In this paper, we investigate the joint offloading, communication and computation resource allocation for IRS-assisted NOMA MEC system.
no code implementations • 6 Feb 2022 • Kaiyi Zhang, Ximing Yang, Yuan Wu, Cheng Jin
Besides, the missing patterns are diverse in reality, but existing methods can only handle fixed ones, which means a poor generalization ability.
no code implementations • 30 Jan 2022 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
Multi-domain text classification (MDTC) aims to leverage all available resources from multiple domains to learn a predictive model that can generalize well on these domains.
no code implementations • 29 Jan 2022 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
Multi-domain text classification (MDTC) has obtained remarkable achievements due to the advent of deep learning.
no code implementations • 28 Dec 2021 • Jinkai Zheng, Tom H. Luan, Longxiang Gao, Yao Zhang, Yuan Wu
In specific, to preserve the precious computing resource at different levels for most appropriate computing tasks, we integrate a learning scheme based on the prediction of futuristic computing tasks in DT.
1 code implementation • 10 Dec 2021 • Kaiyi Zhang, Ximing Yang, Yuan Wu, Cheng Jin
The points generated by AXform do not have the strong 2-manifold constraint, which improves the generation of non-smooth surfaces.
no code implementations • 19 Oct 2021 • Xumin Huang, Peichun Li, Rong Yu, Yuan Wu, Kan Xie, Shengli Xie
In PVEC, different PLOs recruit PVs as edge computing nodes for offloading services through an incentive mechanism, which is designed according to the computation demand and parking capacity constraints derived from FedParking.
no code implementations • 14 Aug 2021 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
Adversarial domain adaptation has made impressive advances in transferring knowledge from the source domain to the target domain by aligning feature distributions of both domains.
no code implementations • EACL (AdaptNLP) 2021 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
We provide theoretical analysis for the CAN framework, showing that CAN's objective is equivalent to minimizing the total divergence among multiple joint distributions of shared features and label predictions.
no code implementations • 31 Jan 2021 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
Using the shared-private paradigm and adversarial training has significantly improved the performances of multi-domain text classification (MDTC) models.
no code implementations • 1 Jan 2021 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
Domain adaptation sets out to address this problem, aiming to leverage labeled data in the source domain to learn a good predictive model for the target domain whose labels are scarce or unavailable.
no code implementations • IEEE 2020 • Ying Chen, Zhiyong Liu, Yongchao Zhang, Yuan Wu, Xin Chen, Lian Zhao
In order to minimize the long-term average delay of the tasks, theoriginal problem is transformed into a Markov decision process (MDP).
no code implementations • ECCV 2020 • Yuan Wu, Diana Inkpen, Ahmed El-Roby
Second, samples from the source and target domains alone are not sufficient for domain-invariant feature extracting in the latent space.
no code implementations • 3 May 2020 • Liang Huang, You Zhang, Weijian Pan, Jinyin Chen, Li Ping Qian, Yuan Wu
Extensive numerical results show both the CNN-based classifier and LSTM-based classifier extract similar radio features relating to modulation reference points.
1 code implementation • 12 Feb 2020 • Tianhui Zhou, Yitong Li, Yuan Wu, David Carlson
We address these challenges by proposing a novel method to capture predictive distributions in regression by defining two neural networks with two distinct loss functions.
no code implementations • 6 Dec 2019 • Liang Huang, Weijian Pan, You Zhang, LiPing Qian, Nan Gao, Yuan Wu
Deep learning has recently been applied to automatically classify the modulation categories of received radio signals without manual experience.
no code implementations • 18 Sep 2019 • Yuan Wu, Yuhong Guo
In this paper we propose a novel dual adversarial co-learning approach for multi-domain text classification (MDTC).
General Classification Multi-Domain Sentiment Classification +4
no code implementations • 4 Sep 2018 • Yuan Wu, Lingling Li, Lian Li
We introduce the chi-square test neural network: a single hidden layer backpropagation neural network using chi-square test theorem to redefine the cost function and the error function.