no code implementations • Findings (EMNLP) 2021 • Haoyu Wang, Fenglong Ma, Yaqing Wang, Jing Gao
We propose to mine outline knowledge of concepts related to given sentences from Wikipedia via BM25 model.
no code implementations • Findings (NAACL) 2022 • Yaqing Wang, Xin Tian, Haoyi Xiong, Yueyang Li, Zeyu Chen, Sheng Guo, Dejing Dou
In this work, we show that Relation Graph augmented Learning (RGL) can improve the performance of few-shot natural language understanding tasks.
no code implementations • 24 Feb 2025 • Longchao Da, Xiaoou Liu, Jiaxin Dai, Lu Cheng, Yaqing Wang, Hua Wei
In this work, we propose a novel framework that quantifies uncertainty in LLM explanations through a reasoning topology perspective.
no code implementations • 8 Oct 2024 • Shiguang Wu, Yaqing Wang, Yatao Bian, Quanming Yao
Meta-learning enables learning systems to adapt quickly to new tasks, similar to humans.
no code implementations • 3 Oct 2024 • Jiale Fu, Yaqing Wang, Simeng Han, Jiaming Fan, Chen Si, Xu Yang
However, the effectiveness of ICL heavily relies on the selection of ICEs, and conventional text-based embedding methods are often inadequate for tasks that require multi-step reasoning, such as mathematical and logical problem solving.
1 code implementation • 28 Jul 2024 • Feijie Wu, Xingchen Wang, Yaqing Wang, Tianci Liu, Lu Su, Jing Gao
In federated learning (FL), accommodating clients' varied computational capacities poses a challenge, often limiting the participation of those with constrained resources in global model training.
no code implementations • 14 Jul 2024 • Yaqing Wang, Hongming Piao, daxiang dong, Quanming Yao, Jingbo Zhou
While existing methods focus on enhancing item ID embeddings for new items within general CTR models, they tend to adopt a global feature interaction approach, often overshadowing new items with sparse data by those with abundant interactions.
no code implementations • 29 Jun 2024 • Quanming Yao, Yongqi Zhang, Yaqing Wang, Nan Yin, James Kwok, Qiang Yang
The brute-force scaleup of training datasets, learnable parameters and computation power, has become a prevalent strategy for developing more robust learning models.
no code implementations • 20 Jun 2024 • Yuan Zhong, Xiaochen Wang, Jiaqi Wang, Xiaokun Zhang, Yaqing Wang, Mengdi Huai, Cao Xiao, Fenglong Ma
To address these limitations, we propose a novel EHR data generation model called EHRPD.
no code implementations • 24 Feb 2024 • Junyu Luo, Xiaochen Wang, Jiaqi Wang, Aofei Chang, Yaqing Wang, Fenglong Ma
Automatic International Classification of Diseases (ICD) coding plays a crucial role in the extraction of relevant information from clinical notes for proper recording and billing.
no code implementations • CVPR 2024 • Jialin Wu, Xia Hu, Yaqing Wang, Bo Pang, Radu Soricut
Large multi-modal models (LMMs) exhibit remarkable performance across numerous tasks.
Ranked #1 on
Visual Question Answering (VQA)
on A-OKVQA
(using extra training data)
1 code implementation • 25 Nov 2023 • Yaqing Wang, Zaifei Yang, Quanming Yao
Thus, the lack of DDIs is implicitly compensated by the enriched drug representations and propagated drug similarities.
no code implementations • 18 Oct 2023 • Yaqing Wang, Jialin Wu, Tanmaya Dabral, Jiageng Zhang, Geoff Brown, Chun-Ta Lu, Frederick Liu, Yi Liang, Bo Pang, Michael Bendersky, Radu Soricut
Intrusive PEFT techniques directly change a model's internal architecture.
no code implementations • 17 Oct 2023 • Yaqing Wang, Jiepu Jiang, Mingyang Zhang, Cheng Li, Yi Liang, Qiaozhu Mei, Michael Bendersky
Personalized text generation presents a specialized mechanism for delivering content that is specific to a user's personal context.
1 code implementation • 11 Oct 2023 • Xiaochen Wang, Junyu Luo, Jiaqi Wang, Ziyi Yin, Suhan Cui, Yuan Zhong, Yaqing Wang, Fenglong Ma
Pretraining has proven to be a powerful technique in natural language processing (NLP), exhibiting remarkable success in various NLP downstream tasks.
1 code implementation • 8 Oct 2023 • Lu Yin, You Wu, Zhenyu Zhang, Cheng-Yu Hsieh, Yaqing Wang, Yiling Jia, Gen Li, Ajay Jaiswal, Mykola Pechenizkiy, Yi Liang, Michael Bendersky, Zhangyang Wang, Shiwei Liu
Large Language Models (LLMs), renowned for their remarkable performance across diverse domains, present a challenge when it comes to practical deployment due to their colossal model size.
no code implementations • 4 Oct 2023 • Yuan Zhong, Suhan Cui, Jiaqi Wang, Xiaochen Wang, Ziyi Yin, Yaqing Wang, Houping Xiao, Mengdi Huai, Ting Wang, Fenglong Ma
Health risk prediction is one of the fundamental tasks under predictive modeling in the medical domain, which aims to forecast the potential health risks that patients may face in the future using their historical Electronic Health Records (EHR).
1 code implementation • 1 Oct 2023 • Shiguang Wu, Yaqing Wang, Quanming Yao
We then adopt a hierarchical adaptation mechanism to adapt the encoder at task-level and the predictor at query-level by the unified GNN adapter.
no code implementations • 29 Sep 2023 • Shengkun Tang, Yaqing Wang, Caiwen Ding, Yi Liang, Yao Li, Dongkuan Xu
Unlike typical adaptive computation challenges that deal with single-step generation problems, diffusion processes with a multi-step generation need to dynamically adjust their computational resource allocation based on the ongoing assessment of each step's importance to the final image output, presenting a unique set of challenges.
no code implementations • 15 Aug 2023 • Cheng Li, Mingyang Zhang, Qiaozhu Mei, Yaqing Wang, Spurthi Amba Hombaiah, Yi Liang, Michael Bendersky
Inspired by the practice of writing education, we develop a multistage and multitask framework to teach LLMs for personalized generation.
1 code implementation • 12 Jul 2023 • Yan Wen, Chen Gao, Lingling Yi, Liwei Qiu, Yaqing Wang, Yong Li
Automated Machine Learning (AutoML) techniques have recently been introduced to design Collaborative Filtering (CF) models in a data-specific manner.
1 code implementation • 6 Jun 2023 • Shiguang Wu, Yaqing Wang, Qinghe Jing, daxiang dong, Dejing Dou, Quanming Yao
Instead of using a fixed modulation function and deciding modulation position by expertise, we propose a modulation framework called ColdNAS for user cold-start problem, where we look for proper modulation structure, including function and position, via neural architecture search.
no code implementations • 5 Mar 2023 • Jiaqi Wang, Shenglai Zeng, Zewei Long, Yaqing Wang, Houping Xiao, Fenglong Ma
This is a new yet practical scenario in federated learning, i. e., labels-at-server semi-supervised federated learning (SemiFL).
no code implementations • 19 Feb 2023 • Tianci Liu, Haoyu Wang, Yaqing Wang, Xiaoqian Wang, Lu Su, Jing Gao
This new framework utilizes data that have similar labels when estimating fairness on a particular label group for better stability, and can unify DP and EOp.
1 code implementation • 8 Jan 2023 • Yan Li, Xinjiang Lu, Yaqing Wang, Dejing Dou
In this work, we propose to address the time series forecasting problem with generative modeling and propose a bidirectional variational auto-encoder (BVAE) equipped with diffusion, denoise, and disentanglement, namely D3VAE.
1 code implementation • CVPR 2023 • Shengkun Tang, Yaqing Wang, Zhenglun Kong, Tianchi Zhang, Yao Li, Caiwen Ding, Yanzhi Wang, Yi Liang, Dongkuan Xu
To handle this challenge, we propose a novel early exiting strategy for unified visual language models, which allows dynamically skip the layers in encoder and decoder simultaneously in term of input layer-wise similarities with multiple times of early exiting, namely \textbf{MuE}.
1 code implementation • 31 Oct 2022 • Yaqing Wang, Sahaj Agarwal, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao
Standard fine-tuning of large pre-trained language models (PLMs) for downstream tasks requires updating hundreds of millions to billions of parameters, and storing a large copy of the PLM weights for every task resulting in increased cost for storing, sharing and serving the models.
2 code implementations • 4 Jul 2022 • Xuhong LI, Haoyi Xiong, Yi Liu, Dingfu Zhou, Zeyu Chen, Yaqing Wang, Dejing Dou
Though image classification datasets could provide the backbone networks with rich visual features and discriminative ability, they are incapable of fully pre-training the target model (i. e., backbone+segmentation modules) in an end-to-end manner.
1 code implementation • 24 May 2022 • Yaqing Wang, Sahaj Agarwal, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao
Standard fine-tuning of large pre-trained language models (PLMs) for downstream tasks requires updating hundreds of millions to billions of parameters, and storing a large copy of the PLM weights for every task resulting in increased cost for storing, sharing and serving the models.
Natural Language Understanding
parameter-efficient fine-tuning
+1
no code implementations • 6 May 2022 • Quanming Yao, Yaqing Wang, Bo Han, James Kwok
While the optimization problem is nonconvex and nonsmooth, we show that its critical points still have good statistical performance on the tensor completion problem.
1 code implementation • EMNLP 2021 • Yaqing Wang, Song Wang, Quanming Yao, Dejing Dou
Short text classification is a fundamental task in natural language processing.
1 code implementation • Findings (NAACL) 2022 • Yaqing Wang, Subhabrata Mukherjee, Xiaodong Liu, Jing Gao, Ahmed Hassan Awadallah, Jianfeng Gao
The first is the use of self-training to leverage large amounts of unlabeled data for prompt-based FN in few-shot settings.
no code implementations • 6 Oct 2021 • Haoran Liu, Haoyi Xiong, Yaqing Wang, Haozhe An, Dongrui Wu, Dejing Dou
Specifically, we design a new metric $\mathcal{P}$-vector to represent the principal subspace of deep features learned in a DNN, and propose to measure angles between the principal subspaces using $\mathcal{P}$-vectors.
no code implementations • 12 Sep 2021 • Liwei Che, Zewei Long, Jiaqi Wang, Yaqing Wang, Houping Xiao, Fenglong Ma
In particular, we propose to use three networks and a dynamic quality control mechanism to generate high-quality pseudo labels for unlabeled data, which are added to the training set.
no code implementations • Findings (EMNLP) 2021 • Yaqing Wang, Haoda Chu, Chao Zhang, Jing Gao
In this work, we study the problem of named entity recognition (NER) in a low resource scenario, focusing on few-shot and zero-shot settings.
no code implementations • 9 Sep 2021 • Zewei Long, Jiaqi Wang, Yaqing Wang, Houping Xiao, Fenglong Ma
Most existing FedSSL methods focus on the classical scenario, i. e, the labeled and unlabeled data are stored at the client side.
no code implementations • NeurIPS 2021 • Yaqing Wang, Abulikemu Abuduweili, Quanming Yao, Dejing Dou
the target property, such that the limited labels can be effectively propagated among similar molecules.
no code implementations • 22 Jun 2021 • Yaqing Wang, Fenglong Ma, Haoyu Wang, Kishlay Jha, Jing Gao
The experimental results show our proposed MetaFEND model can detect fake news on never-seen events effectively and outperform the state-of-the-art methods.
no code implementations • 1 Jan 2021 • Yaqing Wang, Subhabrata Mukherjee, Haoda Chu, Yuancheng Tu, Ming Wu, Jing Gao, Ahmed Hassan Awadallah
Neural sequence labeling is an important technique employed for many Natural Language Processing (NLP) tasks, such as Named Entity Recognition (NER), slot tagging for dialog systems and semantic parsing.
no code implementations • 1 Jan 2021 • Haoran Liu, Haoyi Xiong, Yaqing Wang, Haozhe An, Dongrui Wu, Dejing Dou
While deep learning is effective to learn features/representations from data, the distributions of samples in feature spaces learned by various architectures for different training tasks (e. g., latent layers of AEs and feature vectors in CNN classifiers) have not been well-studied or compared.
no code implementations • 6 Dec 2020 • Zewei Long, Liwei Che, Yaqing Wang, Muchao Ye, Junyu Luo, Jinze Wu, Houping Xiao, Fenglong Ma
In this paper, we focus on designing a general framework FedSiam to tackle different scenarios of federated semi-supervised learning, including four settings in the labels-at-client scenario and two setting in the labels-at-server scenario.
1 code implementation • COLING 2022 • Junyu Luo, Zifei Zheng, Hanzhong Ye, Muchao Ye, Yaqing Wang, Quanzeng You, Cao Xiao, Fenglong Ma
To fairly evaluate the performance, we also propose three specific evaluation metrics.
no code implementations • 7 Oct 2020 • Yaqing Wang, Subhabrata Mukherjee, Haoda Chu, Yuancheng Tu, Ming Wu, Jing Gao, Ahmed Hassan Awadallah
While self-training serves as an effective mechanism to learn from large amounts of unlabeled data -- meta-learning helps in adaptive sample re-weighting to mitigate error propagation from noisy pseudo-labels.
no code implementations • 16 Aug 2020 • Yaqing Wang, Fenglong Ma, Jing Gao
To tackle this challenging task, we propose a cross-graph representation learning framework, i. e., CrossVal, which can leverage an external KG to validate the facts in the target KG efficiently.
no code implementations • 14 Aug 2020 • Yaqing Wang, Quanming Yao, James T. Kwok
Extensive low-rank matrix completion experiments on a number of synthetic and real-world data sets show that the proposed method obtains state-of-the-art recovery performance while being the fastest in comparison to existing low-rank matrix learning methods.
no code implementations • 24 Jun 2020 • Xin Luna Dong, Xiang He, Andrey Kan, Xi-An Li, Yan Liang, Jun Ma, Yifan Ethan Xu, Chenwei Zhang, Tong Zhao, Gabriel Blanco Saldana, Saurabh Deshpande, Alexandre Michetti Manduca, Jay Ren, Surender Pal Singh, Fan Xiao, Haw-Shiuan Chang, Giannis Karamanolakis, Yuning Mao, Yaqing Wang, Christos Faloutsos, Andrew McCallum, Jiawei Han
Can one build a knowledge graph (KG) for all products in the world?
no code implementations • 15 Jun 2020 • Yaqing Wang, Yifan Ethan Xu, Xi-An Li, Xin Luna Dong, Jing Gao
(1) We formalize the problem of validating the textual attribute values of products from a variety of categories as a natural language inference task in the few-shot learning setting, and propose a meta-learning latent variable model to jointly process the signals obtained from product profiles and textual attribute values.
no code implementations • 21 Apr 2020 • Alexander Hanbo Li, Yaqing Wang, Changyou Chen, Jing Gao
Effective inference for a generative adversarial model remains an important and challenging problem.
1 code implementation • 28 Dec 2019 • Yaqing Wang, Weifeng Yang, Fenglong Ma, Jin Xu, Bin Zhong, Qiang Deng, Jing Gao
In order to tackle this challenge, we propose a reinforced weakly-supervised fake news detection framework, i. e., WeFEND, which can leverage users' reports as weak supervision to enlarge the amount of training data for fake news detection.
4 code implementations • 10 Apr 2019 • Yaqing Wang, Quanming Yao, James Kwok, Lionel M. Ni
Machine learning has been highly successful in data-intensive applications but is often hampered when the data set is small.
no code implementations • 8 Mar 2019 • Yaqing Wang, James T. Kwok, Lionel M. Ni
However, existing CSC methods can only model noises from Gaussian distribution, which is restrictive and unrealistic.
no code implementations • 27 Sep 2018 • Hanbo Li, Yaqing Wang, Changyou Chen, Jing Gao
We propose a novel approach, Adversarial Inference by Matching priors and conditionals (AIM), which explicitly matches prior and conditional distributions in both data and code spaces, and puts a direct constraint on the dependency structure of the generative model.
1 code implementation • Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining 2018 • Yaqing Wang, Fenglong Ma, Zhiwei Jin, Ye Yuan, Guangxu Xun, Kishlay Jha, Lu Su, Jing Gao
One of the unique challenges for fake news detection on social media is how to identify fake news on newly emerged events.
no code implementations • ICML 2018 • Yaqing Wang, Quanming Yao, James T. Kwok, Lionel M. Ni
Convolutional sparse coding (CSC) has been popularly used for the learning of shift-invariant dictionaries in image and signal processing.
no code implementations • 21 Jun 2017 • Yaqing Wang, Quanming Yao, James T. Kwok, Lionel M. Ni
Convolutional sparse coding (CSC) improves sparse coding by learning a shift-invariant dictionary from the data.