no code implementations • 8 Aug 2024 • Qirui Jiao, Daoyuan Chen, Yilun Huang, Yaliang Li, Ying Shen
Besides, we investigate alternative methods for generating image difference data through "object removal" and conduct a thorough evaluation to confirm the dataset's diversity, quality, and robustness, presenting several insights on the synthesis of such a contrastive dataset.
Ranked #53 on Visual Question Answering on MM-Vet
1 code implementation • 16 Jul 2024 • Daoyuan Chen, Haibin Wang, Yilun Huang, Ce Ge, Yaliang Li, Bolin Ding, Jingren Zhou
The emergence of large-scale multi-modal generative models has drastically advanced artificial intelligence, introducing unprecedented levels of performance and functionality.
1 code implementation • 11 Jul 2024 • Zhen Qin, Daoyuan Chen, WenHao Zhang, Liuyi Yao, Yilun Huang, Bolin Ding, Yaliang Li, Shuiguang Deng
As LLMs and MLLMs rely on vast amounts of model parameters and data to achieve emergent capabilities, the importance of data is receiving increasingly widespread attention and recognition.
no code implementations • 23 May 2024 • Ce Ge, Zhijian Ma, Daoyuan Chen, Yaliang Li, Bolin Ding
Large language models exhibit exceptional generalization capabilities, primarily attributed to the utilization of diversely sourced data.
1 code implementation • 3 Apr 2024 • Zhe Xu, Daoyuan Chen, Jiayi Kuang, Zihao Yi, Yaliang Li, Ying Shen
Emotional Support Conversation (ESC) systems are pivotal in providing empathetic interactions, aiding users through negative emotional states by understanding and addressing their unique experiences.
1 code implementation • 17 Mar 2024 • Mengsha Liu, Daoyuan Chen, Yaliang Li, Guian Fang, Ying Shen
Data visualization serves as a critical means for presenting data and mining its valuable insights.
1 code implementation • 21 Feb 2024 • Dawei Gao, Zitao Li, Xuchen Pan, Weirui Kuang, Zhijian Ma, Bingchen Qian, Fei Wei, WenHao Zhang, Yuexiang Xie, Daoyuan Chen, Liuyi Yao, Hongyi Peng, Zeyu Zhang, Lin Zhu, Chen Cheng, Hongzhu Shi, Yaliang Li, Bolin Ding, Jingren Zhou
With the rapid advancement of Large Language Models (LLMs), significant progress has been made in multi-agent applications.
no code implementations • 18 Feb 2024 • Jiamu Bai, Daoyuan Chen, Bingchen Qian, Liuyi Yao, Yaliang Li
Federated Learning (FL) has recently been applied to the parameter-efficient fine-tuning of Large Language Models (LLMs).
1 code implementation • 8 Feb 2024 • Zhenqing Ling, Daoyuan Chen, Liuyi Yao, Yaliang Li, Ying Shen
The confluence of Federated Learning (FL) and Large Language Models (LLMs) is ushering in a new era in privacy-preserving natural language processing.
no code implementations • 31 Jan 2024 • Qirui Jiao, Daoyuan Chen, Yilun Huang, Yaliang Li, Ying Shen
Despite the impressive capabilities of Multimodal Large Language Models (MLLMs) in integrating text and image modalities, challenges remain in accurately interpreting detailed visual elements.
Ranked #72 on Visual Question Answering on MM-Vet
2 code implementations • 11 Dec 2023 • Zhen Qin, Daoyuan Chen, Bingchen Qian, Bolin Ding, Yaliang Li, Shuiguang Deng
Pre-trained large language models (LLMs) need fine-tuning to improve their responsiveness to natural language instructions.
2 code implementations • 5 Sep 2023 • Daoyuan Chen, Yilun Huang, Zhijian Ma, Hesen Chen, Xuchen Pan, Ce Ge, Dawei Gao, Yuexiang Xie, Zhaoyang Liu, Jinyang Gao, Yaliang Li, Bolin Ding, Jingren Zhou
A data recipe is a mixture of data from different sources for training LLMs, which plays a vital role in LLMs' performance.
1 code implementation • 1 Sep 2023 • Weirui Kuang, Bingchen Qian, Zitao Li, Daoyuan Chen, Dawei Gao, Xuchen Pan, Yuexiang Xie, Yaliang Li, Bolin Ding, Jingren Zhou
When several entities have similar interested tasks, but their data cannot be shared because of privacy concerns regulations, federated learning (FL) is a mainstream solution to leverage the data of different entities.
2 code implementations • 4 May 2023 • Daoyuan Chen, Liuyi Yao, Dawei Gao, Bolin Ding, Yaliang Li
To overcome these challenges, we propose a novel approach named pFedGate for efficient personalized FL by adaptively and efficiently learning sparse local models.
no code implementations • 23 Mar 2023 • Daoyuan Chen, Dawei Gao, Yuexiang Xie, Xuchen Pan, Zitao Li, Yaliang Li, Bolin Ding, Jingren Zhou
Federated Learning (FL) aims to train high-quality models in collaboration with distributed clients while not uploading their local data, which attracts increasing attention in both academia and industry.
1 code implementation • 3 Feb 2023 • Zeyu Qin, Liuyi Yao, Daoyuan Chen, Yaliang Li, Bolin Ding, Minhao Cheng
We conduct the first study of backdoor attacks in the pFL framework, testing 4 widely used backdoor attacks against 6 pFL methods on benchmark datasets FEMNIST and CIFAR-10, a total of 600 experiments.
1 code implementation • 8 Jun 2022 • Daoyuan Chen, Dawei Gao, Weirui Kuang, Yaliang Li, Bolin Ding
Personalized Federated Learning (pFL), which utilizes and deploys distinct local models, has gained increasing attention in recent years due to its success in handling the statistical heterogeneity of FL clients.
1 code implementation • 7 Jun 2022 • Liuyi Yao, Dawei Gao, Zhen Wang, Yuexiang Xie, Weirui Kuang, Daoyuan Chen, Haohui Wang, Chenhe Dong, Bolin Ding, Yaliang Li
To investigate the heterogeneity in federated learning in real-world scenarios, we generalize the classic federated learning to federated hetero-task learning, which emphasizes the inconsistency across the participants in federated learning in terms of both data distribution and learning tasks.
1 code implementation • 11 Apr 2022 • Yuexiang Xie, Zhen Wang, Dawei Gao, Daoyuan Chen, Liuyi Yao, Weirui Kuang, Yaliang Li, Bolin Ding, Jingren Zhou
Although remarkable progress has been made by existing federated learning (FL) platforms to provide infrastructures for development, these platforms may not well tackle the challenges brought by various types of heterogeneity, including the heterogeneity in participants' local data, resources, behaviors and learning goals.
no code implementations • 29 Sep 2021 • Daoyuan Chen, Wuchao Li, Yaliang Li, Bolin Ding, Kai Zeng, Defu Lian, Jingren Zhou
We theoretically analyze prediction error bounds that link $\epsilon$ with data characteristics for an illustrative learned index method.
no code implementations • 4 Jan 2021 • Yaliang Li, Daoyuan Chen, Bolin Ding, Kai Zeng, Jingren Zhou
In this paper, we propose a formal machine learning based framework to quantify the index learning objective, and study two general and pluggable techniques to enhance the learning efficiency and learning effectiveness for learned indexes.
no code implementations • ACL 2020 • Daoyuan Chen, Yaliang Li, Kai Lei, Ying Shen
Distant supervision based methods for entity and relation extraction have received increasing popularity due to the fact that these methods require light human annotation efforts.
1 code implementation • 13 Jan 2020 • Daoyuan Chen, Yaliang Li, Minghui Qiu, Zhen Wang, Bofang Li, Bolin Ding, Hongbo Deng, Jun Huang, Wei. Lin, Jingren Zhou
Motivated by the necessity and benefits of task-oriented BERT compression, we propose a novel compression method, AdaBERT, that leverages differentiable Neural Architecture Search to automatically compress BERT into task-adaptive small models for specific tasks.
1 code implementation • 22 Nov 2019 • Yang Deng, Wai Lam, Yuexiang Xie, Daoyuan Chen, Yaliang Li, Min Yang, Ying Shen
Community question answering (CQA) gains increasing popularity in both academy and industry recently.
no code implementations • COLING 2018 • Kai Lei, Daoyuan Chen, Yaliang Li, Nan Du, Min Yang, Wei Fan, Ying Shen
Distantly supervised relation extraction greatly reduces human efforts in extracting relational facts from unstructured texts.