no code implementations • 17 Dec 2024 • Jiale Liu, Yifan Zeng, Malte Højmark-Bertelsen, Marie Normann Gadeberg, Huazheng Wang, Qingyun Wu
Traditional enterprises face significant challenges in processing business documents, where tasks like extracting transport references from invoices remain largely manual despite their crucial role in logistics operations.
no code implementations • 3 Nov 2024 • Shaokun Zhang, Jieyu Zhang, Dujian Ding, Mirian Hipolito Garcia, Ankur Mallick, Daniel Madrigal, Menglin Xia, Victor Rühle, Qingyun Wu, Chi Wang
Recent advancements have enabled Large Language Models (LLMs) to function as agents that can perform actions using external tools.
1 code implementation • 31 May 2024 • Yifan Zeng, Ojas Tendolkar, Raymond Baartmans, Qingyun Wu, Lizhong Chen, Huazheng Wang
A common approach to sort the ranking list is by prompting LLMs for a pairwise or setwise comparison which often relies on sorting algorithms.
no code implementations • 29 May 2024 • Linxin Song, Jiale Liu, Jieyu Zhang, Shaokun Zhang, Ao Luo, Shijian Wang, Qingyun Wu, Chi Wang
Leveraging multiple large language model (LLM) agents has shown to be a promising approach for tackling complex tasks, while the effective design of multiple agents for a particular application remains an art.
no code implementations • 9 May 2024 • Zachary Coalson, Huazheng Wang, Qingyun Wu, Sanghyun Hong
In this paper, we study the robustness of "data-centric" approaches to finding neural network architectures (known as neural architecture search) to data distribution shifts.
1 code implementation • 19 Mar 2024 • Xudong Guo, Kaixuan Huang, Jiale Liu, Wenhui Fan, Natalia Vélez, Qingyun Wu, Huazheng Wang, Thomas L. Griffiths, Mengdi Wang
Large Language Models (LLMs) have emerged as integral tools for reasoning, planning, and decision-making, drawing upon their extensive world knowledge and proficiency in language-related tasks.
4 code implementations • 17 Mar 2024 • Yiran Wu, Tianwei Yue, Shaokun Zhang, Chi Wang, Qingyun Wu
In StateFlow, we distinguish between "process grounding" (via state and state transitions) and "sub-task solving" (through actions within a state), enhancing control and interpretability of the task-solving procedure.
1 code implementation • 2 Mar 2024 • Yifan Zeng, Yiran Wu, Xiao Zhang, Huazheng Wang, Qingyun Wu
In this paper, we propose AutoDefense, a multi-agent defense framework that filters harmful responses from LLMs.
1 code implementation • 17 Feb 2024 • Shaokun Zhang, Jieyu Zhang, Jiale Liu, Linxin Song, Chi Wang, Ranjay Krishna, Qingyun Wu
Researchers and practitioners have recently reframed powerful Large Language Models (LLMs) as agents, enabling them to automate complex tasks largely via the use of specialized functions.
no code implementations • 14 Feb 2024 • Negar Arabzadeh, Julia Kiseleva, Qingyun Wu, Chi Wang, Ahmed Awadallah, Victor Dibia, Adam Fourney, Charles Clarke
The rapid development in the field of Large Language Models (LLMs) has led to a surge in applications that facilitate collaboration among multiple agents to assist humans in their daily tasks.
no code implementations • 15 Nov 2023 • Xiaobo Xia, Jiale Liu, Shaokun Zhang, Qingyun Wu, Hongxin Wei, Tongliang Liu
Coreset selection is powerful in reducing computational costs and accelerating data processing for deep learning algorithms.
1 code implementation • 16 Oct 2023 • Shaokun Zhang, Xiaobo Xia, Zhaoqing Wang, Ling-Hao Chen, Jiale Liu, Qingyun Wu, Tongliang Liu
However, since the prompts need to be sampled from a large volume of annotated examples, finding the right prompt may result in high annotation costs.
1 code implementation • 8 Oct 2023 • Rishab Balasubramanian, Jiawei Li, Prasad Tadepalli, Huazheng Wang, Qingyun Wu, Haoyu Zhao
Contrary to prior understanding of multi-armed bandits, our work reveals a surprising fact that the attackability of a specific CMAB instance also depends on whether the bandit instance is known or unknown to the adversary.
2 code implementations • 16 Aug 2023 • Qingyun Wu, Gagan Bansal, Jieyu Zhang, Yiran Wu, Beibin Li, Erkang Zhu, Li Jiang, Xiaoyun Zhang, Shaokun Zhang, Jiale Liu, Ahmed Hassan Awadallah, Ryen W White, Doug Burger, Chi Wang
AutoGen is an open-source framework that allows developers to build LLM applications via multiple agents that can converse with each other to accomplish tasks.
2 code implementations • NeurIPS 2023 • Zeyu Zhang, Yi Su, Hui Yuan, Yiran Wu, Rishab Balasubramanian, Qingyun Wu, Huazheng Wang, Mengdi Wang
Building upon this, we leverage offline RL techniques for off-policy LTR and propose the Click Model-Agnostic Unified Off-policy Learning to Rank (CUOLR) method, which could be easily applied to a wide range of click models.
3 code implementations • 2 Jun 2023 • Yiran Wu, Feiran Jia, Shaokun Zhang, Hangyu Li, Erkang Zhu, Yue Wang, Yin Tat Lee, Richard Peng, Qingyun Wu, Chi Wang
Employing Large Language Models (LLMs) to address mathematical problems is an intriguing research endeavor, considering the abundance of math problems expressed in natural language across numerous science and engineering fields.
no code implementations • 28 May 2023 • Shaokun Zhang, Yiran Wu, Zhonghua Zheng, Qingyun Wu, Chi Wang
In this work, we propose a hyperparameter optimization method named \emph{HyperTime} to find hyperparameters robust to potential temporal distribution shifts in the unseen test data.
no code implementations • 29 Jun 2022 • Kaixuan Huang, Yu Wu, Xuezhou Zhang, Shenyinying Tu, Qingyun Wu, Mengdi Wang, Huazheng Wang
Online influence maximization aims to maximize the influence spread of a content in a social network with unknown network model by selecting a few seed nodes.
Model-based Reinforcement Learning
reinforcement-learning
+2
no code implementations • 11 Nov 2021 • Qingyun Wu, Chi Wang
In this work, we propose an Automated Machine Learning (AutoML) system to search for models not only with good prediction accuracy but also fair.
1 code implementation • 9 Jun 2021 • Qingyun Wu, Chi Wang, John Langford, Paul Mineiro, Marco Rossi
We propose the ChaCha (Champion-Challengers) algorithm for making an online choice of hyperparameters in online learning settings.
no code implementations • 14 Apr 2021 • Chuanhao Li, Qingyun Wu, Hongning Wang
However, all existing collaborative bandit learning solutions impose a stationary assumption about the environment, i. e., both user preferences and the dependency among users are assumed static over time.
no code implementations • ICLR 2021 • Chi Wang, Qingyun Wu, Silu Huang, Amin Saied
We study the problem of using low cost to search for hyperparameter configurations in a large search space with heterogeneous evaluation cost and model quality.
no code implementations • 5 Sep 2020 • Chuanhao Li, Qingyun Wu, Hongning Wang
Non-stationary bandits and online clustering of bandits lift the restrictive assumptions in contextual bandits and provide solutions to many important real-world scenarios.
no code implementations • 16 Jul 2020 • Kanak Mahadik, Qingyun Wu, Shuai Li, Amit Sabne
This algorithm lazily creates clusters in a distributed manner, and dramatically reduces the network data sharing requirement, achieving high scalability.
1 code implementation • 23 May 2020 • Shijun Li, Wenqiang Lei, Qingyun Wu, Xiangnan He, Peng Jiang, Tat-Seng Chua
In this work, we consider the conversational recommendation for cold-start users, where a system can both ask the attributes from and recommend items to a user interactively.
1 code implementation • 4 May 2020 • Qingyun Wu, Chi Wang, Silu Huang
To address this problem, we develop a new cost-frugal HPO solution.
no code implementations • 21 Feb 2020 • Wenqiang Lei, Xiangnan He, Yisong Miao, Qingyun Wu, Richang Hong, Min-Yen Kan, Tat-Seng Chua
Recommender systems are embracing conversational technologies to obtain user preferences dynamically, and to overcome inherent limitations of their static models.
2 code implementations • 12 Nov 2019 • Chi Wang, Qingyun Wu, Markus Weimer, Erkang Zhu
We study the problem of using low computational cost to automate the choices of learners and hyperparameters for an ad-hoc training dataset and error metric, by conducting trials of different configurations on the given training data.
no code implementations • 10 Jun 2019 • Huazheng Wang, Sonwoo Kim, Eric McCord-Snook, Qingyun Wu, Hongning Wang
We prove that the projected gradient is an unbiased estimation of the true gradient, and show that this lower-variance gradient estimation results in significant regret reduction.
1 code implementation • 9 Jun 2019 • Qingyun Wu, Zhige Li, Huazheng Wang, Wei Chen, Hongning Wang
We capitalize on an important property of the influence maximization problem named network assortativity, which is ignored by most existing works in online influence maximization.
1 code implementation • NeurIPS 2018 • Yi Qi, Qingyun Wu, Hongning Wang, Jie Tang, Maosong Sun
Implicit feedback, such as user clicks, although abundant in online information service systems, does not provide substantial evidence on users' evaluation of system's output.
1 code implementation • 23 May 2018 • Qingyun Wu, Naveen Iyer, Hongning Wang
Multi-armed bandit algorithms have become a reference solution for handling the explore/exploit dilemma in recommender systems, and many other important real-world problems, such as display advertisement.