Search Results for author: John C. S. Lui

Found 49 papers, 9 papers with code

Learning Best Paths in Quantum Networks

no code implementations14 Jun 2025 Xuchuang Wang, Maoli Liu, Xutong Liu, Zhuohua Li, Mohammad Hajiesmaili, John C. S. Lui, Don Towsley

We introduce two online learning algorithms, BeQuP-Link and BeQuP-Path, to identify the best path using link-level and path-level feedback, respectively.

Benchmarking

Federated In-Context Learning: Iterative Refinement for Improved Answer Quality

no code implementations9 Jun 2025 Ruhan Wang, Zhiyong Wang, Chengkai Huang, Rui Wang, Tong Yu, Lina Yao, John C. S. Lui, Dongruo Zhou

For question-answering (QA) tasks, in-context learning (ICL) enables language models to generate responses without modifying their parameters by leveraging examples provided in the input.

In-Context Learning Question Answering

Offline Clustering of Linear Bandits: Unlocking the Power of Clusters in Data-Limited Environments

no code implementations25 May 2025 Jingyuan Liu, Zeyu Zhang, Xuchuang Wang, Xutong Liu, John C. S. Lui, Mohammad Hajiesmaili, Carlee Joe-Wong

The key challenge in Off-ClusBand arises from data insufficiency for users: unlike the online case, in the offline case, we have a fixed, limited dataset to work from and thus must determine whether we have enough data to confidently cluster users together.

Clustering Multi-Armed Bandits

Online Clustering of Dueling Bandits

no code implementations4 Feb 2025 Zhiyong Wang, Jiahang Sun, Mingze Kong, Jize Xie, QinGhua Hu, John C. S. Lui, Zhongxiang Dai

In applications involving a large number of users, the performance of contextual MAB can be significantly improved by facilitating collaboration among multiple users.

Clustering Decision Making +4

Offline Learning for Combinatorial Multi-armed Bandits

no code implementations31 Jan 2025 Xutong Liu, Xiangxiang Dai, Jinhang Zuo, Siwei Wang, Carlee-Joe Wong, John C. S. Lui, Wei Chen

The combinatorial multi-armed bandit (CMAB) is a fundamental sequential decision-making framework, extensively studied over the past decade.

Decision Making Language Modeling +5

Multi-Agent Conversational Online Learning for Adaptive LLM Response Identification

1 code implementation3 Jan 2025 Xiangxiang Dai, Yuejin Xie, Maoli Liu, Xuchuang Wang, Zhuohua Li, Huanyu Wang, John C. S. Lui

The remarkable generative capability of large language models (LLMs) has sparked a growing interest in automatically generating responses for different applications.

Demystifying Online Clustering of Bandits: Enhanced Exploration Under Stochastic and Smoothed Adversarial Contexts

no code implementations1 Jan 2025 Zhuohua Li, Maoli Liu, Xiangxiang Dai, John C. S. Lui

context generation used in previous studies, thus enhancing the performance of existing algorithms for online clustering of bandits.

Clustering Online Clustering +1

Unifying KV Cache Compression for Large Language Models with LeanKV

no code implementations4 Dec 2024 Yanqi Zhang, Yuwei Hu, Runyuan Zhao, John C. S. Lui, Haibo Chen

Large language models (LLMs) exhibit exceptional performance but incur significant serving costs due to their substantial memory requirements, with the key-value (KV) cache being a primary bottleneck.

Quantization

Combinatorial Logistic Bandits

1 code implementation22 Oct 2024 Xutong Liu, Xiangxiang Dai, Xuchuang Wang, Mohammad Hajiesmaili, John C. S. Lui

Under the 1-norm triggering probability modulated (TPM) smoothness condition, CLogUCB achieves a regret bound of $\tilde{O}(d\sqrt{\kappa KT})$, where $\tilde{O}$ ignores logarithmic factors, $d$ is the dimension of the feature vector, $\kappa$ represents the nonlinearity of the logistic model, and $K$ is the maximum number of base arms a super arm can trigger.

Computational Efficiency Learning-To-Rank

Stochastic Bandits Robust to Adversarial Attacks

no code implementations16 Aug 2024 Xuchuang Wang, Jinhang Zuo, Xutong Liu, John C. S. Lui, Mohammad Hajiesmaili

For both cases, we devise two types of algorithms with regret bounds having additive or multiplicative $C$ dependence terms.

Model-based RL as a Minimalist Approach to Horizon-Free and Second-Order Bounds

no code implementations16 Aug 2024 Zhiyong Wang, Dongruo Zhou, John C. S. Lui, Wen Sun

Learning a transition model via Maximum Likelihood Estimation (MLE) followed by planning inside the learned model is perhaps the most standard and simplest Model-based Reinforcement Learning (RL) framework.

Model-based Reinforcement Learning Offline RL +1

AxiomVision: Accuracy-Guaranteed Adaptive Visual Model Selection for Perspective-Aware Video Analytics

2 code implementations29 Jul 2024 Xiangxiang Dai, Zeyu Zhang, Peng Yang, Yuedong Xu, Xutong Liu, John C. S. Lui

The rapid evolution of multimedia and computer vision technologies requires adaptive visual model deployment strategies to effectively handle diverse tasks and varying environments.

Edge-computing Model Selection +2

Merit-based Fair Combinatorial Semi-Bandit with Unrestricted Feedback Delays

2 code implementations22 Jul 2024 Ziqun Chen, Kechao Cai, Zhuoyue Chen, Jinbei Zhang, John C. S. Lui

We study the stochastic combinatorial semi-bandit problem with unrestricted feedback delays under merit-based fairness constraints.

Fairness

Combinatorial Multivariant Multi-Armed Bandits with Applications to Episodic Reinforcement Learning and Beyond

no code implementations3 Jun 2024 Xutong Liu, Siwei Wang, Jinhang Zuo, Han Zhong, Xuchuang Wang, Zhiyong Wang, Shuai Li, Mohammad Hajiesmaili, John C. S. Lui, Wei Chen

We introduce a novel framework of combinatorial multi-armed bandits (CMAB) with multivariant and probabilistically triggering arms (CMAB-MT), where the outcome of each arm is a $d$-dimensional multivariant random variable and the feedback follows a general arm triggering process.

Multi-Armed Bandits Reinforcement Learning (RL)

Cost-Effective Online Multi-LLM Selection with Versatile Reward Models

no code implementations26 May 2024 Xiangxiang Dai, Jin Li, Xutong Liu, Anqi Yu, John C. S. Lui

The NP-hard integer linear programming problem for selecting multiple LLMs with trade-off dilemmas is addressed by: i) decomposing the integer problem into a relaxed form by the local server, ii) utilizing a discretization rounding scheme that provides optimal LLM combinations by the scheduling cloud, and iii) continual online updates based on feedback.

Scheduling

FedConPE: Efficient Federated Conversational Bandits with Heterogeneous Clients

no code implementations5 May 2024 Zhuohua Li, Maoli Liu, John C. S. Lui

These systems interactively present queries associated with "key terms" to users and leverage user feedback to estimate user preferences more efficiently.

Management Recommendation Systems

Variance-Dependent Regret Bounds for Non-stationary Linear Bandits

no code implementations15 Mar 2024 Zhiyong Wang, Jize Xie, Yi Chen, John C. S. Lui, Dongruo Zhou

We investigate the non-stationary stochastic linear bandit problem where the reward distribution evolves each round.

Federated Contextual Cascading Bandits with Asynchronous Communication and Heterogeneous Users

no code implementations26 Feb 2024 Hantao Yang, Xutong Liu, Zhiyong Wang, Hong Xie, John C. S. Lui, Defu Lian, Enhong Chen

We study the problem of federated contextual combinatorial cascading bandits, where $|\mathcal{U}|$ agents collaborate under the coordination of a central server to provide tailored recommendations to the $|\mathcal{U}|$ corresponding users.

Fed-CVLC: Compressing Federated Learning Communications with Variable-Length Codes

no code implementations6 Feb 2024 Xiaoxin Su, Yipeng Zhou, Laizhong Cui, John C. S. Lui, Jiangchuan Liu

In Federated Learning (FL) paradigm, a parameter server (PS) concurrently communicates with distributed participating clients for model collection, update aggregation, and model distribution over multiple rounds, without touching private data owned by individual clients.

Federated Learning Model Compression +1

Adversarial Attacks on Cooperative Multi-agent Bandits

no code implementations3 Nov 2023 Jinhang Zuo, Zhiyao Zhang, Xuchuang Wang, Cheng Chen, Shuai Li, John C. S. Lui, Mohammad Hajiesmaili, Adam Wierman

Cooperative multi-agent multi-armed bandits (CMA2B) consider the collaborative efforts of multiple agents in a shared multi-armed bandit game.

Multi-Armed Bandits

A Survey of Federated Evaluation in Federated Learning

no code implementations14 May 2023 Behnaz Soltani, Yipeng Zhou, Venus Haghighi, John C. S. Lui

In traditional machine learning, it is trivial to conduct model evaluation since all data samples are managed centrally by a server.

Federated Learning Survey

Contextual Combinatorial Bandits with Probabilistically Triggered Arms

no code implementations30 Mar 2023 Xutong Liu, Jinhang Zuo, Siwei Wang, John C. S. Lui, Mohammad Hajiesmaili, Adam Wierman, Wei Chen

We study contextual combinatorial bandits with probabilistically triggered arms (C$^2$MAB-T) under a variety of smoothness conditions that capture a wide range of applications, such as contextual cascading bandits and contextual influence maximization bandits.

Efficient Explorative Key-term Selection Strategies for Conversational Contextual Bandits

1 code implementation1 Mar 2023 Zhiyong Wang, Xutong Liu, Shuai Li, John C. S. Lui

To tackle these issues, we first propose ``ConLinUCB", a general framework for conversational bandits with better information incorporation, combining arm-level and key-term-level feedback to estimate user preference in one step at each time.

Computational Efficiency Multi-Armed Bandits +1

On-Demand Communication for Asynchronous Multi-Agent Bandits

no code implementations15 Feb 2023 Yu-Zhen Janice Chen, Lin Yang, Xuchuang Wang, Xutong Liu, Mohammad Hajiesmaili, John C. S. Lui, Don Towsley

We propose ODC, an on-demand communication protocol that tailors the communication of each pair of agents based on their empirical pull times.

Interactive Log Parsing via Light-weight User Feedback

no code implementations28 Jan 2023 Liming Wang, Hong Xie, Ye Li, Jian Tan, John C. S. Lui

Template mining is one of the foundational tasks to support log analysis, which supports the diagnosis and troubleshooting of large scale Web applications.

Log Parsing

Federated Online Clustering of Bandits

1 code implementation31 Aug 2022 Xutong Liu, Haoru Zhao, Tong Yu, Shuai Li, John C. S. Lui

Contextual multi-armed bandit (MAB) is an important sequential decision-making problem in recommendation systems.

Clustering Decision Making +3

Batch-Size Independent Regret Bounds for Combinatorial Semi-Bandits with Probabilistically Triggered Arms or Independent Arms

no code implementations31 Aug 2022 Xutong Liu, Jinhang Zuo, Siwei Wang, Carlee Joe-Wong, John C. S. Lui, Wei Chen

Under this new condition, we propose a BCUCB-T algorithm with variance-aware confidence intervals and conduct regret analysis which reduces the $O(K)$ factor to $O(\log K)$ or $O(\log^2 K)$ in the regret bound, significantly improving the regret bounds for the above applications.

Multiple-Play Stochastic Bandits with Shareable Finite-Capacity Arms

no code implementations17 Jun 2022 Xuchuang Wang, Hong Xie, John C. S. Lui

When the "per-load" reward follows a Gaussian distribution, we prove a sample complexity lower bound of learning the capacity from load-dependent rewards and also a regret lower bound of this new MP-MAB problem.

Multi-Armed Bandits

LPC-AD: Fast and Accurate Multivariate Time Series Anomaly Detection via Latent Predictive Coding

no code implementations5 May 2022 Zhi Qi, Hong Xie, Ye Li, Jian Tan, Feifei Li, John C. S. Lui

LPC-AD is motivated by the ever-increasing needs for fast and accurate MTS anomaly detection methods to support fast troubleshooting in cloud computing, micro-service systems, etc.

Anomaly Detection Cloud Computing +2

Multi-Player Multi-Armed Bandits with Finite Shareable Resources Arms: Learning Algorithms & Applications

no code implementations28 Apr 2022 Xuchuang Wang, Hong Xie, John C. S. Lui

The reward from a shareable arm is equal to the "per-load" reward multiplied by the minimum between the number of players pulling the arm and the arm's maximal shareable resources.

Edge-computing Multi-Armed Bandits

Decentralized Task Offloading in Edge Computing: A Multi-User Multi-Armed Bandit Approach

no code implementations22 Dec 2021 Xiong Wang, Jiancheng Ye, John C. S. Lui

Mobile edge computing facilitates users to offload computation tasks to edge servers for meeting their stringent delay requirements.

Edge-computing

Decentralized Stochastic Proximal Gradient Descent with Variance Reduction over Time-varying Networks

no code implementations20 Dec 2021 Xuanjie Li, Yuedong Xu, Jessie Hui Wang, Xin Wang, John C. S. Lui

By transforming our decentralized algorithm into a centralized inexact proximal gradient algorithm with variance reduction, and controlling the bounds of error sequences, we prove that DPSVRG converges at the rate of $O(1/T)$ for general convex objectives plus a non-smooth term with $T$ as the number of iterations, while DSPG converges at the rate $O(\frac{1}{\sqrt{T}})$.

Cooperative Stochastic Bandits with Asynchronous Agents and Constrained Feedback

no code implementations NeurIPS 2021 Lin Yang, Yu-Zhen Janice Chen, Stephen Pasteris, Mohammad Hajiesmaili, John C. S. Lui, Don Towsley

This paper studies a cooperative multi-armed bandit problem with $M$ agents cooperating together to solve the same instance of a $K$-armed stochastic bandit problem with the goal of maximizing the cumulative reward of agents.

Decision Making

Multi-layered Network Exploration via Random Walks: From Offline Optimization to Online Learning

no code implementations9 Jun 2021 Xutong Liu, Jinhang Zuo, Xiaowei Chen, Wei Chen, John C. S. Lui

For the online learning setting, neither the network structure nor the node weights are known initially.

Adversarial Bandits with Corruptions: Regret Lower Bound and No-regret Algorithm

no code implementations NeurIPS 2020 Lin Yang, Mohammad Hajiesmaili, Mohammad Sadegh Talebi, John C. S. Lui, Wing Shing Wong

We characterize the regret of ExpRb as a function of the corruption budget and show that for the case of a known corruption budget, the regret of ExpRb is tight.

Restless-UCB, an Efficient and Low-complexity Algorithm for Online Restless Bandits

no code implementations NeurIPS 2020 Siwei Wang, Longbo Huang, John C. S. Lui

Compared to existing algorithms, our result eliminates the exponential factor (in $M, N$) in the regret upper bound, due to a novel exploitation of the sparsity in transitions in general restless bandit problems.

Online Competitive Influence Maximization

no code implementations24 Jun 2020 Jinhang Zuo, Xutong Liu, Carlee Joe-Wong, John C. S. Lui, Wei Chen

In this paper, we introduce a new Online Competitive Influence Maximization (OCIM) problem, where two competing items (e. g., products, news stories) propagate in the same network and influence probabilities on edges are unknown.

Walking with Perception: Efficient Random Walk Sampling via Common Neighbor Awareness

1 code implementation ‏‏‎ ‎ 2020 Yongkun Li, Zhiyong Wu, Shuai Lin, Hong Xie, Min Lv, Yinlong Xu, John C. S. Lui

Random walk is widely applied to sample large-scale graphs due to its simplicity of implementation and solid theoretical foundations of bias analysis.

Computational Efficiency

Combining Offline Causal Inference and Online Bandit Learning for Data Driven Decision

no code implementations16 Jan 2020 Li Ye, Yishi Lin, Hong Xie, John C. S. Lui

A typical alternative is offline causal inference, which analyzes logged data alone to make decisions.

Causal Inference

Conversational Contextual Bandit: Algorithm and Application

no code implementations4 Jun 2019 Xiaoying Zhang, Hong Xie, Hang Li, John C. S. Lui

Here, a key-term can relate to a subset of arms, for example, a category of articles in news recommendation.

Articles News Recommendation +1

Community Exploration: From Offline Optimization to Online Learning

no code implementations NeurIPS 2018 Xiaowei Chen, Weiran Huang, Wei Chen, John C. S. Lui

We introduce the community exploration problem that has many real-world applications such as online advertising.

Graphlet Count Estimation via Convolutional Neural Networks

1 code implementation7 Oct 2018 Xutong Liu, Yu-Zhen Janice Chen, John C. S. Lui, Konstantin Avrachenkov

The number of each graphlet, called graphlet count, is a signature which characterizes the local network structure of a given graph.

Beyond the Click-Through Rate: Web Link Selection with Multi-level Feedback

no code implementations4 May 2018 Kun Chen, Kechao Cai, Longbo Huang, John C. S. Lui

The web link selection problem is to select a small subset of web links from a large web link pool, and to place the selected links on a web page that can only accommodate a limited number of links, e. g., advertisements, recommendations, or news feeds.

Multi-level Feedback Web Links Selection Problem: Learning and Optimization

no code implementations8 Sep 2017 Kechao Cai, Kun Chen, Longbo Huang, John C. S. Lui

To our best knowledge, we are the first to model the links selection problem as a constrained multi-armed bandit problem and design an effective links selection algorithm by learning the links' multi-level structure with provable \emph{sub-linear} regret and violation bounds.

Monet: A User-oriented Behavior-based Malware Variants Detection System for Android

no code implementations10 Dec 2016 Mingshen Sun, Xiaolei Li, John C. S. Lui, Richard T. B. Ma, Zhenkai Liang

We present the design and implementation of MONET, which has a client and a backend server module.

Cryptography and Security

DroidAnalytics: A Signature Based Analytic System to Collect, Extract, Analyze and Associate Android Malware

no code implementations28 Feb 2013 Min Zheng, Mingshen Sun, John C. S. Lui

In this paper, we present the design and implementation of DroidAnalytics, a signature based analytic system to automatically collect, manage, analyze and extract android malware.

Cryptography and Security

Online Robust Subspace Tracking from Partial Information

1 code implementation18 Sep 2011 Jun He, Laura Balzano, John C. S. Lui

This paper presents GRASTA (Grassmannian Robust Adaptive Subspace Tracking Algorithm), an efficient and robust online algorithm for tracking subspaces from highly incomplete information.

Matrix Completion

Cannot find the paper you are looking for? You can Submit a new open access paper.