Search Results for author: Zhi Zhou

Found 50 papers, 19 papers with code

Injecting Adrenaline into LLM Serving: Boosting Resource Utilization and Throughput via Attention Disaggregation

1 code implementation26 Mar 2025 Yunkai Liang, Zhangyu Chen, Pengfei Zuo, Zhi Zhou, Xu Chen, Zhou Yu

The memory-bound nature of decoding-phase attention computation inherently enables an effective offloading strategy, yielding two complementary advantages: 1) improved memory capacity and bandwidth utilization in prefill instances, and 2) increased decoding batch sizes that enhance compute utilization in decoding instances, collectively boosting overall system performance.

Large Language Model Scheduling

Micro Text Classification Based on Balanced Positive-Unlabeled Learning

1 code implementation17 Mar 2025 Lin-Han Jia, Lan-Zhe Guo, Zhi Zhou, Si-Ye Han, Zi-Wen Li, Yu-Feng Li

In real-world text classification tasks, negative texts often contain a minimal proportion of negative content, which is especially problematic in areas like text quality control, legal risk screening, and sensitive information interception.

text-classification Text Classification

LawGPT: Knowledge-Guided Data Generation and Its Application to Legal LLM

1 code implementation10 Feb 2025 Zhi Zhou, Kun-Yang Yu, Shi-Yu Tian, Jiang-Xin Shi, Xiao-Wen Yang, Pengxiao Song, Yi-Xuan Jin, Lan-Zhe Guo, Yu-Feng Li

To address these limitations, we study data generation for legal reasoning to improve the legal reasoning performance of open-source LLMs with the help of proprietary LLMs.

Legal Reasoning

Step Back to Leap Forward: Self-Backtracking for Boosting Reasoning of Language Models

1 code implementation6 Feb 2025 Xiao-Wen Yang, Xuan-Yi Zhu, Wen-Da Wei, Ding-Chu Zhang, Jie-Jing Shao, Zhi Zhou, Lan-Zhe Guo, Yu-Feng Li

The integration of slow-thinking mechanisms into large language models (LLMs) offers a promising way toward achieving Level 2 AGI Reasoners, as exemplified by systems like OpenAI's o1.

Online Location Planning for AI-Defined Vehicles: Optimizing Joint Tasks of Order Serving and Spatio-Temporal Heterogeneous Model Fine-Tuning

no code implementations6 Feb 2025 Bokeng Zheng, Bo Rao, Tianxiang Zhu, Chee Wei Tan, Jingpu Duan, Zhi Zhou, Xu Chen, Xiaoxi Zhang

Advances in artificial intelligence (AI) including foundation models (FMs), are increasingly transforming human society, with smart city driving the evolution of urban living. Meanwhile, vehicle crowdsensing (VCS) has emerged as a key enabler, leveraging vehicles' mobility and sensor-equipped capabilities.

Multi-agent Reinforcement Learning point of interests

Bridging Internal Probability and Self-Consistency for Effective and Efficient LLM Reasoning

no code implementations1 Feb 2025 Zhi Zhou, Tan Yuhao, Zenan Li, Yuan YAO, Lan-Zhe Guo, Xiaoxing Ma, Yu-Feng Li

In this paper, we present the first theoretical error decomposition analysis of these techniques, breaking down their error into estimation error and model error.

Contrast-Aware Calibration for Fine-Tuned CLIP: Leveraging Image-Text Alignment

no code implementations31 Jan 2025 Song-Lin Lv, Yu-Yang Chen, Zhi Zhou, Yu-Feng Li, Lan-Zhe Guo

Vision-language models (VLMs), such as CLIP, have demonstrated exceptional generalization capabilities and can quickly adapt to downstream tasks through prompt fine-tuning.

TabFSBench: Tabular Benchmark for Feature Shifts in Open Environment

1 code implementation31 Jan 2025 Zi-Jian Cheng, Zi-Yi Jia, Zhi Zhou, Lan-Zhe Guo, Yu-Feng Li

TabFSBench evaluates impacts of four distinct feature-shift scenarios on four tabular model categories across various datasets and assesses the performance of large language models (LLMs) and tabular LLMs in the tabular benchmark for the first time.

Vision-Language Model Selection and Reuse for Downstream Adaptation

no code implementations30 Jan 2025 Hao-Zhe Tan, Zhi Zhou, Yu-Feng Li, Lan-Zhe Guo

The proposal is highly computationally efficient and growable since the model labeling process is completed target task independent and the ability could grow with the number of candidate VLMs.

Language Modeling Language Modelling +1

CGI: Identifying Conditional Generative Models with Example Images

no code implementations23 Jan 2025 Zhi Zhou, Hao-Zhe Tan, Peng-Xiao Song, Lan-Zhe Guo

In this paper, we propose Conditional Generative Model Identification (CGI), which aims to provide an effective way to identify the most suitable model using user-provided example images rather than requiring users to manually review a large number of models with example images.

Text Matching

BMIP: Bi-directional Modality Interaction Prompt Learning for VLM

no code implementations14 Jan 2025 Song-Lin Lv, Yu-Yang Chen, Zhi Zhou, Ming Yang, Lan-Zhe Guo

Vision-language models (VLMs) have exhibited remarkable generalization capabilities, and prompt learning for VLMs has attracted great attention for the ability to adapt pre-trained VLMs to specific downstream tasks.

Domain Generalization Prompt Learning

You Only Submit One Image to Find the Most Suitable Generative Model

no code implementations16 Dec 2024 Zhi Zhou, Lan-Zhe Guo, Peng-Xiao Song, Yu-Feng Li

In this paper, we propose a novel setting called Generative Model Identification (GMI), which aims to enable the user to identify the most appropriate generative model(s) for the user's requirements from a large number of candidate models efficiently.

Image Generation Text Matching

Fully Test-time Adaptation for Tabular Data

no code implementations14 Dec 2024 Zhi Zhou, Kun-Yang Yu, Lan-Zhe Guo, Yu-Feng Li

To this end, we propose the Fully Test-time Adaptation for Tabular data, namely FTAT, which enables FTTA methods to robustly optimize the label distribution of predictions, adapt to shifted covariate distributions, and suit a variety of tasks and models effectively.

Data Augmentation Test-time Adaptation

Neuro-Symbolic Data Generation for Math Reasoning

no code implementations6 Dec 2024 Zenan Li, Zhi Zhou, Yuan YAO, Yu-Feng Li, Chun Cao, Fan Yang, Xian Zhang, Xiaoxing Ma

A critical question about Large Language Models (LLMs) is whether their apparent deficiency in mathematical reasoning is inherent, or merely a result of insufficient exposure to high-quality mathematical data.

Diversity Math +1

Enabling Small Models for Zero-Shot Selection and Reuse through Model Label Learning

no code implementations21 Aug 2024 Jia Zhang, Zhi Zhou, Lan-Zhe Guo, Yu-Feng Li

In this paper, we attempt to demonstrate that by constructing a model hub and aligning models with their functionalities using model labels, new tasks can be solved in a zero-shot manner by effectively selecting and reusing models in the hub.

Image Classification Zero-Shot Learning

Point Source Identification Using Singularity Enriched Neural Networks

1 code implementation17 Aug 2024 Tianhao Hu, Bangti Jin, Zhi Zhou

The inverse problem of recovering point sources represents an important class of applied inverse problems.

LawGPT: A Chinese Legal Knowledge-Enhanced Large Language Model

2 code implementations7 Jun 2024 Zhi Zhou, Jiang-Xin Shi, Peng-Xiao Song, Xiao-Wen Yang, Yi-Xuan Jin, Lan-Zhe Guo, Yu-Feng Li

Large language models (LLMs), including both proprietary and open-source models, have showcased remarkable capabilities in addressing a wide range of downstream tasks.

Language Modeling Language Modelling +1

Robustness Assessment of Mathematical Reasoning in the Presence of Missing and Contradictory Conditions

no code implementations7 Jun 2024 Shi-Yu Tian, Zhi Zhou, Lin-Han Jia, Lan-Zhe Guo, Yu-Feng Li

To further study this problem, we develop a benchmark called Problems with Missing and Contradictory conditions (PMC) and introduce two novel metrics to evaluate the performance of few-shot prompting methods in these scenarios.

Hallucination Mathematical Reasoning

DeCoOp: Robust Prompt Tuning with Out-of-Distribution Detection

1 code implementation1 Jun 2024 Zhi Zhou, Ming Yang, Jiang-Xin Shi, Lan-Zhe Guo, Yu-Feng Li

In this paper, we explore a problem setting called Open-world Prompt Tuning (OPT), which involves tuning prompts on base classes and evaluating on a combination of base and new classes.

Out-of-Distribution Detection

Online Resource Allocation for Edge Intelligence with Colocated Model Retraining and Inference

1 code implementation25 May 2024 Huaiguang Cai, Zhi Zhou, Qianyi Huang

This highlights the advantages and applicable scenarios of colocating model retraining and inference.

Opara: Exploiting Operator Parallelism for Expediting DNN Inference on GPUs

1 code implementation16 Dec 2023 Aodong Chen, Fei Xu, Li Han, Yuan Dong, Li Chen, Zhi Zhou, Fangming Liu

GPUs have become the \emph{defacto} hardware devices for accelerating Deep Neural Network (DNN) inference workloads.

Scheduling

DYNAMITE: Dynamic Interplay of Mini-Batch Size and Aggregation Frequency for Federated Learning with Static and Streaming Dataset

no code implementations20 Oct 2023 Weijie Liu, Xiaoxi Zhang, Jingpu Duan, Carlee Joe-Wong, Zhi Zhou, Xu Chen

Federated Learning (FL) is a distributed learning paradigm that can coordinate heterogeneous edge devices to perform model training without sharing private data.

Federated Learning Navigate

Long-Tail Learning with Foundation Model: Heavy Fine-Tuning Hurts

1 code implementation18 Sep 2023 Jiang-Xin Shi, Tong Wei, Zhi Zhou, Jie-Jing Shao, Xin-Yan Han, Yu-Feng Li

The fine-tuning paradigm in addressing long-tail learning tasks has sparked significant interest since the emergence of foundation models.

 Ranked #1 on Long-tail Learning on CIFAR-100-LT (ρ=100) (using extra training data)

Fine-Grained Image Classification Long-tail learning with class descriptors

Solving Elliptic Optimal Control Problems via Neural Networks and Optimality System

no code implementations23 Aug 2023 Yongcheng Dai, Bangti Jin, Ramesh Sau, Zhi Zhou

In this work, we investigate a neural network based solver for optimal control problems (without / with box constraint) for linear and semilinear second-order elliptic problems.

Serving Graph Neural Networks With Distributed Fog Servers For Smart IoT Services

no code implementations4 Jul 2023 Liekang Zeng, Xu Chen, Peng Huang, Ke Luo, Xiaoxi Zhang, Zhi Zhou

Graph Neural Networks (GNNs) have gained growing interest in miscellaneous applications owing to their outstanding ability in extracting latent representation on graph structures.

Miscellaneous

Towards Carbon-Neutral Edge Computing: Greening Edge AI by Harnessing Spot and Future Carbon Markets

no code implementations22 Apr 2023 Huirong Ma, Zhi Zhou, Xiaoxi Zhang, Xu Chen

Provisioning dynamic machine learning (ML) inference as a service for artificial intelligence (AI) applications of edge devices faces many challenges, including the trade-off among accuracy loss, carbon emission, and unknown future costs.

Edge-computing

Electrical Impedance Tomography with Deep Calderón Method

1 code implementation18 Apr 2023 Siyu Cen, Bangti Jin, Kwancheol Shin, Zhi Zhou

Electrical impedance tomography (EIT) is a noninvasive medical imaging modality utilizing the current-density/voltage data measured on the surface of the subject.

Conductivity Imaging from Internal Measurements with Mixed Least-Squares Deep Neural Networks

1 code implementation29 Mar 2023 Bangti Jin, Xiyao Li, Qimeng Quan, Zhi Zhou

In this work we develop a novel approach using deep neural networks to reconstruct the conductivity distribution in elliptic problems from one measurement of the solution over the whole domain.

GNN at the Edge: Cost-Efficient Graph Neural Network Processing over Distributed Edge Servers

no code implementations31 Oct 2022 Liekang Zeng, Chongyu Yang, Peng Huang, Zhi Zhou, Shuai Yu, Xu Chen

Edge intelligence has arisen as a promising computing paradigm for supporting miscellaneous smart applications that rely on machine learning techniques.

Graph Neural Network Miscellaneous +1

Solving Elliptic Problems with Singular Sources using Singularity Splitting Deep Ritz Method

1 code implementation7 Sep 2022 Tianhao Hu, Bangti Jin, Zhi Zhou

Extensive numerical experiments in two- and multi-dimensional spaces with point sources, line sources or their combinations are presented to illustrate the efficiency of the proposed approach, and a comparative study with several existing approaches based on neural networks is also given, which shows clearly its competitiveness for the specific class of problems.

LAMDA-SSL: Semi-Supervised Learning in Python

1 code implementation9 Aug 2022 Lin-Han Jia, Lan-Zhe Guo, Zhi Zhou, Yu-Feng Li

The second part shows the usage of LAMDA-SSL by abundant examples in detail.

Robust Deep Semi-Supervised Learning: A Brief Introduction

no code implementations12 Feb 2022 Lan-Zhe Guo, Zhi Zhou, Yu-Feng Li

Semi-supervised learning (SSL) is the branch of machine learning that aims to improve learning performance by leveraging unlabeled data when labels are insufficient.

Edge Robotics: Edge-Computing-Accelerated Multi-Robot Simultaneous Localization and Mapping

no code implementations25 Dec 2021 Peng Huang, Liekang Zeng, Xu Chen, Ke Luo, Zhi Zhou, Shuai Yu

With the wide penetration of smart robots in multifarious fields, Simultaneous Localization and Mapping (SLAM) technique in robotics has attracted growing attention in the community.

Edge-computing Simultaneous Localization and Mapping

STEP: Out-of-Distribution Detection in the Presence of Limited In-Distribution Labeled Data

no code implementations NeurIPS 2021 Zhi Zhou, Lan-Zhe Guo, Zhanzhan Cheng, Yu-Feng Li, ShiLiang Pu

However, in many real-world applications, it is desirable to have SSL algorithms that not only classify the samples drawn from the same distribution of labeled data but also detect out-of-distribution (OOD) samples drawn from an unknown distribution.

Out-of-Distribution Detection Out of Distribution (OOD) Detection

Deep Reinforcement Learning with Spatio-temporal Traffic Forecasting for Data-Driven Base Station Sleep Control

no code implementations21 Jan 2021 Qiong Wu, Xu Chen, Zhi Zhou, Liang Chen, Junshan Zhang

To meet the ever increasing mobile traffic demand in 5G era, base stations (BSs) have been densely deployed in radio access networks (RANs) to increase the network coverage and capacity.

Deep Reinforcement Learning

FedHome: Cloud-Edge based Personalized Federated Learning for In-Home Health Monitoring

1 code implementation14 Dec 2020 Qiong Wu, Xu Chen, Zhi Zhou, Junshan Zhang

In this paper, we propose FedHome, a novel cloud-edge based federated learning framework for in-home health monitoring, which learns a shared global model in the cloud from multiple homes at the network edges and achieves data privacy protection by keeping user data locally.

Human Activity Recognition Personalized Federated Learning

CoEdge: Cooperative DNN Inference with Adaptive Workload Partitioning over Heterogeneous Edge Devices

no code implementations6 Dec 2020 Liekang Zeng, Xu Chen, Zhi Zhou, Lei Yang, Junshan Zhang

CoEdge utilizes available computation and communication resources at the edge and dynamically partitions the DNN inference workload adaptive to devices' computing capabilities and network conditions.

When Deep Reinforcement Learning Meets Federated Learning: Intelligent Multi-Timescale Resource Management for Multi-access Edge Computing in 5G Ultra Dense Network

no code implementations22 Sep 2020 Shuai Yu, Xu Chen, Zhi Zhou, Xiaowen Gong, Di wu

Ultra-dense edge computing (UDEC) has great potential, especially in the 5G era, but it still faces challenges in its current solutions, such as the lack of: i) efficient utilization of multiple 5G resources (e. g., computation, communication, storage and service resources); ii) low overhead offloading decision making and resource allocation strategies; and iii) privacy and security protection schemes.

Decision Making Deep Reinforcement Learning +3

The energy technique for the six-step BDF method

no code implementations17 Jul 2020 Georgios Akrivis, Minghua Chen, Fan Yu, Zhi Zhou

In combination with the Grenander--Szeg\"o theorem, we observe that a relaxed positivity condition on multipliers, milder than the basic %fundamental requirement of the Nevanlinna--Odeh multipliers that the sum of the absolute values of their components is strictly less than $1$, makes the energy technique applicable to the stability analysis of BDF methods for parabolic equations with selfadjoint elliptic part.

Numerical Analysis Numerical Analysis

HierTrain: Fast Hierarchical Edge AI Learning with Hybrid Parallelism in Mobile-Edge-Cloud Computing

no code implementations22 Mar 2020 Deyin Liu, Xu Chen, Zhi Zhou, Qing Ling

We develop a novel \textit{hybrid parallelism} method, which is the key to HierTrain, to adaptively assign the DNN model layers and the data samples across the three levels of edge device, edge server and cloud center.

Cloud Computing Scheduling

DeepCP: Deep Learning Driven Cascade Prediction Based Autonomous Content Placement in Closed Social Network

no code implementations9 Mar 2020 Qiong Wu, Muhong Wu, Xu Chen, Zhi Zhou, Kaiwen He, Liang Chen

Accordingly, we further propose a novel autonomous content placement mechanism CP-GAN which adopts the generative adversarial network (GAN) for agile placement decision making to reduce the content access latency and enhance users' QoE.

Decision Making Generative Adversarial Network

HFEL: Joint Edge Association and Resource Allocation for Cost-Efficient Hierarchical Federated Edge Learning

no code implementations26 Feb 2020 Siqi Luo, Xu Chen, Qiong Wu, Zhi Zhou, Shuai Yu

We further formulate a joint computation and communication resource allocation and edge association problem for device users under HFEL framework to achieve global cost minimization.

Distributed, Parallel, and Cluster Computing

Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing

no code implementations4 Oct 2019 En Li, Liekang Zeng, Zhi Zhou, Xu Chen

As a key technology of enabling Artificial Intelligence (AI) applications in 5G era, Deep Neural Networks (DNNs) have quickly attracted widespread attention.

Change Point Detection Collaborative Inference +1

Follow Me at the Edge: Mobility-Aware Dynamic Service Placement for Mobile Edge Computing

no code implementations14 Sep 2018 Tao Ouyang, Zhi Zhou, Xu Chen

To address this challenge in terms of the performance-cost trade-off, in this paper we study the mobile edge service performance optimization problem under long-term cost budget constraint.

Cloud Computing Edge-computing

Edge Intelligence: On-Demand Deep Learning Model Co-Inference with Device-Edge Synergy

no code implementations20 Jun 2018 En Li, Zhi Zhou, Xu Chen

As the backbone technology of machine learning, deep neural networks (DNNs) have have quickly ascended to the spotlight.

Cannot find the paper you are looking for? You can Submit a new open access paper.