Search Results for author: Zhisheng Niu

Found 22 papers, 2 papers with code

Mobility Accelerates Learning: Convergence Analysis on Hierarchical Federated Learning in Vehicular Networks

no code implementations18 Jan 2024 Tan Chen, Jintao Yan, Yuxuan Sun, Sheng Zhou, Deniz Gündüz, Zhisheng Niu

Hierarchical federated learning (HFL) enables distributed training of models across multiple devices with the help of several edge servers and a cloud edge server in a privacy-preserving manner.

Federated Learning Privacy Preserving

Data-Heterogeneous Hierarchical Federated Learning with Mobility

no code implementations19 Jun 2023 Tan Chen, Jintao Yan, Yuxuan Sun, Sheng Zhou, Deniz Gunduz, Zhisheng Niu

Federated learning enables distributed training of machine learning (ML) models across multiple devices in a privacy-preserving manner.

Federated Learning Privacy Preserving

MASS: Mobility-Aware Sensor Scheduling of Cooperative Perception for Connected Automated Driving

no code implementations25 Feb 2023 Yukuan Jia, Ruiqing Mao, Yuxuan Sun, Sheng Zhou, Zhisheng Niu

Specifically, we design a mobility-aware sensor scheduling (MASS) algorithm based on the restless multi-armed bandit (RMAB) theory to maximize the expected average perception gain.

Scheduling

SMDP-Based Dynamic Batching for Efficient Inference on GPU-Based Platforms

no code implementations30 Jan 2023 Yaodan Xu, Jingzhou Sun, Sheng Zhou, Zhisheng Niu

In particular, parallel computing resources on the platforms, such as graphics processing units (GPUs), have higher computational and energy efficiency with larger batch sizes.

Edge-computing

MOB-FL: Mobility-Aware Federated Learning for Intelligent Connected Vehicles

no code implementations7 Dec 2022 Bowen Xie, Yuxuan Sun, Sheng Zhou, Zhisheng Niu, Yang Xu, Jingran Chen, Deniz Gündüz

Federated learning (FL) is a promising approach to enable the future Internet of vehicles consisting of intelligent connected vehicles (ICVs) with powerful sensing, computing and communication capabilities.

Federated Learning Trajectory Prediction

MEET: Mobility-Enhanced Edge inTelligence for Smart and Green 6G Networks

no code implementations27 Oct 2022 Yuxuan Sun, Bowen Xie, Sheng Zhou, Zhisheng Niu

Accordingly, base stations (BSs) and edge servers (ESs) need to be densely deployed, leading to huge deployment and operation costs, in particular the energy costs.

DOLPHINS: Dataset for Collaborative Perception enabled Harmonious and Interconnected Self-driving

1 code implementation15 Jul 2022 Ruiqing Mao, Jingyu Guo, Yukuan Jia, Yuxuan Sun, Sheng Zhou, Zhisheng Niu

In this work, we release DOLPHINS: Dataset for cOllaborative Perception enabled Harmonious and INterconnected Self-driving, as a new simulated large-scale various-scenario multi-view multi-modality autonomous driving dataset, which provides a ground-breaking benchmark platform for interconnected autonomous driving.

Autonomous Driving Object Detection

Multi-user Co-inference with Batch Processing Capable Edge Server

no code implementations3 Jun 2022 Wenqi Shi, Sheng Zhou, Zhisheng Niu, Miao Jiang, Lu Geng

To deal with the coupled offloading and scheduling introduced by concurrent batch processing, we first consider an offline problem with a constant edge inference latency and the same latency constraint.

Scheduling

Time-Correlated Sparsification for Efficient Over-the-Air Model Aggregation in Wireless Federated Learning

no code implementations17 Feb 2022 Yuxuan Sun, Sheng Zhou, Zhisheng Niu, Deniz Gündüz

In this work, we propose time-correlated sparsification with hybrid aggregation (TCS-H) for communication-efficient FEEL, which exploits jointly the power of model compression and over-the-air computation.

Federated Learning Model Compression +1

Online V2X Scheduling for Raw-Level Cooperative Perception

no code implementations12 Feb 2022 Yukuan Jia, Ruiqing Mao, Yuxuan Sun, Sheng Zhou, Zhisheng Niu

Cooperative perception of connected vehicles comes to the rescue when the field of view restricts stand-alone intelligence.

Scheduling

Coded Computation across Shared Heterogeneous Workers with Communication Delay

no code implementations23 Sep 2021 Yuxuan Sun, Fan Zhang, Junlin Zhao, Sheng Zhou, Zhisheng Niu, Deniz Gündüz

In this work, we consider a multi-master heterogeneous-worker distributed computing scenario, where multiple matrix multiplication tasks are encoded and allocated to workers for parallel computation.

Distributed Computing

Dynamic Scheduling for Over-the-Air Federated Edge Learning with Energy Constraints

no code implementations31 May 2021 Yuxuan Sun, Sheng Zhou, Zhisheng Niu, Deniz Gündüz

In this work, we consider an over-the-air FEEL system with analog gradient aggregation, and propose an energy-aware dynamic device scheduling algorithm to optimize the training performance under energy constraints of devices, where both communication energy for gradient aggregation and computation energy for local training are included.

Scheduling

Joint Device Scheduling and Resource Allocation for Latency Constrained Wireless Federated Learning

no code implementations14 Jul 2020 Wenqi Shi, Sheng Zhou, Zhisheng Niu, Miao Jiang, Lu Geng

Then, a greedy device scheduling algorithm is introduced, which in each step selects the device consuming the least updating time obtained by the optimal bandwidth allocation, until the lower bound begins to increase, meaning that scheduling more devices will degrade the model accuracy.

Federated Learning Scheduling

Device Scheduling with Fast Convergence for Wireless Federated Learning

no code implementations3 Nov 2019 Wenqi Shi, Sheng Zhou, Zhisheng Niu

In each iteration of FL (called round), the edge devices update local models based on their own data and contribute to the global training by uploading the model updates via wireless channels.

Federated Learning Scheduling

Improving Device-Edge Cooperative Inference of Deep Learning via 2-Step Pruning

1 code implementation8 Mar 2019 Wenqi Shi, Yunzhong Hou, Sheng Zhou, Zhisheng Niu, Yang Zhang, Lu Geng

Since the output data size of a DNN layer can be larger than that of the raw data, offloading intermediate data between layers can suffer from high transmission latency under limited wireless bandwidth.

Distributed Policy Learning Based Random Access for Diversified QoS Requirements

no code implementations6 Mar 2019 Zhiyuan Jiang, Sheng Zhou, Zhisheng Niu

Future wireless access networks need to support diversified quality of service (QoS) metrics required by various types of Internet-of-Things (IoT) devices, e. g., age of information (AoI) for status generating sources and ultra low latency for safety information in vehicular networks.

A Two-Step Learning and Interpolation Method for Location-Based Channel Database

no code implementations4 Dec 2018 Ruichen Deng, Zhiyuan Jiang, Sheng Zhou, Shuguang Cui, Zhisheng Niu

Timely and accurate knowledge of channel state information (CSI) is necessary to support scheduling operations at both physical and network layers.

Scheduling

Time-Sequence Channel Inference for Beam Alignment in Vehicular Networks

no code implementations4 Dec 2018 Sheng Chen, Zhiyuan Jiang, Sheng Zhou, Zhisheng Niu

In this paper, we propose a learning-based low-overhead beam alignment method for vehicle-to-infrastructure communication in vehicular networks.

Neural Network simulation

Exploiting Wireless Channel State Information Structures Beyond Linear Correlations: A Deep Learning Approach

no code implementations3 Dec 2018 Zhiyuan Jiang, Sheng Chen, Andreas F. Molisch, Rath Vannithamby, Sheng Zhou, Zhisheng Niu

Knowledge of information about the propagation channel in which a wireless system operates enables better, more efficient approaches for signal transmissions.

Dimensionality Reduction

A Block Regression Model for Short-Term Mobile Traffic Forecasting

no code implementations17 Nov 2015 Huimin Pan, Jingchu Liu, Sheng Zhou, Zhisheng Niu

Based on these characteristics, we propose a \emph{Block Regression} ({BR}) model for mobile traffic forecasting.

regression

Cannot find the paper you are looking for? You can Submit a new open access paper.