Search Results for author: Mingxuan Yuan

Found 67 papers, 23 papers with code

Beyond Standard MoE: Mixture of Latent Experts for Resource-Efficient Language Models

no code implementations29 Mar 2025 Zehua Liu, Han Wu, Ruifeng She, Xiaojin Fu, Xiongwei Han, Tao Zhong, Mingxuan Yuan

In this paper, we introduce Mixture of Latent Experts (MoLE), a novel parameterization methodology that facilitates the mapping of specific experts into a shared latent space.

Computational Efficiency

Unlocking Efficient Long-to-Short LLM Reasoning with Model Merging

1 code implementation26 Mar 2025 Han Wu, Yuxuan Yao, Shuqi Liu, Zehua Liu, Xiaojin Fu, Xiongwei Han, Xing Li, Hui-Ling Zhen, Tao Zhong, Mingxuan Yuan

Model merging, on the other hand, offers a cost-effective and robust alternative by integrating the quick-thinking capabilities of System 1 models with the methodical reasoning of System 2 models.

Prompt Engineering Reinforcement Learning (RL)

Open3DBench: Open-Source Benchmark for 3D-IC Backend Implementation and PPA Evaluation

1 code implementation17 Mar 2025 Yunqi Shi, Chengrui Gao, Wanqi Ren, Siyuan Xu, Ke Xue, Mingxuan Yuan, Chao Qian, Zhi-Hua Zhou

This work introduces Open3DBench, an open-source 3D-IC backend implementation benchmark built upon the OpenROAD-flow-scripts framework, enabling comprehensive evaluation of power, performance, area, and thermal metrics.

Timing-Driven Global Placement by Efficient Critical Path Extraction

no code implementations28 Feb 2025 Yunqi Shi, Siyuan Xu, Shixiong Kai, Xi Lin, Ke Xue, Mingxuan Yuan, Chao Qian

Timing optimization during the global placement of integrated circuits has been a significant focus for decades, yet it remains a complex, unresolved issue.

ARS: Automatic Routing Solver with Large Language Models

1 code implementation21 Feb 2025 Kai Li, Fei Liu, Zhenkun Wang, Xialiang Tong, Xiongwei Han, Mingxuan Yuan

Real-world Vehicle Routing Problems (VRPs) are characterized by a variety of practical constraints, making manual solver design both knowledge-intensive and time-consuming.

Language Modeling Language Modelling +1

Sens-Merging: Sensitivity-Guided Parameter Balancing for Merging Large Language Models

no code implementations18 Feb 2025 Shuqi Liu, Han Wu, Bowei He, Xiongwei Han, Mingxuan Yuan, Linqi Song

Notably, when combined with existing merging techniques, our method enables merged models to outperform specialized fine-tuned models, particularly in code generation tasks.

Code Generation General Knowledge +1

1bit-Merging: Dynamic Quantized Merging for Large Language Models

no code implementations15 Feb 2025 Shuqi Liu, Han Wu, Bowei He, Zehua Liu, Xiongwei Han, Mingxuan Yuan, Linqi Song

Recent advances in large language models have led to specialized models excelling in specific domains, creating a need for efficient model merging techniques.

Code Generation Math +1

LoRE-Merging: Exploring Low-Rank Estimation For Large Language Model Merging

no code implementations15 Feb 2025 Zehua Liu, Han Wu, Yuxuan Yao, Ruifeng She, Xiongwei Han, Tao Zhong, Mingxuan Yuan

While most current approaches rely on further training techniques, such as fine-tuning or reinforcement learning, to enhance model capacities, model merging stands out for its ability of improving models without requiring any additional training.

Language Modeling Language Modelling +1

Certifying Language Model Robustness with Fuzzed Randomized Smoothing: An Efficient Defense Against Backdoor Attacks

no code implementations9 Feb 2025 Bowei He, Lihao Yin, Hui-Ling Zhen, Jianping Zhang, Lanqing Hong, Mingxuan Yuan, Chen Ma

The widespread deployment of pre-trained language models (PLMs) has exposed them to textual backdoor attacks, particularly those planted during the pre-training stage.

Language Modeling Language Modelling

CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference

1 code implementation6 Feb 2025 Zehua Pei, Lancheng Zou, Hui-Ling Zhen, Xianzhi Yu, Wulong Liu, Sinno Jialin Pan, Mingxuan Yuan, Bei Yu

Large language models (LLMs) achieve impressive performance by scaling model parameters, but this comes with significant inference overhead.

AttentionPredictor: Temporal Pattern Matters for Efficient LLM Inference

1 code implementation6 Feb 2025 Qingyue Yang, Jie Wang, Xing Li, Zhihai Wang, Chen Chen, Lei Chen, Xianzhi Yu, Wulong Liu, Jianye Hao, Mingxuan Yuan, Bin Li

With the development of large language models (LLMs), efficient inference through Key-Value (KV) cache compression has attracted considerable attention, especially for long-context generation.

KVTuner: Sensitivity-Aware Layer-wise Mixed Precision KV Cache Quantization for Efficient and Nearly Lossless LLM Inference

1 code implementation6 Feb 2025 Xing Li, Zeyu Xing, Yiming Li, Linping Qu, Hui-Ling Zhen, Wulong Liu, Yiwu Yao, Sinno Jialin Pan, Mingxuan Yuan

KV cache quantization can improve Large Language Models (LLMs) inference throughput and latency in long contexts and large batch-size scenarios while preserving LLMs effectiveness.

Mathematical Reasoning Quantization

E2ESlack: An End-to-End Graph-Based Framework for Pre-Routing Slack Prediction

no code implementations13 Jan 2025 Saurabh Bodhe, Zhanguang Zhang, Atia Hamidizadeh, Shixiong Kai, Yingxue Zhang, Mingxuan Yuan

The framework includes a TimingParser that supports DEF, SDF and LIB files for feature extraction and graph construction, an arrival time prediction model and a fast RAT estimation module.

graph construction Prediction

MixPE: Quantization and Hardware Co-design for Efficient LLM Inference

no code implementations25 Nov 2024 Yu Zhang, Mingzi Wang, Lancheng Zou, Wulong Liu, Hui-Ling Zhen, Mingxuan Yuan, Bei Yu

Transformer-based large language models (LLMs) have achieved remarkable success as model sizes continue to grow, yet their deployment remains challenging due to significant computational and memory demands.

Quantization

FuseGPT: Learnable Layers Fusion of Generative Pre-trained Transformers

1 code implementation21 Nov 2024 Zehua Pei, Hui-Ling Zhen, Xianzhi Yu, Sinno Jialin Pan, Mingxuan Yuan, Bei Yu

In this paper, we propose FuseGPT, a novel methodology to recycle the pruned transformer blocks to further recover the model performance.

The Graph's Apprentice: Teaching an LLM Low Level Knowledge for Circuit Quality Estimation

no code implementations30 Oct 2024 Reza Moravej, Saurabh Bodhe, Zhanguang Zhang, Didier Chetelat, Dimitrios Tsaras, Yingxue Zhang, Hui-Ling Zhen, Jianye Hao, Mingxuan Yuan

Logic synthesis is a crucial phase in the circuit design process, responsible for transforming hardware description language (HDL) designs into optimized netlists.

Knowledge Distillation

SeaDAG: Semi-autoregressive Diffusion for Conditional Directed Acyclic Graph Generation

no code implementations21 Oct 2024 Xinyi Zhou, Xing Li, Yingzhao Lian, Yiwen Wang, Lei Chen, Mingxuan Yuan, Jianye Hao, Guangyong Chen, Pheng Ann Heng

We explicitly train the model to learn graph conditioning with a condition loss, which enhances the diffusion model's capacity to generate graphs that are both realistic and aligned with specified properties.

Decoder Denoising +1

Logic Synthesis Optimization with Predictive Self-Supervision via Causal Transformers

no code implementations16 Sep 2024 Raika Karimi, Faezeh Faez, Yingxue Zhang, Xing Li, Lei Chen, Mingxuan Yuan, Mahdi Biparva

Contemporary hardware design benefits from the abstraction provided by high-level logic gates, streamlining the implementation of logic circuits.

Reinforcement Learning (RL)

MTLSO: A Multi-Task Learning Approach for Logic Synthesis Optimization

no code implementations9 Sep 2024 Faezeh Faez, Raika Karimi, Yingxue Zhang, Xing Li, Lei Chen, Mingxuan Yuan, Mahdi Biparva

On the other hand, we employ a hierarchical graph representation learning strategy to improve the model's capacity for learning expressive graph-level representations of large AIGs, surpassing traditional plain GNNs.

Graph Classification Graph Representation Learning +1

RTLRewriter: Methodologies for Large Models aided RTL Code Optimization

1 code implementation4 Sep 2024 Xufeng Yao, Yiwen Wang, Xing Li, Yingzhao Lian, Ran Chen, Lei Chen, Mingxuan Yuan, Hong Xu, Bei Yu

Our comparative analysis with established compilers such as Yosys and E-graph demonstrates significant improvements, highlighting the benefits of integrating large models into the early stages of circuit design.

Benchmarking

ShortCircuit: AlphaZero-Driven Circuit Design

no code implementations19 Aug 2024 Dimitrios Tsaras, Antoine Grosnit, Lei Chen, Zhiyao Xie, Haitham Bou-Ammar, Mingxuan Yuan

Chip design relies heavily on generating Boolean circuits, such as AND-Inverter Graphs (AIGs), from functional descriptions like truth tables.

Benchmarking End-To-End Performance of AI-Based Chip Placement Algorithms

no code implementations3 Jul 2024 Zhihai Wang, Zijie Geng, Zhaojie Tu, Jie Wang, Yuxi Qian, Zhexuan Xu, Ziyan Liu, Siyuan Xu, Zhentao Tang, Shixiong Kai, Mingxuan Yuan, Jianye Hao, Bin Li, Yongdong Zhang, Feng Wu

We executed six state-of-the-art AI-based chip placement algorithms on these designs and plugged the results of each single-point algorithm into the physical implementation workflow to obtain the final PPA results.

Benchmarking

UDC: A Unified Neural Divide-and-Conquer Framework for Large-Scale Combinatorial Optimization Problems

2 code implementations29 Jun 2024 Zhi Zheng, Changliang Zhou, Tong Xialiang, Mingxuan Yuan, Zhenkun Wang

Employing a high-efficiency Graph Neural Network (GNN) for global instance dividing and a fixed-length sub-path solver for conquering divided sub-problems, the proposed UDC framework demonstrates extensive applicability, achieving superior performance in 10 representative large-scale CO problems.

Combinatorial Optimization Graph Neural Network

Logic Synthesis with Generative Deep Neural Networks

no code implementations7 Jun 2024 Xihan Li, Xing Li, Lei Chen, Xing Zhang, Mingxuan Yuan, Jun Wang

While deep learning has achieved significant success in various domains, its application to logic circuit design has been limited due to complex constraints and strict feasibility requirement.

SmoothGNN: Smoothing-aware GNN for Unsupervised Node Anomaly Detection

no code implementations27 May 2024 Xiangyu Dong, Xingyi Zhang, Yanni Sun, Lei Chen, Mingxuan Yuan, Sibo Wang

We introduce Individual Smoothing Patterns (ISP) and Neighborhood Smoothing Patterns (NSP), which indicate that the representations of anomalous nodes are harder to smooth than those of normal ones.

Anomaly Detection Graph Learning +1

DPN: Decoupling Partition and Navigation for Neural Solvers of Min-max Vehicle Routing Problems

1 code implementation27 May 2024 Zhi Zheng, Shunyu Yao, Zhenkun Wang, Xialiang Tong, Mingxuan Yuan, Ke Tang

The min-max vehicle routing problem (min-max VRP) traverses all given customers by assigning several routes and aims to minimize the length of the longest route.

Reinforcement Learning (RL)

Prompt Learning for Generalized Vehicle Routing

1 code implementation20 May 2024 Fei Liu, Xi Lin, Weiduo Liao, Zhenkun Wang, Qingfu Zhang, Xialiang Tong, Mingxuan Yuan

To be concrete, we propose a novel prompt learning method to facilitate fast zero-shot adaptation of a pre-trained model to solve routing problem instances from different distributions.

Combinatorial Optimization Zero-shot Generalization

GraSS: Combining Graph Neural Networks with Expert Knowledge for SAT Solver Selection

no code implementations17 May 2024 Zhanguang Zhang, Didier Chetelat, Joseph Cotnareanu, Amur Ghose, Wenyi Xiao, Hui-Ling Zhen, Yingxue Zhang, Jianye Hao, Mark Coates, Mingxuan Yuan

In this paper we present GraSS, a novel approach for automatic SAT solver selection based on tripartite graph representations of instances and a heterogeneous graph neural network (GNN) model.

Graph Neural Network

Instance-Conditioned Adaptation for Large-scale Generalization of Neural Combinatorial Optimization

no code implementations3 May 2024 Changliang Zhou, Xi Lin, Zhenkun Wang, Xialiang Tong, Mingxuan Yuan, Qingfu Zhang

The neural combinatorial optimization (NCO) approach has shown great potential for solving routing problems without the requirement of expert knowledge.

Combinatorial Optimization

Learning to Cut via Hierarchical Sequence/Set Model for Efficient Mixed-Integer Programming

no code implementations19 Apr 2024 Jie Wang, Zhihai Wang, Xijun Li, Yufei Kuang, Zhihao Shi, Fangzhou Zhu, Mingxuan Yuan, Jia Zeng, Yongdong Zhang, Feng Wu

Moreover, we observe that (P3) what order of selected cuts to prefer significantly impacts the efficiency of MILP solvers as well.

Self-Improved Learning for Scalable Neural Combinatorial Optimization

no code implementations28 Mar 2024 Fu Luo, Xi Lin, Zhenkun Wang, Xialiang Tong, Mingxuan Yuan, Qingfu Zhang

The end-to-end neural combinatorial optimization (NCO) method shows promising performance in solving complex combinatorial optimization problems without the need for expert design.

Combinatorial Optimization

Circuit Transformer: A Transformer That Preserves Logical Equivalence

1 code implementation14 Mar 2024 Xihan Li, Xing Li, Lei Chen, Xing Zhang, Mingxuan Yuan, Jun Wang

In this study, we introduce a generative neural model, the "Circuit Transformer", which eliminates such wrong predictions and produces logic circuits strictly equivalent to given Boolean functions.

Hallucination

IB-Net: Initial Branch Network for Variable Decision in Boolean Satisfiability

no code implementations6 Mar 2024 Tsz Ho Chan, Wenyi Xiao, Junhua Huang, HuiLing Zhen, Guangji Tian, Mingxuan Yuan

Boolean Satisfiability problems are vital components in Electronic Design Automation, particularly within the Logic Equivalence Checking process.

Multi-Task Learning for Routing Problem with Cross-Problem Zero-Shot Generalization

1 code implementation23 Feb 2024 Fei Liu, Xi Lin, Zhenkun Wang, Qingfu Zhang, Xialiang Tong, Mingxuan Yuan

The results show that the unified model demonstrates superior performance in the eleven VRPs, reducing the average gap to around 5% from over 20% in the existing approach and achieving a significant performance boost on benchmark datasets as well as a real-world logistics application.

Attribute Combinatorial Optimization +2

DiLA: Enhancing LLM Tool Learning with Differential Logic Layer

no code implementations19 Feb 2024 Yu Zhang, Hui-Ling Zhen, Zehua Pei, Yingzhao Lian, Lihao Yin, Mingxuan Yuan, Bei Yu

In this paper, we propose a novel differential logic layer-aided language modeling (DiLA) approach, where logical constraints are integrated into the forward and backward passes of a network layer, to provide another option for LLM tool learning.

Language Modeling Language Modelling +1

BetterV: Controlled Verilog Generation with Discriminative Guidance

no code implementations3 Feb 2024 Zehua Pei, Hui-Ling Zhen, Mingxuan Yuan, Yu Huang, Bei Yu

In this work, we propose a Verilog generation framework, BetterV, which fine-tunes the large language models (LLMs) on processed domain-specific datasets and incorporates generative discriminators for guidance on particular design demands.

Text Generation

Machine Learning Insides OptVerse AI Solver: Design Principles and Applications

no code implementations11 Jan 2024 Xijun Li, Fangzhou Zhu, Hui-Ling Zhen, Weilin Luo, Meng Lu, Yimin Huang, Zhenan Fan, Zirui Zhou, Yufei Kuang, Zhihai Wang, Zijie Geng, Yang Li, Haoyang Liu, Zhiwu An, Muming Yang, Jianshu Li, Jie Wang, Junchi Yan, Defeng Sun, Tao Zhong, Yong Zhang, Jia Zeng, Mingxuan Yuan, Jianye Hao, Jun Yao, Kun Mao

To this end, we present a comprehensive study on the integration of machine learning (ML) techniques into Huawei Cloud's OptVerse AI Solver, which aims to mitigate the scarcity of real-world mathematical programming instances, and to surpass the capabilities of traditional optimization techniques.

Decision Making Management

LLM4EDA: Emerging Progress in Large Language Models for Electronic Design Automation

1 code implementation28 Dec 2023 RuiZhe Zhong, Xingbo Du, Shixiong Kai, Zhentao Tang, Siyuan Xu, Hui-Ling Zhen, Jianye Hao, Qiang Xu, Mingxuan Yuan, Junchi Yan

Since circuit can be represented with HDL in a textual format, it is reasonable to question whether LLMs can be leveraged in the EDA field to achieve fully automated chip design and generate circuits with improved power, performance, and area (PPA).

Answer Generation Chatbot +1

Algorithm Evolution Using Large Language Model

4 code implementations26 Nov 2023 Fei Liu, Xialiang Tong, Mingxuan Yuan, Qingfu Zhang

In this paper, we propose a novel approach called Algorithm Evolution using Large Language Model (AEL).

Language Modeling Language Modelling +2

Large Language Model for Multi-objective Evolutionary Optimization

1 code implementation19 Oct 2023 Fei Liu, Xi Lin, Zhenkun Wang, Shunyu Yao, Xialiang Tong, Mingxuan Yuan, Qingfu Zhang

It is also promising to see the operator only learned from a few instances can have robust generalization performance on unseen problems with quite different patterns and settings.

Evolutionary Algorithms Language Modeling +5

A Circuit Domain Generalization Framework for Efficient Logic Synthesis in Chip Design

1 code implementation22 Aug 2023 Zhihai Wang, Lei Chen, Jie Wang, Xing Li, Yinqi Bai, Xijun Li, Mingxuan Yuan, Jianye Hao, Yongdong Zhang, Feng Wu

In particular, we notice that the runtime of the Resub and Mfs2 operators often dominates the overall runtime of LS optimization processes.

Domain Generalization

DeepGate2: Functionality-Aware Circuit Representation Learning

1 code implementation25 May 2023 Zhengyuan Shi, Hongyang Pan, Sadaf Khan, Min Li, Yi Liu, Junhua Huang, Hui-Ling Zhen, Mingxuan Yuan, Zhufei Chu, Qiang Xu

Circuit representation learning aims to obtain neural representations of circuit elements and has emerged as a promising research direction that can be applied to various EDA and logic reasoning tasks.

Graph Neural Network Representation Learning

Conflict-driven Structural Learning Towards Higher Coverage Rate in ATPG

no code implementations4 Mar 2023 Hui-Ling Zhen, Naixing Wang, Junhua Huang, Xinyue Huang, Mingxuan Yuan, Yu Huang

(2) Conflict-driven implication and justification have been applied to increase decision accuracy and solving efficiency.

Heuristics for Vehicle Routing Problem: A Survey and Recent Advances

no code implementations1 Mar 2023 Fei Liu, Chengyu Lu, Lin Gui, Qingfu Zhang, Xialiang Tong, Mingxuan Yuan

Vehicle routing is a well-known optimization research topic with significant practical importance.

Survey

HardSATGEN: Understanding the Difficulty of Hard SAT Formula Generation and A Strong Structure-Hardness-Aware Baseline

1 code implementation4 Feb 2023 Yang Li, Xinyan Chen, Wenxuan Guo, Xijun Li, Wanqian Luo, Junhua Huang, Hui-Ling Zhen, Mingxuan Yuan, Junchi Yan

On top of the observations that industrial formulae exhibit clear community structure and oversplit substructures lead to the difficulty in semantic formation of logical structures, we propose HardSATGEN, which introduces a fine-grained control mechanism to the neural split-merge paradigm for SAT formula generation to better recover the structural and computational properties of the industrial benchmarks.

Learning Cut Selection for Mixed-Integer Linear Programming via Hierarchical Sequence Model

no code implementations1 Feb 2023 Zhihai Wang, Xijun Li, Jie Wang, Yufei Kuang, Mingxuan Yuan, Jia Zeng, Yongdong Zhang, Feng Wu

Cut selection -- which aims to select a proper subset of the candidate cuts to improve the efficiency of solving MILPs -- heavily depends on (P1) which cuts should be preferred, and (P2) how many cuts should be selected.

SATformer: Transformer-Based UNSAT Core Learning

no code implementations2 Sep 2022 Zhengyuan Shi, Min Li, Yi Liu, Sadaf Khan, Junhua Huang, Hui-Ling Zhen, Mingxuan Yuan, Qiang Xu

This paper introduces SATformer, a novel Transformer-based approach for the Boolean Satisfiability (SAT) problem.

Graph Neural Network Multi-Task Learning

LQoCo: Learning to Optimize Cache Capacity Overloading in Storage Systems

no code implementations21 Mar 2022 Ji Zhang, Xijun Li, Xiyao Zhou, Mingxuan Yuan, Zhuo Cheng, Keji Huang, YiFan Li

Cache plays an important role to maintain high and stable performance (i. e. high throughput, low tail latency and throughput jitter) in storage systems.

Management

Machine Learning Methods in Solving the Boolean Satisfiability Problem

no code implementations2 Mar 2022 Wenxuan Guo, Junchi Yan, Hui-Ling Zhen, Xijun Li, Mingxuan Yuan, Yaohui Jin

This paper reviews the recent literature on solving the Boolean satisfiability problem (SAT), an archetypal NP-complete problem, with the help of machine learning techniques.

BIG-bench Machine Learning

A Data-Driven Column Generation Algorithm For Bin Packing Problem in Manufacturing Industry

no code implementations25 Feb 2022 Jiahui Duan, Xialiang Tong, Fei Ni, Zhenan He, Lei Chen, Mingxuan Yuan

The bin packing problem exists widely in real logistic scenarios (e. g., packing pipeline, express delivery), with its goal to improve the packing efficiency and reduce the transportation cost.

Combinatorial Optimization

Learning to Reformulate for Linear Programming

no code implementations17 Jan 2022 Xijun Li, Qingyu Qu, Fangzhou Zhu, Jia Zeng, Mingxuan Yuan, Kun Mao, Jie Wang

In the past decades, a serial of traditional operation research algorithms have been proposed to obtain the optimum of a given LP in a fewer solving time.

A Hierarchical Reinforcement Learning Based Optimization Framework for Large-scale Dynamic Pickup and Delivery Problems

no code implementations NeurIPS 2021 Yi Ma, Xiaotian Hao, Jianye Hao, Jiawen Lu, Xing Liu, Tong Xialiang, Mingxuan Yuan, Zhigang Li, Jie Tang, Zhaopeng Meng

To address this problem, existing methods partition the overall DPDP into fixed-size sub-problems by caching online generated orders and solve each sub-problem, or on this basis to utilize the predicted future orders to optimize each sub-problem further.

Hierarchical Reinforcement Learning

Learning-Aided Heuristics Design for Storage System

no code implementations14 Jun 2021 Yingtian Tang, Han Lu, Xijun Li, Lei Chen, Mingxuan Yuan, Jia Zeng

Computer systems such as storage systems normally require transparent white-box algorithms that are interpretable for human experts.

Deep Reinforcement Learning reinforcement-learning +1

Learning to Select Cuts for Efficient Mixed-Integer Programming

no code implementations28 May 2021 Zeren Huang, Kerong Wang, Furui Liu, Hui-Ling Zhen, Weinan Zhang, Mingxuan Yuan, Jianye Hao, Yong Yu, Jun Wang

In the online A/B testing of the product planning problems with more than $10^7$ variables and constraints daily, Cut Ranking has achieved the average speedup ratio of 12. 42% over the production solver without any accuracy loss of solution.

Multiple Instance Learning

Learning to Optimize Industry-Scale Dynamic Pickup and Delivery Problems

no code implementations27 May 2021 Xijun Li, Weilin Luo, Mingxuan Yuan, Jun Wang, Jiawen Lu, Jie Wang, Jinhu Lu, Jia Zeng

Our method is entirely data driven and thus adaptive, i. e., the relational representation of adjacent vehicles can be learned and corrected by ST-DDGN from data periodically.

Graph Embedding Management +1

Bilevel Learning Model Towards Industrial Scheduling

no code implementations10 Aug 2020 Longkang Li, Hui-Ling Zhen, Mingxuan Yuan, Jiawen Lu, XialiangTong, Jia Zeng, Jun Wang, Dirk Schnieders

In this paper, we propose a Bilevel Deep reinforcement learning Scheduler, \textit{BDS}, in which the higher level is responsible for exploring an initial global sequence, whereas the lower level is aiming at exploitation for partial sequence refinements, and the two levels are connected by a sliding-window sampling mechanism.

Deep Reinforcement Learning model +1

Transfer Learning-Based Outdoor Position Recovery with Telco Data

no code implementations10 Dec 2019 Yige Zhang, Aaron Yi Ding, Jorg Ott, Mingxuan Yuan, Jia Zeng, Kun Zhang, Weixiong Rao

In this paper, by leveraging the recently developed transfer learning techniques, we design a novel Telco position recovery framework, called TLoc, to transfer good models in the carefully selected source domains (those fine-grained small subareas) to a target one which originally suffers from poor localization accuracy.

Position Transfer Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.