Search Results for author: Linnan Wang

Found 18 papers, 10 papers with code

BLASX: A High Performance Level-3 BLAS Library for Heterogeneous Multi-GPU Computing

1 code implementation16 Oct 2015 Linnan Wang, Wei Wu, Jianxiong Xiao, Yi Yang

Basic Linear Algebra Subprograms (BLAS) are a set of low level linear algebra kernels widely adopted by applications involved with the deep learning and scientific computing.

Distributed, Parallel, and Cluster Computing

Large Scale Artificial Neural Network Training Using Multi-GPUs

no code implementations13 Nov 2015 Linnan Wang, Wei Wu, Jianxiong Xiao, Yang Yi

This paper describes a method for accelerating large scale Artificial Neural Networks (ANN) training using multi-GPUs by reducing the forward and backward passes to matrix multiplication.

Accelerating Deep Neural Network Training with Inconsistent Stochastic Gradient Descent

no code implementations17 Mar 2016 Linnan Wang, Yi Yang, Martin Renqiang Min, Srimat Chakradhar

Then we present the study of ISGD batch size to the learning rate, parallelism, synchronization cost, system saturation and scalability.

Simple and Efficient Parallelization for Probabilistic Temporal Tensor Factorization

no code implementations11 Nov 2016 Guangxi Li, Zenglin Xu, Linnan Wang, Jinmian Ye, Irwin King, Michael Lyu

Probabilistic Temporal Tensor Factorization (PTTF) is an effective algorithm to model the temporal tensor data.

Learning Compact Recurrent Neural Networks with Block-Term Tensor Decomposition

no code implementations CVPR 2018 Jinmian Ye, Linnan Wang, Guangxi Li, Di Chen, Shandian Zhe, Xinqi Chu, Zenglin Xu

On three challenging tasks, including Action Recognition in Videos, Image Captioning and Image Generation, BT-RNN outperforms TT-RNN and the standard RNN in terms of both prediction accuracy and convergence rate.

Action Recognition In Videos Image Captioning +3

SuperNeurons: Dynamic GPU Memory Management for Training Deep Neural Networks

no code implementations13 Jan 2018 Linnan Wang, Jinmian Ye, Yiyang Zhao, Wei Wu, Ang Li, Shuaiwen Leon Song, Zenglin Xu, Tim Kraska

Given the limited GPU DRAM, SuperNeurons not only provisions the necessary memory for the training, but also dynamically allocates the memory for convolution workspaces to achieve the high performance.

Management Scheduling

Neural Architecture Search using Deep Neural Networks and Monte Carlo Tree Search

2 code implementations18 May 2018 Linnan Wang, Yiyang Zhao, Yuu Jinnai, Yuandong Tian, Rodrigo Fonseca

Neural Architecture Search (NAS) has shown great success in automating the design of neural networks, but the prohibitive amount of computations behind current NAS methods requires further investigations in improving the sample efficiency and the network evaluation cost to get better results in a shorter time.

Image Captioning Neural Architecture Search +4

Sample-Efficient Neural Architecture Search by Learning Action Space for Monte Carlo Tree Search

1 code implementation1 Jan 2019 Linnan Wang, Saining Xie, Teng Li, Rodrigo Fonseca, Yuandong Tian

To improve the sample efficiency, this paper proposes Latent Action Neural Architecture Search (LaNAS), which learns actions to recursively partition the search space into good or bad regions that contain networks with similar performance metrics.

Evolutionary Algorithms Image Classification +1

AlphaX: eXploring Neural Architectures with Deep Neural Networks and Monte Carlo Tree Search

1 code implementation26 Mar 2019 Linnan Wang, Yiyang Zhao, Yuu Jinnai, Yuandong Tian, Rodrigo Fonseca

Neural Architecture Search (NAS) has shown great success in automating the design of neural networks, but the prohibitive amount of computations behind current NAS methods requires further investigations in improving the sample efficiency and the network evaluation cost to get better results in a shorter time.

Image Captioning Neural Architecture Search +4

Sample-Efficient Neural Architecture Search by Learning Action Space

1 code implementation17 Jun 2019 Linnan Wang, Saining Xie, Teng Li, Rodrigo Fonseca, Yuandong Tian

To improve the sample efficiency, this paper proposes Latent Action Neural Architecture Search (LaNAS), which learns actions to recursively partition the search space into good or bad regions that contain networks with similar performance metrics.

Evolutionary Algorithms Neural Architecture Search

Neural Architecture Search by Learning Action Space for Monte Carlo Tree Search

no code implementations25 Sep 2019 Linnan Wang, Saining Xie, Teng Li, Rodrigo Fonseca, Yuandong Tian

As a result, using manually designed action space to perform NAS often leads to sample-inefficient explorations of architectures and thus can be sub-optimal.

Bayesian Optimization Neural Architecture Search

Few-shot Neural Architecture Search

2 code implementations11 Jun 2020 Yiyang Zhao, Linnan Wang, Yuandong Tian, Rodrigo Fonseca, Tian Guo

supernet, to approximate the performance of every architecture in the search space via weight-sharing.

Neural Architecture Search Transfer Learning

Learning Search Space Partition for Black-box Optimization using Monte Carlo Tree Search

2 code implementations NeurIPS 2020 Linnan Wang, Rodrigo Fonseca, Yuandong Tian

If the nonlinear partition function and the local model fits well with ground-truth black-box function, then good partitions and candidates can be reached with much fewer samples.

Bayesian Optimization Neural Architecture Search

Learning Space Partitions for Path Planning

2 code implementations NeurIPS 2021 Kevin Yang, Tianjun Zhang, Chris Cummins, Brandon Cui, Benoit Steiner, Linnan Wang, Joseph E. Gonzalez, Dan Klein, Yuandong Tian

Path planning, the problem of efficiently discovering high-reward trajectories, often requires optimizing a high-dimensional and multimodal reward function.

Multi-objective Optimization by Learning Space Partition

no code implementations ICLR 2022 Yiyang Zhao, Linnan Wang, Kevin Yang, Tianjun Zhang, Tian Guo, Yuandong Tian

In this paper, we propose LaMOO, a novel multi-objective optimizer that learns a model from observed samples to partition the search space and then focus on promising regions that are likely to contain a subset of the Pareto frontier.

Neural Architecture Search

Multi-objective Optimization by Learning Space Partitions

1 code implementation7 Oct 2021 Yiyang Zhao, Linnan Wang, Kevin Yang, Tianjun Zhang, Tian Guo, Yuandong Tian

In this paper, we propose LaMOO, a novel multi-objective optimizer that learns a model from observed samples to partition the search space and then focus on promising regions that are likely to contain a subset of the Pareto frontier.

Neural Architecture Search

Searching the Deployable Convolution Neural Networks for GPUs

no code implementations CVPR 2022 Linnan Wang, Chenhan Yu, Satish Salian, Slawomir Kierat, Szymon Migacz, Alex Fit Florea

To achieve this goal, we build a distributed NAS system to search on a novel search space that consists of prominent factors to impact latency and accuracy.

Neural Architecture Search

GPUNet: Searching the Deployable Convolution Neural Networks for GPUs

1 code implementation26 Apr 2022 Linnan Wang, Chenhan Yu, Satish Salian, Slawomir Kierat, Szymon Migacz, Alex Fit Florea

This paper intends to expedite the model customization with a model hub that contains the optimized models tiered by their inference latency using Neural Architecture Search (NAS).

Neural Architecture Search

Cannot find the paper you are looking for? You can Submit a new open access paper.