Search Results for author: Jiacheng Yang

Found 11 papers, 6 papers with code

MAgent: A Many-Agent Reinforcement Learning Platform for Artificial Collective Intelligence

3 code implementations2 Dec 2017 Lianmin Zheng, Jiacheng Yang, Han Cai, Wei-Nan Zhang, Jun Wang, Yong Yu

Unlike previous research platforms on single or multi-agent reinforcement learning, MAgent focuses on supporting the tasks and the applications that require hundreds to millions of agents.

Multi-agent Reinforcement Learning reinforcement-learning +1

Learning to Design Circuits

no code implementations5 Dec 2018 Hanrui Wang, Jiacheng Yang, Hae-Seung Lee, Song Han

We propose Learning to Design Circuits (L2DC) to leverage reinforcement learning that learns to efficiently generate new circuits data and to optimize circuits.

Bayesian Optimization

Towards Making the Most of BERT in Neural Machine Translation

2 code implementations15 Aug 2019 Jiacheng Yang, Mingxuan Wang, Hao Zhou, Chengqi Zhao, Yong Yu, Wei-Nan Zhang, Lei LI

Our experiments in machine translation show CTNMT gains of up to 3 BLEU score on the WMT14 English-German language pair which even surpasses the previous state-of-the-art pre-training aided NMT by 1. 4 BLEU score.

Machine Translation NMT +2

AOG-LSTM: An adaptive attention neural network for visual storytelling

no code implementations Neurocomputing 2023 Hanqing Liu, Jiacheng Yang, Chia-Hao Chang, Wei Wang, Hai-Tao Zheng, Yong Jiang, Hui Wang, Rui Xie, and Wei Wu

Moreover, the existing method of alleviating error accumulation based on replacing reference words does not take into account the different effects of each word.

Visual Storytelling

Minuet: Accelerating 3D Sparse Convolutions on GPUs

1 code implementation1 Dec 2023 Jiacheng Yang, Christina Giannoula, Jun Wu, Mostafa Elhoushi, James Gleeson, Gennady Pekhimenko

Minuet proposes to (i) replace the hash tables used in the Map step with a novel segmented sorting double-traversed binary search algorithm that highly utilizes the on-chip memory hierarchy of GPUs, (ii) use a lightweight scheme to autotune the tile size in the Gather and Scatter operations of the GMaS step, such that to adapt the execution to the particular characteristics of each SC layer, dataset, and GPU architecture, and (iii) employ a padding-efficient GEMM grouping approach that reduces both memory padding and kernel launching overheads.

Accelerating Graph Neural Networks on Real Processing-In-Memory Systems

no code implementations26 Feb 2024 Christina Giannoula, Peiming Yang, Ivan Fernandez Vega, Jiacheng Yang, Yu Xin Li, Juan Gomez Luna, Mohammad Sadrosadati, Onur Mutlu, Gennady Pekhimenko

Graph Neural Network (GNN) execution involves both compute-intensive and memory-intensive kernels, the latter dominates the total time, being significantly bottlenecked by data movement between memory and processors.

Cannot find the paper you are looking for? You can Submit a new open access paper.