no code implementations • 19 Oct 2023 • Youngkyu Lee, Jongho Park, Chang-Ock Lee
The performance of neural networks has been significantly improved by increasing the number of channels in convolutional layers.
no code implementations • 11 Oct 2021 • Youngkyu Lee, Jongho Park, Chang-Ock Lee
In this paper, we propose a new convolution methodology called ``two-level'' group convolution that is robust with respect to the increase of the number of groups and suitable for multi-GPU parallel computation.
no code implementations • 16 Mar 2021 • Chang-Ock Lee, Youngkyu Lee, Jongho Park
We observe that layers of DNN can be interpreted as the time step of a time-dependent problem and can be parallelized by emulating a parallel-in-time algorithm called parareal.