[2023.04.27(Thu.)]Artificial Intelligence & AI Convergence Network Colloquium | |||||||||
작성자 | 김성민 | 분류 | 행사 | ||||||
---|---|---|---|---|---|---|---|---|---|
공지대상 |
![]() |
등록일 | 2023-04-19 | 마감일 | 2023-04-27 | 마감여부 | ![]() |
조회수 | 301 |
공지부서 | 소프트웨어융합대학교학팀 | ||||||||
< Artificial Intelligence & AI Convergence Network Colloquium >![]() ![]() ![]() ![]() ![]()
acceleration. While they improve the performance, GPUs are underutilized during the training. This paper proposes
out-of-order (ooo) back-prop, an effective scheduling technique for neural network training. By exploiting the
dependencies of gradient computations, ooo backprop enables to reorder their executions to make the most of
the GPU resources. We show that the GPU utilization in single- and multi-GPU training can be commonly
improved by applying ooo backprop and prioritizing critical operations. We propose three scheduling
algorithms based on ooo backprop. For single-GPU training, we schedule with multi-stream ooo computation
to mask the kernel launch overhead. In data-parallel training, we reorder the gradient computations to
maximize the overlapping of computation and parameter communication; in pipeline-parallel training, we
prioritize critical gradient computations to reduce the pipeline stalls. We evaluate our optimizations with twelve
neural networks and five public datasets. Compared to the respective state of the art training systems, our
algorithms improve the training throughput by 1.03--1.58× for single-GPU training, by 1.10--1.27× for data-
parallel training, and by 1.41--1.99× for pipeline-parallel training.
![]()
systems and big data systems.
|
이전 | [2023.05.04(Thu.)] AI Convergence Network & Artificial Intelligence Colloquium |
---|---|
다음 | [네이버 커넥트재단] 부스트코스 코칭스터디 - 봄 과정 모집 안내 |