Ä¿¹Â´ÏƼ

¾ÆÁÖ´ëÇб³ ¼ÒÇÁÆ®¿þ¾îÇаú¿¡ ¿À½Å °ÍÀ» ȯ¿µÇÕ´Ï´Ù.

Çаú°øÁö»çÇ×

[2023.04.27(¸ñ)] Artificial Intelligence & AI Convergence Network Colloquium °³ÃÖ ¾È³»
  • ±Û¾´ÀÌ °ü¸®ÀÚ
  • ÀÛ¼ºÀÏ 2023-04-19 14:32:05
  • Á¶È¸¼ö 171

´ëÇпø ÀΰøÁö´ÉÇаú & AIÀ¶ÇÕ³×Æ®¿öÅ©Çаú¿¡¼­´Â Artificial Intelligence & AI Convergence Network ColloquiumÀ» 

4¿ù 27ÀÏ(¸ñ) ¿ÀÈÄ 1½Ã 30ºÐ¿¡ °³ÃÖÇÏ°íÀÚ ÇÏ¿À´Ï ¸¹Àº Âü¿© ºÎŹµå¸³´Ï´Ù.
 
¢º When : 2023³â 4¿ù 27ÀÏ(¸ñ) ¿ÀÈÄ 1½Ã 30ºÐ
¢º Where : ÆÈ´Þ°ü 407È£ 
¢º Speaker : ¼­Áö¿ø ±³¼ö(ÇѾç´ëÇб³ ÄÄÇ»ÅͼÒÇÁÆ®¿þ¾îÇкÎ)
¢º Title : Out-Of-Order BackProp: An Effective Scheduling Technique for Deep Learning

¢º Abstract : Neural network training requires a large amount of computation and thus GPUs are often used for the acceleration. While they improve the performance, GPUs are underutilized during the training. This paper proposes out-of-order (ooo) back-prop, an effective scheduling technique for neural network training. By exploiting the dependencies of gradient computations, ooo backprop enables to reorder their executions to make the most of the GPU resources. We show that the GPU utilization in single- and multi-GPU training can be commonly improved by applying ooo backprop and prioritizing critical operations. We propose three scheduling algorithms based on ooo backprop. For single-GPU training, we schedule with multi-stream ooo computation to mask the kernel launch overhead. In data-parallel training, we reorder the gradient computations to maximize the overlapping of computation and parameter communication; in pipeline-parallel training, we prioritize critical gradient computations to reduce the pipeline stalls. We evaluate our optimizations with twelve neural networks and five public datasets. Compared to the respective state of the art training systems, our algorithms improve the training throughput by 1.03--1.58¡¿ for single-GPU training, by 1.10--1.27¡¿ for data-parallel training, and by 1.41--1.99¡¿ for pipeline-parallel training.


¢º Bio : Jiwon Seo is an assistant professor at the department of computer science in Hanyang university, Korea. He received his PhD in electrical engineering from Stanford in 2015. His research interests include machine learning systems and big data systems.


¢º Host : ¼ÒÇÁÆ®¿þ¾îÇаú ¾ÈÁ¤¼· ±³¼ö(jsahn@ajou.ac.kr)


 
¸ñ·Ï





ÀÌÀü±Û ¹è¸®¾îÇÁ¸® ¾Û °³¹ß ÄÜÅ×½ºÆ®
´ÙÀ½±Û Á¦5ȸ 2023 ¿¬±¸°³¹ßƯ±¸ AI SPARK 縰Áö