深度學習(Deep Learning, DL)在許多領域已取得突破性發展,然而高效能運算 是實現人工智能應用的關鍵。先前研究發現深度神經網路(Deep Neural Network, DNN)中有許多零或非常接近零的權重(Weight)。在深度學習硬體加速器(Deep Learning Accelerator)設計時,將這些權重刪除可以大幅提升運算效能,即為權 重剪枝(Weight Pruning)。但是即使相同的神經網路模型在不同的應用下,模型 的參數也會有所差異。這些差異會造成硬體設計上的不同,而有不同的工作排程 (Job Scheduling)需求。為了縮小硬體設計時間成本,以自動化方法分析並歸納 出適當的工作排程顯得十分重要。我們以權重剪枝(Weight Pruning)為基礎,實 現硬體資源的最佳化問題,並且探討硬體架構下的效能指標,以及提出工作排程 問題(Job Scheduling Problem)的解決方法。
Application of Deep Learning (DL) has achieved a huge breakthrough in many fields. Many innovative DL applications require efficient computation. Previous work has found that neural networks of DL have many zero and near to zero weights. These weights can be deleted, i.e. weight pruning, to improve computation efficiency of the deep neural networks (DNNs). Also, different neural network model varies from one to one. This leads to the difficulty in the hardware design and job scheduling. Thus, using an automation technology to analyze and support the hardware accelerator design flow may be helpful. In this work, we study an optimization problem based on weight pruning technology, discuss the performance of hardware design, and propose a solution to a job scheduling problem.