現今,動態電壓調配機制(Dynamic Voltage Scaling)被廣泛地用於解決時下電路設計之電源耗電問題。決定每個運行時間下的最佳操作電壓需考量工作負荷量變異、製程變異以及環境因素變異。在晶片設計階段,上述變異議題的預測實屬困難,故過去有許多文獻提出基於增強式學習法的動態電壓調配機制。其中,電壓緩降法則能基於臨界機率容忍時序錯誤,進而達到電源節省。然而,並沒有動態電壓調配機制的研究同時考慮電壓緩降法則。在本論文中,我們提出結合「Q適應學習法(Q-Learning Algorithm)」的動態電壓緩降調配機制,透過關鍵電路路徑監控器和基於時序容錯率的電壓預測,同時考慮溫度、電壓以及工作負荷量的晶片變異議題,達到效能最佳化及電源節省的目的。為評估其機制之優良,我們比較了兩種決定型動態電壓調配機制:步進式機制、統計模型式機制。相較於此兩種傳統機制,實驗結果顯示,針對三種45奈米工業製程晶片設計,我們所提出的Q適應學習法動態電壓緩降調配機制(搭配0.01的時序容忍機率值)可分別節省至多83.9%、29.1%的電源耗損。此研究為首度探討在電壓緩降法則下,基於增強式學習法的電壓緩降調配機制。
Dynamic voltage scaling (DVS) has been widely used to suppress power consumption in modern designs. The decision of optimal operating voltage at runtime should consider the variations in workload, process as well as environment. As these variations are hard to predict accurately at design time, various reinforcement learning based DVS schemes have been proposed in the literature. However, none of them can be readily applied to designs with graceful degradation, where timing errors are allowed with bounded probability to trade for further power reduction. In this thesis, we propose a Q-learning based DVS scheme dedicated to the designs with graceful degradation. We compare it with two deterministic DVS schemes, i.e., a stepping based scheme and a statistical modeling based scheme. Experimental results on three 45nm industrial designs show that the proposed Q-learning based scheme can achieve up to 83.9% and 29.1% power reduction respectively with 0.01 timing error probability bound. This is the first in-depth work to explore reinforcement learning based DVS schemes for designs with graceful degradation.