卷積神經網絡(convolution neural network)已廣泛應用於電腦視覺任務(computer vision tasks)的領域,然而標準的神經網絡需要大量的運算和參數,這對嵌入式設備而言是個挑戰。因此前人提出了一種新穎的神經網路架構MobileNets,MobileNets採用深度可分離卷積(depthwise separable convolution)代替標準卷積,使其運算量和參數大幅減少且精度損失有限。而MobileNets中主要有兩種不同的計算方法pointwise和depthwise,如果用傳統的加速器來計算這兩種不同的運算,會因為運算參數和方式的不同而造成硬體利用率低下。除此之外,常見降低神經網路計算負擔的方法還有量化(quantization),其透過減少位寬(bit width)或採用不同位寬來降低計算負荷,但如果用相同精度的硬體來計算不同位寬的資料,則無法有效的節省運算時間。基於MobileNets和量化網路,本文提出了一種可以有效計算量MobileNets的新型計算架構,以達到加速運算和節省面積的效果。;Convolution neural network (CNN) has been widely applied in the fields of computer vision applications. However, conventional neural network computations require a lot of operations and parameters, which becomes a challenge for embedded devices. MobileNets, a novel CNN which adopts depthwise separable convolution to replace the standard convolution, has substantially reduced operations and parameters with only limited loss in accuracy. There are mainly two different calculation methods in MobileNets, pointwise and depthwise. If the same accelerator is used to perform these two different operations, the accelerator may not able to be fully exploited due to different operation parameters. In addition, there are some methods for neural network quantization, which limit the bit width to reduce computing energy and parameters. If the same precision hardware is used to calculate quantized operations, the maximum benefit cannot be achieved. Therefore, A novel architecture which can effectively calculate quantized MobileNets is proposed in this thesis.