14.7 A 288µW programmable deep-learning processor with 270KB on-chip weight storage using non-uniform memory hierarchy for mobile intelligence.

ISSCC(2017)

引用 126|浏览332
暂无评分
摘要
Deep learning has proven to be a powerful tool for a wide range of applications, such as speech recognition and object detection, among others. Recently there has been increased interest in deep learning for mobile IoT [1] to enable intelligence at the edge and shield the cloud from a deluge of data by only forwarding meaningful events. This hierarchical intelligence thereby enhances radio bandwidth and power efficiency by trading-off computation and communication at edge devices. Since many mobile applications are “always-on” (e.g., voice commands), low power is a critical design constraint. However, prior works have focused on high performance reconfigurable processors [2–3] optimized for large-scale deep neural networks (DNNs) that consume u003e50mW. Off-chip weight storage in DRAM is also common in the prior works [2–3], which implies significant additional power consumption due to intensive off-chip data movement.
更多
查看译文
关键词
programmable deep-learning processor,on-chip weight storage,nonuniform memory hierarchy,mobile intelligence,deep learning,speech recognition,object detection,Internet of Things,mobile IoT,hierarchical intelligence,radio bandwidth,power efficiency,edge devices,mobile applications,reconfigurable processors,deep neural networks,DNN,DRAM,power consumption,power 288 muW
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要