Файл:Accelerating Deep Convolutional Neural Networks Using Specialized Hardware CNN20Whitepaper.pdf

Материал из Материалы по машинному обучению
Перейти к: навигация, поиск
Accelerating_Deep_Convolutional_Neural_Networks_Using_Specialized_Hardware_CNN20Whitepaper.pdf(0 × 0 пикселей, размер файла: 665 КБ, MIME-тип: application/pdf)

Kalin Ovtcharov, Olatunji Ruwase, Joo-Young Kim, Jeremy Fowers, Karin Strauss, Eric S. Chung Microsoft Research



Recent breakthroughs in the development of multi-layer convolutional neural networks have led to stateof-the-art improvements in the accuracy of non-trivial recognition tasks such as large-category image classification and automatic speech recognition [1]. These many-layered neural networks are large, complex, and require substantial computing resources to train and evaluate [2]. Unfortunately, these demands come at an inopportune moment due to the recent slowing of gains in commodity processor performance.

Hardware specialization in the form of GPGPUs, FPGAs, and ASICs1 offers a promising path towards major leaps in processing capability while achieving high energy efficiency. To harness specialization, an effort is underway at Microsoft to accelerate Deep Convolutional Neural Networks (CNN) using servers augmented with FPGAs—similar to the hardware that is being integrated into some of Microsoft’s datacenters [3]. Initial efforts to implement a single-node CNN accelerator on a mid-range FPGA show significant promise, resulting in respectable performance relative to prior FPGA designs and high-end GPGPUs, at a fraction of the power. In the future, combining multiple FPGAs over a low-latency communication fabric offers further opportunity to train and evaluate models of unprecedented size and quality.

История файла

Нажмите на дату/время, чтобы просмотреть, как тогда выглядел файл.

текущий12:15, 28 декабря 20160 × 0 (665 КБ)Slikos (обсуждение | вклад)
  • Вы не можете перезаписать этот файл.

Следующая 1 страница ссылается на данный файл: