Show simple item record

dc.identifier.urihttp://hdl.handle.net/11401/77267
dc.description.sponsorshipThis work is sponsored by the Stony Brook University Graduate School in compliance with the requirements for completion of degree.en_US
dc.formatMonograph
dc.format.mediumElectronic Resourceen_US
dc.language.isoen_US
dc.publisherThe Graduate School, Stony Brook University: Stony Brook, NY.
dc.typeThesis
dcterms.abstractDeep convolutional neural networks (CNNs) are rapidly becoming the domi-nant approach to computer vision and a major component of many other pervasivemachine learning tasks, such as speech recognition, natural language processing,and fraud detection. As research and development of CNNs progresses, the size ofthe networks grows, leading to large increases in the computation and bandwidthrequired to evaluate these networks. Typical CNNs in use today already exceedthe capabilities of general-purpose CPUs, resulting in rapid adoption and activeresearch of CNN hardware accelerators such as GPUs, FPGAs, and ASICs. Inthis work, we develop a novel CNN accelerator architecture and design method-ology that breaks away from the commonly accepted practice of processing thenetworks layer by layer. By modifying the order in which the original input dataare brought on chip, changing it to a pyramid-shaped multi-layer sliding window,our architecture enables effective on-chip caching during CNN evaluation. Thecaching in turn reduces the off-chip memory bandwidth requirements, which is aprimary bottleneck in many CNN environments.
dcterms.available2017-09-20T16:52:19Z
dcterms.contributorFerdman, Michaelen_US
dcterms.contributorHonarmand, Nimaen_US
dcterms.contributorSamaras, Dimitrisen_US
dcterms.contributorBerg, Alex.en_US
dcterms.creatorAlwani, Manoj
dcterms.dateAccepted2017-09-20T16:52:19Z
dcterms.dateSubmitted2017-09-20T16:52:19Z
dcterms.descriptionDepartment of Computer Science.en_US
dcterms.extent54 pg.en_US
dcterms.formatMonograph
dcterms.formatApplication/PDFen_US
dcterms.identifierhttp://hdl.handle.net/11401/77267
dcterms.issued2015-12-01
dcterms.languageen_US
dcterms.provenanceMade available in DSpace on 2017-09-20T16:52:19Z (GMT). No. of bitstreams: 1 Alwani_grad.sunysb_0771M_12631.pdf: 777045 bytes, checksum: 666c7e7605c6fc0d8cee47b2ba640e3d (MD5) Previous issue date: 1en
dcterms.publisherThe Graduate School, Stony Brook University: Stony Brook, NY.
dcterms.subjectComputer science
dcterms.subjectConvolutional Neural Network, Deep Learning, FPGA, High Level Synthesis
dcterms.titleFused Convolutional Neural Network Accelerators
dcterms.typeThesis


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record