Withdraw
Loading…
ACCELERATING CONVOLUTION IN DEEP NEURAL NETWORKS ON ACAPI-BASED FPGA
Ma, Yuan
Loading…
Permalink
https://hdl.handle.net/2142/125074
Description
- Title
- ACCELERATING CONVOLUTION IN DEEP NEURAL NETWORKS ON ACAPI-BASED FPGA
- Author(s)
- Ma, Yuan
- Issue Date
- 2020-05-01
- Keyword(s)
- Convolutional neural networks (CNNs), field-programmable gate array (FPGA)
- Abstract
- Convolutional neural networks (CNNs) have emerged as a crucial part in many applications ranging from self-driving cars to voice-activated assistants. Numerous cloud computing providers, such as Amazon (AWS), IBM (SoftLayer), and Microsoft (Azure), choose to use heterogeneous computing systems to off-load the CNN computations from the CPU to a dedicated hardware since such hardware provides significant improvements in both computing throughput and energy savings. In this senior thesis, the author presents a weight- stationary systolic convolution kernel design for a field-programmable gate array (FPGA) and its implementation targeting Nallantech 250s+ card that is enabled for the coherent accelerator processor interface (CAPI). CAPI is an interface for heterogeneous systems that allows accelerators to access I/O devices as CPU peers. Systolic array architecture has shown advantages for accelerators in tasks that involve vector dot-product calculations, such as matrix multiplication and convolution. The proposed hardware is synthesized by the High-Level Synthesis tool targeting Kintex UltraScale+ XCKU15P FPGA and can provide high throughput (4.6×CPU) computations for real-time applications. In the future, the author plans to extend this kernel to a complete CNN by supporting CPU-FPGA task partitioning using the coherent memory space enabled by CAPI.
- Type of Resource
- text
- Language
- eng
Owning Collections
Senior Theses - Electrical and Computer Engineering PRIMARY
The best of ECE undergraduate researchManage Files
Loading…
Edit Collection Membership
Loading…
Edit Metadata
Loading…
Edit Properties
Loading…
Embargoes
Loading…