A Hardware Acceleration Platform for AI-Based Inference at the Edge
Creators
- 1. Hellenic Mediterranean University, Heraklion, Greece
- 2. National Institute of Telecommunications, Warsaw, Poland
- 3. University of Nicosia, Nicosia, Cyprus
Description
Machine learning (ML) algorithms are already transforming the way data are collected and processed in the data center, where some form of AI has permeated most areas of computing. The integration of AI algorithms at the edge is the next logical step which is already under investigation. However, harnessing such algorithms at the edge will require more computing power than what current platforms offer. In this paper, we present an FPGA system-on-chip-based architecture that supports the acceleration of ML algorithms in an edge environment. The system supports dynamic deployment of ML functions driven either locally or remotely, thus achieving a remarkable degree of flexibility . We demonstrate the efficacy of this architecture by executing a version of the well-known YOLO classifier which demonstrates competitive performance while requiring a reasonable amount of resources on the device.
Files
Karras2020_Article_AHardwareAccelerationPlatformF.pdf
Files
(699.9 kB)
Name | Size | Download all |
---|---|---|
md5:39659dd5cfd12acf4619aa1f2ff039a5
|
699.9 kB | Preview Download |