Published March 8, 2017 | Version v1
Working paper Open

Performance Portability of OpenCL with Application to Neural Networks

  • 1. High Performance Computing Section, IT Dept., NTNU

Contributors

  • 1. Faculty of Medicine, Kavli Institute for Systems Neuroscience / Centre for Neural Computation, NTNU

Description

This whitepaper investigates the parallel performance of a sample application that implements an approximate expectation-maximization method for inferring the network structure and time varying states of a hidden population within the framework of the kinetic Ising model. The size of networks that can yield informative results can be made arbitrarily large, and the long-running computational demand is highly localized, making the application a strong candidate for future exascale platforms.
Previous investigations using OpenMP on the Intel Xeon Phi architecture have suggested that the class of accelerator unit may play a significant part in attainable application performance. An OpenCL parallelization enables experiments with a variety of accelerator units. We examine how this programming model affects the performance of a portable implementation, and use it to compare accelerator technologies in terms of their suitability for future extreme-scale computations.

Files

WP231.pdf

Files (645.6 kB)

Name Size Download all
md5:6955f8bde342ad6e2fbaaf3f7c2d0910
645.6 kB Preview Download

Additional details

Funding

PRACE-4IP – PRACE 4th Implementation Phase Project 653838
European Commission