Published September 28, 2023 | Version v1
Report Open

Inference of ML models on Intel GPUs with SYCL and Intel OneAPI using SOFIE

  • 1. CERN openlab summer student
  • 2. CERN

Description

TMVA provides a fast inference system that takes an ONNX model as input and produces compilationready standalone C++ scripts as output which can be compiled and executed on CPU architectures. The idea of this project is to extend this capability to generate from the TMVA SOFIE model representation code that can be run also on Intel GPUs using both SYCL and Intel OneAPI libraries. These will allow for a more efficient evaluation of these models on Intel accelerator hardware.

Files

Inference_of_ML_Models_on_Intel_GPUs_with_SYCL_and_Intel_OneAPI_using_SOFIE.pdf