Fabio Baruffa


Contact  fabio.baruffa@remove-this.intel.com

Enhance Machine Learning Performance with Intel® Software tools

Date: Tuesday 27.08.2019


9:00 - 10:30
Morning session: 
Artificial Intelligence on Intel Hardware Platforms

  • Intel’s Hardware and Software directions for Artificial Intelligence (AI)
            Machine Learning (ML) and Deep Learning (DL)
  • Hardware Accelerated Deep Learning instructions and implementations
            DL Boost, VNNI instructions

13:30 - 11:00  Coffee break

11:00 - 12:30

  • Performance optimized Python
            Hands-on Labs with Python focus on Classical Machine Learning examples and algorithms

12.30 - 14:00 Lunch break

14:00 - 17:00
Afternoon session: Optimized Deep Learning Frameworks

  • Performance optimized Frameworks solutions from Intel
            Tensorflow, Keras, Caffe, Pytorch, BigDL and others
  • Performance acceleration with Intel MKL and Intel MKL-DNN for Deep Neural Network

15:30 - 16:00 Coffee break
16:00 - 17:00

Afternoon session: Distributed Deep Learning Solutions on HPC systems

  • Accelerate Training and Inference of Distributed solutions on HPC (MPI) environments using Xeon (x86)
            Distributed Tensorflow with Horovod
            Distributed Machine Learning with Daal4py


The use of data analytics techniques, such as Machine Learning and Deep Learning, has become the key for gaining insight into the incredible amount of data generated by scientific investigations (simulations and observations). Therefore it is crucial for the scientific community to incorporate these new tools in their workflows, in order to make full use of modern and upcoming data sets. In this tutorial we will provide an overview on the most known machine learning algorithms for supervised and unsupervised learning. With small example codes we show how to implement such algorithms using the Intel® Distribution for Python*, and which performance benefit can be obtained with minimal effort from the developer perspective.
Furthermore, the demand of using Deep Learning techniques in many scientific domains is rapidly emerging and the requirements for large compute and memory resources is increasing. One of the consequences is the need of the high-performance computing capability for processing and inferring the valuable information inherent in the data.

We cover also how to accelerate the training of deep neural networks with Tensorflow, thanks to the highly optimized Intel® Math Kernel Library (Intel® MKL). We also demonstrate techniques on how to leverage deep neural network training on multiple nodes on a HPC system.


Fabio Barffa is a senior software technical consulting engineer at Intel. He provides customer support in the high performance computing (HPC) area and artificial intelligence software solutions at large scale.

Prior at Intel, he has been working as HPC application specialist and developer in the largest supercomputing centers in Europe, mainly the Leibniz Supercomputing Center and the Max-Plank Computing and Data Facility in Munich, as well as Cineca in Italy. He has been involved in software development, analysis of scientific code and optimization for HPC systems. He holds a PhD in Physics from University of Regensburg for his research in the area of spintronics devices and quantum computing.




We use cookies in order to design and continuously improve our website for you. By continuing to use the website, you agree to the use of cookies. You can find further information on this in our privacy policy.