Scaling CUDA C++ Applications to Multiple Nodes (SCCAMN)

 

Resumen del Curso

Present-day high-performance computing (HPC) and deep learning applications benefit from, and even require, cluster-scale GPU compute power. Writing CUDA applications that can correctly and efficiently utilize GPUs across a cluster requires a distinct set of skills. In this workshop, you’ll learn the tools and techniques needed to write CUDA C++ applications that can scale efficiently to clusters of NVIDIA GPUs.

Prerrequisitos

Intermediate experience writing CUDA C/C++ applications.

Suggested materials to satisfy the prerequisites:

  • Fundamentals of Accelerated Computing with CUDA C/C++
  • Accelerating CUDA C++ Applications with Multiple GPUs
  • Accelerating CUDA C++ Applications with Concurrent Streams
  • Scaling Workloads Across Multiple GPUs with CUDA C++

Objetivos del curso

By participating in this workshop, you’ll:

  • Learn several methods for writing multi-GPU CUDA C++ applications
  • Use a variety of multi-GPU communication patterns and understand their tradeoffs
  • Write portable, scalable CUDA code with the single-program multiple-data (SPMD) paradigm using CUDA-aware MPI and NVSHMEM
  • Improve multi-GPU SPMD code with NVSHMEM’s symmetric memory model and its ability to perform GPU-initiated data transfers
  • Get practice with common multi-GPU coding paradigms like domain decomposition and halo exchanges

Follow On Courses

Precios & Delivery methods

Entrenamiento en línea

Duración
1 día

Precio
  • Consulta precio y disponibilidad
Classroom training

Duración
1 día

Precio
  • Consulta precio y disponibilidad

Calendario

Por el momento no hay fechas programadas para este curso