
- Clarify reference to Shared Memory section in CUDA programming guide - Update cuBLAS interface version description - Add hyperlink to Shared Memory documentation
matrixMul - Matrix Multiplication (CUDA Runtime API Version)
Description
This sample implements matrix multiplication and is exactly the same as the second example of the Shared Memory section of the programming guide. It has been written for clarity of exposition to illustrate various CUDA programming principles, not with the goal of providing the most performant generic kernel for matrix multiplication. To illustrate GPU performance for matrix multiply, this sample also shows how to use the CUDA 4.0+ interface for cuBLAS to demonstrate high-performance performance for matrix multiplication.
Key Concepts
CUDA Runtime API, Linear Algebra
Supported SM Architectures
SM 5.0 SM 5.2 SM 5.3 SM 6.0 SM 6.1 SM 7.0 SM 7.2 SM 7.5 SM 8.0 SM 8.6 SM 8.7 SM 8.9 SM 9.0
Supported OSes
Linux, Windows
Supported CPU Architecture
x86_64, armv7l, aarch64
CUDA APIs involved
CUDA Runtime API
cudaStreamCreateWithFlags, cudaProfilerStop, cudaMalloc, cudaFree, cudaMallocHost, cudaProfilerStart, cudaEventSynchronize, cudaEventRecord, cudaFreeHost, cudaStreamSynchronize, cudaEventDestroy, cudaEventElapsedTime, cudaMemcpyAsync, cudaEventCreate
Prerequisites
Download and install the CUDA Toolkit 12.5 for your corresponding platform.