inferenceAmiVersion
Specifies an option from a collection of preconfigured Amazon Machine Image (AMI) images. Each image is configured by Amazon Web Services with a set of software and driver versions. Amazon Web Services optimizes these configurations for different machine learning workloads.
By selecting an AMI version, you can ensure that your inference environment is compatible with specific software requirements, such as CUDA driver versions, Linux kernel versions, or Amazon Web Services Neuron driver versions.
The AMI version names, and their configurations, are the following:
al2-ami-sagemaker-inference-gpu-2
Accelerator: GPU
NVIDIA driver version: 535
CUDA version: 12.2
al2-ami-sagemaker-inference-gpu-2-1
Accelerator: GPU
NVIDIA driver version: 535
CUDA version: 12.2
NVIDIA Container Toolkit with disabled CUDA-compat mounting
al2-ami-sagemaker-inference-gpu-3-1
Accelerator: GPU
NVIDIA driver version: 550
CUDA version: 12.4
NVIDIA Container Toolkit with disabled CUDA-compat mounting
al2-ami-sagemaker-inference-neuron-2
Accelerator: Inferentia2 and Trainium
Neuron driver version: 2.19