Interface InferenceComponentComputeResourceRequirements.Builder

  • Method Details

    • numberOfCpuCoresRequired

      InferenceComponentComputeResourceRequirements.Builder numberOfCpuCoresRequired(Float numberOfCpuCoresRequired)

      The number of CPU cores to allocate to run a model that you assign to an inference component.

      Parameters:
      numberOfCpuCoresRequired - The number of CPU cores to allocate to run a model that you assign to an inference component.
      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • numberOfAcceleratorDevicesRequired

      InferenceComponentComputeResourceRequirements.Builder numberOfAcceleratorDevicesRequired(Float numberOfAcceleratorDevicesRequired)

      The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and Amazon Web Services Inferentia.

      Parameters:
      numberOfAcceleratorDevicesRequired - The number of accelerators to allocate to run a model that you assign to an inference component. Accelerators include GPUs and Amazon Web Services Inferentia.
      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • minMemoryRequiredInMb

      InferenceComponentComputeResourceRequirements.Builder minMemoryRequiredInMb(Integer minMemoryRequiredInMb)

      The minimum MB of memory to allocate to run a model that you assign to an inference component.

      Parameters:
      minMemoryRequiredInMb - The minimum MB of memory to allocate to run a model that you assign to an inference component.
      Returns:
      Returns a reference to this object so that method calls can be chained together.
    • maxMemoryRequiredInMb

      InferenceComponentComputeResourceRequirements.Builder maxMemoryRequiredInMb(Integer maxMemoryRequiredInMb)

      The maximum MB of memory to allocate to run a model that you assign to an inference component.

      Parameters:
      maxMemoryRequiredInMb - The maximum MB of memory to allocate to run a model that you assign to an inference component.
      Returns:
      Returns a reference to this object so that method calls can be chained together.