InferenceConfiguration
Base inference parameters to pass to a model in a call to Converse or ConverseStream. For more information, see Inference parameters for foundation models.
If you need to pass additional parameters that the model supports, use the additionalModelRequestFields
request field in the call to Converse
or ConverseStream
. For more information, see Model parameters.
Types
Properties
The maximum number of tokens to allow in the generated response. The default value is the maximum allowed value for the model that you are using. For more information, see Inference parameters for foundation models.
A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response.
The likelihood of the model selecting higher-probability options while generating a response. A lower value makes the model more likely to choose higher-probability options, while a higher value makes the model more likely to choose lower-probability options.