InferenceConfiguration
Base inference parameters to pass to a model in a call to Converse or ConverseStream. For more information, see Inference parameters for foundation models.
If you need to pass additional parameters that the model
supports, use the additionalModelRequestFields
request field in the call to Converse
or ConverseStream
.
For more information, see Model parameters.
Contents
 maxTokens

The maximum number of tokens to allow in the generated response. The default value is the maximum allowed value for the model that you are using. For more information, see Inference parameters for foundation models.
Type: Integer
Valid Range: Minimum value of 1.
Required: No
 stopSequences

A list of stop sequences. A stop sequence is a sequence of characters that causes the model to stop generating the response.
Type: Array of strings
Array Members: Minimum number of 0 items. Maximum number of 4 items.
Length Constraints: Minimum length of 1.
Required: No
 temperature

The likelihood of the model selecting higherprobability options while generating a response. A lower value makes the model more likely to choose higherprobability options, while a higher value makes the model more likely to choose lowerprobability options.
The default value is the default value for the model that you are using. For more information, see Inference parameters for foundation models.
Type: Float
Valid Range: Minimum value of 0. Maximum value of 1.
Required: No
 topP

The percentage of mostlikely candidates that the model considers for the next token. For example, if you choose a value of 0.8 for
topP
, the model selects from the top 80% of the probability distribution of tokens that could be next in the sequence.The default value is the default value for the model that you are using. For more information, see Inference parameters for foundation models.
Type: Float
Valid Range: Minimum value of 0. Maximum value of 1.
Required: No
See Also
For more information about using this API in one of the languagespecific AWS SDKs, see the following: