Wählen Sie Ihre Cookie-Einstellungen aus

Wir verwenden essentielle Cookies und ähnliche Tools, die für die Bereitstellung unserer Website und Services erforderlich sind. Wir verwenden Performance-Cookies, um anonyme Statistiken zu sammeln, damit wir verstehen können, wie Kunden unsere Website nutzen, und Verbesserungen vornehmen können. Essentielle Cookies können nicht deaktiviert werden, aber Sie können auf „Anpassen“ oder „Ablehnen“ klicken, um Performance-Cookies abzulehnen.

Wenn Sie damit einverstanden sind, verwenden AWS und zugelassene Drittanbieter auch Cookies, um nützliche Features der Website bereitzustellen, Ihre Präferenzen zu speichern und relevante Inhalte, einschließlich relevanter Werbung, anzuzeigen. Um alle nicht notwendigen Cookies zu akzeptieren oder abzulehnen, klicken Sie auf „Akzeptieren“ oder „Ablehnen“. Um detailliertere Entscheidungen zu treffen, klicken Sie auf „Anpassen“.

Troubleshooting tool calls

Fokusmodus
Troubleshooting tool calls - Amazon Nova
Diese Seite wurde nicht in Ihre Sprache übersetzt. Übersetzung anfragen

You might see different errors when working with tools and Amazon Nova models. Here are a few examples and tips to help you troubleshoot.

An error occurred (ModelErrorException) when calling the Converse operation: The system encountered an unexpected error during processing. Try your request again.

Validate your inference parameters - Amazon Nova models have more success calling tools when using greedy decoding. To enable greedy decoding, set the top P, top K, and temperature parameters to 1.

inferenceConfig={ "topP": 1, "temperature": 1 } additionalModelRequestFields={"inferenceConfig": {"topK": 1}}

Increase the maximum token count - It is common for tool outputs to require a large token output, ensure that the max tokens set is large enough to accommodate the expected return schema. If the model response is larger than your max token count, it will trigger an exception. You can set the maximum tokens in the inferenceConfig parameter:

inferenceConfig={ "maxTokens": 3000 }

Review the system prompt - To improve the accuracy of tool calling, Amazon Nova uses chain of thought reasoning when calling a tool. You will see this output in the response in <thinking> tags. We do not recommended trying to remove this functionality. Instead, we recommend that you drop the output if you do not need it in your application.

An error occurred (ModelErrorException) when calling the Converse operation: The system encountered an unexpected error during processing. Try your request again.

Validate your inference parameters - Amazon Nova models have more success calling tools when using greedy decoding. To enable greedy decoding, set the top P, top K, and temperature parameters to 1.

inferenceConfig={ "topP": 1, "temperature": 1 } additionalModelRequestFields={"inferenceConfig": {"topK": 1}}

Increase the maximum token count - It is common for tool outputs to require a large token output, ensure that the max tokens set is large enough to accommodate the expected return schema. If the model response is larger than your max token count, it will trigger an exception. You can set the maximum tokens in the inferenceConfig parameter:

inferenceConfig={ "maxTokens": 3000 }

Review the system prompt - To improve the accuracy of tool calling, Amazon Nova uses chain of thought reasoning when calling a tool. You will see this output in the response in <thinking> tags. We do not recommended trying to remove this functionality. Instead, we recommend that you drop the output if you do not need it in your application.

An error occurred (ValidationException) when calling the Converse operation

The passed tool configuration does not comply to the required conventions. Refer to the Converse API spec for specifications for all parameters.

An error occurred (ValidationException) when calling the Converse operation

The passed tool configuration does not comply to the required conventions. Refer to the Converse API spec for specifications for all parameters.

If it looks like tools are being ignored by the model, make sure you're following the correct JSON schema for the tool config:

  • Top level schema must of of type Object

  • Only three fields are supported in top-level Object - type (must be set to 'object'), properties, and required

  • Common unsupported fields at the top level are: $schema, description, title, and additionalProperties

If it looks like tools are being ignored by the model, make sure you're following the correct JSON schema for the tool config:

  • Top level schema must of of type Object

  • Only three fields are supported in top-level Object - type (must be set to 'object'), properties, and required

  • Common unsupported fields at the top level are: $schema, description, title, and additionalProperties

You might receive a successful response from Amazon Nova, but the model does not acknowledge that the needed tool is not present.

Validate your inference parameters - Amazon Nova models have more success calling tools when using greedy decoding. To enable greedy decoding, set the top P, top K, and temperature parameters to 1.

inferenceConfig={ "topP": 1, "temperature": 1 } additionalModelRequestFields={"inferenceConfig": {"topK": 1}}

Review the system prompt - To improve the accuracy of tool calling, Amazon Nova uses chain of thought reasoning when calling a tool. You will see this output in the response in <thinking> tags. We do not recommended trying to remove this functionality. Instead, we recommend that you drop the output if you do not need it in your application.

You might receive a successful response from Amazon Nova, but the model does not acknowledge that the needed tool is not present.

Validate your inference parameters - Amazon Nova models have more success calling tools when using greedy decoding. To enable greedy decoding, set the top P, top K, and temperature parameters to 1.

inferenceConfig={ "topP": 1, "temperature": 1 } additionalModelRequestFields={"inferenceConfig": {"topK": 1}}

Review the system prompt - To improve the accuracy of tool calling, Amazon Nova uses chain of thought reasoning when calling a tool. You will see this output in the response in <thinking> tags. We do not recommended trying to remove this functionality. Instead, we recommend that you drop the output if you do not need it in your application.

DatenschutzNutzungsbedingungen für die WebsiteCookie-Einstellungen
© 2025, Amazon Web Services, Inc. oder Tochtergesellschaften. Alle Rechte vorbehalten.