Class: Aws::Bedrock::Types::GuardrailManagedWords
- Inherits:
-
Struct
- Object
- Struct
- Aws::Bedrock::Types::GuardrailManagedWords
- Defined in:
- gems/aws-sdk-bedrock/lib/aws-sdk-bedrock/types.rb
Overview
The managed word list that was configured for the guardrail. (This is a list of words that are pre-defined and managed by guardrails only.)
Constant Summary collapse
- SENSITIVE =
[:input_action, :output_action]
Instance Attribute Summary collapse
-
#input_action ⇒ String
The action to take when harmful content is detected in the input.
-
#input_enabled ⇒ Boolean
Indicates whether guardrail evaluation is enabled on the input.
-
#output_action ⇒ String
The action to take when harmful content is detected in the output.
-
#output_enabled ⇒ Boolean
Indicates whether guardrail evaluation is enabled on the output.
-
#type ⇒ String
ManagedWords$type The managed word type that was configured for the guardrail.
Instance Attribute Details
#input_action ⇒ String
The action to take when harmful content is detected in the input. Supported values include:
BLOCK
– Block the content and replace it with blocked messaging.NONE
– Take no action but return detection information in the trace response.
4070 4071 4072 4073 4074 4075 4076 4077 4078 |
# File 'gems/aws-sdk-bedrock/lib/aws-sdk-bedrock/types.rb', line 4070 class GuardrailManagedWords < Struct.new( :type, :input_action, :output_action, :input_enabled, :output_enabled) SENSITIVE = [:input_action, :output_action] include Aws::Structure end |
#input_enabled ⇒ Boolean
Indicates whether guardrail evaluation is enabled on the input. When disabled, you aren't charged for the evaluation. The evaluation doesn't appear in the response.
4070 4071 4072 4073 4074 4075 4076 4077 4078 |
# File 'gems/aws-sdk-bedrock/lib/aws-sdk-bedrock/types.rb', line 4070 class GuardrailManagedWords < Struct.new( :type, :input_action, :output_action, :input_enabled, :output_enabled) SENSITIVE = [:input_action, :output_action] include Aws::Structure end |
#output_action ⇒ String
The action to take when harmful content is detected in the output. Supported values include:
BLOCK
– Block the content and replace it with blocked messaging.NONE
– Take no action but return detection information in the trace response.
4070 4071 4072 4073 4074 4075 4076 4077 4078 |
# File 'gems/aws-sdk-bedrock/lib/aws-sdk-bedrock/types.rb', line 4070 class GuardrailManagedWords < Struct.new( :type, :input_action, :output_action, :input_enabled, :output_enabled) SENSITIVE = [:input_action, :output_action] include Aws::Structure end |
#output_enabled ⇒ Boolean
Indicates whether guardrail evaluation is enabled on the output. When disabled, you aren't charged for the evaluation. The evaluation doesn't appear in the response.
4070 4071 4072 4073 4074 4075 4076 4077 4078 |
# File 'gems/aws-sdk-bedrock/lib/aws-sdk-bedrock/types.rb', line 4070 class GuardrailManagedWords < Struct.new( :type, :input_action, :output_action, :input_enabled, :output_enabled) SENSITIVE = [:input_action, :output_action] include Aws::Structure end |
#type ⇒ String
ManagedWords$type The managed word type that was configured for the guardrail. (For now, we only offer profanity word list)
4070 4071 4072 4073 4074 4075 4076 4077 4078 |
# File 'gems/aws-sdk-bedrock/lib/aws-sdk-bedrock/types.rb', line 4070 class GuardrailManagedWords < Struct.new( :type, :input_action, :output_action, :input_enabled, :output_enabled) SENSITIVE = [:input_action, :output_action] include Aws::Structure end |