You are viewing documentation for version 2 of the AWS SDK for Ruby. Version 3 documentation can be found here.

Class: Aws::Rekognition::Types::IndexFacesRequest

Inherits:
Struct
  • Object
show all
Defined in:
(unknown)

Overview

Note:

When passing IndexFacesRequest as input to an Aws::Client method, you can use a vanilla Hash:

{
  collection_id: "CollectionId", # required
  image: { # required
    bytes: "data",
    s3_object: {
      bucket: "S3Bucket",
      name: "S3ObjectName",
      version: "S3ObjectVersion",
    },
  },
  external_image_id: "ExternalImageId",
  detection_attributes: ["DEFAULT"], # accepts DEFAULT, ALL
  max_faces: 1,
  quality_filter: "NONE", # accepts NONE, AUTO
}

Instance Attribute Summary collapse

Instance Attribute Details

#collection_idString

The ID of an existing collection to which you want to add the faces that are detected in the input images.

Returns:

  • (String)

    The ID of an existing collection to which you want to add the faces that are detected in the input images.

#detection_attributesArray<String>

An array of facial attributes that you want to be returned. This can be the default list of attributes or all attributes. If you don\'t specify a value for Attributes or if you specify ["DEFAULT"], the API returns the following subset of facial attributes: BoundingBox, Confidence, Pose, Quality, and Landmarks. If you provide ["ALL"], all facial attributes are returned, but the operation takes longer to complete.

If you provide both, ["ALL", "DEFAULT"], the service uses a logical AND operator to determine which attributes to return (in this case, all attributes).

Returns:

  • (Array<String>)

    An array of facial attributes that you want to be returned.

#external_image_idString

The ID you want to assign to all the faces detected in the image.

Returns:

  • (String)

    The ID you want to assign to all the faces detected in the image.

#imageTypes::Image

The input image as base64-encoded bytes or an S3 object. If you use the AWS CLI to call Amazon Rekognition operations, passing base64-encoded image bytes isn\'t supported.

Returns:

  • (Types::Image)

    The input image as base64-encoded bytes or an S3 object.

#max_facesInteger

The maximum number of faces to index. The value of MaxFaces must be greater than or equal to 1. IndexFaces returns no more than 100 detected faces in an image, even if you specify a larger value for MaxFaces.

If IndexFaces detects more faces than the value of MaxFaces, the faces with the lowest quality are filtered out first. If there are still more faces than the value of MaxFaces, the faces with the smallest bounding boxes are filtered out (up to the number that\'s needed to satisfy the value of MaxFaces). Information about the unindexed faces is available in the UnindexedFaces array.

The faces that are returned by IndexFaces are sorted by the largest face bounding box size to the smallest size, in descending order.

MaxFaces can be used with a collection associated with any version of the face model.

Returns:

  • (Integer)

    The maximum number of faces to index.

#quality_filterString

A filter that specifies how much filtering is done to identify faces that are detected with low quality. Filtered faces aren\'t indexed. If you specify AUTO, filtering prioritizes the identification of faces that don’t meet the required quality bar chosen by Amazon Rekognition. The quality bar is based on a variety of common use cases. Low-quality detections can occur for a number of reasons. Some examples are an object that\'s misidentified as a face, a face that\'s too blurry, or a face with a pose that\'s too extreme to use. If you specify NONE, no filtering is performed. The default value is AUTO.

To use quality filtering, the collection you are using must be associated with version 3 of the face model.

Possible values:

  • NONE
  • AUTO

Returns:

  • (String)

    A filter that specifies how much filtering is done to identify faces that are detected with low quality.