PNG  IHDR;IDATxܻn0K )(pA 7LeG{ §㻢|ذaÆ 6lذaÆ 6lذaÆ 6lom$^yذag5bÆ 6lذaÆ 6lذa{ 6lذaÆ `}HFkm,mӪôô! x|'ܢ˟;E:9&ᶒ}{v]n&6 h_tڠ͵-ҫZ;Z$.Pkž)!o>}leQfJTu іچ\X=8Rن4`Vwl>nG^is"ms$ui?wbs[m6K4O.4%/bC%t Mז -lG6mrz2s%9s@-k9=)kB5\+͂Zsٲ Rn~GRC wIcIn7jJhۛNCS|j08yiHKֶۛkɈ+;SzL/F*\Ԕ#"5m2[S=gnaPeғL lذaÆ 6l^ḵaÆ 6lذaÆ 6lذa; _ذaÆ 6lذaÆ 6lذaÆ RIENDB` package Paws::Rekognition::IndexFacesResponse; use Moose; has FaceModelVersion => (is => 'ro', isa => 'Str'); has FaceRecords => (is => 'ro', isa => 'ArrayRef[Paws::Rekognition::FaceRecord]'); has OrientationCorrection => (is => 'ro', isa => 'Str'); has UnindexedFaces => (is => 'ro', isa => 'ArrayRef[Paws::Rekognition::UnindexedFace]'); has _request_id => (is => 'ro', isa => 'Str'); ### main pod documentation begin ### =head1 NAME Paws::Rekognition::IndexFacesResponse =head1 ATTRIBUTES =head2 FaceModelVersion => Str The version number of the face detection model that's associated with the input collection (C). =head2 FaceRecords => ArrayRef[L] An array of faces detected and added to the collection. For more information, see Searching Faces in a Collection in the Amazon Rekognition Developer Guide. =head2 OrientationCorrection => Str If your collection is associated with a face detection model that's later than version 3.0, the value of C is always null and no orientation information is returned. If your collection is associated with a face detection model that's version 3.0 or earlier, the following applies: =over =item * If the input image is in .jpeg format, it might contain exchangeable image file format (Exif) metadata that includes the image's orientation. Amazon Rekognition uses this orientation information to perform image correction - the bounding box coordinates are translated to represent object locations after the orientation information in the Exif metadata is used to correct the image orientation. Images in .png format don't contain Exif metadata. The value of C is null. =item * If the image doesn't contain orientation information in its Exif metadata, Amazon Rekognition returns an estimated orientation (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnEt perform image correction for images. The bounding box coordinates aren't translated and represent the object locations before the image is rotated. =back Bounding box information is returned in the C array. You can get the version of the face detection model by calling DescribeCollection. Valid values are: C<"ROTATE_0">, C<"ROTATE_90">, C<"ROTATE_180">, C<"ROTATE_270"> =head2 UnindexedFaces => ArrayRef[L] An array of faces that were detected in the image but weren't indexed. They weren't indexed because the quality filter identified them as low quality, or the C request parameter filtered them out. To use the quality filter, you specify the C request parameter. =head2 _request_id => Str =cut 1;