This repository has been archived on 2025-09-14. You can view files and clone it, but cannot push or open issues or pull requests.
Files
pxz-hos-client-cpp-module/support/aws-sdk-cpp-master/aws-cpp-sdk-rekognition/include/aws/rekognition/model/IndexFacesResult.h

325 lines
17 KiB
C
Raw Normal View History

/**
* Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
* SPDX-License-Identifier: Apache-2.0.
*/
#pragma once
#include <aws/rekognition/Rekognition_EXPORTS.h>
#include <aws/core/utils/memory/stl/AWSVector.h>
#include <aws/rekognition/model/OrientationCorrection.h>
#include <aws/core/utils/memory/stl/AWSString.h>
#include <aws/rekognition/model/FaceRecord.h>
#include <aws/rekognition/model/UnindexedFace.h>
#include <utility>
namespace Aws
{
template<typename RESULT_TYPE>
class AmazonWebServiceResult;
namespace Utils
{
namespace Json
{
class JsonValue;
} // namespace Json
} // namespace Utils
namespace Rekognition
{
namespace Model
{
class AWS_REKOGNITION_API IndexFacesResult
{
public:
IndexFacesResult();
IndexFacesResult(const Aws::AmazonWebServiceResult<Aws::Utils::Json::JsonValue>& result);
IndexFacesResult& operator=(const Aws::AmazonWebServiceResult<Aws::Utils::Json::JsonValue>& result);
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline const Aws::Vector<FaceRecord>& GetFaceRecords() const{ return m_faceRecords; }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline void SetFaceRecords(const Aws::Vector<FaceRecord>& value) { m_faceRecords = value; }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline void SetFaceRecords(Aws::Vector<FaceRecord>&& value) { m_faceRecords = std::move(value); }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& WithFaceRecords(const Aws::Vector<FaceRecord>& value) { SetFaceRecords(value); return *this;}
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& WithFaceRecords(Aws::Vector<FaceRecord>&& value) { SetFaceRecords(std::move(value)); return *this;}
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& AddFaceRecords(const FaceRecord& value) { m_faceRecords.push_back(value); return *this; }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& AddFaceRecords(FaceRecord&& value) { m_faceRecords.push_back(std::move(value)); return *this; }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline const OrientationCorrection& GetOrientationCorrection() const{ return m_orientationCorrection; }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline void SetOrientationCorrection(const OrientationCorrection& value) { m_orientationCorrection = value; }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline void SetOrientationCorrection(OrientationCorrection&& value) { m_orientationCorrection = std::move(value); }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline IndexFacesResult& WithOrientationCorrection(const OrientationCorrection& value) { SetOrientationCorrection(value); return *this;}
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline IndexFacesResult& WithOrientationCorrection(OrientationCorrection&& value) { SetOrientationCorrection(std::move(value)); return *this;}
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline const Aws::String& GetFaceModelVersion() const{ return m_faceModelVersion; }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline void SetFaceModelVersion(const Aws::String& value) { m_faceModelVersion = value; }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline void SetFaceModelVersion(Aws::String&& value) { m_faceModelVersion = std::move(value); }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline void SetFaceModelVersion(const char* value) { m_faceModelVersion.assign(value); }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline IndexFacesResult& WithFaceModelVersion(const Aws::String& value) { SetFaceModelVersion(value); return *this;}
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline IndexFacesResult& WithFaceModelVersion(Aws::String&& value) { SetFaceModelVersion(std::move(value)); return *this;}
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline IndexFacesResult& WithFaceModelVersion(const char* value) { SetFaceModelVersion(value); return *this;}
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline const Aws::Vector<UnindexedFace>& GetUnindexedFaces() const{ return m_unindexedFaces; }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline void SetUnindexedFaces(const Aws::Vector<UnindexedFace>& value) { m_unindexedFaces = value; }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline void SetUnindexedFaces(Aws::Vector<UnindexedFace>&& value) { m_unindexedFaces = std::move(value); }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& WithUnindexedFaces(const Aws::Vector<UnindexedFace>& value) { SetUnindexedFaces(value); return *this;}
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& WithUnindexedFaces(Aws::Vector<UnindexedFace>&& value) { SetUnindexedFaces(std::move(value)); return *this;}
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& AddUnindexedFaces(const UnindexedFace& value) { m_unindexedFaces.push_back(value); return *this; }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& AddUnindexedFaces(UnindexedFace&& value) { m_unindexedFaces.push_back(std::move(value)); return *this; }
private:
Aws::Vector<FaceRecord> m_faceRecords;
OrientationCorrection m_orientationCorrection;
Aws::String m_faceModelVersion;
Aws::Vector<UnindexedFace> m_unindexedFaces;
};
} // namespace Model
} // namespace Rekognition
} // namespace Aws