This repository has been archived on 2025-09-14. You can view files and clone it, but cannot push or open issues or pull requests.
Files
pxz-hos-client-cpp-module/support/aws-sdk-cpp-master/aws-cpp-sdk-rekognition/include/aws/rekognition/model/IndexFacesResult.h

325 lines
17 KiB
C++
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

/**
* Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved.
* SPDX-License-Identifier: Apache-2.0.
*/
#pragma once
#include <aws/rekognition/Rekognition_EXPORTS.h>
#include <aws/core/utils/memory/stl/AWSVector.h>
#include <aws/rekognition/model/OrientationCorrection.h>
#include <aws/core/utils/memory/stl/AWSString.h>
#include <aws/rekognition/model/FaceRecord.h>
#include <aws/rekognition/model/UnindexedFace.h>
#include <utility>
namespace Aws
{
template<typename RESULT_TYPE>
class AmazonWebServiceResult;
namespace Utils
{
namespace Json
{
class JsonValue;
} // namespace Json
} // namespace Utils
namespace Rekognition
{
namespace Model
{
class AWS_REKOGNITION_API IndexFacesResult
{
public:
IndexFacesResult();
IndexFacesResult(const Aws::AmazonWebServiceResult<Aws::Utils::Json::JsonValue>& result);
IndexFacesResult& operator=(const Aws::AmazonWebServiceResult<Aws::Utils::Json::JsonValue>& result);
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline const Aws::Vector<FaceRecord>& GetFaceRecords() const{ return m_faceRecords; }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline void SetFaceRecords(const Aws::Vector<FaceRecord>& value) { m_faceRecords = value; }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline void SetFaceRecords(Aws::Vector<FaceRecord>&& value) { m_faceRecords = std::move(value); }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& WithFaceRecords(const Aws::Vector<FaceRecord>& value) { SetFaceRecords(value); return *this;}
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& WithFaceRecords(Aws::Vector<FaceRecord>&& value) { SetFaceRecords(std::move(value)); return *this;}
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& AddFaceRecords(const FaceRecord& value) { m_faceRecords.push_back(value); return *this; }
/**
* <p>An array of faces detected and added to the collection. For more information,
* see Searching Faces in a Collection in the Amazon Rekognition Developer Guide.
* </p>
*/
inline IndexFacesResult& AddFaceRecords(FaceRecord&& value) { m_faceRecords.push_back(std::move(value)); return *this; }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline const OrientationCorrection& GetOrientationCorrection() const{ return m_orientationCorrection; }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline void SetOrientationCorrection(const OrientationCorrection& value) { m_orientationCorrection = value; }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline void SetOrientationCorrection(OrientationCorrection&& value) { m_orientationCorrection = std::move(value); }
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline IndexFacesResult& WithOrientationCorrection(const OrientationCorrection& value) { SetOrientationCorrection(value); return *this;}
/**
* <p>If your collection is associated with a face detection model that's later
* than version 3.0, the value of <code>OrientationCorrection</code> is always null
* and no orientation information is returned.</p> <p>If your collection is
* associated with a face detection model that's version 3.0 or earlier, the
* following applies:</p> <ul> <li> <p>If the input image is in .jpeg format, it
* might contain exchangeable image file format (Exif) metadata that includes the
* image's orientation. Amazon Rekognition uses this orientation information to
* perform image correction - the bounding box coordinates are translated to
* represent object locations after the orientation information in the Exif
* metadata is used to correct the image orientation. Images in .png format don't
* contain Exif metadata. The value of <code>OrientationCorrection</code> is
* null.</p> </li> <li> <p>If the image doesn't contain orientation information in
* its Exif metadata, Amazon Rekognition returns an estimated orientation
* (ROTATE_0, ROTATE_90, ROTATE_180, ROTATE_270). Amazon Rekognition doesnt
* perform image correction for images. The bounding box coordinates aren't
* translated and represent the object locations before the image is rotated.</p>
* </li> </ul> <p>Bounding box information is returned in the
* <code>FaceRecords</code> array. You can get the version of the face detection
* model by calling <a>DescribeCollection</a>. </p>
*/
inline IndexFacesResult& WithOrientationCorrection(OrientationCorrection&& value) { SetOrientationCorrection(std::move(value)); return *this;}
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline const Aws::String& GetFaceModelVersion() const{ return m_faceModelVersion; }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline void SetFaceModelVersion(const Aws::String& value) { m_faceModelVersion = value; }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline void SetFaceModelVersion(Aws::String&& value) { m_faceModelVersion = std::move(value); }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline void SetFaceModelVersion(const char* value) { m_faceModelVersion.assign(value); }
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline IndexFacesResult& WithFaceModelVersion(const Aws::String& value) { SetFaceModelVersion(value); return *this;}
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline IndexFacesResult& WithFaceModelVersion(Aws::String&& value) { SetFaceModelVersion(std::move(value)); return *this;}
/**
* <p>The version number of the face detection model that's associated with the
* input collection (<code>CollectionId</code>).</p>
*/
inline IndexFacesResult& WithFaceModelVersion(const char* value) { SetFaceModelVersion(value); return *this;}
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline const Aws::Vector<UnindexedFace>& GetUnindexedFaces() const{ return m_unindexedFaces; }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline void SetUnindexedFaces(const Aws::Vector<UnindexedFace>& value) { m_unindexedFaces = value; }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline void SetUnindexedFaces(Aws::Vector<UnindexedFace>&& value) { m_unindexedFaces = std::move(value); }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& WithUnindexedFaces(const Aws::Vector<UnindexedFace>& value) { SetUnindexedFaces(value); return *this;}
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& WithUnindexedFaces(Aws::Vector<UnindexedFace>&& value) { SetUnindexedFaces(std::move(value)); return *this;}
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& AddUnindexedFaces(const UnindexedFace& value) { m_unindexedFaces.push_back(value); return *this; }
/**
* <p>An array of faces that were detected in the image but weren't indexed. They
* weren't indexed because the quality filter identified them as low quality, or
* the <code>MaxFaces</code> request parameter filtered them out. To use the
* quality filter, you specify the <code>QualityFilter</code> request
* parameter.</p>
*/
inline IndexFacesResult& AddUnindexedFaces(UnindexedFace&& value) { m_unindexedFaces.push_back(std::move(value)); return *this; }
private:
Aws::Vector<FaceRecord> m_faceRecords;
OrientationCorrection m_orientationCorrection;
Aws::String m_faceModelVersion;
Aws::Vector<UnindexedFace> m_unindexedFaces;
};
} // namespace Model
} // namespace Rekognition
} // namespace Aws