[ aws . rekognition ]
Detects custom labels in a supplied image by using an Amazon Rekognition Custom Labels model.
You specify which version of a model version to use by using the ProjectVersionArn
input parameter.
You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file.
For each object that the model version detects on an image, the API returns a (CustomLabel
) object in an array (CustomLabels
). Each CustomLabel
object provides the label name (Name
), the level of confidence that the image contains the object (Confidence
), and object location information, if it exists, for the label on the image (Geometry
).
During training model calculates a threshold value that determines if a prediction for a label is true. By default, DetectCustomLabels
doesn’t return labels whose confidence value is below the model’s calculated threshold value. To filter labels that are returned, specify a value for MinConfidence
that is higher than the model’s calculated threshold. You can get the model’s calculated threshold from the model’s training results shown in the Amazon Rekognition Custom Labels console. To get all labels, regardless of confidence, specify a MinConfidence
value of 0.
You can also add the MaxResults
parameter to limit the number of labels returned.
This is a stateless API operation. That is, the operation does not persist any data.
This operation requires permissions to perform the rekognition:DetectCustomLabels
action.
See also: AWS API Documentation
See ‘aws help’ for descriptions of global parameters.
detect-custom-labels
--project-version-arn <value>
[--image <value>]
[--max-results <value>]
[--min-confidence <value>]
[--image-bytes <value>]
[--cli-input-json | --cli-input-yaml]
[--generate-cli-skeleton <value>]
--project-version-arn
(string)
The ARN of the model version that you want to use.
--image
(structure)
Provides the input image either as bytes or an S3 object.
You pass image bytes to an Amazon Rekognition API operation by using the
Bytes
property. For example, you would use theBytes
property to pass an image loaded from a local file system. Image bytes passed by using theBytes
property must be base64-encoded. Your code may not need to encode image bytes if you are using an AWS SDK to call Amazon Rekognition API operations.For more information, see Analyzing an Image Loaded from a Local File System in the Amazon Rekognition Developer Guide.
You pass images stored in an S3 bucket to an Amazon Rekognition API operation by using the
S3Object
property. Images stored in an S3 bucket do not need to be base64-encoded.The region for the S3 bucket containing the S3 object must match the region you use for Amazon Rekognition operations.
If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes using the Bytes property is not supported. You must first upload the image to an Amazon S3 bucket and then call the operation using the S3Object property.
For Amazon Rekognition to process an S3 object, the user must have permission to access the S3 object. For more information, see Resource Based Policies in the Amazon Rekognition Developer Guide.
To specify a local file use
--image-bytes
instead.Bytes -> (blob)
Blob of image bytes up to 5 MBs.
S3Object -> (structure)
Identifies an S3 object as the image source.
Bucket -> (string)
Name of the S3 bucket.
Name -> (string)
S3 object key name.
Version -> (string)
If the bucket is versioning enabled, you can specify the object version.
Shorthand Syntax:
Bytes=blob,S3Object={Bucket=string,Name=string,Version=string}
JSON Syntax:
{
"Bytes": blob,
"S3Object": {
"Bucket": "string",
"Name": "string",
"Version": "string"
}
}
--max-results
(integer)
Maximum number of results you want the service to return in the response. The service returns the specified number of highest confidence labels ranked from highest confidence to lowest.
--min-confidence
(float)
Specifies the minimum confidence level for the labels to return. Amazon Rekognition doesn’t return any labels with a confidence lower than this specified value. If you specify a value of 0, all labels are return, regardless of the default thresholds that the model version applies.
--image-bytes
(blob)
The content of the image to be uploaded. To specify the content of a local file use the fileb:// prefix. Example: fileb://image.png
--cli-input-json
| --cli-input-yaml
(string)
Reads arguments from the JSON string provided. The JSON string follows the format provided by --generate-cli-skeleton
. If other arguments are provided on the command line, those values will override the JSON-provided values. It is not possible to pass arbitrary binary values using a JSON-provided value as the string will be taken literally. This may not be specified along with --cli-input-yaml
.
--generate-cli-skeleton
(string)
Prints a JSON skeleton to standard output without sending an API request. If provided with no value or the value input
, prints a sample input JSON that can be used as an argument for --cli-input-json
. Similarly, if provided yaml-input
it will print a sample input YAML that can be used with --cli-input-yaml
. If provided with the value output
, it validates the command inputs and returns a sample output JSON for that command.
See ‘aws help’ for descriptions of global parameters.
CustomLabels -> (list)
An array of custom labels detected in the input image.
(structure)
A custom label detected in an image by a call to DetectCustomLabels .
Name -> (string)
The name of the custom label.
Confidence -> (float)
The confidence that the model has in the detection of the custom label. The range is 0-100. A higher value indicates a higher confidence.
Geometry -> (structure)
The location of the detected object on the image that corresponds to the custom label. Includes an axis aligned coarse bounding box surrounding the object and a finer grain polygon for more accurate spatial information.
BoundingBox -> (structure)
An axis-aligned coarse representation of the detected item’s location on the image.
Width -> (float)
Width of the bounding box as a ratio of the overall image width.
Height -> (float)
Height of the bounding box as a ratio of the overall image height.
Left -> (float)
Left coordinate of the bounding box as a ratio of overall image width.
Top -> (float)
Top coordinate of the bounding box as a ratio of overall image height.
Polygon -> (list)
Within the bounding box, a fine-grained polygon around the detected item.
(structure)
The X and Y coordinates of a point on an image. The X and Y values returned are ratios of the overall image size. For example, if the input image is 700x200 and the operation returns X=0.5 and Y=0.25, then the point is at the (350,50) pixel coordinate on the image.
An array of
Point
objects,Polygon
, is returned by DetectText and by DetectCustomLabels .Polygon
represents a fine-grained polygon around a detected item. For more information, see Geometry in the Amazon Rekognition Developer Guide.X -> (float)
The value of the X coordinate for a point on a
Polygon
.Y -> (float)
The value of the Y coordinate for a point on a
Polygon
.