2023/02/10 - Amazon SageMaker Service - 3 updated api methods
Changes Amazon SageMaker Autopilot adds support for selecting algorithms in CreateAutoMLJob API.
{'AutoMLJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': [{'AutoMLAlgorithms': ['xgboost ' '| ' 'linear-learner ' '| ' 'mlp ' '| ' 'lightgbm ' '| ' 'catboost ' '| ' 'randomforest ' '| ' 'extra-trees ' '| ' 'nn-torch ' '| ' 'fastai']}]}}}
Creates an Autopilot job.
Find the best-performing model after you run an Autopilot job by calling .
For information about how to use Autopilot, see Automate Model Development with Amazon SageMaker Autopilot .
See also: AWS API Documentation
Request Syntax
client.create_auto_ml_job( AutoMLJobName='string', InputDataConfig=[ { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix', 'S3Uri': 'string' } }, 'CompressionType': 'None'|'Gzip', 'TargetAttributeName': 'string', 'ContentType': 'string', 'ChannelType': 'training'|'validation' }, ], OutputDataConfig={ 'KmsKeyId': 'string', 'S3OutputPath': 'string' }, ProblemType='BinaryClassification'|'MulticlassClassification'|'Regression', AutoMLJobObjective={ 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro' }, AutoMLJobConfig={ 'CompletionCriteria': { 'MaxCandidates': 123, 'MaxRuntimePerTrainingJobInSeconds': 123, 'MaxAutoMLJobRuntimeInSeconds': 123 }, 'SecurityConfig': { 'VolumeKmsKeyId': 'string', 'EnableInterContainerTrafficEncryption': True|False, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] } }, 'DataSplitConfig': { 'ValidationFraction': ... }, 'CandidateGenerationConfig': { 'FeatureSpecificationS3Uri': 'string', 'AlgorithmsConfig': [ { 'AutoMLAlgorithms': [ 'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai', ] }, ] }, 'Mode': 'AUTO'|'ENSEMBLING'|'HYPERPARAMETER_TUNING' }, RoleArn='string', GenerateCandidateDefinitionsOnly=True|False, Tags=[ { 'Key': 'string', 'Value': 'string' }, ], ModelDeployConfig={ 'AutoGenerateEndpointName': True|False, 'EndpointName': 'string' } )
string
[REQUIRED]
Identifies an Autopilot job. The name must be unique to your account and is case insensitive.
list
[REQUIRED]
An array of channel objects that describes the input data and its location. Each channel is a named input source. Similar to InputDataConfig supported by . Format(s) supported: CSV, Parquet. A minimum of 500 rows is required for the training dataset. There is not a minimum number of rows required for the validation dataset.
(dict) --
A channel is a named input source that training algorithms can consume. The validation dataset size is limited to less than 2 GB. The training dataset size must be less than 100 GB. For more information, see .
Note
A validation dataset must contain the same headers as the training dataset.
DataSource (dict) -- [REQUIRED]
The data source for an AutoML channel.
S3DataSource (dict) -- [REQUIRED]
The Amazon S3 location of the input data.
S3DataType (string) -- [REQUIRED]
The data type.
A ManifestFile should have the format shown below:
[ {"prefix": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/DOC-EXAMPLE-PREFIX/"},
"DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-1",
"DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-2",
... "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-N" ]
An S3Prefix should have the following format:
s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER-OR-FILE
S3Uri (string) -- [REQUIRED]
The URL to the Amazon S3 data source.
CompressionType (string) --
You can use Gzip or None . The default value is None .
TargetAttributeName (string) -- [REQUIRED]
The name of the target variable in supervised learning, usually represented by 'y'.
ContentType (string) --
The content type of the data from the input source. You can use text/csv;header=present or x-application/vnd.amazon+parquet . The default value is text/csv;header=present .
ChannelType (string) --
The channel type (optional) is an enum string. The default value is training . Channels for training and validation must share the same ContentType and TargetAttributeName . For information on specifying training and validation channel types, see ` How to specify training and validation datasets https://docs.aws.amazon.com/sagemaker/latest/dg/autopilot-datasets-problem-types.html#autopilot-data-sources-training-or-validation`__ .
dict
[REQUIRED]
Provides information about encryption and the Amazon S3 output path needed to store artifacts from an AutoML job. Format(s) supported: CSV.
KmsKeyId (string) --
The Key Management Service (KMS) encryption key ID.
S3OutputPath (string) -- [REQUIRED]
The Amazon S3 output path. Must be 128 characters or less.
string
Defines the type of supervised learning available for the candidates. For more information, see Amazon SageMaker Autopilot problem types and algorithm support .
dict
Defines the objective metric used to measure the predictive quality of an AutoML job. You provide an AutoMLJobObjective$MetricName and Autopilot infers whether to minimize or maximize it.
MetricName (string) -- [REQUIRED]
The name of the objective metric used to measure the predictive quality of a machine learning system. This metric is optimized during training to provide the best estimate for model parameter values from data.
Here are the options:
Accuracy
The ratio of the number of correctly classified items to the total number of (correctly and incorrectly) classified items. It is used for both binary and multiclass classification. Accuracy measures how close the predicted class values are to the actual values. Values for accuracy metrics vary between zero (0) and one (1). A value of 1 indicates perfect accuracy, and 0 indicates perfect inaccuracy.
AUC
The area under the curve (AUC) metric is used to compare and evaluate binary classification by algorithms that return probabilities, such as logistic regression. To map the probabilities into classifications, these are compared against a threshold value.
The relevant curve is the receiver operating characteristic curve (ROC curve). The ROC curve plots the true positive rate (TPR) of predictions (or recall) against the false positive rate (FPR) as a function of the threshold value, above which a prediction is considered positive. Increasing the threshold results in fewer false positives, but more false negatives.
AUC is the area under this ROC curve. Therefore, AUC provides an aggregated measure of the model performance across all possible classification thresholds. AUC scores vary between 0 and 1. A score of 1 indicates perfect accuracy, and a score of one half (0.5) indicates that the prediction is not better than a random classifier.
BalancedAccuracy
BalancedAccuracy is a metric that measures the ratio of accurate predictions to all predictions. This ratio is calculated after normalizing true positives (TP) and true negatives (TN) by the total number of positive (P) and negative (N) values. It is used in both binary and multiclass classification and is defined as follows: 0.5*((TP/P)+(TN/N)), with values ranging from 0 to 1. BalancedAccuracy gives a better measure of accuracy when the number of positives or negatives differ greatly from each other in an imbalanced dataset. For example, when only 1% of email is spam.
F1
The F1 score is the harmonic mean of the precision and recall, defined as follows: F1 = 2 * (precision * recall) / (precision + recall). It is used for binary classification into classes traditionally referred to as positive and negative. Predictions are said to be true when they match their actual (correct) class, and false when they do not.
Precision is the ratio of the true positive predictions to all positive predictions, and it includes the false positives in a dataset. Precision measures the quality of the prediction when it predicts the positive class.
Recall (or sensitivity) is the ratio of the true positive predictions to all actual positive instances. Recall measures how completely a model predicts the actual class members in a dataset.
F1 scores vary between 0 and 1. A score of 1 indicates the best possible performance, and 0 indicates the worst.
F1macro
The F1macro score applies F1 scoring to multiclass classification problems. It does this by calculating the precision and recall, and then taking their harmonic mean to calculate the F1 score for each class. Lastly, the F1macro averages the individual scores to obtain the F1macro score. F1macro scores vary between 0 and 1. A score of 1 indicates the best possible performance, and 0 indicates the worst.
MAE
The mean absolute error (MAE) is a measure of how different the predicted and actual values are, when they're averaged over all values. MAE is commonly used in regression analysis to understand model prediction error. If there is linear regression, MAE represents the average distance from a predicted line to the actual value. MAE is defined as the sum of absolute errors divided by the number of observations. Values range from 0 to infinity, with smaller numbers indicating a better model fit to the data.
MSE
The mean squared error (MSE) is the average of the squared differences between the predicted and actual values. It is used for regression. MSE values are always positive. The better a model is at predicting the actual values, the smaller the MSE value is
Precision
Precision measures how well an algorithm predicts the true positives (TP) out of all of the positives that it identifies. It is defined as follows: Precision = TP/(TP+FP), with values ranging from zero (0) to one (1), and is used in binary classification. Precision is an important metric when the cost of a false positive is high. For example, the cost of a false positive is very high if an airplane safety system is falsely deemed safe to fly. A false positive (FP) reflects a positive prediction that is actually negative in the data.
PrecisionMacro
The precision macro computes precision for multiclass classification problems. It does this by calculating precision for each class and averaging scores to obtain precision for several classes. PrecisionMacro scores range from zero (0) to one (1). Higher scores reflect the model's ability to predict true positives (TP) out of all of the positives that it identifies, averaged across multiple classes.
R2
R2, also known as the coefficient of determination, is used in regression to quantify how much a model can explain the variance of a dependent variable. Values range from one (1) to negative one (-1). Higher numbers indicate a higher fraction of explained variability. R2 values close to zero (0) indicate that very little of the dependent variable can be explained by the model. Negative values indicate a poor fit and that the model is outperformed by a constant function. For linear regression, this is a horizontal line.
Recall
Recall measures how well an algorithm correctly predicts all of the true positives (TP) in a dataset. A true positive is a positive prediction that is also an actual positive value in the data. Recall is defined as follows: Recall = TP/(TP+FN), with values ranging from 0 to 1. Higher scores reflect a better ability of the model to predict true positives (TP) in the data, and is used in binary classification.
Recall is important when testing for cancer because it's used to find all of the true positives. A false positive (FP) reflects a positive prediction that is actually negative in the data. It is often insufficient to measure only recall, because predicting every output as a true positive yield a perfect recall score.
RecallMacro
The RecallMacro computes recall for multiclass classification problems by calculating recall for each class and averaging scores to obtain recall for several classes. RecallMacro scores range from 0 to 1. Higher scores reflect the model's ability to predict true positives (TP) in a dataset. Whereas, a true positive reflects a positive prediction that is also an actual positive value in the data. It is often insufficient to measure only recall, because predicting every output as a true positive yields a perfect recall score.
RMSE
Root mean squared error (RMSE) measures the square root of the squared difference between predicted and actual values, and it's averaged over all values. It is used in regression analysis to understand model prediction error. It's an important metric to indicate the presence of large model errors and outliers. Values range from zero (0) to infinity, with smaller numbers indicating a better model fit to the data. RMSE is dependent on scale, and should not be used to compare datasets of different sizes.
If you do not specify a metric explicitly, the default behavior is to automatically use:
MSE : for regression.
F1 : for binary classification
Accuracy : for multiclass classification.
dict
A collection of settings used to configure an AutoML job.
CompletionCriteria (dict) --
How long an AutoML job is allowed to run, or how many candidates a job is allowed to generate.
MaxCandidates (integer) --
The maximum number of times a training job is allowed to run.
MaxRuntimePerTrainingJobInSeconds (integer) --
The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the used by the action.
MaxAutoMLJobRuntimeInSeconds (integer) --
The maximum runtime, in seconds, an AutoML job has to complete.
If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.
SecurityConfig (dict) --
The security configuration for traffic encryption or Amazon VPC settings.
VolumeKmsKeyId (string) --
The key used to encrypt stored data.
EnableInterContainerTrafficEncryption (boolean) --
Whether to use traffic encryption between the container layers.
VpcConfig (dict) --
The VPC configuration.
SecurityGroupIds (list) -- [REQUIRED]
The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.
(string) --
Subnets (list) -- [REQUIRED]
The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones .
(string) --
DataSplitConfig (dict) --
The configuration for splitting the input training dataset.
Type: AutoMLDataSplitConfig
ValidationFraction (float) --
The validation fraction (optional) is a float that specifies the portion of the training dataset to be used for validation. The default value is 0.2, and values must be greater than 0 and less than 1. We recommend setting this value to be less than 0.5.
CandidateGenerationConfig (dict) --
The configuration for generating a candidate for an AutoML job (optional).
FeatureSpecificationS3Uri (string) --
A URL to the Amazon S3 data source containing selected features from the input data source to run an Autopilot job. You can input FeatureAttributeNames (optional) in JSON format as shown below:
{ "FeatureAttributeNames":["col1", "col2", ...] } .
You can also specify the data type of the feature (optional) in the format shown below:
{ "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }
Note
These column keys may not include the target column.
In ensembling mode, Autopilot only supports the following data types: numeric , categorical , text , and datetime . In HPO mode, Autopilot can support numeric , categorical , text , datetime , and sequence .
If only FeatureDataTypes is provided, the column keys (col1 , col2 ,..) should be a subset of the column names in the input data.
If both FeatureDataTypes and FeatureAttributeNames are provided, then the column keys should be a subset of the column names provided in FeatureAttributeNames .
The key name FeatureAttributeNames is fixed. The values listed in ["col1", "col2", ...] are case sensitive and should be a list of strings containing unique values that are a subset of the column names in the input data. The list of columns provided must not include the target column.
AlgorithmsConfig (list) --
Stores the configuration information for the selection of algorithms used to train the model candidates.
The list of available algorithms to choose from depends on the training mode set in ` AutoMLJobConfig.Mode https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_AutoMLJobConfig.html`__ .
AlgorithmsConfig should not be set in AUTO training mode.
When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, AutoMLCandidateGenerationConfig uses the full set of algorithms for the given training mode.
When AlgorithmsConfig is not provided, AutoMLCandidateGenerationConfig uses the full set of algorithms for the given training mode.
For the list of all algorithms per training mode, see .
For more information on each algorithm, see the Algorithm support section in Autopilot developer guide.
(dict) --
The collection of algorithms run on a dataset for training the model candidates of an Autopilot job.
AutoMLAlgorithms (list) -- [REQUIRED]
The selection of algorithms run on a dataset to train the model candidates of an Autopilot job.
Note
Selected algorithms must belong to the list corresponding to the training mode set in `` AutoMLJobConfig.Mode `` (ENSEMBLING or HYPERPARAMETER_TUNING ). Choose a minimum of 1 algorithm.
In ENSEMBLING mode:
"catboost"
"extra-trees"
"fastai"
"lightgbm"
"linear-learner"
"nn-torch"
"randomforest"
"xgboost"
In HYPERPARAMETER_TUNING mode:
"linear-learner"
"mlp"
"xgboost"
(string) --
Mode (string) --
The method that Autopilot uses to train the data. You can either specify the mode manually or let Autopilot choose for you based on the dataset size by selecting AUTO . In AUTO mode, Autopilot chooses ENSEMBLING for datasets smaller than 100 MB, and HYPERPARAMETER_TUNING for larger ones.
The ENSEMBLING mode uses a multi-stack ensemble model to predict classification and regression tasks directly from your dataset. This machine learning mode combines several base models to produce an optimal predictive model. It then uses a stacking ensemble method to combine predictions from contributing members. A multi-stack ensemble model can provide better performance over a single model by combining the predictive capabilities of multiple models. See Autopilot algorithm support for a list of algorithms supported by ENSEMBLING mode.
The HYPERPARAMETER_TUNING (HPO) mode uses the best hyperparameters to train the best version of a model. HPO automatically selects an algorithm for the type of problem you want to solve. Then HPO finds the best hyperparameters according to your objective metric. See Autopilot algorithm support for a list of algorithms supported by HYPERPARAMETER_TUNING mode.
string
[REQUIRED]
The ARN of the role that is used to access the data.
boolean
Generates possible candidates without training the models. A candidate is a combination of data preprocessors, algorithms, and algorithm parameter settings.
list
Each tag consists of a key and an optional value. Tag keys must be unique per resource.
(dict) --
A tag object that consists of a key and an optional value, used to manage metadata for SageMaker Amazon Web Services resources.
You can add tags to notebook instances, training jobs, hyperparameter tuning jobs, batch transform jobs, models, labeling jobs, work teams, endpoint configurations, and endpoints. For more information on adding tags to SageMaker resources, see AddTags .
For more information on adding metadata to your Amazon Web Services resources with tagging, see Tagging Amazon Web Services resources . For advice on best practices for managing Amazon Web Services resources with tagging, see Tagging Best Practices: Implement an Effective Amazon Web Services Resource Tagging Strategy .
Key (string) -- [REQUIRED]
The tag key. Tag keys must be unique per resource.
Value (string) -- [REQUIRED]
The tag value.
dict
Specifies how to generate the endpoint name for an automatic one-click Autopilot model deployment.
AutoGenerateEndpointName (boolean) --
Set to True to automatically generate an endpoint name for a one-click Autopilot model deployment; set to False otherwise. The default value is False .
Note
If you set AutoGenerateEndpointName to True , do not specify the EndpointName ; otherwise a 400 error is thrown.
EndpointName (string) --
Specifies the endpoint name to use for a one-click Autopilot model deployment if the endpoint name is not generated automatically.
Note
Specify the EndpointName if and only if you set AutoGenerateEndpointName to False ; otherwise a 400 error is thrown.
dict
Response Syntax
{ 'AutoMLJobArn': 'string' }
Response Structure
(dict) --
AutoMLJobArn (string) --
The unique ARN assigned to the AutoML job when it is created.
{'AutoMLJobConfig': {'CandidateGenerationConfig': {'AlgorithmsConfig': [{'AutoMLAlgorithms': ['xgboost ' '| ' 'linear-learner ' '| ' 'mlp ' '| ' 'lightgbm ' '| ' 'catboost ' '| ' 'randomforest ' '| ' 'extra-trees ' '| ' 'nn-torch ' '| ' 'fastai']}]}}, 'BestCandidate': {'FinalAutoMLJobObjectiveMetric': {'StandardMetricName': 'Accuracy ' '| ' 'MSE ' '| ' 'F1 ' '| ' 'F1macro ' '| ' 'AUC ' '| ' 'RMSE ' '| ' 'MAE ' '| ' 'R2 ' '| ' 'BalancedAccuracy ' '| ' 'Precision ' '| ' 'PrecisionMacro ' '| ' 'Recall ' '| ' 'RecallMacro'}}}
Returns information about an Amazon SageMaker AutoML job.
See also: AWS API Documentation
Request Syntax
client.describe_auto_ml_job( AutoMLJobName='string' )
string
[REQUIRED]
Requests information about an AutoML job using its unique name.
dict
Response Syntax
{ 'AutoMLJobName': 'string', 'AutoMLJobArn': 'string', 'InputDataConfig': [ { 'DataSource': { 'S3DataSource': { 'S3DataType': 'ManifestFile'|'S3Prefix', 'S3Uri': 'string' } }, 'CompressionType': 'None'|'Gzip', 'TargetAttributeName': 'string', 'ContentType': 'string', 'ChannelType': 'training'|'validation' }, ], 'OutputDataConfig': { 'KmsKeyId': 'string', 'S3OutputPath': 'string' }, 'RoleArn': 'string', 'AutoMLJobObjective': { 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro' }, 'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression', 'AutoMLJobConfig': { 'CompletionCriteria': { 'MaxCandidates': 123, 'MaxRuntimePerTrainingJobInSeconds': 123, 'MaxAutoMLJobRuntimeInSeconds': 123 }, 'SecurityConfig': { 'VolumeKmsKeyId': 'string', 'EnableInterContainerTrafficEncryption': True|False, 'VpcConfig': { 'SecurityGroupIds': [ 'string', ], 'Subnets': [ 'string', ] } }, 'DataSplitConfig': { 'ValidationFraction': ... }, 'CandidateGenerationConfig': { 'FeatureSpecificationS3Uri': 'string', 'AlgorithmsConfig': [ { 'AutoMLAlgorithms': [ 'xgboost'|'linear-learner'|'mlp'|'lightgbm'|'catboost'|'randomforest'|'extra-trees'|'nn-torch'|'fastai', ] }, ] }, 'Mode': 'AUTO'|'ENSEMBLING'|'HYPERPARAMETER_TUNING' }, 'CreationTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string', 'PartialFailureReasons': [ { 'PartialFailureMessage': 'string' }, ], 'BestCandidate': { 'CandidateName': 'string', 'FinalAutoMLJobObjectiveMetric': { 'Type': 'Maximize'|'Minimize', 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro', 'Value': ..., 'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro' }, 'ObjectiveStatus': 'Succeeded'|'Pending'|'Failed', 'CandidateSteps': [ { 'CandidateStepType': 'AWS::SageMaker::TrainingJob'|'AWS::SageMaker::TransformJob'|'AWS::SageMaker::ProcessingJob', 'CandidateStepArn': 'string', 'CandidateStepName': 'string' }, ], 'CandidateStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping', 'InferenceContainers': [ { 'Image': 'string', 'ModelDataUrl': 'string', 'Environment': { 'string': 'string' } }, ], 'CreationTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string', 'CandidateProperties': { 'CandidateArtifactLocations': { 'Explainability': 'string', 'ModelInsights': 'string' }, 'CandidateMetrics': [ { 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro', 'Value': ..., 'Set': 'Train'|'Validation'|'Test', 'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro'|'LogLoss'|'InferenceLatency' }, ] } }, 'AutoMLJobStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping', 'AutoMLJobSecondaryStatus': 'Starting'|'AnalyzingData'|'FeatureEngineering'|'ModelTuning'|'MaxCandidatesReached'|'Failed'|'Stopped'|'MaxAutoMLJobRuntimeReached'|'Stopping'|'CandidateDefinitionsGenerated'|'GeneratingExplainabilityReport'|'Completed'|'ExplainabilityError'|'DeployingModel'|'ModelDeploymentError'|'GeneratingModelInsightsReport'|'ModelInsightsError', 'GenerateCandidateDefinitionsOnly': True|False, 'AutoMLJobArtifacts': { 'CandidateDefinitionNotebookLocation': 'string', 'DataExplorationNotebookLocation': 'string' }, 'ResolvedAttributes': { 'AutoMLJobObjective': { 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro' }, 'ProblemType': 'BinaryClassification'|'MulticlassClassification'|'Regression', 'CompletionCriteria': { 'MaxCandidates': 123, 'MaxRuntimePerTrainingJobInSeconds': 123, 'MaxAutoMLJobRuntimeInSeconds': 123 } }, 'ModelDeployConfig': { 'AutoGenerateEndpointName': True|False, 'EndpointName': 'string' }, 'ModelDeployResult': { 'EndpointName': 'string' } }
Response Structure
(dict) --
AutoMLJobName (string) --
Returns the name of the AutoML job.
AutoMLJobArn (string) --
Returns the ARN of the AutoML job.
InputDataConfig (list) --
Returns the input data configuration for the AutoML job.
(dict) --
A channel is a named input source that training algorithms can consume. The validation dataset size is limited to less than 2 GB. The training dataset size must be less than 100 GB. For more information, see .
Note
A validation dataset must contain the same headers as the training dataset.
DataSource (dict) --
The data source for an AutoML channel.
S3DataSource (dict) --
The Amazon S3 location of the input data.
S3DataType (string) --
The data type.
A ManifestFile should have the format shown below:
[ {"prefix": "s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER/DOC-EXAMPLE-PREFIX/"},
"DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-1",
"DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-2",
... "DOC-EXAMPLE-RELATIVE-PATH/DOC-EXAMPLE-FOLDER/DATA-N" ]
An S3Prefix should have the following format:
s3://DOC-EXAMPLE-BUCKET/DOC-EXAMPLE-FOLDER-OR-FILE
S3Uri (string) --
The URL to the Amazon S3 data source.
CompressionType (string) --
You can use Gzip or None . The default value is None .
TargetAttributeName (string) --
The name of the target variable in supervised learning, usually represented by 'y'.
ContentType (string) --
The content type of the data from the input source. You can use text/csv;header=present or x-application/vnd.amazon+parquet . The default value is text/csv;header=present .
ChannelType (string) --
The channel type (optional) is an enum string. The default value is training . Channels for training and validation must share the same ContentType and TargetAttributeName . For information on specifying training and validation channel types, see ` How to specify training and validation datasets https://docs.aws.amazon.com/sagemaker/latest/dg/autopilot-datasets-problem-types.html#autopilot-data-sources-training-or-validation`__ .
OutputDataConfig (dict) --
Returns the job's output data config.
KmsKeyId (string) --
The Key Management Service (KMS) encryption key ID.
S3OutputPath (string) --
The Amazon S3 output path. Must be 128 characters or less.
RoleArn (string) --
The Amazon Resource Name (ARN) of the Identity and Access Management (IAM) role that has read permission to the input data location and write permission to the output data location in Amazon S3.
AutoMLJobObjective (dict) --
Returns the job's objective.
MetricName (string) --
The name of the objective metric used to measure the predictive quality of a machine learning system. This metric is optimized during training to provide the best estimate for model parameter values from data.
Here are the options:
Accuracy
The ratio of the number of correctly classified items to the total number of (correctly and incorrectly) classified items. It is used for both binary and multiclass classification. Accuracy measures how close the predicted class values are to the actual values. Values for accuracy metrics vary between zero (0) and one (1). A value of 1 indicates perfect accuracy, and 0 indicates perfect inaccuracy.
AUC
The area under the curve (AUC) metric is used to compare and evaluate binary classification by algorithms that return probabilities, such as logistic regression. To map the probabilities into classifications, these are compared against a threshold value.
The relevant curve is the receiver operating characteristic curve (ROC curve). The ROC curve plots the true positive rate (TPR) of predictions (or recall) against the false positive rate (FPR) as a function of the threshold value, above which a prediction is considered positive. Increasing the threshold results in fewer false positives, but more false negatives.
AUC is the area under this ROC curve. Therefore, AUC provides an aggregated measure of the model performance across all possible classification thresholds. AUC scores vary between 0 and 1. A score of 1 indicates perfect accuracy, and a score of one half (0.5) indicates that the prediction is not better than a random classifier.
BalancedAccuracy
BalancedAccuracy is a metric that measures the ratio of accurate predictions to all predictions. This ratio is calculated after normalizing true positives (TP) and true negatives (TN) by the total number of positive (P) and negative (N) values. It is used in both binary and multiclass classification and is defined as follows: 0.5*((TP/P)+(TN/N)), with values ranging from 0 to 1. BalancedAccuracy gives a better measure of accuracy when the number of positives or negatives differ greatly from each other in an imbalanced dataset. For example, when only 1% of email is spam.
F1
The F1 score is the harmonic mean of the precision and recall, defined as follows: F1 = 2 * (precision * recall) / (precision + recall). It is used for binary classification into classes traditionally referred to as positive and negative. Predictions are said to be true when they match their actual (correct) class, and false when they do not.
Precision is the ratio of the true positive predictions to all positive predictions, and it includes the false positives in a dataset. Precision measures the quality of the prediction when it predicts the positive class.
Recall (or sensitivity) is the ratio of the true positive predictions to all actual positive instances. Recall measures how completely a model predicts the actual class members in a dataset.
F1 scores vary between 0 and 1. A score of 1 indicates the best possible performance, and 0 indicates the worst.
F1macro
The F1macro score applies F1 scoring to multiclass classification problems. It does this by calculating the precision and recall, and then taking their harmonic mean to calculate the F1 score for each class. Lastly, the F1macro averages the individual scores to obtain the F1macro score. F1macro scores vary between 0 and 1. A score of 1 indicates the best possible performance, and 0 indicates the worst.
MAE
The mean absolute error (MAE) is a measure of how different the predicted and actual values are, when they're averaged over all values. MAE is commonly used in regression analysis to understand model prediction error. If there is linear regression, MAE represents the average distance from a predicted line to the actual value. MAE is defined as the sum of absolute errors divided by the number of observations. Values range from 0 to infinity, with smaller numbers indicating a better model fit to the data.
MSE
The mean squared error (MSE) is the average of the squared differences between the predicted and actual values. It is used for regression. MSE values are always positive. The better a model is at predicting the actual values, the smaller the MSE value is
Precision
Precision measures how well an algorithm predicts the true positives (TP) out of all of the positives that it identifies. It is defined as follows: Precision = TP/(TP+FP), with values ranging from zero (0) to one (1), and is used in binary classification. Precision is an important metric when the cost of a false positive is high. For example, the cost of a false positive is very high if an airplane safety system is falsely deemed safe to fly. A false positive (FP) reflects a positive prediction that is actually negative in the data.
PrecisionMacro
The precision macro computes precision for multiclass classification problems. It does this by calculating precision for each class and averaging scores to obtain precision for several classes. PrecisionMacro scores range from zero (0) to one (1). Higher scores reflect the model's ability to predict true positives (TP) out of all of the positives that it identifies, averaged across multiple classes.
R2
R2, also known as the coefficient of determination, is used in regression to quantify how much a model can explain the variance of a dependent variable. Values range from one (1) to negative one (-1). Higher numbers indicate a higher fraction of explained variability. R2 values close to zero (0) indicate that very little of the dependent variable can be explained by the model. Negative values indicate a poor fit and that the model is outperformed by a constant function. For linear regression, this is a horizontal line.
Recall
Recall measures how well an algorithm correctly predicts all of the true positives (TP) in a dataset. A true positive is a positive prediction that is also an actual positive value in the data. Recall is defined as follows: Recall = TP/(TP+FN), with values ranging from 0 to 1. Higher scores reflect a better ability of the model to predict true positives (TP) in the data, and is used in binary classification.
Recall is important when testing for cancer because it's used to find all of the true positives. A false positive (FP) reflects a positive prediction that is actually negative in the data. It is often insufficient to measure only recall, because predicting every output as a true positive yield a perfect recall score.
RecallMacro
The RecallMacro computes recall for multiclass classification problems by calculating recall for each class and averaging scores to obtain recall for several classes. RecallMacro scores range from 0 to 1. Higher scores reflect the model's ability to predict true positives (TP) in a dataset. Whereas, a true positive reflects a positive prediction that is also an actual positive value in the data. It is often insufficient to measure only recall, because predicting every output as a true positive yields a perfect recall score.
RMSE
Root mean squared error (RMSE) measures the square root of the squared difference between predicted and actual values, and it's averaged over all values. It is used in regression analysis to understand model prediction error. It's an important metric to indicate the presence of large model errors and outliers. Values range from zero (0) to infinity, with smaller numbers indicating a better model fit to the data. RMSE is dependent on scale, and should not be used to compare datasets of different sizes.
If you do not specify a metric explicitly, the default behavior is to automatically use:
MSE : for regression.
F1 : for binary classification
Accuracy : for multiclass classification.
ProblemType (string) --
Returns the job's problem type.
AutoMLJobConfig (dict) --
Returns the configuration for the AutoML job.
CompletionCriteria (dict) --
How long an AutoML job is allowed to run, or how many candidates a job is allowed to generate.
MaxCandidates (integer) --
The maximum number of times a training job is allowed to run.
MaxRuntimePerTrainingJobInSeconds (integer) --
The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the used by the action.
MaxAutoMLJobRuntimeInSeconds (integer) --
The maximum runtime, in seconds, an AutoML job has to complete.
If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.
SecurityConfig (dict) --
The security configuration for traffic encryption or Amazon VPC settings.
VolumeKmsKeyId (string) --
The key used to encrypt stored data.
EnableInterContainerTrafficEncryption (boolean) --
Whether to use traffic encryption between the container layers.
VpcConfig (dict) --
The VPC configuration.
SecurityGroupIds (list) --
The VPC security group IDs, in the form sg-xxxxxxxx. Specify the security groups for the VPC that is specified in the Subnets field.
(string) --
Subnets (list) --
The ID of the subnets in the VPC to which you want to connect your training job or model. For information about the availability of specific instance types, see Supported Instance Types and Availability Zones .
(string) --
DataSplitConfig (dict) --
The configuration for splitting the input training dataset.
Type: AutoMLDataSplitConfig
ValidationFraction (float) --
The validation fraction (optional) is a float that specifies the portion of the training dataset to be used for validation. The default value is 0.2, and values must be greater than 0 and less than 1. We recommend setting this value to be less than 0.5.
CandidateGenerationConfig (dict) --
The configuration for generating a candidate for an AutoML job (optional).
FeatureSpecificationS3Uri (string) --
A URL to the Amazon S3 data source containing selected features from the input data source to run an Autopilot job. You can input FeatureAttributeNames (optional) in JSON format as shown below:
{ "FeatureAttributeNames":["col1", "col2", ...] } .
You can also specify the data type of the feature (optional) in the format shown below:
{ "FeatureDataTypes":{"col1":"numeric", "col2":"categorical" ... } }
Note
These column keys may not include the target column.
In ensembling mode, Autopilot only supports the following data types: numeric , categorical , text , and datetime . In HPO mode, Autopilot can support numeric , categorical , text , datetime , and sequence .
If only FeatureDataTypes is provided, the column keys (col1 , col2 ,..) should be a subset of the column names in the input data.
If both FeatureDataTypes and FeatureAttributeNames are provided, then the column keys should be a subset of the column names provided in FeatureAttributeNames .
The key name FeatureAttributeNames is fixed. The values listed in ["col1", "col2", ...] are case sensitive and should be a list of strings containing unique values that are a subset of the column names in the input data. The list of columns provided must not include the target column.
AlgorithmsConfig (list) --
Stores the configuration information for the selection of algorithms used to train the model candidates.
The list of available algorithms to choose from depends on the training mode set in ` AutoMLJobConfig.Mode https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_AutoMLJobConfig.html`__ .
AlgorithmsConfig should not be set in AUTO training mode.
When AlgorithmsConfig is provided, one AutoMLAlgorithms attribute must be set and one only. If the list of algorithms provided as values for AutoMLAlgorithms is empty, AutoMLCandidateGenerationConfig uses the full set of algorithms for the given training mode.
When AlgorithmsConfig is not provided, AutoMLCandidateGenerationConfig uses the full set of algorithms for the given training mode.
For the list of all algorithms per training mode, see .
For more information on each algorithm, see the Algorithm support section in Autopilot developer guide.
(dict) --
The collection of algorithms run on a dataset for training the model candidates of an Autopilot job.
AutoMLAlgorithms (list) --
The selection of algorithms run on a dataset to train the model candidates of an Autopilot job.
Note
Selected algorithms must belong to the list corresponding to the training mode set in `` AutoMLJobConfig.Mode `` (ENSEMBLING or HYPERPARAMETER_TUNING ). Choose a minimum of 1 algorithm.
In ENSEMBLING mode:
"catboost"
"extra-trees"
"fastai"
"lightgbm"
"linear-learner"
"nn-torch"
"randomforest"
"xgboost"
In HYPERPARAMETER_TUNING mode:
"linear-learner"
"mlp"
"xgboost"
(string) --
Mode (string) --
The method that Autopilot uses to train the data. You can either specify the mode manually or let Autopilot choose for you based on the dataset size by selecting AUTO . In AUTO mode, Autopilot chooses ENSEMBLING for datasets smaller than 100 MB, and HYPERPARAMETER_TUNING for larger ones.
The ENSEMBLING mode uses a multi-stack ensemble model to predict classification and regression tasks directly from your dataset. This machine learning mode combines several base models to produce an optimal predictive model. It then uses a stacking ensemble method to combine predictions from contributing members. A multi-stack ensemble model can provide better performance over a single model by combining the predictive capabilities of multiple models. See Autopilot algorithm support for a list of algorithms supported by ENSEMBLING mode.
The HYPERPARAMETER_TUNING (HPO) mode uses the best hyperparameters to train the best version of a model. HPO automatically selects an algorithm for the type of problem you want to solve. Then HPO finds the best hyperparameters according to your objective metric. See Autopilot algorithm support for a list of algorithms supported by HYPERPARAMETER_TUNING mode.
CreationTime (datetime) --
Returns the creation time of the AutoML job.
EndTime (datetime) --
Returns the end time of the AutoML job.
LastModifiedTime (datetime) --
Returns the job's last modified time.
FailureReason (string) --
Returns the failure reason for an AutoML job, when applicable.
PartialFailureReasons (list) --
Returns a list of reasons for partial failures within an AutoML job.
(dict) --
The reason for a partial failure of an AutoML job.
PartialFailureMessage (string) --
The message containing the reason for a partial failure of an AutoML job.
BestCandidate (dict) --
The best model candidate selected by SageMaker Autopilot using both the best objective metric and lowest InferenceLatency for an experiment.
CandidateName (string) --
The name of the candidate.
FinalAutoMLJobObjectiveMetric (dict) --
The best candidate result from an AutoML training job.
Type (string) --
The type of metric with the best result.
MetricName (string) --
The name of the metric with the best result. For a description of the possible objective metrics, see AutoMLJobObjective$MetricName .
Value (float) --
The value of the metric with the best result.
StandardMetricName (string) --
The name of the standard metric. For a description of the standard metrics, see Autopilot candidate metrics .
ObjectiveStatus (string) --
The objective's status.
CandidateSteps (list) --
Information about the candidate's steps.
(dict) --
Information about the steps for a candidate and what step it is working on.
CandidateStepType (string) --
Whether the candidate is at the transform, training, or processing step.
CandidateStepArn (string) --
The ARN for the candidate's step.
CandidateStepName (string) --
The name for the candidate's step.
CandidateStatus (string) --
The candidate's status.
InferenceContainers (list) --
Information about the inference container definitions.
(dict) --
A list of container definitions that describe the different containers that make up an AutoML candidate. For more information, see .
Image (string) --
The Amazon Elastic Container Registry (Amazon ECR) path of the container. For more information, see .
ModelDataUrl (string) --
The location of the model artifacts. For more information, see .
Environment (dict) --
The environment variables to set in the container. For more information, see .
(string) --
(string) --
CreationTime (datetime) --
The creation time.
EndTime (datetime) --
The end time.
LastModifiedTime (datetime) --
The last modified time.
FailureReason (string) --
The failure reason.
CandidateProperties (dict) --
The properties of an AutoML candidate job.
CandidateArtifactLocations (dict) --
The Amazon S3 prefix to the artifacts generated for an AutoML candidate.
Explainability (string) --
The Amazon S3 prefix to the explainability artifacts generated for the AutoML candidate.
ModelInsights (string) --
The Amazon S3 prefix to the model insight artifacts generated for the AutoML candidate.
CandidateMetrics (list) --
Information about the candidate metrics for an AutoML job.
(dict) --
Information about the metric for a candidate produced by an AutoML job.
MetricName (string) --
The name of the metric.
Value (float) --
The value of the metric.
Set (string) --
The dataset split from which the AutoML job produced the metric.
StandardMetricName (string) --
The name of the standard metric.
Note
For definitions of the standard metrics, see ` Autopilot candidate metrics https://docs.aws.amazon.com/sagemaker/latest/dg/autopilot-model-support-validation.html#autopilot-metrics`__ .
AutoMLJobStatus (string) --
Returns the status of the AutoML job.
AutoMLJobSecondaryStatus (string) --
Returns the secondary status of the AutoML job.
GenerateCandidateDefinitionsOnly (boolean) --
Indicates whether the output for an AutoML job generates candidate definitions only.
AutoMLJobArtifacts (dict) --
Returns information on the job's artifacts found in AutoMLJobArtifacts .
CandidateDefinitionNotebookLocation (string) --
The URL of the notebook location.
DataExplorationNotebookLocation (string) --
The URL of the notebook location.
ResolvedAttributes (dict) --
Contains ProblemType , AutoMLJobObjective , and CompletionCriteria . If you do not provide these values, they are auto-inferred. If you do provide them, the values used are the ones you provide.
AutoMLJobObjective (dict) --
Specifies a metric to minimize or maximize as the objective of a job.
MetricName (string) --
The name of the objective metric used to measure the predictive quality of a machine learning system. This metric is optimized during training to provide the best estimate for model parameter values from data.
Here are the options:
Accuracy
The ratio of the number of correctly classified items to the total number of (correctly and incorrectly) classified items. It is used for both binary and multiclass classification. Accuracy measures how close the predicted class values are to the actual values. Values for accuracy metrics vary between zero (0) and one (1). A value of 1 indicates perfect accuracy, and 0 indicates perfect inaccuracy.
AUC
The area under the curve (AUC) metric is used to compare and evaluate binary classification by algorithms that return probabilities, such as logistic regression. To map the probabilities into classifications, these are compared against a threshold value.
The relevant curve is the receiver operating characteristic curve (ROC curve). The ROC curve plots the true positive rate (TPR) of predictions (or recall) against the false positive rate (FPR) as a function of the threshold value, above which a prediction is considered positive. Increasing the threshold results in fewer false positives, but more false negatives.
AUC is the area under this ROC curve. Therefore, AUC provides an aggregated measure of the model performance across all possible classification thresholds. AUC scores vary between 0 and 1. A score of 1 indicates perfect accuracy, and a score of one half (0.5) indicates that the prediction is not better than a random classifier.
BalancedAccuracy
BalancedAccuracy is a metric that measures the ratio of accurate predictions to all predictions. This ratio is calculated after normalizing true positives (TP) and true negatives (TN) by the total number of positive (P) and negative (N) values. It is used in both binary and multiclass classification and is defined as follows: 0.5*((TP/P)+(TN/N)), with values ranging from 0 to 1. BalancedAccuracy gives a better measure of accuracy when the number of positives or negatives differ greatly from each other in an imbalanced dataset. For example, when only 1% of email is spam.
F1
The F1 score is the harmonic mean of the precision and recall, defined as follows: F1 = 2 * (precision * recall) / (precision + recall). It is used for binary classification into classes traditionally referred to as positive and negative. Predictions are said to be true when they match their actual (correct) class, and false when they do not.
Precision is the ratio of the true positive predictions to all positive predictions, and it includes the false positives in a dataset. Precision measures the quality of the prediction when it predicts the positive class.
Recall (or sensitivity) is the ratio of the true positive predictions to all actual positive instances. Recall measures how completely a model predicts the actual class members in a dataset.
F1 scores vary between 0 and 1. A score of 1 indicates the best possible performance, and 0 indicates the worst.
F1macro
The F1macro score applies F1 scoring to multiclass classification problems. It does this by calculating the precision and recall, and then taking their harmonic mean to calculate the F1 score for each class. Lastly, the F1macro averages the individual scores to obtain the F1macro score. F1macro scores vary between 0 and 1. A score of 1 indicates the best possible performance, and 0 indicates the worst.
MAE
The mean absolute error (MAE) is a measure of how different the predicted and actual values are, when they're averaged over all values. MAE is commonly used in regression analysis to understand model prediction error. If there is linear regression, MAE represents the average distance from a predicted line to the actual value. MAE is defined as the sum of absolute errors divided by the number of observations. Values range from 0 to infinity, with smaller numbers indicating a better model fit to the data.
MSE
The mean squared error (MSE) is the average of the squared differences between the predicted and actual values. It is used for regression. MSE values are always positive. The better a model is at predicting the actual values, the smaller the MSE value is
Precision
Precision measures how well an algorithm predicts the true positives (TP) out of all of the positives that it identifies. It is defined as follows: Precision = TP/(TP+FP), with values ranging from zero (0) to one (1), and is used in binary classification. Precision is an important metric when the cost of a false positive is high. For example, the cost of a false positive is very high if an airplane safety system is falsely deemed safe to fly. A false positive (FP) reflects a positive prediction that is actually negative in the data.
PrecisionMacro
The precision macro computes precision for multiclass classification problems. It does this by calculating precision for each class and averaging scores to obtain precision for several classes. PrecisionMacro scores range from zero (0) to one (1). Higher scores reflect the model's ability to predict true positives (TP) out of all of the positives that it identifies, averaged across multiple classes.
R2
R2, also known as the coefficient of determination, is used in regression to quantify how much a model can explain the variance of a dependent variable. Values range from one (1) to negative one (-1). Higher numbers indicate a higher fraction of explained variability. R2 values close to zero (0) indicate that very little of the dependent variable can be explained by the model. Negative values indicate a poor fit and that the model is outperformed by a constant function. For linear regression, this is a horizontal line.
Recall
Recall measures how well an algorithm correctly predicts all of the true positives (TP) in a dataset. A true positive is a positive prediction that is also an actual positive value in the data. Recall is defined as follows: Recall = TP/(TP+FN), with values ranging from 0 to 1. Higher scores reflect a better ability of the model to predict true positives (TP) in the data, and is used in binary classification.
Recall is important when testing for cancer because it's used to find all of the true positives. A false positive (FP) reflects a positive prediction that is actually negative in the data. It is often insufficient to measure only recall, because predicting every output as a true positive yield a perfect recall score.
RecallMacro
The RecallMacro computes recall for multiclass classification problems by calculating recall for each class and averaging scores to obtain recall for several classes. RecallMacro scores range from 0 to 1. Higher scores reflect the model's ability to predict true positives (TP) in a dataset. Whereas, a true positive reflects a positive prediction that is also an actual positive value in the data. It is often insufficient to measure only recall, because predicting every output as a true positive yields a perfect recall score.
RMSE
Root mean squared error (RMSE) measures the square root of the squared difference between predicted and actual values, and it's averaged over all values. It is used in regression analysis to understand model prediction error. It's an important metric to indicate the presence of large model errors and outliers. Values range from zero (0) to infinity, with smaller numbers indicating a better model fit to the data. RMSE is dependent on scale, and should not be used to compare datasets of different sizes.
If you do not specify a metric explicitly, the default behavior is to automatically use:
MSE : for regression.
F1 : for binary classification
Accuracy : for multiclass classification.
ProblemType (string) --
The problem type.
CompletionCriteria (dict) --
How long a job is allowed to run, or how many candidates a job is allowed to generate.
MaxCandidates (integer) --
The maximum number of times a training job is allowed to run.
MaxRuntimePerTrainingJobInSeconds (integer) --
The maximum time, in seconds, that each training job executed inside hyperparameter tuning is allowed to run as part of a hyperparameter tuning job. For more information, see the used by the action.
MaxAutoMLJobRuntimeInSeconds (integer) --
The maximum runtime, in seconds, an AutoML job has to complete.
If an AutoML job exceeds the maximum runtime, the job is stopped automatically and its processing is ended gracefully. The AutoML job identifies the best model whose training was completed and marks it as the best-performing model. Any unfinished steps of the job, such as automatic one-click Autopilot model deployment, are not completed.
ModelDeployConfig (dict) --
Indicates whether the model was deployed automatically to an endpoint and the name of that endpoint if deployed automatically.
AutoGenerateEndpointName (boolean) --
Set to True to automatically generate an endpoint name for a one-click Autopilot model deployment; set to False otherwise. The default value is False .
Note
If you set AutoGenerateEndpointName to True , do not specify the EndpointName ; otherwise a 400 error is thrown.
EndpointName (string) --
Specifies the endpoint name to use for a one-click Autopilot model deployment if the endpoint name is not generated automatically.
Note
Specify the EndpointName if and only if you set AutoGenerateEndpointName to False ; otherwise a 400 error is thrown.
ModelDeployResult (dict) --
Provides information about endpoint for the model deployment.
EndpointName (string) --
The name of the endpoint to which the model has been deployed.
Note
If model deployment fails, this field is omitted from the response.
{'Candidates': {'FinalAutoMLJobObjectiveMetric': {'StandardMetricName': 'Accuracy ' '| MSE ' '| F1 ' '| ' 'F1macro ' '| AUC ' '| ' 'RMSE ' '| MAE ' '| R2 ' '| ' 'BalancedAccuracy ' '| ' 'Precision ' '| ' 'PrecisionMacro ' '| ' 'Recall ' '| ' 'RecallMacro'}}}
List the candidates created for the job.
See also: AWS API Documentation
Request Syntax
client.list_candidates_for_auto_ml_job( AutoMLJobName='string', StatusEquals='Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping', CandidateNameEquals='string', SortOrder='Ascending'|'Descending', SortBy='CreationTime'|'Status'|'FinalObjectiveMetricValue', MaxResults=123, NextToken='string' )
string
[REQUIRED]
List the candidates created for the job by providing the job's name.
string
List the candidates for the job and filter by status.
string
List the candidates for the job and filter by candidate name.
string
The sort order for the results. The default is Ascending .
string
The parameter by which to sort the results. The default is Descending .
integer
List the job's candidates up to a specified limit.
string
If the previous response was truncated, you receive this token. Use it in your next request to receive the next set of results.
dict
Response Syntax
{ 'Candidates': [ { 'CandidateName': 'string', 'FinalAutoMLJobObjectiveMetric': { 'Type': 'Maximize'|'Minimize', 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro', 'Value': ..., 'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro' }, 'ObjectiveStatus': 'Succeeded'|'Pending'|'Failed', 'CandidateSteps': [ { 'CandidateStepType': 'AWS::SageMaker::TrainingJob'|'AWS::SageMaker::TransformJob'|'AWS::SageMaker::ProcessingJob', 'CandidateStepArn': 'string', 'CandidateStepName': 'string' }, ], 'CandidateStatus': 'Completed'|'InProgress'|'Failed'|'Stopped'|'Stopping', 'InferenceContainers': [ { 'Image': 'string', 'ModelDataUrl': 'string', 'Environment': { 'string': 'string' } }, ], 'CreationTime': datetime(2015, 1, 1), 'EndTime': datetime(2015, 1, 1), 'LastModifiedTime': datetime(2015, 1, 1), 'FailureReason': 'string', 'CandidateProperties': { 'CandidateArtifactLocations': { 'Explainability': 'string', 'ModelInsights': 'string' }, 'CandidateMetrics': [ { 'MetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro', 'Value': ..., 'Set': 'Train'|'Validation'|'Test', 'StandardMetricName': 'Accuracy'|'MSE'|'F1'|'F1macro'|'AUC'|'RMSE'|'MAE'|'R2'|'BalancedAccuracy'|'Precision'|'PrecisionMacro'|'Recall'|'RecallMacro'|'LogLoss'|'InferenceLatency' }, ] } }, ], 'NextToken': 'string' }
Response Structure
(dict) --
Candidates (list) --
Summaries about the AutoMLCandidates .
(dict) --
Information about a candidate produced by an AutoML training job, including its status, steps, and other properties.
CandidateName (string) --
The name of the candidate.
FinalAutoMLJobObjectiveMetric (dict) --
The best candidate result from an AutoML training job.
Type (string) --
The type of metric with the best result.
MetricName (string) --
The name of the metric with the best result. For a description of the possible objective metrics, see AutoMLJobObjective$MetricName .
Value (float) --
The value of the metric with the best result.
StandardMetricName (string) --
The name of the standard metric. For a description of the standard metrics, see Autopilot candidate metrics .
ObjectiveStatus (string) --
The objective's status.
CandidateSteps (list) --
Information about the candidate's steps.
(dict) --
Information about the steps for a candidate and what step it is working on.
CandidateStepType (string) --
Whether the candidate is at the transform, training, or processing step.
CandidateStepArn (string) --
The ARN for the candidate's step.
CandidateStepName (string) --
The name for the candidate's step.
CandidateStatus (string) --
The candidate's status.
InferenceContainers (list) --
Information about the inference container definitions.
(dict) --
A list of container definitions that describe the different containers that make up an AutoML candidate. For more information, see .
Image (string) --
The Amazon Elastic Container Registry (Amazon ECR) path of the container. For more information, see .
ModelDataUrl (string) --
The location of the model artifacts. For more information, see .
Environment (dict) --
The environment variables to set in the container. For more information, see .
(string) --
(string) --
CreationTime (datetime) --
The creation time.
EndTime (datetime) --
The end time.
LastModifiedTime (datetime) --
The last modified time.
FailureReason (string) --
The failure reason.
CandidateProperties (dict) --
The properties of an AutoML candidate job.
CandidateArtifactLocations (dict) --
The Amazon S3 prefix to the artifacts generated for an AutoML candidate.
Explainability (string) --
The Amazon S3 prefix to the explainability artifacts generated for the AutoML candidate.
ModelInsights (string) --
The Amazon S3 prefix to the model insight artifacts generated for the AutoML candidate.
CandidateMetrics (list) --
Information about the candidate metrics for an AutoML job.
(dict) --
Information about the metric for a candidate produced by an AutoML job.
MetricName (string) --
The name of the metric.
Value (float) --
The value of the metric.
Set (string) --
The dataset split from which the AutoML job produced the metric.
StandardMetricName (string) --
The name of the standard metric.
Note
For definitions of the standard metrics, see ` Autopilot candidate metrics https://docs.aws.amazon.com/sagemaker/latest/dg/autopilot-model-support-validation.html#autopilot-metrics`__ .
NextToken (string) --
If the previous response was truncated, you receive this token. Use it in your next request to receive the next set of results.