EMR Serverless

2023/04/11 - EMR Serverless - 1 updated api methods

Changes  This release extends GetJobRun API to return job run timeout (executionTimeoutMinutes) specified during StartJobRun call (or default timeout of 720 minutes if none was specified).

GetJobRun (updated) Link ΒΆ
Changes (response)
{'jobRun': {'executionTimeoutMinutes': 'long'}}

Displays detailed information about a job run.

See also: AWS API Documentation

Request Syntax

client.get_job_run(
    applicationId='string',
    jobRunId='string'
)
type applicationId

string

param applicationId

[REQUIRED]

The ID of the application on which the job run is submitted.

type jobRunId

string

param jobRunId

[REQUIRED]

The ID of the job run.

rtype

dict

returns

Response Syntax

{
    'jobRun': {
        'applicationId': 'string',
        'jobRunId': 'string',
        'name': 'string',
        'arn': 'string',
        'createdBy': 'string',
        'createdAt': datetime(2015, 1, 1),
        'updatedAt': datetime(2015, 1, 1),
        'executionRole': 'string',
        'state': 'SUBMITTED'|'PENDING'|'SCHEDULED'|'RUNNING'|'SUCCESS'|'FAILED'|'CANCELLING'|'CANCELLED',
        'stateDetails': 'string',
        'releaseLabel': 'string',
        'configurationOverrides': {
            'applicationConfiguration': [
                {
                    'classification': 'string',
                    'properties': {
                        'string': 'string'
                    },
                    'configurations': {'... recursive ...'}
                },
            ],
            'monitoringConfiguration': {
                's3MonitoringConfiguration': {
                    'logUri': 'string',
                    'encryptionKeyArn': 'string'
                },
                'managedPersistenceMonitoringConfiguration': {
                    'enabled': True|False,
                    'encryptionKeyArn': 'string'
                }
            }
        },
        'jobDriver': {
            'sparkSubmit': {
                'entryPoint': 'string',
                'entryPointArguments': [
                    'string',
                ],
                'sparkSubmitParameters': 'string'
            },
            'hive': {
                'query': 'string',
                'initQueryFile': 'string',
                'parameters': 'string'
            }
        },
        'tags': {
            'string': 'string'
        },
        'totalResourceUtilization': {
            'vCPUHour': 123.0,
            'memoryGBHour': 123.0,
            'storageGBHour': 123.0
        },
        'networkConfiguration': {
            'subnetIds': [
                'string',
            ],
            'securityGroupIds': [
                'string',
            ]
        },
        'totalExecutionDurationSeconds': 123,
        'executionTimeoutMinutes': 123
    }
}

Response Structure

  • (dict) --

    • jobRun (dict) --

      The output displays information about the job run.

      • applicationId (string) --

        The ID of the application the job is running on.

      • jobRunId (string) --

        The ID of the job run.

      • name (string) --

        The optional job run name. This doesn't have to be unique.

      • arn (string) --

        The execution role ARN of the job run.

      • createdBy (string) --

        The user who created the job run.

      • createdAt (datetime) --

        The date and time when the job run was created.

      • updatedAt (datetime) --

        The date and time when the job run was updated.

      • executionRole (string) --

        The execution role ARN of the job run.

      • state (string) --

        The state of the job run.

      • stateDetails (string) --

        The state details of the job run.

      • releaseLabel (string) --

        The EMR release associated with the application your job is running on.

      • configurationOverrides (dict) --

        The configuration settings that are used to override default configuration.

        • applicationConfiguration (list) --

          The override configurations for the application.

          • (dict) --

            A configuration specification to be used when provisioning an application. A configuration consists of a classification, properties, and optional nested configurations. A classification refers to an application-specific configuration file. Properties are the settings you want to change in that file.

            • classification (string) --

              The classification within a configuration.

            • properties (dict) --

              A set of properties specified within a configuration classification.

              • (string) --

                • (string) --

            • configurations (list) --

              A list of additional configurations to apply within a configuration object.

        • monitoringConfiguration (dict) --

          The override configurations for monitoring.

          • s3MonitoringConfiguration (dict) --

            The Amazon S3 configuration for monitoring log publishing.

            • logUri (string) --

              The Amazon S3 destination URI for log publishing.

            • encryptionKeyArn (string) --

              The KMS key ARN to encrypt the logs published to the given Amazon S3 destination.

          • managedPersistenceMonitoringConfiguration (dict) --

            The managed log persistence configuration for a job run.

            • enabled (boolean) --

              Enables managed logging and defaults to true. If set to false, managed logging will be turned off.

            • encryptionKeyArn (string) --

              The KMS key ARN to encrypt the logs stored in managed log persistence.

      • jobDriver (dict) --

        The job driver for the job run.

        • sparkSubmit (dict) --

          The job driver parameters specified for Spark.

          • entryPoint (string) --

            The entry point for the Spark submit job run.

          • entryPointArguments (list) --

            The arguments for the Spark submit job run.

            • (string) --

          • sparkSubmitParameters (string) --

            The parameters for the Spark submit job run.

        • hive (dict) --

          The job driver parameters specified for Hive.

          • query (string) --

            The query for the Hive job run.

          • initQueryFile (string) --

            The query file for the Hive job run.

          • parameters (string) --

            The parameters for the Hive job run.

      • tags (dict) --

        The tags assigned to the job run.

        • (string) --

          • (string) --

      • totalResourceUtilization (dict) --

        The aggregate vCPU, memory, and storage resources used from the time job start executing till the time job is terminated, rounded up to the nearest second.

        • vCPUHour (float) --

          The aggregated vCPU used per hour from the time job start executing till the time job is terminated.

        • memoryGBHour (float) --

          The aggregated memory used per hour from the time job start executing till the time job is terminated.

        • storageGBHour (float) --

          The aggregated storage used per hour from the time job start executing till the time job is terminated.

      • networkConfiguration (dict) --

        The network configuration for customer VPC connectivity.

        • subnetIds (list) --

          The array of subnet Ids for customer VPC connectivity.

          • (string) --

        • securityGroupIds (list) --

          The array of security group Ids for customer VPC connectivity.

          • (string) --

      • totalExecutionDurationSeconds (integer) --

        The job run total execution duration in seconds. This field is only available for job runs in a COMPLETED , FAILED , or CANCELLED state.

      • executionTimeoutMinutes (integer) --

        Maximum duration for the job run to run. If the job run runs beyond this duration, it will be automatically cancelled.