Note: You are viewing the documentation for an older major version of the AWS CLI (version 1).

AWS CLI version 2, the latest major version of AWS CLI, is now stable and recommended for general use. To view this page for the AWS CLI version 2, click here. For more information see the AWS CLI version 2 installation instructions and migration guide.

[ aws . forecast ]

describe-dataset-import-job

Description

Describes a dataset import job created using the CreateDatasetImportJob operation.

In addition to listing the parameters provided in the CreateDatasetImportJob request, this operation includes the following properties:

  • CreationTime
  • LastModificationTime
  • DataSize
  • FieldStatistics
  • Status
  • Message - If an error occurred, information about the error.

See also: AWS API Documentation

Synopsis

  describe-dataset-import-job
--dataset-import-job-arn <value>
[--cli-input-json <value>]
[--generate-cli-skeleton <value>]
[--debug]
[--endpoint-url <value>]
[--no-verify-ssl]
[--no-paginate]
[--output <value>]
[--query <value>]
[--profile <value>]
[--region <value>]
[--version <value>]
[--color <value>]
[--no-sign-request]
[--ca-bundle <value>]
[--cli-read-timeout <value>]
[--cli-connect-timeout <value>]

Options

--dataset-import-job-arn (string)

The Amazon Resource Name (ARN) of the dataset import job.

--cli-input-json (string) Performs service operation based on the JSON string provided. The JSON string follows the format provided by --generate-cli-skeleton. If other arguments are provided on the command line, the CLI values will override the JSON-provided values. It is not possible to pass arbitrary binary values using a JSON-provided value as the string will be taken literally.

--generate-cli-skeleton (string) Prints a JSON skeleton to standard output without sending an API request. If provided with no value or the value input, prints a sample input JSON that can be used as an argument for --cli-input-json. If provided with the value output, it validates the command inputs and returns a sample output JSON for that command.

Global Options

--debug (boolean)

Turn on debug logging.

--endpoint-url (string)

Override command's default URL with the given URL.

--no-verify-ssl (boolean)

By default, the AWS CLI uses SSL when communicating with AWS services. For each SSL connection, the AWS CLI will verify SSL certificates. This option overrides the default behavior of verifying SSL certificates.

--no-paginate (boolean)

Disable automatic pagination.

--output (string)

The formatting style for command output.

  • json
  • text
  • table

--query (string)

A JMESPath query to use in filtering the response data.

--profile (string)

Use a specific profile from your credential file.

--region (string)

The region to use. Overrides config/env settings.

--version (string)

Display the version of this tool.

--color (string)

Turn on/off color output.

  • on
  • off
  • auto

--no-sign-request (boolean)

Do not sign requests. Credentials will not be loaded if this argument is provided.

--ca-bundle (string)

The CA certificate bundle to use when verifying SSL certificates. Overrides config/env settings.

--cli-read-timeout (int)

The maximum socket read time in seconds. If the value is set to 0, the socket read will be blocking and not timeout. The default value is 60 seconds.

--cli-connect-timeout (int)

The maximum socket connect time in seconds. If the value is set to 0, the socket connect will be blocking and not timeout. The default value is 60 seconds.

Output

DatasetImportJobName -> (string)

The name of the dataset import job.

DatasetImportJobArn -> (string)

The ARN of the dataset import job.

DatasetArn -> (string)

The Amazon Resource Name (ARN) of the dataset that the training data was imported to.

TimestampFormat -> (string)

The format of timestamps in the dataset. The format that you specify depends on the DataFrequency specified when the dataset was created. The following formats are supported

  • "yyyy-MM-dd" For the following data frequencies: Y, M, W, and D
  • "yyyy-MM-dd HH:mm:ss" For the following data frequencies: H, 30min, 15min, and 1min; and optionally, for: Y, M, W, and D

TimeZone -> (string)

The single time zone applied to every item in the dataset

UseGeolocationForTimeZone -> (boolean)

Whether TimeZone is automatically derived from the geolocation attribute.

GeolocationFormat -> (string)

The format of the geolocation attribute. Valid Values:"LAT_LONG" and "CC_POSTALCODE" .

DataSource -> (structure)

The location of the training data to import and an AWS Identity and Access Management (IAM) role that Amazon Forecast can assume to access the data.

If encryption is used, DataSource includes an AWS Key Management Service (KMS) key.

S3Config -> (structure)

The path to the data stored in an Amazon Simple Storage Service (Amazon S3) bucket along with the credentials to access the data.

Path -> (string)

The path to an Amazon Simple Storage Service (Amazon S3) bucket or file(s) in an Amazon S3 bucket.

RoleArn -> (string)

The ARN of the AWS Identity and Access Management (IAM) role that Amazon Forecast can assume to access the Amazon S3 bucket or files. If you provide a value for the KMSKeyArn key, the role must allow access to the key.

Passing a role across AWS accounts is not allowed. If you pass a role that isn't in your account, you get an InvalidInputException error.

KMSKeyArn -> (string)

The Amazon Resource Name (ARN) of an AWS Key Management Service (KMS) key.

EstimatedTimeRemainingInMinutes -> (long)

The estimated time remaining in minutes for the dataset import job to complete.

FieldStatistics -> (map)

Statistical information about each field in the input data.

key -> (string)

value -> (structure)

Provides statistics for each data field imported into to an Amazon Forecast dataset with the CreateDatasetImportJob operation.

Count -> (integer)

The number of values in the field. If the response value is -1, refer to CountLong .

CountDistinct -> (integer)

The number of distinct values in the field. If the response value is -1, refer to CountDistinctLong .

CountNull -> (integer)

The number of null values in the field. If the response value is -1, refer to CountNullLong .

CountNan -> (integer)

The number of NAN (not a number) values in the field. If the response value is -1, refer to CountNanLong .

Min -> (string)

For a numeric field, the minimum value in the field.

Max -> (string)

For a numeric field, the maximum value in the field.

Avg -> (double)

For a numeric field, the average value in the field.

Stddev -> (double)

For a numeric field, the standard deviation.

CountLong -> (long)

The number of values in the field. CountLong is used instead of Count if the value is greater than 2,147,483,647.

CountDistinctLong -> (long)

The number of distinct values in the field. CountDistinctLong is used instead of CountDistinct if the value is greater than 2,147,483,647.

CountNullLong -> (long)

The number of null values in the field. CountNullLong is used instead of CountNull if the value is greater than 2,147,483,647.

CountNanLong -> (long)

The number of NAN (not a number) values in the field. CountNanLong is used instead of CountNan if the value is greater than 2,147,483,647.

DataSize -> (double)

The size of the dataset in gigabytes (GB) after the import job has finished.

Status -> (string)

The status of the dataset import job. States include:

  • ACTIVE
  • CREATE_PENDING , CREATE_IN_PROGRESS , CREATE_FAILED
  • DELETE_PENDING , DELETE_IN_PROGRESS , DELETE_FAILED
  • CREATE_STOPPING , CREATE_STOPPED

Message -> (string)

If an error occurred, an informational message about the error.

CreationTime -> (timestamp)

When the dataset import job was created.

LastModificationTime -> (timestamp)

The last time the resource was modified. The timestamp depends on the status of the job:

  • CREATE_PENDING - The CreationTime .
  • CREATE_IN_PROGRESS - The current timestamp.
  • CREATE_STOPPING - The current timestamp.
  • CREATE_STOPPED - When the job stopped.
  • ACTIVE or CREATE_FAILED - When the job finished or failed.

Format -> (string)

The format of the imported data, CSV or PARQUET.