Using batch load with the console - Amazon Timestream

Using batch load with the console

Following are steps for using batch load with the AWS Management Console. You can download a sample CSV at sample CSV.

Access batch load

Follow these steps to access batch load using the AWS Management Console.

  1. Open the Amazon Timestream console.

  2. In the navigation pane, choose Management Tools, and then choose Batch load tasks.

  3. From here, you can view the list of batch load tasks and drill into a given task for more details. You can also create and resume tasks.

Create a batch load task

Follow these steps to create a batch load task using the AWS Management Console.

  1. Open the Amazon Timestream console.

  2. In the navigation pane, choose Management Tools, and then choose Batch load tasks.

  3. Choose Create batch load task.

  4. In Import destination, choose the following.

    • Target database – Select the name of the database created in Create a database.

    • Target table – Select the name of the table created in Create a table.

    If necessary, you can add a table from this panel with the Create new table button.

  5. From Data source S3 location in Data source, select the S3 bucket where the source data is stored. Use the Browse S3 button to view S3 resources the active AWS account has access to, or enter the S3 location URL. The data source must be located in the same region.

  6. In File format settings (expandable section), you can use the default settings to parse input data. You can also choose Advanced settings. From there you can choose CSV format parameters, and select parameters to parse input data. For information about these parameters, see CSV format parameters.

  7. From Configure data model mapping, configure the data model. For additional data model guidance, see Data model mappings for batch load

    • From Data model mapping, choose Mapping configuration input, and choose one of the following.

      • Visual builder – To map data visually, choose TargetMultiMeasureName or MeasureNameColumn. Then from Visual builder, map the columns.

        Visual builder automatically detects and loads the source column headers from the data source file when a single CSV file is selected as the data source. Choose the attribute and data type to create your mapping.

        For information about using the visual builder, see Using the visual builder.

      • JSON editor – A freeform JSON editor for configuring your data model. Choose this option if you're familiar with Timestream for LiveAnalytics and want to build advanced data model mappings.

      • JSON file from S3 – Select a JSON model file you have stored in S3. Choose this option if you've already configured a data model and want to reuse it for additional batch loads.

  8. From Error logs S3 location in Error log report, select the S3 location that will be used to report errors. For information about how to use this report, see Using batch load error reports.

  9. For Encryption key type, choose one of the following.

    • Amazon S3-managed key (SSE-S3) – An encryption key that Amazon S3 creates, manages, and uses for you.

    • AWS KMS key (SSE-KMS) – An encryption key protected by AWS Key Management Service (AWS KMS).

  10. Choose Next.

  11. On the Review and create page, review the settings and edit as necessary.


    You can't change batch load task settings after the task has been created. Task completion times will vary based on the amount of data being imported.

  12. Choose Create batch load task.

Resume a batch load task

When you select a batch load task with a status of "Progress stopped" which is still resumable, you are prompted to resume the task. There is also a banner with a Resume task button when you view the details for those tasks. Resumable tasks have a "resume by" date. After that date expires, tasks cannot be resumed.

Using the visual builder

You can use the visual builder to map source data columns one or more CSV file(s) stored in an S3 bucket to destination columns in a Timestream for LiveAnalytics table.


Your role will need the SelectObjectContent permission for the file. Without this, you will need to add and delete columns manually.

Auto load source columns mode

Timestream for LiveAnalytics can automatically scan the source CSV file for column names if you specify one bucket only. When there are no existing mappings, you can choose Import source columns.

  1. With the Visual builder option selected from the Mapping configuration input settings, set the Timestamp time input. Milliseconds is the default setting.

  2. Click the Load source columns button to import the column headers found in the source data file. The table will be populated with the source column header names from the data source file.

  3. Choose the Target table column name, Timestream attribute type, and Data type for each source column.

    For details about these columns and possible values, see Mapping fields.

  4. Use the drag-to-fill feature to set the value for multiple columns at once.

Manually add source columns

If you're using a bucket or CSV prefix and not a single CSV, you can add and delete column mappings from the visual editor with the Add column mapping and Delete column mapping buttons. There is also a button to reset mappings.

Mapping fields

  • Source column name – The name of a column in the source file that represents a measure to import. Timestream for LiveAnalytics can populate this value automatically when you use Import source columns.

  • Target table column name – Optional input that indicates the column name for the measure in the target table.

  • Timestream attribute type – The attribute type of the data in the specified source column such as DIMENSION.

    • TIMESTAMP – Specifies when a measure was collected.

    • MULTI – Multiple measures are represented.

    • DIMENSION – Time series metadata.

    • MEASURE_NAME – For single-measure records, this is the measure name.

  • Data type – The type of Timestream column, such as BOOLEAN.

    • BIGINT – A 64-bit integer.

    • BOOLEAN – The two truth values of logic—true and false.

    • DOUBLE – 64-bit variable-precision number.

    • TIMESTAMP – An instance in time that uses nanosecond precision time in UTC, and tracks the time since the Unix epoch.