Configure source and destination - Amazon Data Firehose

Amazon Data Firehose was previously known as Amazon Kinesis Data Firehose

Configure source and destination

  1. Sign in to the AWS Management Console and open the Amazon Data Firehose console at https://console.aws.amazon.com/firehose

  2. Choose Create Firehose stream.

  3. Enter values for the following fields:

    Source
    • Direct PUT: Choose this option to create a Firehose stream that producer applications write to directly. Currently, the following are AWS services and agents and open source services that are integrated with Direct PUT in Amazon Data Firehose:

      • AWS SDK

      • AWS Lambda

      • AWS CloudWatch Logs

      • AWS CloudWatch Events

      • AWS Cloud Metric Streams

      • AWS IOT

      • AWS Eventbridge

      • Amazon Simple Email Service

      • Amazon SNS

      • AWS WAF web ACL logs

      • Amazon API Gateway - Access logs

      • Amazon Pinpoint

      • Amazon MSK Broker Logs

      • Amazon Route 53 Resolver query logs

      • AWS Network Firewall Alerts Logs

      • AWS Network Firewall Flow Logs

      • Amazon Elasticache Redis SLOWLOG

      • Kinesis Agent (linux)

      • Kinesis Tap (windows)

      • Fluentbit

      • Fluentd

      • Apache Nifi

      • Snowflake

    • Kinesis stream: Choose this option to configure a Firehose stream that uses a Kinesis data stream as a data source. You can then use Amazon Data Firehose to read data easily from an existing Kinesis data stream and load it into destinations. For more information about using Kinesis Data Streams as your data source, see Writing to Amazon Data Firehose Using Kinesis Data Streams.

    • Amazon MSK: Choose this option to configure a Firehose stream that uses Amazon MSK as a data source. You can then use Firehose to read data easily from an existing Amazon MSK clusters and load it into specified S3 buckets. For more information about using Amazon MSK as your data source, see Writing to Amazon Data Firehose Using Amazon MSK.

    Firehose stream destination

    The destination of your Firehose stream. Amazon Data Firehose can send data records to various destinations, including Amazon Simple Storage Service (Amazon S3), Amazon Redshift, Amazon OpenSearch Service, and any HTTP endpoint that is owned by you or any of your third-party service providers. The following are the supported destinations:

    • Amazon OpenSearch Service

    • Amazon OpenSearch Serverless

    • Amazon Redshift

    • Amazon S3

    • Coralogix

    • Datadog

    • Dynatrace

    • Elastic

    • HTTP Endpoint

    • Honeycomb

    • Logic Monitor

    • Logz.io

    • MongoDB Cloud

    • New Relic

    • Splunk

    • Splunk Observability Cloud

    • Sumo Logic

    • Snowflake

    Firehose stream name

    The name of your Firehose stream.