Welcome Guest! Log in
Due to some maintenance operations, stambia.org will be switched to read-only mode during the 13th November. It will be possible to read and download, but impossible to post on the forums or create new accounts. For any question please contact the support team.

Loading data into Redshift via S3 buckets

    When you need to load data from your local databases to Amazon Redshift, Stambia DI lets you design a simple mapping which will take care of the following steps automatically:

    1. Export data from your database to a file

    2. Send the file to an Amazon S3 bucket

    3. Load the S3 file into Redshift

    This article shows how to create the required metadata and how to design the mapping.

     

    Pre-requisites

    Please make sure that you have installed the Amazon components as indicated in this article.

    Creating the Amazon S3 metadata

    Simply create an Amazon S3 metadata object:

    468 01

     

    And right click to add a bucket and a folder. You can configure each node:

    468 02

    Creating the Amazon Redshift metadata

    Select the rdbms.AmazonRedshift metadata model:

    468 03

    Fill in the Jdbc URL and credentials, and reverse your Redshift database schema as you would do with any Postgresql database.

    Once this is done, drag and drop the S3 Folder into the Redshift schema, like this:

    468 04

    And rename this node as "S3_STAGING". The Load templates will need this name to connect to this folder. 

     

    Designing the mapping

    Simply design your mapping as usual, using your Redshift table as a target. And make sure that the "LOAD Rdbms to Redshift SQL (with COPY)" and "INTEGRATION Redshift" template are selected:

    468 05

     

     

    Subcategories

    Articles

    Suggest a new Article!