Create the Output Stage based on Your Logical Data Model

The Output Stage is a set of tables and views that serve as a source for loading data into your workspace. The Output Stage is the definition of your analytical data that the GoodData platform will download and process.

The Output Stage is generated for your cloud data warehouse based on the logical date model (LDM; see Create a Logical Data Model in Your Workspace). The GoodData platform scans the LDM and suggests SQL queries that you then execute on your schema. The Output Stage follows the naming convention as described in Naming Convention for Output Stage Objects.

Before creating the Output Stage, make sure that your LDM is ready and there are no unpublished changes.

Creating the Output Stage as described in the following procedure (that is, directly from the LDM Modeler) is available only for cloud data warehouses (Amazon Redshift, Snowflake, Google BigQuery). If you are using the GoodData Agile Data Warehousing Service (ADS), you need to use the API to create the Output Stage.

Steps:

  1. From the Data Integration Console (see Accessing Data Integration Console), click Projects to open the Projects page, and click the name of the project for which you want to create the Output Stage.
  2. Click Model data on the top.
    The LDM Modeler opens displaying your logical data model.
  3. Click the menu button on the top right, and click Create Output Stage.
  4. Select the data source that you want to create the Output Stage for and the SQL command that you want to generate. You can choose from creating or altering tables and creating views.
  5. Click Create.
    The creating process starts. When the process completes, the dialog shows the SQL queries that you can use to create the Output Stage.
  6. Review the suggested queries and modify them as needed. Once done, execute the queries on your schema to create the Output Stage.
    Your LDM and the Output Stage are now synchronized.

If you do not have data load processes set up yet, you can now deploy processes for loading data (see Deploy a Process with the Data Integration Console) and schedule them to load data regularly (see Scheduling a Process).