ReOrc docs
Get ReOrc
English
English
  • About ReOrc
  • Set up and deployment
    • Set up organization
    • Install ReOrc agent
  • Getting started
    • 1. Set up a connection
      • BigQuery setup
    • 2. Create a project
    • 3. Create data models
    • 4. Build models in console
    • 5. Set up a pipeline
  • Connections
    • Destinations
      • Google Service Account
    • Integrations
      • Slack
  • Data modeling
    • Overview
    • Sources
    • Models
      • Model schema
      • Model configurations
    • Jinja templating
      • Variables
      • Macros
    • Materialization
    • Data lineage
    • Data tests
      • Built-in generic tests
      • Custom generic tests
      • Singular tests
  • Semantic modeling
    • Overview
    • Data Modelling vs Semantic Layer
    • Cube
      • Custom Dimension
      • Custom Measure
        • Aggregation Function
        • SQL functions and operators
        • Calculating Period-over-Period Changes
      • Relationship
    • View
      • Primary Dimension
      • Add Shared Fields
    • Shared Fields
    • Integration
      • Guandata Integration
      • Looker Studio
  • Pipeline
    • Overview
    • Modeling pipeline
    • Advanced pipeline
    • Job
  • Health tracking
    • Pipeline health
    • Data quality
  • Data governance
    • Data protection
  • Asset management
    • Console
    • Metadata
    • Version history
    • Packages and dependencies
  • DATA SERVICE
    • Overview
    • Create & edit Data Service
    • Data preview & download
    • Data sharing API
    • Access control
  • AI-powered
    • Rein AI Copilot
  • Settings
    • Organization settings
    • Project settings
    • Profile settings
    • Roles and permissions
  • Platform Specific
    • Doris/SelectDB
Powered by GitBook
On this page
  • Publish the models
  • Create a pipeline
  • Create a job
  1. Getting started

5. Set up a pipeline

Previous4. Build models in consoleNextDestinations

Last updated 15 days ago

After you've done creating and polishing the models, you can finally wrap the workflow in a pipeline and create a job to schedule the execution of the workflow. This ensures your transformations are reproducible and reliable every time they run.

Publish the models

Before creating a pipeline from the models, you need to publish them. Publishing models or any assets before using them in pipelines ensures reproducibility, makes dependencies clear, and enables parallel development across teams.

Open each model in the editor and follow these steps:

  1. In the top-right corner, click Publish.

  2. Provide a version number and description.

    For the first publish, we'll leave it as default.

  3. Click Publish.

The models are now ready for use in a pipeline.

Create a pipeline

Follow these steps:

  1. Switch to the Pipelines tab.

  2. Click on the + icon and select Creating modeling pipeline.

  3. Provide the name for your pipeline.

  4. Select the models that the pipeline should contain.

    Here we add the three models set up in the previous step.

  5. Click Confirm.

    The new modeling pipeline is then added under the Models folder and presented in the DAG view.

  6. Click Publish.

Create a job

Now that the pipeline is created and published, you can create a job to schedule execution time. Follow these steps:

  1. Open the pipeline that you've created.

  2. Switch to the Jobs tab and click + Create a job.

  3. Provide the configuration for the job.

Provide the name, select the target environment, and customize the associated model variables.

Since we've executed and verified the models using the console. We now can select the Production environment to run the transform against the production database.

Provide the time schedule for the job and the trigger type:

  • Standard setup: trigger the run at a specific time between certain intervals.

  • Manually trigger: no scheduling and can be manually triggered in the Pipeline Health dashboard. See: Pipeline health.

Here you can configure notifications on job failure. By default, notifications are sent by email.

  1. Click Create.

The new job will be displayed in the Jobs section. ReOrc will execute the job at the scheduled time.

Congratulations! You have completed building a data workflow with ReOrc.

For the next step, you can explore in detail each feature and module:

Advanced setup: specify the schedule in crontab format, for those familiar with the cron scheduler. See: .

You can the execution status and progress of the job in the

Cron
Pipeline health dashboard.

Connections

Learn about the available connections

Data modeling

Learn about the core assets in data modeling

Health tracking

Explore the health tracking dashboard