👉 Check out our short video tutorials:

Jobs are a way to schedule one off or recurring pipelines runs.

A Job can run multiple iterations of the same pipeline over time or by using different parameters as inputs. For example, you could create a Job which uses the same ETL pipeline but extracts data from a different data source for each pipeline run.

Jobs take a snapshot of your project directory when they are created. Each of the job’s pipeline runs copy the project directory snapshot and execute the files without changing the original snapshot. This means Jobs run consistently throughout their entire lifetime.

Different pipeline runs that are part of the same job are completely isolated from a scheduling perspective and do not affect each others’ state.


💡 Write data and large artifacts to the /data directory. This helps to save disk space by not including them in the project directory snapshot. Alternatively, you can add the artifacts to .gitignore as the ignored patterns are not copied to the snapshot.

Parameterizing pipelines and steps

Jobs run a specific pipeline for a given set of parameters. If you define multiple values for the same parameter, then the job will run the pipeline once for every combination of parameter values.


💡 Unlike environment variables, you can define pipeline and step level parameters with the same name without one (automatically) overwritting the other, you can access both values.

You can define pipeline parameters at two levels:

  • Pipelines: The parameters and their values will be available across every pipeline step.

  • Pipeline steps: The parameters will only be available in which they are defined.

Editing pipeline parameters

  1. Open a pipeline via the Pipelines option in the left menu pane.

  2. Click on SETTINGS in the top right corner.

  3. Towards the top you will find the Pipeline parameters section.

  4. Input some JSON like {"my-param": <param-value>}.

  5. Make sure to Save at the bottom of your screen.

Editing pipeline step parameters

  1. Open a pipeline via the Pipelines option in the left menu pane.

  2. Click on a pipeline step to open its Properties.

  3. Towards the bottom you will find the Parameters section.

  4. Input some JSON like {"my-param": <param-value>}.

Interacting with parameters through code

After you have set parameters for your pipeline and/or steps you can use their values inside your scripts (see parameters API reference).

Let’s say you have set the following parameters on your pipeline:

  "vegetable": "carrot",
  "fruit": "banana"

And for your pipeline step:

  "candy": "chocolate",
  "fruit": "apple"

Then inside the pipeline step you can access the parameters as follows:

import orchest

# Get the parameters of the current step and the pipeline.
fruit = orchest.get_step_param("fruit")               # "apple"
vegetable = orchest.get_pipeline_param("vegetable")   # "carrot"

Specify job parameters with a file

You can easily run a pipeline for multiple parameter configurations by creating a job parameters file.

If you place the file in the same folder as your pipeline file the job parameters file will automatically be detected when creating a job.

For a pipeline called main.orchest the job parameters file should be named main.parameters.json, and be put in the same folder as the pipeline file (both in the project directory).

You can also select a file manually when creating a job.

The JSON file should be formatted as below. Note that wrapping the values in a list is required, even if you’re assigning just one parameter value to a key. It is allowed to omit keys you don’t want to specify.

   "pipeline_parameters": {
      "some_key": ["a", "list", "of", "values"]
   "62a62810-336c-44c4-af6a-35228e8f2028": {
      "some_key": [1, 2, 3],
      "another_key": [1]

You can find the step UUIDs in the pipeline file (e.g. main.orchest), pipelines are regular JSON files.

Running a job

Make sure you have read the previous section on how to parametrize your pipeline. With jobs you get to run the same pipeline for different parameter values. For now you can think of it as a grid search, i.e. looping over all combinations of values for different parameters. To run a job:

  1. Make sure you have defined some parameters or you will only be able to schedule the pipeline as is.

  2. Click on Jobs in the left menu pane.

  3. Click the “+” sign to configure your job.

  4. Choose a Job name and the Pipeline you want to run the job for.

  5. Your default set of parameters are pre-loaded. By clicking on the values a JSON editor opens, allowing you to add additional values you would like the pipeline to run for.

  6. If you would like to schedule the job to run at a specific time have a look at Scheduling. In case you don’t want your job to run every combination of your parameter values, you can deselect them through the Pipeline runs option.

  7. Press Run job.

To inspect the result of your job; click on the job you just created, choose a specific pipeline run (the one you want to inspect) and click on View pipeline. The pipeline is now opened in read-only mode giving you the opportunity to check the logs or to open the HTML version of you notebooks.


💡 Upon job creation, Orchest (under the hood) takes a snapshot of the required environments. This way you can freely iterate on and update your existing environments without worrying about breaking your existing jobs.