Skip to content

Compute Configuration Overview

Compute configurations store compute and job settings that sources are linked to. For performance and cost recommendations, see Production Workload Compute Configuration Recommendations.


Compute List

Navigate to System Configurations → Compute Configurations from the main menu.

2025-07-01_11-00-14.png

The table lists all compute configurations with filtering by name/description and column sorting. Click any row to open its settings page, click the launch icon in the Start column to start a compute manually, or click NEW + to create a new configuration.


Settings

2025-07-01_11-07-18.png

  • Name*: A unique name.
  • Description*: A one sentence summary describing the compute.
  • Default Compute Configuration: A flag that marks the compute as the default. Once active, toggle is disabled until another default compute is selected.
  • Compute Type*: Create either a new Job (default, recommended) or user specified pool.
  • Scale Mode*: The number of workers can be automatically managed by Databricks or can be a fixed value.
  • Job Task Type*: Jobs will either execute a custom notebook in Databricks or the DataForge Jar will be used.
  • Notebook Path*: The full file path to the custom notebook. Only required when custom notebook job task type is selected.

Duplicate (near Save) creates a copy with the same settings named " COPY". The copy is not attached to any objects automatically.


Advanced Parameters

Advanced parameters are shown based on your required-parameter selections; descriptions for each appear in the UI. User-modified parameters are displayed in bold.

A notable parameter is Max Processes Per Compute — by default, each new process starts a new job run of the attached compute configuration.