Support Information and Considerations When Using Pipelines

In Data Integration, a pipeline is a design-time resource for creating a set of tasks connected in a sequence or in parallel to facilitate data processing. It lets you manage and orchestrate the execution of a set of related tasks and activities for a workload instead of running tasks individually and handling run outcomes separately. These pipelines are not designed for low-latency tasks. Occasionally each step might experience several minutes of delay because of network or cloud issues. In between those steps are reconciliation processes that might take a minute or at times even longer.

Ensure that you understand what is supported and any current limitations before you create pipelines in Data Integration.

  • By default, a workspace has a limit of four concurrent task runs. When you have more than four tasks in a pipeline, integration tasks and data loader tasks are queued based on the limit of four concurrent runs per default workspace. SQL, OCI Data Flow, and REST tasks are not queued.
  • The limit of four concurrent runs is across pipelines in a workspace; the limit is not just within a pipeline.
  • When designing a pipeline, remember that while the maximum supported concurrency is 16, only four concurrent task runs can take place at a time during execution.
  • Nesting of pipelines is supported, up to a maximum of three.
  • We don't support running more than 100 tasks in a single pipeline. This limit includes the tasks in nested pipelines.
  • In a pipeline, when an operator's output parameters are used in the next operator, refrain from using the same output parameter multiple times as doing so might result in indeterminate result.

Was this article helpful?