Skip to the content.

Pipeline API

Argo Workflow automation is driven by yaml templates that are desined using the ADSL (Argo Domain Specific Language). More examples for the same can be found here

Pipeline API makes use of the Argo models defined in the Argo Python client repository

Why Pipeline API

All the low level details regarding the image/container details are stored in yaml file

Configuration file :

Config file contains mainly two components

Below is the sample yaml which contains jobprofilerclient to run a sample data profiler using pydeequ and resource template sparkk8sScala is to run a spark job using Spark Operator

ConfigFile

Usage

A quick example to show how to complex DAG can be created

tasks

Once tasks has been added , dag can be compiled and also we can view the dependencies once the dag is compiled , submit the dag to argo workflow

workflow

This code results in a DAG which looks like this when submitted to ArgoWorkflow

dag