Offloading Large Workflows¶
v2.4 and after
Argo stores workflows as Kubernetes resources (i.e. within EtcD). This creates a limit to their size as resources must be under 1MB. Each resource includes the status of each node, which is stored in the
/status/nodes field for the resource. This can be over 1MB. If this happens, we try and compress the node status and store it in
/status/compressedNodes. If the status is still too large, we then try and store it in an SQL database.
To enable this feature, configure a Postgres or MySQL database under
persistence in your configuration and set
Why aren't my workflows appearing in the database?¶
Offloading is expensive and often unneccessary, so we only offload when we need to. Your workflows aren't probably large enough.
Error "Failed to submit workflow: etcdserver: request is too large."¶
You must use the Argo CLI having exported