

A data pipeline moves data from one or more sources through a series of transformations to a destination where it can be analysed or served. Getting this right from the start saves enormous debugging effort later.
Most pipelines follow the Extract → Transform → Load (ETL) pattern:
A variation, ELT, loads raw data first and transforms it inside the destination (common with cloud data warehouses like BigQuery).