Back to glossary

AI GLOSSARY

Data Pipeline

Deployment & Infrastructure

An automated sequence of steps that moves and transforms data from its source to its destination, such as from raw storage through preprocessing into a format ready for model training or inference. Reliable data pipelines are foundational infrastructure for any production AI system.
See also: data preprocessing, inference, MLOps.