Welcome to Dynamic Pipeline’s Documentation!

_images/DynamicPipeline_Logo.png

Dynamic Pipeline is a high-level API toolkit to help data scientists building models in ensemble way, and automating Machine Learning workflow with simple codes. Comparing other popular “AutoML or Automatic Machine Learning” APIs, Dynamic Pipeline is designed as an omni-ensembled ML workflow optimizer with higher-level API targeting to avoid manual repetitive train-along-evaluate experiments in general pipeline building.

To achieve that, Dynamic Pipeline creates Pipeline Cluster Traversal Experiments to assemble all cross-matching pipelines covering major tasks of Machine Learning workflow, and apply traversal-experiment to search the optimal baseline model. Besides, by modularizing all key pipeline components in reusable packages, it allows all components to be custom tunable along with high scalability.

_images/DynamicPipeline_Workflow.PNG

The core concept in Dynamic Pipeline is Pipeline Cluster Traversal Experiments, which is a theory, first raised by Tony Dong during Genpact 2020 GVector Conference, to optimize and automate Machine Learning Workflow using ensemble pipelines algorithm.

Comparing other automatic or classic machine learning workflow’s repetitive experiments using single pipeline, Pipeline Cluster Traversal Experiments is more powerful, with larger coverage scope, to find the best model without manual intervention, and also more flexible with elasticity to cope with unseen data due to its ensemble designs in each component.

_images/PipelineClusterTraversalExperiments.PNG

In summary, Dynamic Pipeline shares a few useful properties for data scientists:

  • Easy & less coding - High-level APIs to implement Pipeline Cluster Traversal Experiments, and each ML component is highly automated and modularized;
  • Well ensembled - Each key component is ensemble of popular algorithms w/ optimal hyperparameters tuning included;
  • Hardly omission - Pipeline Cluster Traversal Experiments are designed to cross-experiment with combined permuted input datasets, feature selection, and model selection;
  • Scalable & Consistency - Each module could add new algorithms easily due to its ensemble & reusable design; no extra needs to modify existing codes for new experiment
  • Adaptable - Pipeline Cluster Traversal Experiments makes it easier to adapt unseen datasets with the right pipeline;
  • Custom Modify Welcomed - Support custom settings to add/remove algorithms or modify hyperparameters for elastic requirements.

Indices and tables