Skip to content

tobigs/sagemaker-batch-ml-pipeline

Repository files navigation

Sagemaker batch ML pipeline

This example demonstrates how to efficiently train and run inference for a large number of models in a single workflow.

Deployment

$ cdk deploy

Usage

  1. https://eu-west-1.console.aws.amazon.com/states/home (Make sure to select the region of deployment)
  2. Open state machine with name like MLTraininingInference*
  3. Press "Start execution" at the top right
  4. No Input is required

Workflow Architecture Overview

  1. Distributed map, which creates a child workflow for every file in the source bucket. All subsequent steps happen for every file in the bucket
  2. Training of a model
  3. Waiting for training to finish
  4. Saving the model to Sagemaker, to make it available for inference
  5. Inference using saved model and data in the target bucket
  6. Waiting for inference to finish
  7. Deletion of saved model

Notice

This is a sample solution intended as a starting point and should not be used in a productive setting without thorough analysis and considerations on the user's side.

Security

See CONTRIBUTING for more information.

License

This library is licensed under the MIT-0 License. See the LICENSE file.

Attribution

The folder images/ is based on this repository from AWS Samples.

About

No description, website, or topics provided.

Resources

License

Code of conduct

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published