aws-samples/sagemaker-distributed-training-workshop

Hands-on workshop for distributed training and hosting on SageMaker

55
/ 100
Established

This workshop helps machine learning engineers and researchers optimize large neural networks for training on AWS. You'll learn how to take your existing model and data, apply distributed training techniques like data and model parallelism, and significantly reduce training time and resource costs. This is for professionals building and deploying AI models at scale.

153 stars.

Use this if you are an ML engineer or researcher struggling with long training times or memory limitations for large deep learning models on AWS.

Not ideal if you are new to machine learning or only working with small, simple models that train quickly on a single machine.

deep-learning model-training performance-optimization cloud-ml large-scale-ai
No Package No Dependents
Maintenance 6 / 25
Adoption 10 / 25
Maturity 16 / 25
Community 23 / 25

How are scores calculated?

Stars

153

Forks

67

Language

Jupyter Notebook

License

Apache-2.0

Last pushed

Nov 04, 2025

Commits (30d)

0

Get this data via API

curl "https://pt-edge.onrender.com/api/v1/quality/ml-frameworks/aws-samples/sagemaker-distributed-training-workshop"

Open to everyone — 100 requests/day, no key needed. Get a free key for 1,000/day.