doc/source/train/tutorials/content/README.md
This tutorial series provides hands-on learning for Ray Train and its ecosystem (Ray Data, Anyscale Workspaces).
The tutorials walk through common ML workload patterns—vision, tabular, time series, generative, policy learning, and recommendation—showing how to scale them from single-node to fully distributed training and inference with minimal code changes.
Pre-install all requirements
!pip install -r requirements.txt
prepare_model / prepare_data_loader.ScalingConfig and RunConfig for scale and checkpointing.train.report.Vision workloads
Real-world computer vision with Food-101, preprocessing with Ray Data, fault-tolerant ResNet training, and scalable inference tasks.
Tabular workloads
Tabular ML with CoverType dataset, XGBoost + Ray Train, checkpoint-aware training, feature importance, and distributed inference.
Time series workloads
New York City taxi demand forecasting with a Transformer model, scaling across GPUs, epoch-level fault tolerance, and remote inference from checkpoints.
Generative computer vision workloads
A mini diffusion pipeline (Food-101-Lite), showcasing Ray Data preprocessing, PyTorch Lightning integration, checkpointing, and image generation.
Policy learning workloads
Diffusion-policy pipeline on Gymnasium's Pendulum-v1 dataset, scaling across GPUs, checkpoint-per-epoch, and direct policy rollout in-notebook.
Recommendation system workloads
Matrix-factorization recommendation system with MovieLens 100K, streaming batches with iter_torch_batches, custom training loop with checkpointing, and modular separation of training/eval/inference.
:hidden:
getting-started/01_02_03_intro_to_ray_train.ipynb
workload-patterns/04a_vision_pattern.ipynb
workload-patterns/04b_tabular_workload_pattern.ipynb
workload-patterns/04c_time_series_workload_pattern.ipynb
workload-patterns/04d1_generative_cv_pattern.ipynb
workload-patterns/04d2_policy_learning_pattern.ipynb
workload-patterns/04e_rec_sys_workload_pattern.ipynb