Distributed training orchestration across clusters. Scales PyTorch/TensorFlow/HuggingFace from laptop to 1000s of nodes. Built-in hyperparameter tuning with Ray Tune, fault tolerance, elastic scaling. Use when training massive models across multiple machines or running distributed hyperparameter sweeps.
/plugin marketplace add zechenzhangAGI/AI-research-SKILLs/plugin install ray-train@zechenzhangAGI/AI-research-SKILLs