doc/source/ray-air/deployment.rst
This page describes how you might use or deploy Ray in your infrastructure. There are two main deployment patterns -- pick and choose, and within existing platforms.
The core idea is that Ray can be complementary to your existing infrastructure and integration tools.
You can pick and choose which Ray AI libraries you want to use.
This is applicable if you are an ML engineer who wants to independently use a Ray library for a specific AI app or service use case and do not need to integrate with existing ML platforms.
For example, Alice wants to use RLlib to train models for her work project. Bob wants to use Ray Serve to deploy his model pipeline. In both cases, Alice and Bob can leverage these libraries independently without any coordination.
This scenario describes most usages of Ray libraries today.
.. https://docs.google.com/drawings/d/1DcrchNda9m_3MH45NuhgKY49ZCRtj2Xny5dgY0X9PCA/edit
.. image:: /images/air_arch_1.svg
In the above diagram:
Ray's many deployment modes <jobs-overview> to launch and manage Ray clusters and Ray applications.You may already have an existing machine learning platform but want to use some subset of Ray's ML libraries. For example, an ML engineer wants to use Ray within the ML Platform their organization has purchased (e.g., SageMaker, Vertex).
Ray can complement existing machine learning platforms by integrating with existing pipeline/workflow orchestrators, storage, and tracking services, without requiring a replacement of your entire ML platform.
.. image:: images/air_arch_2.png
In the above diagram: