Back to Ray

Ray Data Quickstart

doc/source/data/quickstart.rst

1.13.16.9 KB
Original Source

.. _data_quickstart:

Ray Data Quickstart

Get started with Ray Data's :class:Dataset <ray.data.Dataset> abstraction for distributed data processing.

This guide introduces you to the core capabilities of Ray Data:

  • :ref:Loading data <loading_key_concept>
  • :ref:Transforming data <transforming_key_concept>
  • :ref:Consuming data <consuming_key_concept>
  • :ref:Saving data <saving_key_concept>

Datasets

Ray Data's main abstraction is a :class:Dataset <ray.data.Dataset>, which represents a distributed collection of data. Datasets are specifically designed for machine learning workloads and can efficiently handle data collections that exceed a single machine's memory.

.. _loading_key_concept:

Loading data

Create datasets from various sources including local files, Python objects, and cloud storage services like S3 or GCS. Ray Data seamlessly integrates with any filesystem supported by Arrow <http://arrow.apache.org/docs/python/generated/pyarrow.fs.FileSystem.html>__.

.. testcode::

import ray

# Load a CSV dataset directly from S3
ds = ray.data.read_csv("s3://anonymous@air-example-data/iris.csv")

# Preview the first record
ds.show(limit=1)

.. testoutput::

{'sepal length (cm)': 5.1, 'sepal width (cm)': 3.5, 'petal length (cm)': 1.4, 'petal width (cm)': 0.2, 'target': 0}

To learn more about creating datasets from different sources, read :ref:Loading data <loading_data>.

.. _transforming_key_concept:

Transforming data

Apply user-defined functions (UDFs) to transform datasets. Ray automatically parallelizes these transformations across your cluster for better performance.

.. testcode::

from typing import Dict
import numpy as np

# Define a transformation to compute a "petal area" attribute
def transform_batch(batch: Dict[str, np.ndarray]) -> Dict[str, np.ndarray]:
    vec_a = batch["petal length (cm)"]
    vec_b = batch["petal width (cm)"]
    batch["petal area (cm^2)"] = np.round(vec_a * vec_b, 2)
    return batch

# Apply the transformation to our dataset
transformed_ds = ds.map_batches(transform_batch)

# View the updated schema with the new column
# .materialize() will execute all the lazy transformations and
# materialize the dataset into object store memory
print(transformed_ds.materialize())

.. testoutput::

shape: (150, 6)
╭───────────────────┬──────────────────┬───────────────────┬──────────────────┬────────┬───────────────────╮
│ sepal length (cm) ┆ sepal width (cm) ┆ petal length (cm) ┆ petal width (cm) ┆ target ┆ petal area (cm^2) │
│ ---               ┆ ---              ┆ ---               ┆ ---              ┆ ---    ┆ ---               │
│ double            ┆ double           ┆ double            ┆ double           ┆ int64  ┆ double            │
╞═══════════════════╪══════════════════╪═══════════════════╪══════════════════╪════════╪═══════════════════╡
│ 5.1               ┆ 3.5              ┆ 1.4               ┆ 0.2              ┆ 0      ┆ 0.28              │
│ 4.9               ┆ 3.0              ┆ 1.4               ┆ 0.2              ┆ 0      ┆ 0.28              │
│ 4.7               ┆ 3.2              ┆ 1.3               ┆ 0.2              ┆ 0      ┆ 0.26              │
│ 4.6               ┆ 3.1              ┆ 1.5               ┆ 0.2              ┆ 0      ┆ 0.3               │
│ 5.0               ┆ 3.6              ┆ 1.4               ┆ 0.2              ┆ 0      ┆ 0.28              │
│ …                 ┆ …                ┆ …                 ┆ …                ┆ …      ┆ …                 │
│ 6.7               ┆ 3.0              ┆ 5.2               ┆ 2.3              ┆ 2      ┆ 11.96             │
│ 6.3               ┆ 2.5              ┆ 5.0               ┆ 1.9              ┆ 2      ┆ 9.5               │
│ 6.5               ┆ 3.0              ┆ 5.2               ┆ 2.0              ┆ 2      ┆ 10.4              │
│ 6.2               ┆ 3.4              ┆ 5.4               ┆ 2.3              ┆ 2      ┆ 12.42             │
│ 5.9               ┆ 3.0              ┆ 5.1               ┆ 1.8              ┆ 2      ┆ 9.18              │
╰───────────────────┴──────────────────┴───────────────────┴──────────────────┴────────┴───────────────────╯
(Showing 10 of 150 rows)

To explore more transformation capabilities, read :ref:Transforming data <transforming_data>.

.. _consuming_key_concept:

Consuming data

Access dataset contents through convenient methods like :meth:~ray.data.Dataset.take_batch and :meth:~ray.data.Dataset.iter_batches. You can also pass datasets directly to Ray Tasks or Actors for distributed processing.

.. testcode::

# Extract the first 3 rows as a batch for processing
print(transformed_ds.take_batch(batch_size=3))

.. testoutput:: :options: +NORMALIZE_WHITESPACE

{'sepal length (cm)': array([5.1, 4.9, 4.7]),
    'sepal width (cm)': array([3.5, 3. , 3.2]),
    'petal length (cm)': array([1.4, 1.4, 1.3]),
    'petal width (cm)': array([0.2, 0.2, 0.2]),
    'target': array([0, 0, 0]),
    'petal area (cm^2)': array([0.28, 0.28, 0.26])}

For more details on working with dataset contents, see :ref:Iterating over Data <iterating-over-data> and :ref:Saving Data <saving-data>.

.. _saving_key_concept:

Saving data

Export processed datasets to a variety of formats and storage locations using methods like :meth:~ray.data.Dataset.write_parquet, :meth:~ray.data.Dataset.write_csv, and more.

.. testcode:: :hide:

# The number of blocks can be non-deterministic. Repartition the dataset beforehand
# so that the number of written files is consistent.
transformed_ds = transformed_ds.repartition(2)

.. testcode::

import os

# Save the transformed dataset as Parquet files
transformed_ds.write_parquet("/tmp/iris")

# Verify the files were created
print(os.listdir("/tmp/iris"))

.. testoutput:: :options: +MOCK

['..._000000.parquet', '..._000001.parquet']

For more information on saving datasets, see :ref:Saving data <saving-data>.