Back to Feast

Dask offline store

docs/reference/offline-stores/dask.md

0.63.02.3 KB
Original Source

Dask offline store

Description

The Dask offline store provides support for reading FileSources.

{% hint style="warning" %} All data is downloaded and joined using Python and therefore may not scale to production workloads. {% endhint %}

Example

{% code title="feature_store.yaml" %}

yaml
project: my_feature_repo
registry: data/registry.db
provider: local
offline_store:
  type: dask

{% endcode %}

The full set of configuration options is available in DaskOfflineStoreConfig.

Functionality Matrix

The set of functionality supported by offline stores is described in detail here. Below is a matrix indicating which functionality is supported by the dask offline store.

Dask
get_historical_features (point-in-time correct join)yes
pull_latest_from_table_or_query (retrieve latest feature values)yes
pull_all_from_table_or_query (retrieve a saved dataset)yes
offline_write_batch (persist dataframes to offline store)yes
write_logged_features (persist logged features to offline store)yes

Below is a matrix indicating which functionality is supported by DaskRetrievalJob.

Dask
export to dataframeyes
export to arrow tableyes
export to arrow batchesno
export to SQLno
export to data lake (S3, GCS, etc.)no
export to data warehouseno
export as Spark dataframeno
local execution of Python-based on-demand transformsyes
remote execution of Python-based on-demand transformsno
persist results in the offline storeyes
preview the query plan before executionyes
read partitioned datayes

To compare this set of functionality against other offline stores, please see the full functionality matrix.