Dataset.split(n: int, *, equal: bool = False, locality_hints: List[Any] | None = None) List[MaterializedDataset][source]#

Materialize and split the dataset into n disjoint pieces.

This method returns a list of MaterializedDataset that can be passed to Ray Tasks and Actors and used to read the dataset rows in parallel.


This operation will trigger execution of the lazy transformations performed on this dataset.


class Worker:

    def train(self, data_iterator):
        for batch in data_iterator.iter_batches(batch_size=8):

workers = [Worker.remote() for _ in range(4)]
shards = ray.data.range(100).split(n=4, equal=True)
ray.get([w.train.remote(s) for w, s in zip(workers, shards)])

Time complexity: O(1)

  • n – Number of child datasets to return.

  • equal – Whether to guarantee each split has an equal number of records. This might drop records if the rows can’t be divided equally among the splits.

  • locality_hints – [Experimental] A list of Ray actor handles of size n. The system tries to co-locate the blocks of the i-th dataset with the i-th actor to maximize data locality.


A list of n disjoint dataset splits.

See also


Unlike split(), which splits a dataset into approximately equal splits, Dataset.split_proportionately() lets you split a dataset into different sizes.


This method is equivalent to Dataset.split_at_indices() if you compute indices manually.


Unlike split(), streaming_split() doesn’t materialize the dataset in memory.