|  View source on GitHub | 
API to define datasets.
Classes
class BeamBasedBuilder: Beam based Builder.
class BeamMetadataDict: A tfds.core.Metadata object supporting Beam-generated datasets.
class BuilderConfig: Base class for DatasetBuilder data configuration.
class DatasetBuilder: Abstract base class for all datasets.
class DatasetCollectionLoader: Loader class for dataset collections.
class DatasetIdentity: Identity of a dataset that completely identifies a dataset.
class DatasetInfo: Information about a dataset.
class DatasetNotFoundError: Exception raised when the dataset cannot be found.
class Experiment: Experiments which can be enabled/disabled on a per version basis.
class FileFormat: Format of the record files.
class GeneratorBasedBuilder: Base class for datasets with data generation based on file adapter.
class Metadata: Abstract base class for DatasetInfo metadata container.
class MetadataDict: A tfds.core.Metadata object that acts as a dict.
class Path: Abstract base class for pathlib.Path-like API.
class ReadInstruction: Reading instruction for a dataset.
class SequentialWriter: Class to write a TFDS dataset sequentially.
class ShardedFileTemplate: Template to produce filenames for sharded datasets.
class SplitDict: Split info object.
class SplitGenerator: Defines the split information for the generator.
class SplitInfo: Wraps proto.SplitInfo with an additional property.
class Version: Dataset version MAJOR.MINOR.PATCH.
class lazy_imports: Lazy importer for heavy dependencies.
Functions
add_data_dir(...): Registers a new default data_dir to search for datasets.
as_path(...): DEPRECATED. Please use from etils import epath with epath.Path().
gcs_path(...): Returns the GCS URI path.
tfds_path(...): Path to tensorflow_datasets/ root dir.