import datasets import glob import os import numpy as np NUM_SHARDS = 10 _URLS = [ f'https://huggingface.co/datasets/commaai/comma2k19/resolve/main/Chunk_{i}.zip' for i in range(1,NUM_SHARDS+1) ] _DESCRIPTION = """\ comma2k19 is a dataset of over 33 hours of commute in California's 280 highway. This means 2019 segments, 1 minute long each, on a 20km section of highway driving between California's San Jose and San Francisco. comma2k19 is a fully reproducible and scalable dataset. The data was collected using comma EONs that has sensors similar to those of any modern smartphone including a road-facing camera, phone GPS, thermometers and 9-axis IMU. Additionally, the EON captures raw GNSS measurements and all CAN data sent by the car with a comma grey panda. """ class Comma2k19(datasets.GeneratorBasedBuilder): def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features( {"path": datasets.Value("string")} ) ) def _split_generators(self, dl_manager): """Returns SplitGenerators.""" dl_manager.download_config.ignore_url_params = True downloaded_files = dl_manager.download(_URLS) local_extracted_archive = dl_manager.extract(downloaded_files) if not dl_manager.is_streaming else [None]*len(downloaded_files) return [ datasets.SplitGenerator( name=str(i), gen_kwargs={"local_extracted_archive":local_extracted_archive[i], "files": dl_manager.iter_archive(downloaded_files[i])} ) for i in range(len(downloaded_files))] def _generate_examples(self, local_extracted_archive, files): files = [os.path.join(dp, f) for dp, dn, filenames in os.walk(local_extracted_archive) for f in filenames] for path in files: yield path, {'path': path} def _get_examples_iterable_for_split(self, split_generator): for path in split_generator.gen_kwargs['files']: yield path[0], {'path': path[0]}