https://hf.co/datasets/<repo_id>. A warning is shown to let the user know about the custom code, and they can avoid this message in future by passing the argument trust_remote_code=True.trust_remote_code=True will be mandatory to load these datasets from the next major release of datasets.HF_DATASETS_TRUST_REMOTE_CODE=0 you can already disable custom code by default without waiting for the next release of datasetshttps://hf.co/datasets/<repo_id>/tree/refs%2Fconvert%2Fparquetload_dataset step that lists the data files of big repositories (up to x100) but requires huggingface_hub 0.20 or newerload_dataset that used to reload data from cache even if the dataset was updated on Hugging Face~/.cache/huggingface/datasets/username___dataset_name/config_name/version/commit_shadatasets 2.15 (using the old scheme) are still reloaded from cache_get_data_files_patterns by @lhoestq in https://github.com/huggingface/datasets/pull/6343usedforsecurity=False in hashlib methods (FIPS compliance) by @Wauplin in https://github.com/huggingface/datasets/pull/6414ruff for formatting by @mariosasko in https://github.com/huggingface/datasets/pull/6434tqdm wrapper by @mariosasko in https://github.com/huggingface/datasets/pull/6433Table.__getstate__ and Table.__setstate__ by @LZHgrla in https://github.com/huggingface/datasets/pull/6444filelock package for file locking by @mariosasko in https://github.com/huggingface/datasets/pull/6445** by @mariosasko in https://github.com/huggingface/datasets/pull/6449dill logic by @mariosasko in https://github.com/huggingface/datasets/pull/6454push_to_hub by @mariosasko in https://github.com/huggingface/datasets/pull/6461__repr__ by @lhoestq in https://github.com/huggingface/datasets/pull/6480torch.Generator objects by @mariosasko in https://github.com/huggingface/datasets/pull/6502list_files_info with list_repo_tree in push_to_hub by @mariosasko in https://github.com/huggingface/datasets/pull/6510Full Changelog: https://github.com/huggingface/datasets/compare/2.15.0...2.16.0
Fetched April 7, 2026