Today's AI/ML headlines are brought to you by ThreatPerspective

Digital Event Horizon

Revolutionizing Machine Learning: Hugging Face's Latest Streaming Enhancements



Hugging Face has introduced significant improvements to its datasets library, doubling data throughput and reducing startup requests by up to 100x. Learn how these enhancements are revolutionizing machine learning model training and development.

  • Efficient dataset loading is achieved through Hugging Face's updates to their datasets library.
  • The latest changes address startup phase issues and introduce a new streaming phase with improved performance.
  • Persistent Data Files Cache, Optimized Resolution Logic, Prefetching for Parquet, and Configurable Buffering enhance data loading efficiency.
  • Data throughput is doubled, allowing faster training and model development.
  • Xet deduplication technology reduces disk space errors and excessive data transfer issues.


  • Hugging Face has made significant strides in improving the efficiency and speed of loading datasets for machine learning models. The latest updates to their popular library, datasets, aim to reduce the burden of data loading on users, allowing them to focus on model training and development.

    The context reveals that Hugging Face has focused on two phases: startup and streaming. In the startup phase, they addressed the issue of creating a multitude of requests for resolving data files. Two major changes were implemented:

    1. Persistent Data Files Cache: The cache of list of data files is now stored across all DataLoader workers. This virtual elimination of startup requests significantly reduces resolution time.

    2. Optimized Resolution Logic: The initial worker's API calls to fetch the file list are minimized, reducing latency even further.

    In the streaming phase, two new features were introduced:

    1. Prefetching for Parquet: Prefetching is enabled for Parquet datasets, allowing the data pipeline to remain full and ensuring that the GPU never waits for data.

    2. Configurable Buffering: Advanced users can fine-tune streaming performance according to their hardware and network setup. This feature allows for maximum control over optimizing I/O operations.

    With these enhancements, Hugging Face has managed to increase data throughput by doubling it, enabling faster training and more efficient model development.

    To put this into perspective, when using plain S3, users often face issues such as disk space errors or 429 "stop requesting" errors due to excessive data transfer. However, with the introduction of Xet, a deduplication-based storage technology, these issues are significantly reduced. Hugging Face's pyspark_huggingface package includes support for Parquet Content-Defined Chunking and Xet deduplication, accelerating data transfers on the platform dramatically.

    Furthermore, the datasets library has been improved to support custom streaming pipelines. This feature is particularly useful for users who require more control over their streaming setup.

    The impact of these updates can be seen in nanoVLM's training process, where streaming enhancements enable faster performance compared to traditional local SSD setups.

    In conclusion, Hugging Face's latest streaming enhancements have revolutionized the way machine learning models are loaded and trained. With reduced startup requests, improved data resolution time, and enhanced streaming capabilities, users can focus on model development without worrying about tedious data loading processes.

    Related Information:
  • https://www.digitaleventhorizon.com/articles/Revolutionizing-Machine-Learning-Hugging-Faces-Latest-Streaming-Enhancements-deh.shtml

  • https://huggingface.co/blog/streaming-datasets


  • Published: Mon Oct 27 11:44:23 2025 by llama3.2 3B Q4_K_M











    © Digital Event Horizon . All rights reserved.

    Privacy | Terms of Use | Contact Us