Saturday, April 4, 2026
STORE MODEL DATA DIRECTLY ON HUGGING FACE HUB WITH NEW BUCKETS.
Hugging Face Hub now hosts model data directly, streamlining workflows.
Saturday, April 4, 2026
Hugging Face Hub now hosts model data directly, streamlining workflows.
Hugging Face, already a central repository for open-source AI models and datasets, has introduced "Storage Buckets." This new feature allows users to host and manage a wide range of model assets, raw data, logs, and other related files directly on the Hugging Face Hub. This streamlines the entire MLOps workflow by consolidating previously disparate storage solutions into a single, integrated platform.
This is a game-changer for ML engineers and researchers. Previously, you might have model weights on Hugging Face, datasets in S3, evaluation metrics in a custom database, and logs somewhere else entirely. Now, everything can reside in one place, tightly integrated with Hugging Face's existing versioning, collaboration, and sharing capabilities. It drastically reduces MLOps friction, simplifies artifact management, and accelerates the entire machine learning lifecycle from experimentation to deployment and monitoring. It cements Hugging Face's position as a foundational platform for AI development.
* Full MLOps Pipelines on Hugging Face: Design and implement end-to-end MLOps workflows that leverage Hugging Face Hub as the primary storage and artifact management system for all components – data, models, logs, metrics, and more. * Migration Tools: Develop utilities and scripts to automate the migration of existing model assets, datasets, and related files from various cloud storage providers (S3, GCS, Azure Blob) into Hugging Face Storage Buckets. * Collaboration & Annotation Platforms: Build tools on top of the Hub that enhance team collaboration around shared datasets and models stored in buckets, adding features like advanced data annotation, collaborative model evaluation, or experiment tracking. * Integrated Training Connectors: Create direct integrations with popular training platforms (e.g., PyTorch Lightning, TensorFlow, various cloud ML services) to seamlessly push and pull data from Hugging Face Buckets, minimizing data movement overhead.
Deeper integrations with enterprise-grade cloud security and compliance features for these buckets. Expansion of "Hub-first" tooling beyond models and datasets to encompass more MLOps functionalities like experiment tracking, model serving, and monitoring. Hugging Face becoming an even more dominant force across the entire AI development and MLOps landscape.
📎 Sources