Hugging Face: The Open-Source Heart of Modern AI

November 7, 2025

Hugging Face: The Open-Source Heart of Modern AI

If you've worked with machine learning in the past few years, you've almost certainly used something built or hosted by Hugging Face. What started in 2016 as a chatbot app has become the world's largest open-source AI platform—hosting 2 million+ models, 500,000+ datasets, and roughly 1 million Spaces (per Hugging Face's State of Open Source report, Spring 2026). With 13+ million users and over 2,000 paying enterprise customers including Intel, Pfizer, Bloomberg, and eBay, Hugging Face has established itself as the definitive hub for collaborative AI development.

Today, Hugging Face isn't just transforming how we build language models—it's expanding into robotics, multimodal AI, and agentic systems, all while maintaining its core mission: making AI open, affordable, and accessible to everyone.

In this comprehensive guide, we'll explore Hugging Face's journey from chatbot to AI powerhouse, examine the technical innovations behind its ecosystem, and reveal the groundbreaking developments reshaping the company's future.


From Chatbot to AI Powerhouse: The Pivot That Changed Everything

Hugging Face was founded in 2016 by Clément Delangue (CEO), Julien Chaumond (CTO), and Thomas Wolf (Chief Science Officer). The trio originally set out to create a fun, personality-driven chatbot app for teenagers—an "artificial BFF" that launched publicly in March 2017 and quickly gained traction with 100,000 daily active users processing over 1 million messages per day.

But by 2018, the founders recognized a critical insight: while they could improve the underlying natural language processing technology, those improvements weren't translating to user growth. The chatbot app had limited potential, but the NLP infrastructure they'd built had enormous value for the developer community.

The Strategic Pivot

In May 2018, following a $4 million seed funding round led by Ronny Conway, Hugging Face made a bold decision: pivot from consumer chatbot to open-source platform for natural language processing.

The transformation accelerated with the release of the Transformers library—first launched as "pytorch-pretrained-bert" in November 2018, with the formal academic paper "HuggingFace's Transformers: State-of-the-art Natural Language Processing" published on arXiv in October 2019. The library has since become a cornerstone of the field, with the repository now passing 160,000 GitHub stars and over 3 million daily installs.

This single decision—choosing to build for developers rather than consumers, and to do so openly—set Hugging Face on a trajectory that would reshape the entire AI industry.


The Hugging Face Hub: Where the AI Community Builds

At the center of Hugging Face's ecosystem is the Hub, a collaborative platform that has become the definitive repository for machine learning assets (figures from Hugging Face's Spring 2026 State of Open Source report):

  • 2 million+ Models — from compact text classifiers to massive generative transformers; the Hub has been adding new repositories at a rate that put model #2M on the platform less than a year after model #1M
  • 500,000+ Datasets — for training, fine-tuning, and benchmarking across more than 8,000 languages
  • ~1 Million Spaces — interactive demos and applications powered by tools like Gradio and Streamlit
  • Over 2 billion total downloads in 2026, with the top 200 models accounting for roughly half of all traffic

The GitHub for AI

The Hub functions like GitHub for machine learning. Every model and dataset lives in a version-controlled repository. Developers can push updates, fork repositories, collaborate through pull requests, and track changes over time. The result is a thriving ecosystem where individuals, startups, and Fortune 500 companies contribute side by side.

Instant Access to State-of-the-Art Models

With just a few lines of Python, you can pull a model directly from the Hub and start using it:

from transformers import pipeline

# Load a sentiment analysis pipeline from the Hub
classifier = pipeline(
    "sentiment-analysis",
    model="distilbert-base-uncased-finetuned-sst-2-english"
)

result = classifier("Hugging Face makes AI development so much easier!")
print(result)
# [{'label': 'POSITIVE', 'score': 0.9998}]

That's the magic of Hugging Face—instant access to state-of-the-art models without needing to train or configure them from scratch. The platform democratizes AI by removing traditional barriers: expensive compute, complex setup, and deep ML expertise.

Datasets and Spaces: The Complete Ecosystem

Datasets provides a unified interface to hundreds of thousands of datasets across every domain—text, images, audio, video, and multimodal formats. Memory mapping enables efficient handling of massive datasets, while streaming support allows working with datasets larger than available disk space.

from datasets import load_dataset

# Load a dataset with a single line
dataset = load_dataset("imdb")

# Access with simple indexing
print(dataset["train"][0])

Spaces takes models from code to interactive applications. Developers can deploy Gradio or Streamlit demos with automatic hosting, SSL certificates, and collaborative development. From research prototypes to production demos, Spaces makes AI tangible and shareable.


The Transformers Library: Democratizing Deep Learning

When Hugging Face released the Transformers library in late 2018, it fundamentally democratized access to deep learning. Before Transformers, using models like BERT or GPT-2 required complex setup, custom code, and significant compute resources. Transformers changed that by providing a unified, high-level API.

Core Capabilities

PyTorch-First, Unified API: Earlier versions of Transformers worked across PyTorch, TensorFlow, and JAX. With the Transformers v5 release (December 2025), the library narrowed its backend focus to PyTorch as the primary framework, while TensorFlow and Flax support is being sunset for long-term maintainability.

Massive Model Library: Access to 400+ model architectures with over 1 million pretrained checkpoints compatible with Transformers on the Hub. Tasks span text classification, translation, summarization, question answering, generation, and far beyond.

Easy Fine-Tuning: The Trainer API with built-in support for mixed precision (including FP8), torch.compile() optimization, and Flash Attention makes fine-tuning on custom datasets remarkably simple.

Multimodal Evolution: Originally focused on NLP, Transformers now supports text, vision, audio, and multimodal tasks, reflecting the industry's evolution toward unified architectures.

Example: Fine-Tuning in Practice

Here's how simple it is to fine-tune a text classifier:

from transformers import (
    AutoModelForSequenceClassification,
    AutoTokenizer,
    Trainer,
    TrainingArguments
)
from datasets import load_dataset

# Load dataset and tokenizer
dataset = load_dataset("imdb")
tokenizer = AutoTokenizer.from_pretrained("distilbert-base-uncased")

# Tokenize the dataset
def tokenize(batch):
    return tokenizer(batch["text"], padding=True, truncation=True)

tokenized_datasets = dataset.map(tokenize, batched=True)

# Load model
model = AutoModelForSequenceClassification.from_pretrained(
    "distilbert-base-uncased",
    num_labels=2
)

# Configure training
args = TrainingArguments(
    output_dir="./results",
    evaluation_strategy="epoch",
    per_device_train_batch_size=8,
    num_train_epochs=2,
    fp16=True,  # Mixed precision training
)

# Train
trainer = Trainer(
    model=model,
    args=args,
    train_dataset=tokenized_datasets["train"].shuffle(seed=42).select(range(2000)),
    eval_dataset=tokenized_datasets["test"].select(range(500)),
)

trainer.train()

This code fine-tunes a DistilBERT model on a subset of the IMDb dataset—a task that would have required hundreds of lines before Hugging Face. The library's design philosophy prioritizes usability without sacrificing flexibility, enabling both beginners and experts to work efficiently.


Beyond Transformers: The Complete Library Ecosystem

While Transformers is the flagship, Hugging Face has built an entire suite of specialized tools covering the full machine learning lifecycle.

Tokenizers: High-Performance Text Processing

Tokenizers provides blazing-fast implementations in Rust with Python bindings. Supporting BPE, WordPiece, Unigram, and WordLevel algorithms (and interoperable with SentencePiece-style models), it can tokenize roughly 1GB of text in under 20 seconds on a server CPU. Full alignment tracking maps tokens back to original text positions—critical for tasks like named entity recognition.

Available in Python, Node.js, Rust, and Ruby, Tokenizers offers the performance needed for production systems.

Accelerate: Distributed Training Made Simple

Accelerate enables the same PyTorch code to run across any distributed configuration with minimal changes—just four lines of code. Released version 1.0.0 in 2024, it supports 6 hardware accelerators: CPU, GPU, TPU, XPU, NPU, and MLU.

Key features include:

  • Automatic mixed precision (including FP8)
  • FSDP and DeepSpeed support for large-scale training
  • device_map="auto" for big model inference across multiple GPUs
  • Integration with Transformers, Diffusers, PEFT, and TRL

Accelerate democratizes distributed computing, allowing researchers and startups to scale without deep infrastructure expertise.

Diffusers: Generative AI for Images, Video, and Audio

Diffusers (released July 2022) provides state-of-the-art pretrained diffusion models for generating images, videos, and audio. With over 10,000 compatible pipelines on the Hub, it's become the standard library for generative AI beyond text.

from diffusers import DiffusionPipeline
import torch

# Load a text-to-image pipeline
pipe = DiffusionPipeline.from_pretrained(
    "stabilityai/stable-diffusion-2-1",
    torch_dtype=torch.float16
)
pipe = pipe.to("cuda")

# Generate an image
image = pipe("A futuristic city at sunset, digital art").images[0]
image.save("generated_city.png")

Diffusers supports adapters like LoRA for efficient fine-tuning, and optimizations including offloading and quantization for memory-constrained devices. The design prioritizes usability, simplicity, and customizability—core Hugging Face values.

Datasets: The ML Data Hub

The Datasets library, which originated as a fork of TensorFlow Datasets, now powers access to the Hub's 500,000+ datasets covering text, image, audio, video, and multimodal formats across more than 8,000 languages. Memory mapping and streaming support enable working with datasets far larger than available RAM or disk space.

Integration spans PyTorch 2.0+, TensorFlow 2.6+, JAX 3.14+, as well as PyArrow, Pandas, Polars, and Spark. The simple load_dataset() function and efficient map() operations make data handling remarkably straightforward.

Evaluate: Measuring What Matters

Evaluate provides dozens of popular metrics covering NLP, computer vision, and audio. Three main categories organize the ecosystem:

  • Metrics: Measure model performance against ground truth (accuracy, F1, BLEU, etc.)
  • Comparisons: Analyze differences between models
  • Measurements: Assess dataset properties

The API is elegantly simple:

import evaluate

# Load a metric
accuracy = evaluate.load("accuracy")

# Compute scores
results = accuracy.compute(references=[0, 1, 2], predictions=[0, 1, 1])
print(results)  # {'accuracy': 0.6667}

Note for LLM evaluation: Hugging Face now recommends LightEval as a newer, more actively maintained alternative specifically for large language model evaluation.


Enterprise Solutions: AI at Scale

While open source remains at Hugging Face's core, the company offers enterprise-grade services enabling organizations to deploy and scale AI securely.

Inference Endpoints: Production-Ready Model Deployment

Inference Endpoints provides fully managed deployment of models as secure, scalable APIs. Available since 2022 with continuous improvements, the service offers:

  • Auto-scaling with scale-to-zero — pay only for what you use
  • Multi-cloud flexibility — choose your cloud provider (AWS, Azure, GCP), region, and hardware
  • Three security tiers:
    • Protected Endpoints — authentication required
    • Public Endpoints — open access for public demos
    • Private Endpoints — PrivateLink integration for VPC connectivity
  • Optimized inference engines — integration with vLLM, Text Generation Inference (TGI), and Text Embeddings Inference (TEI)
  • Broad model support across Transformers, Diffusers, Sentence Transformers, and more

Per Hugging Face's published pricing, CPU instances start at $0.033/hr (1 vCPU, 2GB RAM on AWS) and GPU instances range from $0.50/hr for an NVIDIA T4 up to roughly $10/hr for a single H100 and higher for larger configurations, billed by the minute. Enterprise plans add dedicated support and SLAs.

Example API usage:

curl https://api-inference.huggingface.co/models/distilbert-base-uncased-finetuned-sst-2-english \
  -X POST \
  -d '{"inputs": "Hugging Face is transforming AI!"}' \
  -H "Authorization: Bearer YOUR_HF_API_TOKEN"

The response returns JSON predictions, enabling seamless integration into any application without managing infrastructure.

Enterprise Hub: Private AI Infrastructure

Enterprise Hub (formerly Private Hub, announced August 2022) provides isolated environments where teams can host proprietary models and datasets with the same collaborative tools as the public Hub. This bridges the gap between open research and enterprise confidentiality requirements.

Three deployment options:

  1. Managed Private Hub — runs in segregated VPCs owned by Hugging Face
  2. On-Cloud Private Hub — runs in customer's own cloud account (AWS, Azure, GCP)
  3. On-Premise — deployment on customer infrastructure for strict compliance

Enterprise features:

  • SSO with SAML integration — centralized authentication
  • Resource Groups — granular role-based access control
  • Storage Regions — GDPR compliance across Europe, North America, and Asia Pacific
  • Comprehensive audit logs — full traceability
  • 1TB private storage per organization member ($25/month per extra TB)
  • 5x more ZeroGPU quota — expanded compute access
  • Advanced security policies — organization-wide enforcement

The platform is SOC2 Type 2 certified and GDPR compliant, meeting enterprise security and privacy requirements.


Strategic Partnerships: The Neutral Switzerland for AI

Hugging Face's positioning as the "neutral Switzerland for AI" is reflected in its diverse strategic partnerships spanning major cloud providers, chip manufacturers, and enterprise software leaders.

AWS: The Preferred Cloud Provider

Amazon Web Services serves as Hugging Face's designated preferred cloud provider—a critical partnership often underappreciated in public discussion. AWS invested in the August 2023 Series D round with a revenue-sharing agreement.

Integration highlights:

  • Hugging Face Deep Learning Containers (DLCs) — pre-configured environments for SageMaker
  • SageMaker JumpStart — one-click deployment of 10,000+ models
  • AWS Custom Silicon — Trainium for training, plus Inferentia/Inferentia2 for inference. AWS markets Inferentia2 as delivering up to 4x higher throughput and 10x lower latency than Inferentia1 (vendor benchmark, comparing the two AWS generations rather than vs. all GPUs)
  • Enterprise Hub via AWS Marketplace — direct billing through AWS accounts
  • Full AWS integration — support across EC2, S3, Lambda, and AWS Data Exchange

This partnership makes Hugging Face's ecosystem deeply integrated with the world's largest cloud platform, enabling enterprises to leverage AI without vendor lock-in.

Google Cloud: Strategic Collaboration and Investment

Google participated as both an investor in the $235M Series D round and a strategic cloud partner (announced January 25, 2024).

Key integrations:

  • One-click deployment from Hub to Vertex AI
  • Google Kubernetes Engine (GKE) support with Hugging Face containers
  • Cloud TPU v5e access — 2.5x better price-performance than previous generation
  • A3 VMs with NVIDIA H100 GPUs and C3 VMs with Intel Sapphire Rapids
  • 10,000+ models in Google Cloud Model Garden
  • Enterprise Hub subscriptions managed via Google Cloud accounts

Microsoft Azure / Microsoft Foundry: Deepening Multimodal Integration

The Microsoft partnership has evolved through multiple expansion phases (initial Azure ML announcement, the May 2024 Build expansion, and a deeper integration announced in 2025). Note: at Microsoft Ignite (November 2025) Microsoft rebranded Azure AI Foundry to Microsoft Foundry, with the new branding becoming effective January 1, 2026.

Highlights of the deeper integration (2025):

  • 10,000+ Hugging Face models in Microsoft Foundry (formerly Azure AI Foundry) with day-0 releases
  • Multimodal support — text, vision, speech models
  • One-click deployment from Hub to Azure Machine Learning
  • Azure-hosted model weights — secure deployment with no external egress
  • OpenAI Chat Completion API compatibility — drop-in replacement
  • Security-first approach — vulnerability scanning on all models

May 2024 additions:

  • 20+ new LLMs including Meta Llama variants
  • AMD MI300X GPU integration
  • Phi-3 integration in HuggingChat
  • Spaces Dev Mode with VS Code — full development environments in the browser

NVIDIA: Training, Inference, and Robotics

The NVIDIA partnership (announced August 2023, expanded November 2024 and January 2025) spans multiple dimensions:

Training infrastructure:

  • Training Cluster as a Service powered by NVIDIA DGX Cloud
  • Each instance features 8x A100 or H100 GPUs with 640GB GPU memory
  • Pricing starts at $36,999/month for multi-node AI supercomputing
  • January 2025 GTC Paris expansion added DGX Cloud Lepton integration with access to latest Hopper and GB200 GPUs

Inference optimization:

  • NVIDIA NIM integration — up to 5x higher throughput on H100 GPUs
  • Optimum-NVIDIA library — hardware-specific optimizations

Robotics collaboration (November 2024):

  • LeRobot integration with NVIDIA Isaac Lab
  • Support for GR00T and Jetson platforms
  • Real-time robotics inference capabilities

Model training partnership:

  • NVIDIA trained StarCoder2-15B using NeMo framework (released February 2024)

ServiceNow: Open-Source Code AI

ServiceNow co-leads the BigCode Project with Hugging Face—an open scientific collaboration for responsible LLM development for code (announced September 2022).

Major releases:

  • StarCoder (May 2023) — 15.5B-parameter model trained on roughly 1 trillion tokens from The Stack v1.2 dataset, covering 80+ programming languages, released under the BigCode OpenRAIL-M v1 license (not MIT — earlier coverage occasionally got this wrong)
  • StarCoder2 (February 2024) — three model sizes (3B, 7B, 15B), with the 15B variant trained on 600+ programming languages and 4.3T tokens, also under BigCode OpenRAIL-M

ServiceNow's Now LLM builds on StarCoder foundation, powering text-to-code, workflow generation, code completion, summarization, and snippet retrieval.

Additional Major Partnerships

Dell Technologies (announced November 2023; Dell Enterprise Hub portal globally available May 2024) — Designated preferred on-premises infrastructure partner for Enterprise Hub, with support for NVIDIA, AMD, and Intel Gaudi accelerators.

Cerebras (March 2025) — Cerebras inference is now available as a provider on the Hub. Cerebras reports more than 2,200 tokens/second for Llama 3.3 70B on its CS-3 Wafer-Scale Engine-3 hardware — a figure the company describes as roughly 70x typical GPU-based deployments (vendor benchmark, not independently audited beyond Cerebras's own and Artificial Analysis's measurements).

Cloudflare (April 2025) — FastRTC, Hugging Face's WebRTC library for AI apps, gets enterprise-grade Cloudflare TURN infrastructure with 10GB of free monthly streaming for users with a Hugging Face token.

JFrog (March 2025) — Advanced security scanning with "JFrog Certified" badges for verified models.

DigitalOcean (2024) — 1-Click Models powered by HUGS on GPU Droplets with NVIDIA H100.

This ecosystem of partnerships positions Hugging Face as the neutral platform where the entire AI industry collaborates, avoiding single-vendor capture while maintaining open-source principles.


BigScience: A Landmark in Open, Collaborative AI Research

Hugging Face helped launch BigScience in May 2021—a one-year research workshop representing one of the most ambitious open-science collaborations in AI history. Over 1,000 researchers from 60+ countries and 250+ institutions participated in building a massive language model transparently and responsibly.

The BLOOM Model: Open AI at Scale

BigScience produced BLOOM (BigScience Large Open-science Open-access Multilingual Language Model), released July 11, 2022 after training from March 11 to July 6, 2022 (117 days).

Technical specifications:

  • 176 billion parameters (176,247,271,424 precisely)
  • Decoder-only Transformer based on modified Megatron-LM GPT2
  • 70 layers, 112 attention heads, 14,336-dimensional hidden layers
  • 2,048-token sequence length with ALiBi positional encodings
  • 250,680-token vocabulary
  • 46 natural languages and 13 programming languages (Java, PHP, C++, Python, JavaScript, C#, Ruby, TypeScript, Lua, Go, Scala, Rust)
  • Among the first publicly-released 100B+ parameter open-access models with broad coverage of languages like Spanish, French, and Arabic

Training infrastructure:

  • 1.6TB preprocessed text from the ROOTS corpus (~366 billion tokens seen during training)
  • 384 NVIDIA A100 80GB GPUs (48 nodes) plus 32 reserve GPUs
  • France's Jean Zay supercomputer
  • Estimated cost: $2-5 million equivalent in cloud computing (€3M compute grant from CNRS and GENCI)
  • Released under RAIL License v1.0 (Responsible AI License)

BigScience's Legacy

Important clarification: BigScience was a time-limited one-year research workshop that ran May 2021–May 2022, not an ongoing organization. The ACL 2022 workshop session served as the closing event. BLOOM, however, is still hosted on the Hub and continues to be downloaded and studied, and the BigScience legacy lives on through derivative projects like BigLAM (focused on cultural heritage data).

BigScience demonstrated that large-scale AI research can be done openly, transparently, and inclusively—establishing new standards for documentation, governance, and ethical considerations.


The Robotics Revolution: Hugging Face's Bold New Frontier

Perhaps no recent development better illustrates Hugging Face's ambition than its aggressive expansion into robotics—positioning the company to democratize physical AI the same way it democratized language models.

The Vision: Open, Affordable, Private Robotics

Hugging Face's robotics mission emphasizes making the field "open, affordable, and private"—extending its core values into the physical world. The strategy combines affordable hardware with open-source software, creating an ecosystem where anyone can build, train, and deploy robots.

Hardware Platforms

HopeJR (announced May 2025, co-developed with The Robot Studio) — A full humanoid robot with:

  • 66 actuated degrees of freedom
  • Walking and manipulation capabilities
  • Open-source, 3D-printable design with dexterous hands
  • Target price: under $3,000 — dramatically undercutting competitors like Tesla's Optimus or Figure AI's robots
  • A waitlist is open; first units were originally targeted for late 2025, with shipments rolling out gradually

Reachy Mini (released following the Pollen Robotics acquisition, April 2025) — Desktop expressive robots:

  • Programmable in Python, JavaScript, and Scratch
  • Starting price around $299 (Lite version), with a more capable wireless/compute variant at higher tiers
  • Designed for AI experimentation, education, and human-robot interaction in a compact form factor

SO-101 Robotic Arm (updated from SO-100 in 2024) — Entry-level 3D-printed programmable manipulation platform for experimentation and education.

Software Ecosystem: LeRobot and SmolVLA

LeRobot Platform — Open-source framework for robotics AI with:

  • Pre-trained models for manipulation tasks
  • Community datasets for robot learning
  • Integration with NVIDIA Isaac Lab (November 2024)
  • Support for GR00T and Jetson platforms

SmolVLA (June 3, 2025) — A compact vision-language-action model:

  • ~450 million total parameters (a pretrained VLM perception module plus a ~100M-parameter action expert)
  • Runs on MacBooks or single consumer GPUs
  • Trained on LeRobot Community Datasets
  • Asynchronous inference for ~30% faster response and 2x task throughput
  • Enables real-time visual reasoning and action planning on edge devices

The robotics initiative represents Hugging Face's most distinctive recent development, extending the company's democratization mission from digital AI to physical embodied intelligence.


Multimodal AI: Beyond Text

While Hugging Face built its reputation on natural language processing, the platform has evolved dramatically to support multimodal AI across vision, audio, video, and combinations thereof.

SmolVLM: State-of-the-Art Vision-Language Models

SmolVLM (initial release November 26, 2024) delivers state-of-the-art vision-language capabilities in compact models optimized for edge devices — laptops and consumer GPUs. The original 2B-parameter SmolVLM was followed by smaller siblings, with the family now spanning 256M, 500M, and 2.2B parameter variants.

Architectural highlights:

  • Uses SmolLM2 (1.7B) as the language backbone, replacing the heavier Llama 3.1 8B used by predecessor models
  • Vision backbone based on shape-optimized SigLIP
  • More aggressive (~9x) pixel-shuffle visual compression vs. Idefics3's 4x

Capabilities:

  • Image captioning, visual question answering, document understanding, and OCR
  • Released under Apache 2.0 license with full open weights, datasets, and training recipes

Explosive Growth in Multimodal Models

The Hub now hosts over 2,000 multimodal models supporting:

  • Image-text models (CLIP, BLIP, LLaVA, Idefics)
  • Audio-text models (Whisper, Wav2Vec2)
  • Video-text models (VideoMAE, TimeSformer)
  • Document understanding (LayoutLM, Donut)

Recent (2025) developments:

  • 128k token context windows in models like Google's Gemma 3 family (released March 2025), enabling long-context visual reasoning
  • Multilingual VLMs with broad language support
  • Small VLMs (sub-2B parameters) optimized for edge deployment
  • Improving long-video understanding capabilities

IDEFICS: Open Multimodal Powerhouse

The IDEFICS family (Image-aware Decoder Enhanced à la Flamingo with Interleaved Cross-attentionS) is Hugging Face's open-access reproduction of DeepMind's Flamingo, with 9B and 80B-parameter variants that accept interleaved sequences of images and text.

Trained on the OBELICS dataset (115B tokens, 141M documents, 353M images), IDEFICS demonstrated that open-source multimodal AI could approach proprietary alternatives in capability while maintaining transparency and accessibility. Hugging Face has since released Idefics2 (8B parameters, April 2024) and continued to evolve the line.


Agentic AI and Developer Platforms: The Next Wave

Hugging Face is pioneering the next generation of AI applications through agentic systems and deeply integrated developer tools.

HuggingChat: The Open Alternative to ChatGPT (and Its Sunset)

HuggingChat launched in April 2023 as Hugging Face's free, open-source answer to ChatGPT, with a critical difference: users could pick from a pool of open-source models rather than being locked to a single proprietary system.

In mid-2025, Hugging Face announced it was sunsetting HuggingChat in its original form to make way for next-generation, more deeply integrated agentic interfaces — a reminder that even successful experiments get retired when the product strategy shifts. The philosophy of user choice and open alternatives, however, lives on across the rest of the ecosystem (see Inference Providers, GitHub Copilot Chat integration, etc.).

smolagents: Lightweight Agentic Framework

smolagents (December 2024/January 2025) offers a lightweight framework for creating agentic systems where LLMs control task flow dynamically.

Key features:

  • Tool integration (search engines, APIs, custom functions)
  • Dynamic task planning and execution
  • Hugging Face Hub interoperability
  • Minimal dependencies and simple API

Agents represent the next evolution beyond static prompting, enabling AI systems to reason about complex tasks, use tools, and adapt behavior based on intermediate results.

HUGS: Enterprise Deployment Platform

HUGS (Hugging Face Generative AI Services, late 2024) enables offline deployment and training of AI models in personalized enterprise environments.

Capabilities:

  • Production-ready optimization
  • Air-gapped deployment for maximum security
  • Custom model training on private data
  • Integration with enterprise infrastructure

OpenEnv: Standardizing Agent Environments

OpenEnv (launched November 2025 in collaboration with Meta's PyTorch team) provides an open-source platform for standardizing AI agent environments.

Components:

  • OpenEnv 0.1 specification — standard format for agent environments
  • Environment Hub — repository of secure sandboxes for agent development
  • Framework integrations — TorchForge, verl, TRL, SkyRL support

OpenEnv addresses the fragmentation in agentic AI development, creating common standards that accelerate research and production deployment.

GitHub Copilot Chat Integration

September 2025 marked a major milestone: Inference Providers now integrate with GitHub Copilot Chat, enabling developers to access open-source LLMs directly in VS Code version 1.104.0+.

Supported models include:

  • Kimi K2
  • DeepSeek V3.1
  • GLM 4.5
  • And many more from Hugging Face's ecosystem

This integration positions Hugging Face's open models as viable alternatives to proprietary coding assistants, reaching developers in their primary workflow environments—a strategic expansion of the company's reach.


Environmental Leadership: Making AI Sustainable

Hugging Face has emerged as a leader in environmental responsibility within the AI industry, developing tools and standards for measuring and reducing carbon emissions.

Carbon Emissions Tracking

CodeCarbon integration enables automatic emissions tracking directly in the Transformers library via automatic CodeCarbonCallback during training. The Hub supports filtering models by carbon footprint with an emissions_threshold parameter in HfApi.

Models can display CO2 emissions data in model cards, promoting transparency and encouraging efficiency improvements.

BLOOM Carbon Analysis: Setting Standards

The comprehensive carbon analysis of BLOOM training (2022-2023) found approximately 25 metric tons of direct CO2 emissions (~50 metric tons total including infrastructure and manufacturing)—significantly lower than comparable models.

This efficiency resulted from France's nuclear-powered computing grid at the Jean Zay supercomputer, demonstrating that infrastructure choices profoundly impact AI's environmental footprint.

Research and Advocacy

Sasha Luccioni, Hugging Face's AI and Climate Lead, has established:

  • Carbon efficiency classification systems
  • Tools for measuring environmental impact
  • Standards for documenting model emissions
  • Educational resources through the "Environmental Impacts of AI Primer" blog series

The research paper "Exploring the Carbon Footprint of Hugging Face's ML Models" (2023) provides comprehensive analysis of the platform's environmental impact.

This work establishes Hugging Face as a thought leader in sustainable AI, not just providing tools but actively researching and advocating for reduced environmental impact across the industry.


Funding and Growth: Becoming the AI Switzerland

Hugging Face's $235 million Series D round closed in August 2023 at a $4.5 billion valuation—doubling from $2 billion in 2022. The round was led by Salesforce Ventures with participation from:

  • Google
  • Amazon (AWS)
  • NVIDIA
  • Intel
  • AMD
  • Qualcomm
  • IBM
  • Sound Ventures

An additional smaller round from Premji Invest and Bossanova Investimentos occurred in January 2024. Total funding exceeds $400 million across nine rounds.

The "Neutral Switzerland" Strategy

The diverse investor base—spanning competing cloud providers, chip manufacturers, and enterprise software leaders—positions Hugging Face as the neutral platform where the entire AI industry collaborates.

Unlike platforms controlled by single vendors, Hugging Face avoids vendor lock-in, enabling users to choose their preferred:

  • Cloud provider (AWS, Azure / Microsoft Foundry, GCP, or on-premises)
  • Hardware (NVIDIA, AMD, Intel, Cerebras, and more)
  • Framework (PyTorch first-class with Transformers v5; the wider Hub still supports older TensorFlow/JAX checkpoints)
  • Model licensing (Apache 2.0, MIT, OpenRAIL-M, and other custom licenses)

This neutrality is a core strategic advantage, making Hugging Face the natural choice for open collaboration.

Revenue and Scale

Revenue grew from roughly $70 million ARR (2023) to approximately $130 million (2024), with 2,000+ paying enterprise customers as of mid-2025 — including Intel, Pfizer, Bloomberg, eBay, and many more. Third-party research (e.g., Sacra) put 2025 ARR around $190 million, though Hugging Face has not officially confirmed exact figures.

Platform metrics demonstrate extraordinary scale (Hugging Face State of Open Source, Spring 2026):

  • 13+ million users
  • 2 million+ models, 500,000+ datasets, ~1 million Spaces
  • Over 2 billion total downloads in 2026, with the top 200 models capturing ~half of all download volume
  • A modality split that skews heavily NLP (~58%), then computer vision (~21%) and audio (~15%)

The Philosophy: Democratizing Machine Learning

From day one, Hugging Face's mission has been about democratizing AI—making it accessible, transparent, and collaborative. The founders describe the company as the "GitHub of machine learning," but it's evolved beyond a simple repository into a movement toward open science.

Core Principles

Open by default: All tools released under permissive licenses (Apache 2.0, MIT). Research published openly. Models shared freely.

Community-driven: Success measured not by proprietary advantages but by community adoption and contribution. Over 1 million repositories created by the global community.

Accessible: Removing barriers of cost, complexity, and expertise. A high school student can access the same models as a Fortune 500 company.

Ethical: Transparent documentation through model cards and data cards. Active participation in responsible AI initiatives. Carbon emissions tracking built into tools.

Neutral: Avoiding vendor lock-in and single-platform capture. Supporting diverse cloud providers, hardware, and frameworks.


Community Recognition and Impact

The AI community has recognized Hugging Face's transformative impact through numerous accolades:

Emerge's 2024 Project of the Year — Named for transformative role in AI and commitment to democratization.

Academic adoption: Researchers worldwide use Hugging Face to share reproducible models and datasets, enabling peer verification and accelerating scientific progress.

Startup enablement: Young companies leverage pretrained models to build products faster without massive compute budgets, reducing barriers to AI entrepreneurship.

Enterprise transformation: Large organizations integrate Hugging Face models into customer service, content moderation, analytics systems, and internal tools.


Additional Notable Initiatives

ZeroGPU Project

Hugging Face offered $10 million worth of GPU compute power to the community, expanding AI training access for researchers and developers who lack expensive infrastructure.

HuggingSnap

HuggingSnap (2025) provides an iPhone application for on-device video understanding using SmolVLM models, demonstrating that powerful multimodal AI can run entirely on mobile devices without cloud dependence.

IBM & NASA Collaboration

In August 2025, IBM and NASA released the Surya foundation model on Hugging Face for solar weather prediction, part of the Prithvi family of geospatial, weather, and solar models. This demonstrates Hugging Face's expansion into scientific computing domains beyond traditional NLP and computer vision.

Inference Providers Integration

Inference Providers launched a unified API for accessing hundreds of ML models with:

  • Zero vendor lock-in — switch providers instantly
  • Pay-as-you-go pricing with no markup
  • Free tier availability for experimentation

The Road Ahead: Continuing Innovation

Hugging Face shows no signs of slowing down. Recent initiatives point to continued expansion across multiple frontiers:

Robotics scale-up: Continued rollout of HopeJR and Reachy Mini, with plans to scale production of affordable open-source humanoid robots.

Multimodal advancement: Further improvements in vision-language models, video understanding, and cross-modal reasoning.

Agent ecosystems: Continued development of agentic frameworks and standardization through OpenEnv.

Hardware optimization: Ongoing collaboration with chipmakers (NVIDIA, AMD, Intel, Cerebras) to improve inference speed and efficiency on diverse hardware.

Enterprise expansion: Deeper cloud provider integrations and enhanced security/compliance features for regulated industries.

Responsible AI standards: Continued leadership in transparency, documentation, bias mitigation, and environmental impact measurement.

The company's growth trajectory suggests it will remain at the center of AI innovation, evolving alongside the technology itself.


Conclusion: The Open-Source Heartbeat of AI

Hugging Face started as a chatbot experiment and became the backbone of modern machine learning. Its libraries—Transformers, Datasets, Tokenizers, Accelerate, Diffusers, Evaluate—form the foundation for countless AI projects. Its Hub has become the world's largest repository of models and datasets. And its commitment to ethical, open research has made it a moral compass in the fast-moving AI landscape.

With 2 million+ models, 500,000+ datasets, 13+ million users, and ambitious expansions into robotics and multimodal AI, Hugging Face is not just documenting the AI revolution—it's actively shaping it.

In a world where AI is transforming everything from art to medicine, robotics to climate science, Hugging Face reminds us that collaboration, transparency, and community matter just as much as code. The future of AI is open, and Hugging Face is building the infrastructure to make that future accessible to everyone.


Resources



FREE WEEKLY NEWSLETTER

Stay on the Nerd Track

One email per week — courses, deep dives, tools, and AI experiments.

No spam. Unsubscribe anytime.