Research

GPT-OSS Adoption Tracker, May 2026

Adoption tracker for OpenAI's GPT-OSS open-weight release in 2026. Hugging Face downloads, fine-tune count, hosted endpoints (Groq, Together, Fireworks), and how GPT-OSS competes with Llama, Qwen, and DeepSeek.

By Ramanath, CTO & Co-Founder at Presenc AI · Last updated: May 2026

OpenAI's First Major Open-Weight Release, One Cycle In

OpenAI released GPT-OSS (its first significant open-weight family) in two variants, 20B and 120B parameters, in mid-2025. The release represented OpenAI's strategic pivot toward open-weight participation after years of closed-API-only deployment. By May 2026, the family has accumulated meaningful adoption metrics. This page tracks where GPT-OSS stands against the dominant open-weight families (Llama, Qwen, DeepSeek).

Headline Adoption Metrics (May 2026)

Metricgpt-oss-20bgpt-oss-120b
Hugging Face Downloads (all-time)~7.3 million~4.6 million
Hugging Face Likes4,6044,772
Like-to-Download Ratio~0.06%~0.10% (very high)
Hosted Endpoints (Groq, Together, Fireworks, etc.)20+10+
Groq Pricing$0.075 / $0.30 per 1M (in/out)$0.15 / $0.60 per 1M (in/out)
Inference Speed (Groq LPU)~1,000 tps~500 tps

Position vs Major Open-Weight Families

FamilyTop Variant Downloads (HF, all-time)Top-20 Text-Gen Position
Qwen (Alibaba)Qwen 3.5 0.6B: 19.0M11 of top 20
Llama (Meta)Llama 3.1-8B-Instruct: 9.8M3 of top 20
GPT-OSS (OpenAI)gpt-oss-20b: 7.3M2 of top 20
DeepSeekDeepSeek V3.2: 4.1M2 of top 20
MistralMistral-7B-Instruct-v0.2: 3.2M1 of top 20

Six Things the Adoption Data Tells You

  1. GPT-OSS placed in the open-weight top 20 within a year. Combined ~11.9 million downloads is enough to make GPT-OSS the 4th-largest open-weight family on Hugging Face by deployment, behind only Qwen, Llama, and the legacy giants (GPT-2, OPT-125m). Strong debut for a brand-new entrant.
  2. The 120B variant has the highest like-to-download ratio in the cohort. 0.10 percent versus 0.06 percent for gpt-oss-20b and ~0.04 percent typical. The pattern reflects strong quality perception among the developers who chose to deploy it.
  3. Groq is the leading inference partner. gpt-oss-20b at ~1,000 tps and gpt-oss-120b at ~500 tps on Groq's LPU hardware. Pricing is competitive with the cost-leader tier (Llama 3.1-8B at $0.05 / $0.08). Together, Fireworks, and others host both variants.
  4. Qwen still dominates open-weight deployment. 11 of 20 most-downloaded text-generation models on HF are Qwen variants; Llama is at 3, GPT-OSS at 2, DeepSeek at 2. The "Chinese open-weight families lead deployment" narrative is structurally robust regardless of OpenAI's entry.
  5. GPT-OSS is not threatening Llama's position. Llama 3.1-8B alone (9.8M downloads) exceeds the entire GPT-OSS family combined (11.9M across two variants). Meta's Llama 4 ecosystem reach (Meta AI consumer surface + Llama API + multi-hoster) substantially exceeds OpenAI's open-weight reach.
  6. The strategic value is partly defensive. OpenAI's GPT-OSS release was partly a hedge against developer concern about closed-API dependency and partly a response to DeepSeek V3's competitive cost-quality position. Treating GPT-OSS as a pure commercial-revenue play would understate the strategic signal.

What This Means for AI Visibility

GPT-OSS adoption matters for brand visibility in two specific ways. First, agent frameworks that route to multiple open-weight models now include GPT-OSS in their model menus; brand recall on GPT-OSS may differ from brand recall on Llama or Qwen due to OpenAI's distinct training-data composition. Second, GPT-OSS deployment is concentrated among self-hosting enterprise developers, which is a high-value demographic for B2B brand visibility. Brands tracking only the cloud-API surface miss the GPT-OSS self-hosted footprint.

Methodology

Hugging Face download and like data pulled from the HF Hub API on May 14, 2026. Hosted endpoint count from Groq, Together AI, Fireworks AI, Replicate, and other inference-provider directories. Comparative open-weight ranking from our companion Hugging Face Most-Downloaded Models page. Refreshed quarterly.

How Presenc AI Helps

Presenc AI tracks brand-mention rates across open-weight model deployments alongside cloud-API surfaces. When a brand performs well on GPT-OSS but underperforms on Llama or Qwen (or vice versa), the gap signals training-data composition differences that matter for downstream agent stacks. For brands with enterprise self-hosting exposure, this open-weight tracking is structural to total AI visibility.

Frequently Asked Questions

Combined ~11.9 million Hugging Face downloads across gpt-oss-20b (~7.3M) and gpt-oss-120b (~4.6M), making GPT-OSS the 4th-largest open-weight family on Hugging Face by deployment. Hosted on Groq, Together AI, Fireworks AI, and 20+ other inference providers. Strong debut for OpenAI's first major open-weight release.
Smaller. Llama 3.1-8B-Instruct alone (~9.8M downloads) exceeds the entire GPT-OSS family combined. Meta's Llama also benefits from the Meta AI consumer surface distribution which OpenAI's open-weight release lacks. GPT-OSS is competitive on inference economics via Groq but not on raw deployment intensity yet.
For hosted inference, Groq leads on speed (~1,000 tps for 20B, ~500 tps for 120B) and price ($0.075/$0.30 and $0.15/$0.60 per 1M tokens). Together AI and Fireworks AI offer competitive pricing with broader feature support. For self-hosting, both variants run on consumer hardware (20B fits in 16-24GB GPU memory at Q4 quantization; 120B needs 80GB+ or multi-GPU).
Three reasons commonly cited: (1) defensive response to DeepSeek V3's competitive cost-quality position in early 2025; (2) developer-trust hedge against closed-API dependency concerns; (3) strategic positioning for hybrid deployments where enterprises run self-hosted inference for sensitive workloads alongside cloud API calls. Treating GPT-OSS as pure commercial play would miss the strategic signal.

Track Your AI Visibility

See how your brand appears across ChatGPT, Claude, Perplexity, and other AI platforms. Start monitoring today.