⚡ What’s New at Groq: October Highlights

Hey everyone!

It’s been a busy month at Groq- new models, better pricing, SDK upgrades, and community updates galore. Here’s a roundup of everything that dropped in October :backhand_index_pointing_down:


:glowing_star: New Model Drop: GPT-OSS-Safeguard 20B

OpenAI’s first safety-focused open-weights model landed on Groq with day-zero support!

Safeguard 20B lets you build customizable Trust & Safety AI: bring your own taxonomy and policies to classify content however you need.

Why it’s awesome:

  • 131K token context window and 65K max output tokens
  • Prompt caching enabled → 50 % cost savings on cached input tokens
  • Full reasoning traces for audits & debugging
  • Supports tool use, browser search, code execution, and JSON modes
  • Runs at 1000 + t/s on Groq, same price as the base GPT-OSS-20B model

:backhand_index_pointing_right: Get started with GPT-OSS-Safeguard 20B


:money_with_wings: Price Drops + Prompt Caching

Good news for your wallet - GPT-OSS models are now even cheaper and faster on Groq:

Plus, automatic prompt caching is now live for GPT-OSS 120b. Cached hits automatically provide:

  • 50 % cost savings on cached input tokens
  • Lower latency
  • Higher effective rate limits (cached tokens don’t count toward rate limits)
  • Zero setup, caching activates automatically when requests share common prefixes

:link: Learn more about prompt caching here


:toolbox: SDKs, Tools, & Ecosystem Updates

That’s it for last month’s updates! Keep building FAST!

3 Likes