Prompt Caching

Caching is currently enabled for Kimi-K2 and gpt-oss-20b according to Prompt Caching - GroqDocs

Can you roll out support for all models please? We’re using qwen3-32b a lot, and caching would be a huge improvement

Yep, we’re rolling it out for more and more models!

Do you have a timeline for the rollout to more models? We’d love to get it on gpt-oss-120b ASAP too

We’re doing rolling testing and releasing to more and more models currently, should be soon but I don’t have an exact rollout timeline.