Caching is currently enabled for Kimi-K2 and gpt-oss-20b according to Prompt Caching - GroqDocs
Can you roll out support for all models please? We’re using qwen3-32b a lot, and caching would be a huge improvement
Caching is currently enabled for Kimi-K2 and gpt-oss-20b according to Prompt Caching - GroqDocs
Can you roll out support for all models please? We’re using qwen3-32b a lot, and caching would be a huge improvement
Yep, we’re rolling it out for more and more models!
Do you have a timeline for the rollout to more models? We’d love to get it on gpt-oss-120b ASAP too
We’re doing rolling testing and releasing to more and more models currently, should be soon but I don’t have an exact rollout timeline.