How to use a 10M context window? Rate limit issue

Unfortunately, Llama Scout currently only runs with a 131k token context window. In fact, I don’t believe any provider runs it with the whole 10M context window. If you’d still like to get more full context window requests in per minute, consider using "service_tier": "flex" or requesting a rate limit increase from Chat With Us.