Are the models hosted on Groq platform isolated or shared?
We mostly have shared inference endpoints, though we do have custom/private models for higher tier Enterprise levels.
To get a sense of our architecture, this might help: Inside the LPU: Deconstructing Groq’s Speed | Groq is fast, low cost inference.