About
Groq is an inference platform that uses custom LPUs to deliver low-latency responses for large language models. It targets developers and enterprises needing fast, affordable model execution via an OpenAI-compatible API.
Groq is an inference platform that uses custom LPUs to deliver low-latency responses for large language models. It targets developers and enterprises needing fast, affordable model execution via an OpenAI-compatible API.