Fast LLM inference, OpenAI-compatible. Simple to integrate, easy to scale. Start building in minutes.
import OpenAI from "openai"; const client = new OpenAI({ apiKey: process.env.GROQ_API_KEY, baseURL: "https://api.groq.com/openai/v1", }); const response = await client.responses.create({ model: "openai/gpt-oss-20b", input: "Explain the importance of fast language models", }); console.log(response.output_text);from openai import OpenAI import os client = OpenAI( api_key=os.environ.get("GROQ_API_KEY"), base_url="https://api.groq.com/openai/v1", ) response = client.responses.create( input="Explain the importance of fast language models", model="openai/gpt-oss-20b", ) print(response.output_text)curl -X POST https://api.groq.com/openai/v1/responses \ -H "Authorization: Bearer $GROQ_API_KEY" \ -H "Content-Type: application/json" \ -d '{ "model": "openai/gpt-oss-20b", "input": "Explain the importance of fast language models" }'Take a quick video tour on how to get started building your app on Groq.