AI Engineer Guide

How to use OpenAI’s gpt-oss models in Groq

OpenAI’s open model LLM gpt-oss are available in Groq

Both the 20b and 120b parameters are supported as well.

AI ModelCurrent Speed (Tokens per Second)Input Token Price (Per Million Tokens)Output Token Price (Per Million Tokens)
GPT OSS 20B 128k1,000$0.10 (10M / $1)*$0.50 (2M / $1)*
GPT OSS 120B 128k500$0.15 (6.67M / $1)*$0.75 (1.33M / $1)*

Usage

Replace GROQ_API_KEY with your API key.

curl "https://api.groq.com/openai/v1/chat/completions" \
  -X POST \
  -H "Content-Type: application/json" \
  -H "Authorization: Bearer ${GROQ_API_KEY}" \
  -d '{
         "messages": [
           {
             "role": "user",
             "content": ""
           }
         ],
         "model": "openai/gpt-oss-120b",
         "temperature": 1,
         "max_completion_tokens": 8192,
         "top_p": 1,
         "stream": true,
         "reasoning_effort": "medium",
         "stop": null
       }'
  

For sdk code snippets you can refer to it in the API playground in Groq

Happy faster inference

#Gpt-Oss #Groq

Stay Updated

Get the latest AI engineering insights delivered to your inbox.

No spam. Unsubscribe at any time.