llama-3.1-8b-instant — Free API Specifications
How to Configure llama-3.1-8b-instant for Free
https://api.groq.com/openai/v1 One-Click Config for Claude Code, Cursor & More
Claude Code
# Claude Code works via OpenRouter's Anthropic-compatible API.
# Note: Only paid Anthropic Claude models are supported (e.g. claude-sonnet-4.6, claude-opus-4).
# Browse available Claude models at: https://openrouter.ai/models?q=anthropic
# Add to ~/.zshrc or ~/.bashrc
export OPENROUTER_API_KEY="<your-openrouter-api-key>" # Get at https://openrouter.ai/settings/keys
export ANTHROPIC_BASE_URL="https://openrouter.ai/api"
export ANTHROPIC_AUTH_TOKEN="$OPENROUTER_API_KEY"
export ANTHROPIC_API_KEY="" # Must be explicitly empty to avoid conflicts
# Optional: pin specific models for each role
# export ANTHROPIC_DEFAULT_SONNET_MODEL="anthropic/claude-sonnet-4.6"
# export ANTHROPIC_DEFAULT_HAIKU_MODEL="anthropic/claude-haiku-4.5"
# Then simply run: claude Cursor
# Cursor → Settings (⚙️) → Models → Add Model
# Enter the model name exactly as shown, then fill in:
# Override OpenAI Base URL: https://api.groq.com/openai/v1
# OpenAI API Key: <your-api-key> # Get at https://console.groq.com/keys
# Click "Verify" to confirm the connection, then enable the model.
#
# Model name to add: llama-3.1-8b-instant Codex
# Add to ~/.zshrc or ~/.bashrc
export OPENAI_BASE_URL="https://api.groq.com/openai/v1"
export OPENAI_API_KEY="<your-api-key>" # Get at https://console.groq.com/keys
# Then run:
codex --model "llama-3.1-8b-instant" Gemini CLI
# ~/.gemini/settings.json
{
"apiKey": "<your-api-key>",
"model": "llama-3.1-8b-instant"
}
# Get API key at https://console.groq.com/keys OpenCode
// ~/.config/opencode/opencode.json
{
"$schema": "https://opencode.ai/config.json",
"provider": {
"free-llm": {
"npm": "@ai-sdk/openai-compatible",
"name": "Free LLM",
"options": {
"baseURL": "https://api.groq.com/openai/v1",
"apiKey": "<your-api-key>"
},
"models": {
"llama-3.1-8b-instant": { "name": "llama-3.1-8b-instant" }
}
}
}
}
// Get API key at https://console.groq.com/keys Hermes
# Step 1 — Edit config.yaml
# Windows: C:\Users\<you>\AppData\Local\hermes\config.yaml
# macOS/Linux: ~/.config/hermes/config.yaml
model:
default: llama-3.1-8b-instant
provider: custom
base_url: ${CUSTOM_BASE_URL}
api_key: ${CUSTOM_API_KEY}
model_aliases:
llama-3.1-8b-instant:
model: "llama-3.1-8b-instant"
provider: "custom"
# Step 2 — Edit .env (same directory as config.yaml)
# Windows: C:\Users\<you>\AppData\Local\hermes\.env
# macOS/Linux: ~/.config/hermes/.env
# ========================
# Custom API (OpenAI-compatible)
# ========================
CUSTOM_API_KEY=<your-api-key> # Get at https://console.groq.com/keys
CUSTOM_BASE_URL=https://api.groq.com/openai/v1 OpenClaw
// ~/.openclaw/openclaw.json (JSON5 format)
{
"agents": {
"defaults": {
"model": {
"primary": "llama-3.1-8b-instant",
},
},
},
"models": {
"providers": {
// Option A — Built-in provider (OpenAI, Anthropic, Google…)
// Just add apiKey; OpenClaw handles the baseUrl automatically
// "openai": { "apiKey": "<your-api-key>" },
// Option B — Custom OpenAI-compatible base URL (e.g. OpenRouter, NVIDIA)
"free-llm": {
"baseUrl": "https://api.groq.com/openai/v1",
"apiKey": "<your-api-key>", // Get at https://console.groq.com/keys
"api": "openai-completions", // openai-completions | anthropic-messages | …
"models": [
{ "id": "llama-3.1-8b-instant", "name": "llama-3.1-8b-instant" },
],
},
},
},
}
// Apply: openclaw gateway restart
// Verify: openclaw doctor --fix Frequently Asked Questions about llama-3.1-8b-instant
Is llama-3.1-8b-instant free to use?
Yes. llama-3.1-8b-instant is available on a permanently free tier via Groq. A credit card may be required to activate the free tier. The free tier includes a rate limit of 30 RPM, 14,400 RPD.
What is llama-3.1-8b-instant best for?
llama-3.1-8b-instant is optimized for chat tasks. It supports text modalities, with a context window of 131K tokens and a maximum output of 131K tokens. llama-3.1-8b-instant — free model from Groq.
Is llama-3.1-8b-instant OpenAI-compatible?
Yes. llama-3.1-8b-instant uses an OpenAI-compatible API endpoint at https://api.groq.com/openai/v1. You can use it with the OpenAI Python/JS SDK, or any tool that accepts a custom baseURL — including Claude Code (cc), Cursor, Codex, and OpenCode.
How do I get an API key for llama-3.1-8b-instant?
Visit Groq's API key page to register and generate a free API key. Once you have the key, use the configuration snippets above to set up Claude Code, Cursor, or your preferred AI coding tool.