Nexus

Groq

Use Groq's ultra-fast LPU inference with Nexus.

Groq provides ultra-fast inference via their custom LPU (Language Processing Unit) hardware. Uses an OpenAI-compatible API.

Installation

import "github.com/xraph/nexus/providers/groq"

Quick Start

provider := groq.New(os.Getenv("GROQ_API_KEY"))

gw := nexus.New(
    nexus.WithProvider(provider),
)

Options

OptionDescription
groq.WithBaseURL(url)Override the API base URL (default: https://api.groq.com/openai/v1)

Capabilities

CapabilitySupported
ChatYes
StreamingYes
EmbeddingsNo
VisionYes
ToolsYes
ThinkingNo

Models

ModelContextMax OutputInput PriceOutput Price
llama-3.3-70b-versatile128K32,768$0.59/M$0.79/M
llama-3.1-8b-instant131K8,192$0.05/M$0.08/M
mixtral-8x7b-3276832,76832,768$0.24/M$0.24/M
gemma2-9b-it8,1928,192$0.20/M$0.20/M

On this page