Overview

Groq provides ultra-fast language model inference through their specialized AI inference hardware and optimized model serving infrastructure. This service offers high-speed text generation with dramatically reduced latency compared to traditional cloud providers, making it ideal for real-time applications and high-throughput scenarios.

Service Information

Property Value
Service Name Groq
Status Enabled
Compatible Nodes Call LLM, HyperFlow LLM PDF transformer

When to Use This Service

Groq is ideal for:

API Requirements

Available Models

Model availability depends on your Groq API access. Check the model dropdown in Call LLM node for currently available options.

Groq typically offers optimized versions of popular open-source models with significantly enhanced inference speed.

Parameters

Parameter Type Default Description
Model Choice (varies) Select Groq-optimized language model
Temperature Number 0.7 Response creativity control (0.0-2.0)
Max tokens Number 1000 Maximum response length
System message Text (empty) System-level instructions for model behavior

Key Features