Groq provides ultra-fast language model inference through their specialized AI inference hardware and optimized model serving infrastructure. This service offers high-speed text generation with dramatically reduced latency compared to traditional cloud providers, making it ideal for real-time applications and high-throughput scenarios.
| Property | Value | 
|---|---|
| Service Name | Groq | 
| Status | Enabled | 
| Compatible Nodes | Call LLM, HyperFlow LLM PDF transformer | 
Groq is ideal for:
Model availability depends on your Groq API access. Check the model dropdown in Call LLM node for currently available options.
Groq typically offers optimized versions of popular open-source models with significantly enhanced inference speed.
| Parameter | Type | Default | Description | 
|---|---|---|---|
| Model | Choice | (varies) | Select Groq-optimized language model | 
| Temperature | Number | 0.7 | Response creativity control (0.0-2.0) | 
| Max tokens | Number | 1000 | Maximum response length | 
| System message | Text | (empty) | System-level instructions for model behavior |