Learn how to use the Groq slash command for fast LLM inference
prompt
- Your instructions or questionsmodel
- Specific model to use (defaults to meta-llama/llama-4-maverick-17b-128e-instruct)
system prompt
- Override the default system prompt
files
- File URLs to include in the request (see LLM File Type Support for supported formats)
meta-llama/llama-4-scout-17b-16e-instruct
- General purpose with balanced performance (16,384 tokens, Very Fast)meta-llama/llama-4-maverick-17b-128e-instruct
(default) - Long context tasks and document analysis (131,072 tokens, Fast)deepseek-r1-distill-llama-70b
- Advanced reasoning and complex problem solving (8,192 tokens, Fast)llama-3.3-70b-versatile
- General purpose with high quality output (32,768 tokens, Fast)llama-3.1-70b-versatile
- Long context and versatile tasks (131,072 tokens, Fast)llama-3.1-8b-instant
- Quick responses with high throughput (131,072 tokens, Very Fast)mixtral-8x7b-32768
- Multi-task capabilities with balanced performance (32,768 tokens, Fast)gemma2-9b-it
- Efficient processing for general tasks (8,192 tokens, Very Fast)