Overview
Standard Mode is the simplest execution mode for the AI Agent Node. The node assembles a prompt from the system message, conversation history, and user input, makes a single LLM call, and returns the response. There are no intermediate reasoning steps, no tool calls, and no self-critique loops. This is the default mode and should be your starting point for any new workflow. Upgrade to a more complex mode only when Standard does not meet your quality requirements.How It Works
Assemble Prompt
The node combines the system prompt, conversation history (if memory is enabled), and the current user message into a single prompt.
Call LLM
The assembled prompt is sent to the configured model. The response streams token-by-token via SSE
llm_token events.Configuration
| Parameter | Type | Default | Description |
|---|---|---|---|
agent_mode | string | — | Must be "standard" |
model | string | — | Model identifier (e.g., "gpt-4o", "claude-sonnet-4-20250514", "ollama/llama3") |
system_prompt | string | "" | Instructions that define the agent’s behavior and persona |
temperature | float | 0.7 | Controls randomness (0 = deterministic, 2 = highly random) |
max_tokens | int | 4096 | Maximum tokens in the response |
Additional Model Settings
| Parameter | Type | Default | Description |
|---|---|---|---|
top_p | float | 1.0 | Nucleus sampling threshold |
frequency_penalty | float | 0.0 | Penalize frequently appearing tokens (-2.0 to 2.0) |
presence_penalty | float | 0.0 | Penalize tokens that have appeared at all (-2.0 to 2.0) |
stop_sequences | string[] | [] | Sequences that cause the model to stop generating |
SSE Events
Standard mode emits the following events during execution:| Event | When | Payload |
|---|---|---|
node_started | Node begins execution | { node_id, node_type } |
llm_token | Each token is generated | { token, node_id } |
llm_finished | LLM generation completes | { node_id, total_tokens } |
node_finished | Node completes | { node_id, status } |
llm_thinking, llm_tool_call, or agent_reflection events.
Use Cases
Simple Q&A
Answer user questions based on the system prompt and conversation history. No external data or tools needed.
Content Generation
Generate blog posts, emails, marketing copy, or other text content from a prompt.
Translation
Translate text between languages. Set a low temperature (0.2-0.3) for consistent, accurate translations.
Summarization
Summarize documents, articles, or conversation threads. Works well when the full text fits within the context window.
Classification
Classify text into categories (sentiment, topic, language). Pair with a Condition Node downstream to branch on the result.
Formatting & Extraction
Reformat data, extract structured fields from unstructured text, or convert between formats (JSON, CSV, Markdown).
Example: RAG with Standard Mode
Standard mode works well for RAG workflows where the knowledge retrieval is handled by upstream nodes: The Search Knowledge Node retrieves relevant context, and the AI Agent Node in Standard mode uses that context to generate a response. No complex reasoning or tool use is needed because the retrieval is already done.System Prompt for RAG
Performance Characteristics
| Metric | Standard Mode |
|---|---|
| LLM calls per execution | 1 |
| Latency | Lowest (single round-trip) |
| Token usage | Lowest (no overhead from reasoning, reflection, or tool calls) |
| Quality ceiling | Moderate (limited by single-pass generation) |
Standard mode is typically 2-5x faster and 2-5x cheaper than multi-step modes like ReAct or Tree of Thoughts. Always benchmark Standard mode first before moving to a more complex strategy.
When to Upgrade
Consider switching to a more advanced mode when:- Accuracy on reasoning tasks is poor — Try Chain of Thought for step-by-step reasoning.
- The agent needs to call tools — Use ReAct or Function Calling mode.
- Output quality needs iterative improvement — Use Reflection for self-critique.
- Multiple approaches should be explored — Use Tree of Thoughts for parallel reasoning.
Best Practices
Write detailed system prompts
Write detailed system prompts
In Standard mode, the system prompt is your primary lever for controlling output quality. Be specific about the desired format, tone, length, and any constraints.
Use low temperature for factual tasks
Use low temperature for factual tasks
Set temperature to 0.0-0.3 for translation, extraction, and factual Q&A. Use 0.7-1.0 for creative tasks where variety is desirable.
Leverage upstream context
Leverage upstream context
Use Search Knowledge, Database, or Variable Nodes upstream to provide the AI Agent with all the context it needs. Standard mode excels when the input is well-prepared.
Set appropriate max_tokens
Set appropriate max_tokens
Avoid using very large
max_tokens values for tasks that need short answers. This wastes budget on unused capacity and can lead to unnecessarily verbose responses.