Overview
The AI Agent Node is the most important node in the Nadoo AI workflow engine. It is the primary interface for invoking large language models and orchestrating complex reasoning patterns. Built on the Agent 2.0 architecture, it supports six distinct execution modes that cover everything from simple Q&A to multi-step tool use and self-reflective generation. Every AI Agent Node is configured with a model, a system prompt, and an execution mode. The mode determines how the node interacts with the LLM — whether it makes a single call, chains multiple reasoning steps, loops with tools, or evaluates parallel reasoning paths.Agent 2.0 Execution Modes
- Standard
- Chain of Thought
- ReAct
- Function Calling
- Reflection
- Tree of Thoughts
Standard Mode
The simplest mode. The node sends the user’s message (plus conversation history and system prompt) to the LLM and returns the response directly. No loops, no tool calls, no extra reasoning steps.Best for: Simple Q&A, content generation, translation, summarization.Configuration:- Assemble the prompt (system + history + user message)
- Call the LLM
- Return the response
Context Window Management
Long conversations can exceed a model’s context window. The AI Agent Node provides three strategies for handling this:| Strategy | Behavior |
|---|---|
truncate | Remove the oldest messages until the conversation fits within the context window |
summarize | Use the LLM to summarize older messages, replacing them with a condensed version |
error | Fail with an error if the context window is exceeded (useful for debugging) |
Memory Integration
Enable conversational memory so the AI Agent retains context across multiple turns within a session.- message_window — Number of recent messages to include in each LLM call
- include_system_prompt — Whether to prepend the system prompt to every call
Model Settings
Fine-tune the LLM’s behavior with these parameters:| Parameter | Type | Default | Description |
|---|---|---|---|
temperature | float | 0.7 | Controls randomness (0 = deterministic, 2 = very random) |
max_tokens | int | 4096 | Maximum number of tokens in the response |
top_p | float | 1.0 | Nucleus sampling threshold |
frequency_penalty | float | 0.0 | Penalize tokens that appear frequently (range -2.0 to 2.0) |
presence_penalty | float | 0.0 | Penalize tokens that have appeared at all (range -2.0 to 2.0) |
stop_sequences | string[] | [] | Sequences that cause the model to stop generating |