The orchestrator is the intelligence layer that manages agent interactions. It:
- Interprets user intent
- Selects the appropriate agent(s)
- Coordinates multi-agent workflows
- Resolves conflicts between outputs
- Delivers unified responses
Orchestration Patterns
Choose a pattern based on your use case complexity.- Single Agent: One agent handles all requests. Best for focused, well-defined domains.
- Supervisor: Central orchestrator coordinates multiple specialized agents. Best for complex, parallelizable tasks.
- Adaptive Network: Agents dynamically hand off to each other. Best for sequential, multi-domain workflows.
Pattern Comparison
| Aspect | Single Agent | Supervisor | Adaptive Network |
|---|---|---|---|
| Complexity | Low | Medium | Medium-High |
| Agents | 1 | Multiple | Multiple |
| Coordination | None | Centralized | Decentralized |
| Execution | Sequential | Parallel | Sequential |
| Latency | Lowest | Medium | Low |
| Best for | Simple tasks | Complex decomposition | Dynamic hand-offs |
How Orchestration Works
Choosing the Right Pattern
Use Single Agent when:- Your app has one primary capability
- Tasks don’t require coordination between specialists
- You want minimal orchestration overhead
- Response latency is critical
- Tasks can be broken into independent subtasks
- You need parallel execution for speed
- Multiple specialists should contribute to responses
- You want centralized control and conflict resolution
- Tasks flow naturally between domains
- You need dynamic routing based on context
- Agents should autonomously decide when to hand off
- Sequential expertise is required
Orchestrator Responsibilities
Task Decomposition
Breaking complex requests into manageable subtasks:Agent Delegation
Routing tasks to appropriate specialists:Conflict Resolution
Handling inconsistencies between agent outputs:Context Management
Maintaining conversation state across agents:Orchestrator Configuration
Navigate to App > Orchestrator. For each orchestration pattern, configure the following.Default AI Model
AI model to be used for operations across the app. Select any of the configured models. The default settings of the model are shown. Click the Settings icon to update the default settings.Voice-to-Voice Interactions
Enable this field to allow users to interact with the app via real-time voice conversations. Once enabled, also provide the AI model that processes speech and generates voice responses. The platform supports various models. See the list of supported models and how to add an external model to the platform.Adaptive Network Orchestration does not support Gemini real time models.
Voice AI Model Settings
| Setting | Description | Key Notes |
|---|---|---|
| Voice | Voice used for audio responses | Depends on model/provider |
| Input Audio Format | Format of incoming audio | Example: pcm16, must match client input |
| Output Audio Format | Format of generated audio | Must match playback capability |
| Speech Speed | Speed of generated speech | 1.0 = default |
| Max Response Output Tokens | Max tokens per response | Controls response length & latency |
| Temperature | Controls randomness/creativity | Lower = deterministic, higher = creative |
| Max Tokens | Max tokens generated | Limits total response size |
| Noise Reduction Type | Filters input audio noise | Near Field (close mic), Far Field (room audio) |
| VAD Type | Speech detection method | Example: Server VAD |
| Threshold | Sensitivity of detection | Lower = more sensitive |
| Prefix Padding | Audio before speech detection | Prevents clipping |
| Silence Duration | Silence before speech ends | Lower = faster response |
| Create Response | Auto-generate response | True/False |
| Interrupt Response | Allow interruption | True/False |
| Transcription Language | Language for speech-to-text | Default: Auto-detect; improves accuracy if set |
| Transcription Prompt | Context for ASR model | Helps recognize domain-specific terms and improve accuracy. Learn More. |
Behavioral Instructions
Use this section to set the guidelines for the agent’s behavior. These instructions will be added to the orchestrator and the system prompt of each agent. Click Modify Instructions, then enter the prompt.Response Processor
The Response Processor is an application-level feature that lets you run a custom script on every agent response before it is delivered to the end user. It executes as the final stage in response generation, after the agent generates its output and before the output leaves the platform. Because it is configured at the app level on the Orchestrator page, the script applies consistently across all agents in the app, regardless of which agent handled the request. Using the Response Processor to Generate Artifacts Developers can write artifact payloads directly inside the Response Processor code. When the processor runs, it constructs the desired payload, writes it into the artifacts key, and the platform appends it for delivery. This is particularly useful when:- The artifact needs to be assembled from multiple tool outputs or session variables rather than a single tool response.
- The payload structure depends on business logic better handled centrally at the application level.
- No tool is needed — the processor can produce artifacts independently based on input context alone.
- Reorder elements to control which artifact renders first.
- Filter artifacts based on channel, user segment, or business logic.
- Transform or enrich payload data before delivery to the client.
- Merge multiple tool outputs into a single consolidated artifact.
- Add metadata, wrapper keys, or channel-specific formatting.
When a Response Processor is active, streaming is not supported. Artifacts and the text response are delivered as a complete payload after the processor finishes. If the processor fails, the original untransformed response is returned, and the error is logged.
Adding a Response Processor
Click Add Script to open the script editor. The following are available as input to the script:- Input — The original user input that triggered this agent run.
- Output — The agent’s response generated by the agent.
- Artifacts — Array of tool outputs and structured data returned during the run by different tools.
- Access input, output or artifacts in the script using $ prefix.
- Access environment variables using env keyword :
env.<variable-name>. - Access content variables as:
content.<variable-name>. - Access memory stores in the script using these methods.
Single Agent Configuration
In a Single Agent setup, all user requests are routed directly to the agent. Since no supervisor agent is involved, the agent’s prompt serves as the primary instruction set for the underlying model. When processing a request, the platform constructs a single consolidated prompt by combining the following components in order, and sends it to the model:- Agent Prompt — The core instructions that define the agent’s role and behavior.
- Behavioral Instructions — Guidelines that control tone, constraints, and response style.
- Tools Assigned to the Agent — Tool definitions available for the agent to invoke.
- Events Enabled in the Application — Event-related context.
Supervisor Configuration
In addition to the configurations discussed above, configure the following for the Supervisor pattern:- Orchestrator Prompt — A set of instructions for the supervisor of the app. This includes instructions and requirements that guide the orchestrator’s decision-making process.
- Orchestration Prompt for Voice-to-Voice Interactions — This prompt serves as instructions for the supervisor in case of voice interactions.
Adaptive Network Configuration
In addition to the configurations discussed above, configure the following for the Adaptive Network pattern:- Initial Agent — Select the agent that serves as the first point of contact for each task. This agent receives the user’s request, processes the initial requirements, and begins task execution.