Welcome to FlowZap, the App to diagram with Speed, Clarity and Control.

AI-Native Generator-Critic Architecture

Architecture

A competitive / generator-critic architecture where multiple generators produce independent answers, then an evaluator agent scores and selects the best output. This approach improves quality and reduces single-model brittleness. It's costlier (multiple LLM calls) but pays off when correctness or creativity matters more than latency.

Full FlowZap Code

Input { # Input
n1: circle label:"Start"
n2: rectangle label:"Creative brief submitted"
n1.handle(right) -> n2.handle(left)
n2.handle(bottom) -> Generators.n3.handle(top) [label="Brief"]
n2.handle(bottom) -> Generators.n4.handle(top) [label="Brief"]
n2.handle(bottom) -> Generators.n5.handle(top) [label="Brief"]
}
Generators { # Generator Agents
n3: rectangle label:"Agent A: Draft option 1"
n4: rectangle label:"Agent B: Draft option 2"
n5: rectangle label:"Agent C: Draft option 3"
n3.handle(bottom) -> Evaluator.n6.handle(top) [label="Option 1"]
n4.handle(bottom) -> Evaluator.n6.handle(top) [label="Option 2"]
n5.handle(bottom) -> Evaluator.n6.handle(top) [label="Option 3"]
}
Evaluator { # Evaluator Agent
n6: rectangle label:"Score all options"
n7: diamond label:"Quality threshold met?"
n8: rectangle label:"Select best output"
n9: circle label:"Done"
n6.handle(right) -> n7.handle(left)
n7.handle(right) -> n8.handle(left) [label="Yes"]
n7.handle(top) -> Generators.n3.handle(bottom) [label="No - Refine"]
n8.handle(right) -> n9.handle(left)
loop [refine until quality threshold met] n3 n4 n5 n6 n7
}

Related templates

AI-Native Orchestrator-Worker Architecture

Architecture

An orchestrator-worker architecture where an orchestrator agent breaks a goal into subtasks, dispatches to specialized workers, then synthesizes a final response. This is the most common 'agent orchestration' architecture—powerful but the orchestrator can become a bottleneck as the number of workers grows. Frameworks like LangGraph focus on explicit routing/state to manage this.

AI-Native Hierarchical Architecture

Architecture

A hierarchical multi-agent architecture that scales orchestration by stacking supervisors and team leads (a tree structure), which mirrors enterprise org structures and helps partition context. This is the 'enterprise-grade agentic AI architecture' when a single orchestrator cannot manage all workers directly. Ideal for large enterprises and multi-domain workflows.

AI-Native Single Agent Architecture

Architecture

A single-agent AI architecture where one agent handles everything: parsing requests, reasoning, calling tools via MCP, and generating responses. This is the default architecture for prototypes and simple automations—easy to debug but hits context-window limits quickly and is hard to parallelize. Ideal for MVPs and solo builders shipping fast.

AI-Native Sequential Pipeline Architecture

Architecture

A sequential pipeline architecture chaining multiple agents in a fixed order (parse → enrich → analyze → format), which is a common 'LLM microservices' setup when each step can be isolated. This structure is often used in document processing and ETL-like workflows because each step is testable and predictable.

AI-Native Parallel Fan-Out Architecture

Architecture

A parallel fan-out architecture that runs multiple agents simultaneously on independent checks (style, security, performance) and then merges results. This is a standard multi-agent design approach for throughput, mapping cleanly to CI/CD, incident response, and research. Fan-in reconciliation becomes the subtle part.

AI-Native Event-Driven Kafka Architecture

Architecture

An event-driven agentic AI architecture that replaces the central orchestrator with Kafka/PubSub topics: agents subscribe, react, and publish new events. This aligns multi-agent systems with proven microservices choreography and is ideal for real-time, high-throughput systems and 'agent mesh' setups.

Back to all templates