Designing Multi-Agent Systems
Reba Habib

The dominant mental model for AI in product design has been, until recently, a single model responding to a single user. A user asks a question; an AI answers it. A user uploads a document; an AI summarizes it. A user describes a task; an AI completes it. This model is intuitive, it maps cleanly onto conventional interface design, and it has been sufficient for the first generation of consumer AI products. It is not sufficient for what is coming next.
The frontier of AI product development is shifting rapidly toward multi-agent systems: architectures in which multiple AI models work together, each with specialized capabilities and defined responsibilities, coordinating to accomplish tasks that no single model could accomplish alone. These systems are already in production at significant scale. GitHub Copilot Workspace uses multiple agents to plan, implement, and review code changes. Google's Gemini ecosystem coordinates specialized models for different modalities and task types. Enterprise AI platforms from Salesforce, Microsoft, and ServiceNow are built on agent orchestration frameworks that coordinate multiple AI components across complex business workflows. The shift from single-model to multi-agent architectures is not a future possibility; it is a present reality that product and design leaders need to understand and design for now.
What makes multi-agent systems particularly challenging from a design perspective is that they introduce a new class of complexity that is qualitatively different from single-model complexity. When a single model fails, the failure is localized and relatively straightforward to diagnose. When a multi-agent system fails, the failure may be the result of an emergent interaction between agents that performed correctly in isolation but produced an incorrect result when combined. When a single model produces an uncertain output, the interface can communicate that uncertainty directly to the user. When a multi-agent system produces an uncertain output, that uncertainty has been propagated through multiple steps of inference, and communicating it meaningfully requires understanding how uncertainty accumulates and transforms as it passes through the system. These are genuinely hard design problems, and they require design thinking that is grounded in an understanding of how multi-agent architectures work.
The Architecture of Multi-Agent Systems
Understanding the design challenges of multi-agent systems requires a working understanding of their architecture. This is not a demand for deep technical knowledge; it is a demand for the kind of conceptual literacy that allows design leaders to participate meaningfully in architectural decisions that have significant design implications.
Multi-agent systems are typically organized around three architectural roles, and the design implications differ significantly across all of them.
The orchestrator is the agent responsible for decomposing a complex task into subtasks, assigning those subtasks to specialized agents, and synthesizing the results into a coherent output. The orchestrator does not typically execute tasks directly; it manages the execution of tasks by other agents. From a design perspective, the orchestrator is the primary point of contact with the human user and the primary source of the user's experience of the system's intelligence and coherence. The design of the orchestrator's interface, including how it communicates what it is doing, what it has found, where it is uncertain, and when it needs human input, determines most of what the user experiences as the multi-agent system's quality.
Specialist agents are the agents that execute specific subtasks: a web search agent that retrieves information from the internet, a code execution agent that runs and tests code, a data analysis agent that processes structured data, a writing agent that produces natural language outputs. Each specialist agent has a bounded capability and a defined interface with the orchestrator. From a design perspective, specialist agents are largely invisible to end users, but the design of their interfaces with the orchestrator, including how they communicate results, uncertainties, and failures, significantly affects the quality of the orchestrator's outputs and therefore the user's experience.
Human-in-the-loop checkpoints are the points at which the multi-agent system pauses to request human input, approval, or correction. The design of these checkpoints is one of the most critical and most underinvested aspects of multi-agent system design. When should the system pause and ask for human input? What information does the human need to make a meaningful decision at that point? How is the system's progress communicated so that the human can evaluate it intelligently? These are design questions with significant consequences for both the quality of the system's outputs and the user's experience of participating in the system's workflow.
Why Multi-Agent Systems Create New Design Challenges
The design challenges introduced by multi-agent systems are not simply scaled-up versions of single-model design challenges. They are qualitatively different in several ways that require new design thinking.
The first new challenge is opacity at the system level. A well-designed single-model interface can communicate meaningfully about the model's reasoning, uncertainty, and limitations because those properties are relatively well-defined for a single model. In a multi-agent system, the reasoning that produced an output is distributed across multiple agents, each of which made its own inferences, applied its own judgment, and produced its own intermediate outputs. Reconstructing a coherent explanation of why the system produced a particular final output requires tracing a chain of reasoning that spans multiple models, multiple steps, and potentially multiple data sources. This is technically difficult and communicatively demanding, and most current multi-agent systems handle it poorly, either by providing no explanation at all or by providing explanations that are technically accurate but practically incomprehensible to users.
OpenAI's research on AI transparency has identified this as one of the central unsolved problems in AI user experience design. As AI systems become more complex, the gap between the complexity of their internal processes and users' ability to understand those processes grows wider, not narrower. Multi-agent systems accelerate this divergence. Designing meaningful transparency in multi-agent systems requires moving away from the goal of explaining every step of the system's reasoning, which is both technically difficult and likely to overwhelm users, toward designing transparency that is calibrated to what users actually need to know to evaluate the system's outputs and make informed decisions about whether and how to act on them.
The second new challenge is error propagation. In a multi-agent system, errors made by specialist agents propagate through the system and can compound in ways that are difficult to predict or detect. A web search agent that retrieves slightly outdated information passes that information to a synthesis agent, which incorporates it into an analysis, which the orchestrator presents to the user as a confident conclusion. The final output is wrong, but the error is not attributable to any single agent's failure; it is the result of an interaction between agents that each performed within their design parameters. Designing systems that are robust to this kind of error propagation requires both technical mechanisms, including validation steps, confidence thresholds, and cross-checking between agents, and interface design that communicates the system's confidence in its outputs in ways that are calibrated to the actual error profile of the multi-agent pipeline.
The third new challenge is the design of human oversight in long-running agent workflows. When a multi-agent system is executing a complex, multi-step task that takes minutes or hours to complete, the design of the human oversight experience is fundamentally different from the design of a single-turn AI interaction. Users need to be able to monitor progress, intervene when the system goes in a wrong direction, approve consequential actions before they are taken, and resume tasks that have been interrupted. These requirements map onto a different class of interface design, closer to process management tools than to conversational AI interfaces, and they require design thinking that most UX organizations have not yet developed.
Designing the Orchestration Layer
Because the orchestrator is the primary point of human-system contact in a multi-agent architecture, the design of the orchestration layer deserves particular attention. The orchestrator is, in effect, the face of the multi-agent system, and its design determines most of what users experience as the system's intelligence, coherence, and trustworthiness.
The most important design decision at the orchestration layer is the level of transparency about the system's internal process. There is a spectrum of possible approaches here. At one extreme, the orchestrator presents only its final outputs, presenting the multi-agent system as a black box that takes inputs and produces results. At the other extreme, the orchestrator provides a complete log of every agent's actions, every intermediate output, and every decision point in the workflow. Neither extreme is optimal. The black box approach fails to give users the information they need to evaluate the system's outputs or to identify when the system has gone wrong. The complete log approach produces information overload that is practically useless to most users.
The right design point on this spectrum depends on the use case and the user population, but the general principle is that transparency should be progressive: the orchestrator should present a high-level summary of its process and provide mechanisms for users who need more detail to access it at increasing levels of granularity. This is analogous to the progressive disclosure principle in information architecture, applied to AI process transparency rather than content organization.
Microsoft's Copilot Studio, which allows enterprise developers to build multi-agent workflows on top of Microsoft's AI infrastructure, has grappled explicitly with this design challenge. The design team's published reflections on Copilot Studio's development describe the tension between providing enough transparency to build user trust and avoiding the cognitive overhead that comes from exposing the full complexity of multi-agent orchestration. The solution they arrived at, a step-by-step progress indicator that summarizes agent actions at a high level with the option to expand any step for more detail, is an instructive example of progressive transparency design in a multi-agent context.
The second important design decision at the orchestration layer is the design of task decomposition visibility. When an orchestrator decomposes a complex user request into subtasks, communicating that decomposition to the user serves two purposes: it gives the user an opportunity to correct misunderstandings before significant work is done, and it sets appropriate expectations for what the system will produce. The design challenge is communicating task decomposition in terms that are meaningful to users without requiring them to understand the technical structure of the multi-agent workflow.
GitHub Copilot Workspace's approach to this is worth examining. When a user submits a coding task to Copilot Workspace, the system presents a structured plan of the steps it intends to take before executing any of them. The plan is written in natural language, organized around the user's goals rather than the system's internal processes, and includes explicit points at which the user can modify or redirect the plan. This design approach treats task decomposition visibility not as a technical transparency requirement but as a collaborative planning tool, which is both more useful to users and more aligned with how skilled human collaborators would approach complex task delegation.
Human-in-the-Loop Design for Multi-Agent Systems
The design of human oversight checkpoints is where multi-agent system design intersects most directly with the broader questions of human-AI collaboration that Stanford HAI and other research organizations have been studying. The fundamental question is: at what points in a multi-agent workflow should the system pause and involve a human, and what does meaningful human involvement look like at those points?
The naive answer is that humans should be involved at every consequential decision point. But in a complex multi-agent workflow, consequential decision points may be numerous, and requiring human approval at each one would make the system so slow and demanding that its value would be largely negated. The sophisticated answer is that the design of human-in-the-loop checkpoints should be calibrated to the risk profile of the actions being taken, the confidence of the system in its outputs, and the reversibility of the consequences.
This calibration principle has concrete design implications. Actions that are irreversible, that have external consequences, or that involve significant resources should always require explicit human approval before execution. These are hard stops in the workflow, not optional review points. Actions that are reversible, that have contained consequences, and that the system is executing with high confidence can proceed autonomously, with the results presented to the user for review after the fact. Actions that fall in between these categories, moderately consequential, moderately reversible, or executed with moderate confidence, are the ones where the design of the checkpoint matters most. These checkpoints should present users with enough information to make a meaningful decision without overwhelming them with the technical details of the system's reasoning.
The design of what is presented at a human-in-the-loop checkpoint is as important as the design of when those checkpoints occur. A checkpoint that presents a complex technical output and asks the user to approve or reject it without context is not meaningful oversight; it is oversight theater. Meaningful oversight requires presenting the system's proposed action, the reasoning behind it, the expected consequences, the alternatives that were considered, and the information that would be needed to evaluate whether the proposed action is appropriate. Designing this presentation in a way that is comprehensible to the users who will encounter it, who may not have deep technical knowledge of the system's internals, is a genuine interface design challenge.
Real-World Examples of Multi-Agent Design
Several production systems provide instructive examples of multi-agent design decisions and their consequences.
Google's Gemini ecosystem, which coordinates specialized models for text, image, code, and other modalities, has addressed the multi-agent design challenge by building a unified interface that abstracts away the complexity of the underlying agent coordination. From the user's perspective, there is a single conversational interface; the routing of requests to appropriate specialist models happens invisibly. This design decision prioritizes simplicity and coherence over transparency about the system's internal structure. It works well for consumer use cases where users are primarily interested in results and less concerned with process, but it creates challenges for enterprise use cases where auditability and process transparency are important requirements.
Salesforce's Agentforce platform takes a different approach, providing enterprise customers with explicit visibility into and control over the multi-agent workflows that execute their business processes. Agentforce allows administrators to define the agents that participate in a workflow, the conditions under which each agent is invoked, the human approval requirements at each step, and the escalation paths when the system encounters situations it cannot handle automatically. This design approach prioritizes control and auditability over simplicity, which is the right tradeoff for enterprise contexts where business process integrity and regulatory compliance are paramount concerns.
Anthropic's own research on multi-agent systems, published in its work on AI safety and alignment, has surfaced a design challenge that is not well-addressed in most current multi-agent products: the problem of agent goal alignment in complex workflows. When multiple agents are coordinating to accomplish a complex task, ensuring that each agent's local objective is aligned with the overall system objective, and that the interaction of multiple aligned agents does not produce misaligned system behavior, is both a technical and a design problem. The design of agent interfaces, the specifications of what each agent is trying to accomplish and what constraints it operates under, directly affects the alignment properties of the system as a whole.
Design Principles for Multi-Agent Systems
Drawing together the threads of this discussion, several design principles emerge for teams working on multi-agent system design.
The principle of layered transparency holds that multi-agent systems should communicate their process at multiple levels of detail, with the default presentation calibrated to what most users need and progressive disclosure mechanisms available for users who need more. This principle applies to both the orchestrator's communication of its overall process and the specialist agents' communication of their individual outputs.
The principle of risk-calibrated autonomy holds that the degree of autonomous action a multi-agent system takes should be calibrated to the risk profile of the actions involved. High-risk, irreversible actions require explicit human approval. Low-risk, reversible actions can proceed autonomously. The design of this calibration is not a single decision but an ongoing process of refinement based on observed system behavior and user feedback.
The principle of recoverable state holds that multi-agent systems should be designed so that users can meaningfully intervene at any point in a workflow and recover to a coherent state. This means that the system's state at any point should be representable in terms that are understandable to users, that the consequences of any intervention should be predictable, and that the system should support graceful recovery from user-initiated changes of direction.
The principle of attribution clarity holds that when a multi-agent system produces an output, it should be possible to understand which agents contributed to that output and in what ways. This is not a requirement for complete technical transparency, which is often neither possible nor useful, but for a level of attribution that allows users to evaluate the output's trustworthiness and to identify the source of problems when the output is incorrect.
Practical Considerations for Design Leaders
For design leaders whose teams are working on multi-agent system design, the practical challenges are significant. The methods, tools, and conceptual frameworks that most UX teams have developed are not well-suited to multi-agent design problems, and building the necessary capabilities requires deliberate investment.
The most immediate practical need is developing design methods for multi-agent workflow visualization. Multi-agent systems cannot be adequately represented in standard user flow diagrams, because those diagrams represent the user's path through an interface but do not capture the system's internal process or the interactions between agents. New visualization methods are needed that can represent the orchestrator-agent relationships, the data flows between agents, the human checkpoint locations, and the error propagation paths in a way that is legible to design teams and useful for design decision-making.
The second practical need is expanding usability testing methods to account for multi-agent complexity. Standard usability testing is designed to evaluate discrete interactions within a session. Multi-agent workflows may span sessions, involve long processing times, and produce outputs through processes that are not directly observable. Testing the usability of these systems requires methods that can evaluate the user's experience of the full workflow, including the experience of waiting, monitoring, intervening, and recovering from errors, not just the experience of the input and output interactions.
The third practical need is developing design collaboration processes that include the technical stakeholders who make architectural decisions with design consequences. In multi-agent systems, the design of the agent interfaces, the orchestration logic, and the error handling mechanisms are all architectural decisions that have significant user experience implications. Design teams that are not involved in those decisions will find themselves designing interfaces for systems whose underlying architecture makes good user experience difficult or impossible to achieve. Getting a seat at the architectural decision table requires both the organizational influence to claim that seat and the technical literacy to contribute meaningfully to the conversations that happen there.
Conclusion
Multi-agent systems represent a genuine frontier in AI product design. They extend the design challenge into new territory: orchestration logic, agent interface design, progressive transparency, risk-calibrated autonomy, and workflow-level human oversight. These are not extensions of existing design problems; they are qualitatively new problems that require new methods, new frameworks, and new forms of cross-functional collaboration.
The organizations that will build the most effective multi-agent systems are not the ones with the most sophisticated AI models. They are the ones that invest in designing the human experience of multi-agent coordination as carefully as they invest in the technical architecture of the agents themselves. The intelligence of a multi-agent system is ultimately expressed through the interface that mediates the human-system relationship, and that interface is a design artifact that requires the same level of craft and rigor as any other consequential design work.
For design leaders, multi-agent systems are both a challenge and an opportunity. The challenge is developing the conceptual literacy, the methods, and the organizational influence to contribute meaningfully to decisions that are currently being made primarily by engineers and researchers. The opportunity is that design is needed at exactly the level where the most important multi-agent design decisions are being made: the level of human oversight, process transparency, and collaborative workflow design that determines whether multi-agent systems are genuinely useful tools or impressive but ultimately frustrating technical demonstrations.