ANypoint Exchange Agent
From Passive Search to Active Intelligence: Transforming Anypoint Exchange into an AI-powered partner for asset management and discovery.
Overview

In late 2025, MuleSoft made a strategic push into the AI and agentic space. The core design teams were fully committed to Agent Fabric, a new agent orchestration platform, which left a gap: the Anypoint Exchange Agent had no dedicated designer. I was brought in from the Salesforce Design team on a three-month loan with a clear mandate: establish the UX strategy, validate it fast, and leave behind a design foundation documented well enough that another designer could carry it forward without losing momentum.

  • Role
    Lead Product Designer

    Practices
    Led end-to-end UX strategy, facilitated cross-team alignment, and established scalable design foundation

    time duration
    Late 2025 - Early 2026

Background

Anypoint Exchange is MuleSoft's API marketplace, where developers and administrators manage hundreds of assets. The Agent was the first conversational AI experience on the platform, with no existing design system or pattern library to draw from. My job was to build the interaction framework from scratch, within an aggressive launch window, on a domain I was ramping on in real time.

The project was already in flight when I joined. The PM had completed initial discovery and written foundational user stories. My role was to bring UX strategy to a team that had momentum but needed a design lead who could move fast, ask the right questions, and make decisions with conviction. The work needed to be left in a state another designer could carry forward.

Understanding the problem

Because discovery was already underway when I joined, my first priority was to understand what the PM had found and figure out where the design thinking needed to go deeper. I reviewed existing research, mapped it against the user stories, and spent time with the PM to pressure-test the assumptions behind them.

What emerged was a clear picture of where users were running into walls. The pain clustered around a few recurring friction points that the existing product architecture could not solve.

  • Manual Housekeeping Burden - Admins were spending hours on repetitive tasks: updating documentation, managing version lifecycles, and keeping catalogs compliant. With 200+ APIs to manage, this was unsustainable at scale.
  • No post-publication intelligence - Once an API was published, there was nothing to flag missing documentation, detect non-conformance, or surface quality issues. Asset quality was inconsistent and invisible.
  • Discovery friction - Standard filters (category, tag, date) couldn't support intent-based search. A developer looking for "payment processing APIs for European markets" had no choice but to manually dig through dozens of assets.
  • No conversational design foundation -The Exchange Agent was MuleSoft's first conversational UI product. There were no existing patterns to reference or extend — I would need to design the interaction framework from zero.

How might we help users manage and discover their API assets through a conversational interface — without sacrificing the scale and control that enterprise administrators depend on?

Usability testing and validation

I designed unmoderated usability tests structured around two personas that reflect the full asset lifecycle:

The Exchange Administrator: Manages catalog quality, enforces compliance standards, and performs "housekeeping" (deprecating old versions, ensuring documentation completeness across hundreds of assets).

The Developer: Seeks fast discovery and reusable assets to accelerate their own work.

To keep pace with the engineering timeline, I made an early call to broaden our recruitment criteria. The initial screening required deep Anypoint expertise, which would have slowed us down significantly. I shifted to participants with general API catalog experience instead, which gave us an immediate pool and let us validate core interaction patterns where domain literacy mattered more than platform familiarity. This kept the design process ahead of development rather than behind it.

Tests were structured around two themes: foundational interactions like opening and closing the panel, understanding AI suggestions, and rediscovering the agent; and complex multi-step workflows covering bulk operations, the double-panel concept, and reviewing AI-generated documentation drafts. Three main findings shaped the final design.

Research details

Chevron
What we found:
  1. Message alignment was causing cognitive friction
    With all messages left-aligned, users had to scan headers to tell their own messages from the agent's responses. The conversation did not read as a conversation. I iterated to a right-aligned user and left-aligned agent pattern, mirroring familiar messaging interfaces.
  2. The double-panel vision did not scale to admin workloads
    Users responded well to the double-panel concept and had no difficulty understanding how the two panels connected. The idea was that when a user asked the agent to address non-conformant APIs, the agent would generate a documentation draft on the spot, surfacing it in a side panel where the user could review and edit it in real time. For individual assets, this felt powerful and intuitive. But when we pushed the scenario to 50 or more APIs missing documentation, the one-at-a-time approach collapsed. Working through each draft individually was too high-touch for the volume administrators were dealing with. The interaction was built for depth, not scale. This finding drove a pivotal conversation with the PM. We agreed to de-scope the double-panel from MVP and prioritize a batch-tagging mechanism instead, letting users apply changes to hundreds of assets through a single natural language command. The double-panel remained documented and validated as a future roadmap item for smaller, more focused tasks.
  3. Mascot Design Built Trust
    Users especially appreciated Agent Max, the friendly mule mascot. His inclusion helped establish an approachable tone, making the AI feel like a helpful assistant rather than a "black-box" algorithm. This was particularly important for easing adoption anxiety around AI.
The solution

Before designing a single new component, I needed to understand what already existed. I audited the MuleSoft design system and Salesforce pattern library, mapping behaviors like panel interactions, confirmation dialogs, and status indicators to conversational use cases. The goal was to borrow before building. Where the existing patterns held up, I adapted them. Where they did not, I designed new ones, always keeping them grounded in the same visual logic as the rest of the platform. The goal was for the agent to feel native to the platform, not like a product bolted on from somewhere else.

Final Product

The final result elevated how users interact with their API catalog, moving away from a static, search-and-browse experience toward something that actively works with them.

Administrators can now manage hundreds of assets through natural conversation, while developers get instant, context-aware answers without digging through documentation. The agent meets users where they are in the platform and responds to what they actually need in that moment.

SUGGESTED PROMPTS AND AGENT MAX

First-time users often don't know where to start with a new AI interface. The welcome screen surfaces three curated prompts giving users an immediate entry point without requiring them to figure out what to ask. These are hand-picked to represent the most common starting points, not AI-generated.

Agent Max, the MuleSoft mascot, anchors the experience from that first screen. Usability testing showed his presence shifted user  
 sentiment, making the experience feel approachable rather than clinical. For an enterprise product asking users to trust AI with
 their data, that first impression matters.

CONTEXTUAL KNOWLEDGE RETRIEVAL

Users can ask natural language questions to instantly surface documentation, platform knowledge, and asset recommendations. Rather than hunting through menus or reading help docs, they get answers in context.

After every agent response, three suggested follow-up recommended prompts appear. The suggestions are AI-driven so they adapt to where the conversation has gone rather than following a scripted path.

BATCH TAGGING

The Exchange Agent proactively identifies assets that don't meet compliance standards due to missing documentation. Rather than flagging them one by one, it surfaces the full set and gives administrators a way to act on all of them at once by applying a      
missing documentation tag. The agent always confirms before executing.

Once approved, the tags are applied across every affected asset, giving administrators a filtered view they can return to when they're ready to do the documentation work. The batch operation turns what would have been hours of manual housekeeping into a single reviewed action.

DISAMBIGUATE & CLARIFY

When a query matches more than one result, the Exchange Agent pauses before acting. It surfaces the matching options and prompts  
the user to confirm which API they mean, so the correct asset is always targeted and never assumed. Once the user makes their selection and clicks "Delete Version," the action executes. The agent then provides a plain-language summary of what changed, giving users a clear picture of the outcome without having to trace it through the interface.

Explicit Approval

The agent always asks before acting. Whether the user is deprecating one asset or hundreds, no action executes without explicit
approval. Once the user approves, the action executes and the button disables, giving immediate feedback that the operation is complete. After every action, the agent provides a plain-language summary of what changed so users don't have to reconstruct the outcome from the interface.

Impact & Learnings

My design work was completed and handed off within the project window. Engineering is currently in the spiking phase, with another designer now carrying the work forward into build. This project pushed me across multiple dimensions at once: no existing conversational AI patterns to draw from, an aggressive timeline, and a team already in motion. Ramping up quickly enough to set strategy, make credible design decisions, and identify opportunities the team had not yet considered, all while the work was already moving, is what made it demanding. These are the outcomes and reflections I'm taking away from it.

Shifting the team's relationship with research- The MuleSoft team had not done much usability testing before this project, largely viewing it as slow and resource-heavy. By introducing UserTesting.com and running quick validation rounds, I showed that feedback could come back in days rather than weeks. That shift in how the team thought about research felt like one of the more lasting things I contributed.

Building a conversational foundation- Prior to this project, MuleSoft lacked established conversational UI patterns. The interaction framework I designed, covering message alignment, thinking states, confirmation flows, and contextual prompts, now serves as the foundational standard for the MuleSoft Agentic Design System.

Shipping a pragmatic MVP without losing sight of the vision- Testing the double-panel concept revealed something more nuanced than a simple pass or fail. For administrators managing dozens of APIs at once, it was not the right fit. The volume made it impractical. But for smaller, more focused tasks requiring careful review and iteration, the concept showed real promise. The batch-tagging solution addressed the most pressing business need and was the right call for MVP. But the double-panel work wasn't wasted. It's documented, the thinking is intact, and I carried it forward. When I returned to my automation team and started working on other AI projects, it became a reference point I drew on directly.

Adapting quickly without losing design quality- API management is a specialized domain with its own vocabulary and mental models. Joining mid-project in an unfamiliar space means making peace with not having full context before making decisions, and you have to ramp up while the work is already moving. The approach that worked was moving fast on the things I could validate quickly, while being explicit with the PM about the things I was still learning. Trust is built through that honesty, not through projecting confidence you don't yet have. Ramping up quickly enough to make credible design decisions and identify opportunities the team hadn't yet considered, on a borrowed timeline, is the thing I'm most proud of from this project.

do the honors,
hover over the button!

Puppy mentorship program
open-source, hackathon
SAMSUNG MOBILE
Consumer app, interaction, accessibility
transaction Infrastructure
data visualization, SaSS, Finance