Back to all work
AI · Amazon Seller Central · 2025
Canvas What if the UI built itself around your question?

Seller Central had been built by dozens of teams over many years. The data existed. The tools existed. But sellers had to navigate a system that reflected Amazon's org chart, not their actual problems. Canvas was a bet that generative AI could change that: ask a question, get a purpose-built workspace in response.

~90% Acceptance Rate Hours to Seconds Tested at Amazon Accelerate 2025 5 Business Streams
Role
Sr. UX Design Manager
Timeline
2025
Scope
Listings, Finance, Compliance, Brand, FBA
01
THE PROBLEM

You could see the org chart in the design.

Amazon Seller Central had grown alongside Amazon itself. Every new business unit that needed to expose data or tools to sellers built their own piece of it. The result was a platform where the information architecture was a direct reflection of Amazon's internal org structure, not how sellers actually thought about running their business.

A seller trying to understand a slump in sales had to visit their advertising dashboard, their inventory page, their pricing tools, and their account health page separately, synthesize the information manually, and form their own conclusions. The platform told you what the data was. It didn't help you understand what it meant or what to do about it.

How sellers experienced it
"I know the data is somewhere in here. I just have to go find it myself."
Sellers spent significant time navigating between disconnected sections of the platform, manually piecing together information from different parts of the system to answer a single business question.
The structural root cause
The UI was a map of Amazon's teams, not a map of seller problems.
Every business unit had built its own section. Listings. Finance. Fulfillment. Compliance. Brand. Each one complete on its own terms, but with no connective tissue between them from the seller's perspective.
Next Gen Selling addressed the information architecture. It reorganized the navigation, grouped data more thoughtfully, and made the platform more coherent. But it couldn't solve the deeper problem: every seller's business is different, and the questions they need to answer are too varied and too personal for any fixed structure to handle.
02
THE INSIGHT

Stop organizing data. Start answering questions.

The insight behind Canvas was simple: instead of asking sellers to navigate to their data, let them ask a question, and have the system build a focused workspace around the answer. Generative AI made this technically possible. But technically possible is not the same as usable at scale.

There was already a lightweight Canvas prototype when my team got involved. The concept worked in a demo. But there were no rules. No framework for what a canvas should contain, how it should behave, what it should allow sellers to do, or how the AI should decide what to surface. It was the wild west. My team's job was to define the framework that would make Canvas work across every business stream, not just the one it was originally designed for.

Start with a real seller question
Every canvas begins with something a seller actually needs to know. Not a metric. Not a data export. A question that a real business owner would ask.
📐
Pull only what's relevant
The canvas assembles data from across the platform, but only the data that answers the question. Not everything available. Just what matters for this problem, right now.
End with a clear action
Every canvas ends with something the seller can do. A button. A decision. A next step. The AI doesn't just explain the situation. It helps resolve it.
03
MY TEAM'S ROLE

One framework. Every business stream.

My team owned day-to-day seller operations across five major business areas: Listings, Finance, Compliance, Brand, and Fulfillment by Amazon. That coverage was what made our involvement in Canvas critical. A framework only works if it works everywhere, not just for one type of seller or one type of problem.

We took the Canvas concept and stress-tested it across every vertical we owned. What does a canvas look like when a seller is facing a policy violation? When they're planning for a product launch? When their sales are declining and they don't know why? When they need to build a promotion strategy across multiple programs simultaneously? Each scenario had different data needs, different action types, and different levels of urgency. The framework had to handle all of them consistently.

01
Listings
Product detail pages, content quality, discoverability. Sellers need to know what's wrong with a listing and how to fix it, fast.
02
Finance
Sales performance, fee analysis, revenue trends. The questions here are complex and the data lives in many places.
03
Compliance
Account health, policy violations, reinstatement. High stakes, time-sensitive, and deeply confusing for sellers who don't know what they did wrong.
04
Brand
Brand analytics, A+ content, promotional strategy. Sellers managing a brand have a different set of questions from those selling single products.
05
Fulfillment by Amazon
Inventory levels, restock recommendations, storage fees. The most operationally complex area, with the highest cost of getting it wrong.
HOW CANVAS WORKS INPUT "What should I restock?" Any question about your business THE FRAMEWORK Intent What is being asked? RULES, NOT TEMPLATES Composition rules Which components fit? Data routing CANVAS Chart For trend questions Metrics For KPI questions Take action → Action Always last Same rules. Same components. Every canvas is different because every question is different.
Tap to expand
WHERE THE UX TEAM HAD IMPACT What we defined Risk without it Component catalog Curated, validated UI building blocks the LLM can select from LLM invents new UI each time No consistency. Sellers can't learn the interface. Trust erodes. Composition rules Which components suit which type of question Right data, wrong component A bar chart for a compliance alert. Context lost. Seller confused. Design guardrails Consistent patterns across all verticals Off-brand, unpredictable outputs Canvas doesn't scale beyond one team Without this work, Canvas generates something different every time. Sellers stop trusting it
Tap to expand
03b
THE FRAMEWORK IN PRACTICE

Individual data points rarely told the whole story. The combination did.

A seller asking why their sales dropped didn't have a finance problem, or a listings problem, or an advertising problem. They had all three, and the story only made sense when the data was read together. Finance showed the decline. Listings showed which products had gone stale. Advertising showed where spend had dropped off. Each stream told a fragment. The canvas told the whole thing.

The UX team's job was to figure out which fragments belonged together for which type of question, and how to display them so the relationship between them was obvious. We validated this the way UX teams do: watching real sellers try to answer real questions. We ran sessions at Accelerate and in earlier testing, observing where sellers got stuck, what they reached for that wasn't there, and what felt immediately legible versus what required explanation.

What emerged were patterns. Not templates. Patterns. Diagnostic questions needed trend data plus contributing factors, displayed together so the cause was visible. Action-oriented questions needed ranked recommendations plus a single clear next step. Urgent questions needed the alert first, always, before anything else. Planning questions needed progress tracked across multiple dimensions at once.

Those patterns became the composition rules. And the rules were what made Canvas scalable. The same framework could handle a restock question for a small seller and a promotional planning question for a brand managing hundreds of products, because the rules adapted the assembly to the question type, not to the seller.

COMPOSITION MODEL, QUESTION TYPE DRIVES ASSEMBLY QUESTION TYPE DATA PULLED CANVAS ASSEMBLES Diagnostic "Why are my sales dropping?" Cause + pattern Finance FBA Listings Ads 4 streams, one answer Trend chart Metric deltas Contributing factors Action-oriented "What should I restock?" Recommendation FBA Supply 2 streams, one plan Ranked table Create shipment → Urgent / fix "Fix my policy violation" Resolution path Compliance Listings 2 streams, clear path ⚠ Alert first Always, for urgent intent Remove bullet 2 Resubmit listing Planning "Prepare for Black Friday" Supply Listings Ads Multi-stream plan Finalize listing Submit deal Confirm inventory The same data streams appear across question types, composition rules determine what to show and how
Tap to expand
04
THE SCENARIOS

Five questions. Five canvases. One framework.

We designed and tested five canvas scenarios that collectively covered the range of seller problems we encountered across business streams. Each one starts from a real question a seller would ask, assembles relevant data from across the platform, and ends with a specific action the seller can take.

FBA · Inventory
What products should I restock?
Pulls 6 months of inventory data, identifies at-risk SKUs, surfaces restock recommendations with quantity guidance, and offers to create a shipment directly.
Finance · Sales
How are my sales trending?
Generates a 6-month sales breakdown with trend analysis, identifies the drivers behind growth or decline, and surfaces ad campaign recommendations.
Compliance · Account Health
Fix my policy violation
Identifies the specific policy that triggered the flag, explains it in plain language, shows the problematic content in the listing, and offers to fix it automatically.
Listings · Launch
Get ready for Black Friday
Builds a product launch and readiness checklist across three categories, tracks progress, and surfaces the most urgent pending actions across listings, operations, and inventory.
Brand · Promotions
What promotions should I run?
Analyzes the seller's catalog and builds a coordinated promotion plan across Coupons, Ads, Best Deals, and Vine. Sellers can review and approve each recommendation.
05
TESTED AT AMAZON ACCELERATE

Real sellers. Real questions. Real feedback.

Amazon Accelerate is Amazon's annual conference for third-party sellers. In 2025, we brought Canvas to the conference and put it in front of real sellers, with real accounts, asking real questions about their actual businesses. Not a focus group, not a usability test with scripted tasks. Sellers at a conference booth, doing what they actually do.

The feedback confirmed the core thesis. Sellers immediately understood the interaction model. The concept of asking a question and getting a focused workspace in return required almost no explanation. What they wanted more of was breadth: more types of questions answered, more business areas covered, more actions available at the end of each canvas.

~90%
Acceptance rate on AI-generated canvas suggestions
Hours
to
Seconds
Time to assemble cross-platform data insights
5
Business streams validated through Accelerate testing
The thing sellers kept saying was: "Why doesn't it do this for everything?" That's not a complaint. That's a product roadmap.
Seller feedback, Amazon Accelerate 2025
06
SEE IT IN ACTION

Try it yourself.

Two prototypes built to the fidelity we used at Amazon Accelerate. The desktop version covers all five canvas scenarios. The mobile version shows how Canvas adapts to a smaller surface without losing the core interaction model.

Desktop
Seller Central: Canvas
Full five-scenario experience: inventory, sales, policy, product launch, and promotions. Ask a question, get a canvas.
Open desktop demo
Mobile
Amazon Seller App: Canvas
The same Canvas framework adapted for mobile. Sellers increasingly manage their business on the go. Canvas had to work there too.
Open mobile demo
Next case study
Seller Assistant: From Q&A to Agentic
View demo →