UX Research

Year

2026

My Role

Principal Product Designer

Defining an AI-Native Design Framework for 2026

Defining an AI-Native Design Framework for 2026

Before building a large SaaS product, I researched and validated the AI design process used by teams like Stripe, Linear, and Ramp. Through deep research and iteration, I synthesized a 14-phase waterfall into 6 overlapping zones: a teachable, tool-grounded framework built for speed, craft, and the design-to-code pipeline that modern teams actually use.

Context

The product design industry is in the middle of a significant shift. AI tools have become widely available and increasingly capable, but adoption patterns vary considerably. Many designers add AI to existing workflows as an enhancement layer, using it to summarize research, generate mood boards, or suggest layouts without changing the underlying structure of how work gets done. The result is faster execution of the same process, not a fundamentally different approach to product development.


Before starting a large SaaS design project, I wanted to understand how the teams doing the most technically ambitious design work were actually integrating AI into their design and development pipelines. Not how they described it in general terms, but what their actual toolchains, handoff processes, and design system requirements looked like in practice.

Solution

The research produced a framework that treats AI as connected infrastructure rather than a collection of point tools. The key finding was that the value of AI in a design workflow compounds when tools are connected: a research session feeds a PRD, which feeds lo-fi generation, which feeds hi-fi refinement, which feeds a design system structured so AI coding tools can read it directly. When any link in that chain is broken, downstream tools have to compensate with guesswork.


The framework organizes the process into 6 overlapping zones rather than sequential phases. Phases imply that one thing must be completed before the next begins. Zones acknowledge that build cycles, testing, and design system work happen concurrently in practice, and that the sequence is a guide for priority rather than a gate for permission.

Approach

The framework emerged from research conducted before beginning a design project. I reviewed published process documentation, interviews, and practitioner accounts from designers and design leaders at companies known for high-quality product design, then cross-referenced job postings to identify the skills and practices these companies were explicitly hiring for.


Key sources included published design process documentation from Stripe and Linear, Intercom's design team blog covering their vibe-coding hackathon, Figma's documentation on MCP server integration and Code Connect, the Design Tokens Community Group stable specification published in October 2025, and multiple practitioner accounts of AI-native design workflows published in 2025 and 2026.


The initial model was a 14-phase sequential process. Through research and validation, it became clear that this structure did not match how teams at the reference companies actually worked. They operated in overlapping activity zones where prototyping, testing, and design system work happened in parallel rather than in sequence.

The real distinction: AI as veneer vs. AI as infrastructure

The real distinction: AI as veneer vs. AI as infrastructure

The real distinction: AI as veneer vs. AI as infrastructure

The clearest way to describe the difference is in terms of connectivity. When AI is used as a surface enhancement summarizing research, generating visual options, and suggesting copy, it operates in isolation. Each tool produces output that a person then carries forward manually. The process is augmented but not changed.

When AI is used as infrastructure, the connections between tools are automated. Research surfaces in the PRD. The PRD informs lo-fi generation. The design system's tokens are structured so AI coding tools can read them directly. Code generated from Figma designs references actual component names rather than approximations. The designer's judgment operates at a higher level since they are evaluating, steering, and refining rather than transcribing.

This distinction is most visible in the design system. A design system built primarily for designers uses components, spacing, and color scales that are visible and applicable in Figma. A design system built as AI infrastructure adds a layer of machine-readable structure: DTCG-compliant tokens exportable as JSON, Code Connect mappings that link Figma components to their code counterparts, and documentation structured so that AI agents can query it programmatically. The visual output may look identical. The difference is in what happens when an AI coding tool tries to implement the design.

14 phases → 6 zones: the structural shift

14 phases → 6 zones: the structural shift

14 phases → 6 zones: the structural shift

The original 14-phase model was built by mapping AI tools to each step of standard product design practice. It was comprehensive and sequential: research, then PRD, then journey mapping, then wireframes, then testing, then hi-fi design, and so on.

The problem was not with the activities since all 14 represent real work that needs to happen. The problem was the implied sequence. Published accounts from the reference companies described a different pattern: framing and research happening in a single AI-assisted session; lo-fi wireframes generated directly from a brief and tested with users before hi-fi began; design system work running in parallel with prototyping rather than after it.

The 6-zone model preserves all 14 activities but reorganizes them to reflect concurrent execution. Frame covers research, PRD, and strategic framing as a connected session. Explore covers journey mapping and lo-fi generation. Build and Test covers hi-fi design, code prototyping, and usability testing as a continuous loop. Systematize covers the design system infrastructure such as tokens, Code Connect, and MCP configuration which runs in parallel with Build and Test. Ship covers designer-authored frontend code, friction logging, and GitHub review. Document covers AI process documentation and portfolio presentation.

The validated framework: 6 zones, end-to-end

The validated framework: 6 zones, end-to-end

The validated framework: 6 zones, end-to-end

The framework is organized into six zones that overlap and run concurrently rather than sequentially. Each zone has a defined scope, a set of AI tools, and a concrete output. Together they cover the full product design lifecycle from initial framing through portfolio documentation.

Zone 1 — Frame. Research, PRD, and strategic framing happen in a single connected session using Claude and Perplexity. The goal is a 1-2 page brief that anchors every downstream decision. When this brief is grounded in real research, AI tools in later zones produce more accurate and relevant output because they have genuine context to work from.

Zone 2 — Explore. Journey maps, task flows, lo-fi wireframes, and AI feature behavior design happen in this zone. FigJam AI supports collaborative mapping. MagicPatterns, Figma Make, and v0 generate rapid wireframes directly from the brief. The sequence is not fixed, journey mapping and lo-fi generation often happen within the same working session.

Zone 3 — Build and Test. Hi-fi design with motion, code prototyping, and usability testing form a continuous loop rather than three separate phases. Figma handles high-fidelity design work. Lovable scaffolds the initial working prototype, Cursor refines it, and Maze facilitates testing. The loop repeats as often as the project requires.

Zone 4 — Systematize. The design system, DTCG-structured tokens, Code Connect configuration, and MCP setup are handled in this zone. It runs in parallel with Zone 3, not after it. This is the infrastructure layer that allows AI coding tools to read design intent directly rather than approximating it. Without this zone in place, the prototype code in Zone 3 operates independently of the design system which reintroduces the same translation errors the framework is designed to eliminate.

Zone 5 — Ship. The designer refines the prototype to production quality, runs a friction log against a defined quality rubric, and submits a GitHub pull request for engineering review. Cursor and Claude Code handle the refinement work. The friction log, a practice documented extensively at Stripe is a structured method for reviewing your own product as a user and scoring it against quality criteria before it reaches engineering.

Zone 6 — Document. AI process documentation and portfolio presentation. This zone produces an artifact that explains which tools were used at each zone, what decisions were made, and what a team would need to know to replicate or build on the approach. Documentation is the output that makes the process transferable rather than personal.

The relationship between zones 4 and 3 is worth emphasizing. MCP configuration in Zone 4 is what connects the design system to the coding environment in Zone 3. When the Figma MCP server is active and Code Connect is configured, Cursor can read component names, token values, and layout context directly from the Figma file rather than reconstructing them from a screenshot. The quality of the resulting code reflects the quality of the design system. A well-structured system produces accurate output. A loosely structured system produces generic output that requires manual correction.

Research

Research

Research

Objective
To build a credible, research-backed AI design framework before beginning a major product project, grounded in published practice from teams with a demonstrated track record of high-quality product design.

Methods
I reviewed published blog posts, design process documentation, and practitioner interviews from Stripe, Linear, Ramp, Intercom, Vercel, and Figma. I read the DTCG specification documentation, Figma's MCP server and Code Connect developer documentation, and multiple independent practitioner accounts published in 2025 and 2026. I cross-referenced findings against job postings from these companies to identify which skills and practices were being explicitly prioritized in hiring.

The most significant technical finding was the relationship between DTCG-structured design tokens and MCP integration. The Figma MCP server allows AI coding tools to read design context directly from a Figma file. For this to work reliably, tokens need to follow the DTCG standard, which reached its first stable version in October 2025 and is now supported by Figma, Tokens Studio, and most major design system tooling.

Key Findings
The research revealed several pain points, including inefficient navigation, limited data accessibility, and a lack of customization options. Frequent errors and insufficient feedback mechanisms further hindered admin workflows, while inconsistent usage patterns underscored the need for a flexible design that catered to both power users and occasional users.

Impact on Design
The quality of AI-generated code output depends directly on design system quality. Teams reporting the strongest results — including those at Affirm and Coinbase — had robust, well-structured design systems before introducing MCP or AI coding tools. MCP amplifies the design system; it does not compensate for a weak one.

Usability testing at fast-moving teams has been compressed into the build cycle rather than treated as a separate gate. Intercom's design team published an account of a process where interactive prototypes built in Lovable or Figma Make are tested with users the same day they are built, with the prototype updated between sessions. This collapses what traditionally operated as a handoff into a continuous loop.

The designer's role in this model shifts from production toward evaluation and steering. AI tools can generate layout, code, and copy quickly. The judgment about what is correct, appropriate, and well-crafted remains a human responsibility.

Objective
To build a credible, research-backed AI design framework before beginning a major product project, grounded in published practice from teams with a demonstrated track record of high-quality product design.

Methods
I reviewed published blog posts, design process documentation, and practitioner interviews from Stripe, Linear, Ramp, Intercom, Vercel, and Figma. I read the DTCG specification documentation, Figma's MCP server and Code Connect developer documentation, and multiple independent practitioner accounts published in 2025 and 2026. I cross-referenced findings against job postings from these companies to identify which skills and practices were being explicitly prioritized in hiring.

The most significant technical finding was the relationship between DTCG-structured design tokens and MCP integration. The Figma MCP server allows AI coding tools to read design context directly from a Figma file. For this to work reliably, tokens need to follow the DTCG standard, which reached its first stable version in October 2025 and is now supported by Figma, Tokens Studio, and most major design system tooling.

Key Findings
The research revealed several pain points, including inefficient navigation, limited data accessibility, and a lack of customization options. Frequent errors and insufficient feedback mechanisms further hindered admin workflows, while inconsistent usage patterns underscored the need for a flexible design that catered to both power users and occasional users.

Impact on Design
The quality of AI-generated code output depends directly on design system quality. Teams reporting the strongest results — including those at Affirm and Coinbase — had robust, well-structured design systems before introducing MCP or AI coding tools. MCP amplifies the design system; it does not compensate for a weak one.

Usability testing at fast-moving teams has been compressed into the build cycle rather than treated as a separate gate. Intercom's design team published an account of a process where interactive prototypes built in Lovable or Figma Make are tested with users the same day they are built, with the prototype updated between sessions. This collapses what traditionally operated as a handoff into a continuous loop.

The designer's role in this model shifts from production toward evaluation and steering. AI tools can generate layout, code, and copy quickly. The judgment about what is correct, appropriate, and well-crafted remains a human responsibility.

Outcome

Outcome

Outcome

The framework is being applied directly to an ongoing SaaS product. Each zone is producing concrete deliverables being documented in a separate case study as the project progresses.

The research resolved a practical question before the project began: whether to structure the work as a traditional process augmented with AI tools, or as a genuinely different approach.

Based on published practice at the reference companies, the answer was clearly the latter. The framework provides a starting point that can be adapted and taught to a team. A process that is explicit enough to be repeatable and flexible enough to accommodate how design work actually unfolds in practice.

More Projects

More Projects

More Projects

Agent Onboarding: The AT&T Digital Guide

Elevating the DirecTV Mobile Apps

Life: Revolutionizing Health Tracking

Agent Onboarding: The AT&T Digital Guide

Elevating the DirecTV Mobile Apps

Life: Revolutionizing Health Tracking

© [UX Kitt]