Experiment #6.
Challenge:
Can an AI agent integrate brand guidelines and operate at a meaningful level of design exploration and implementation without heavy prompting?
Theory:
If agents can generate structurally sound, brand-aligned outputs with minimal instruction, they could accelerate stakeholder reviews and pre-implementation validation.
Assumptions:
Prompt clarity is consistent.
A screenshot + instruction anchors structural reconstruction.
Agents have access to publicly available brand guidelines.
AI Tools:
• Google Notebook
• Chat GPT
• Loveable
Persona driven scenario:
Context: Booked lodging designs, accessibility annotations
Persona: UX designer
Goal: Generate a web-ready screen from an existing screenshot, test dynamic content modules, implement branding and test for accessibility
What I started with
A post-booking lodging confirmation screen (mobile), used as the structural baseline.
Structural Reconstruction
After Figma Make failed to generate an accurate web-ready screen, expectations were low.
Loveable successfully reconstructed a web-ready version of the screen with high structural accuracy. Notably, it correctly identified the experience as an Expedia post-booking page — the first agent in the series to distinguish that context.
Because it recreated the screen correctly, the experiment could continue.
Dynamic Module Testing
I prompted the agent to:
Obfuscate sensitive details
Validate pattern consistency
Insert a dynamic module
Generate two variants: Business Traveler and Family Traveler
Present the variants side-by-side
Execution was fast and technically accurate.
However, from a design perspective, the output was visually neutral (meh) — structurally correct but lacking brand depth or experiential nuance.
Key Observations
Structural reconstruction: Strong
Context recognition: Strong
Speed: Comparable to v0
Visual expression: Generic
Brand integration: Surface-level
The agent could reproduce pattern logic and layout integrity, but did not independently elevate the experience to a brand-distinct level.
Visual Treatment Exploration
After successfully generating the dynamic module variants (Business Traveler and Family Traveler), I wanted to test visual adaptability — not branding per se — but if and how the agent could shift visual treatments within the same structural system.
I prompted the agent to generate three visual treatments of the Business Traveler variant:
Dark
Neutral
No specific direction
Prompt dialog
Accordion
Card Style
Surprise me
Result
All three were structurally correct BUT all three were visually restrained.
The treatments differed primarily in surface styling — color tokens and minor contrast shifts — but did not demonstrate meaningful variation in hierarchy, spacing, emphasis, or experiential tone.
Observation:
The agent can swap tokens.
It does not yet reinterpret visual intent.
This became the pivot point for the next test.
Branding application test
Because Loveable had correctly identified the screen as Expedia, I then tested whether it could apply a distinct brand system (Hotels.com) to:
The Business Traveler screen
A dynamic module variant
The output was technically accurate — control labels updated and color systems shifted — but the result felt surface-level. At one point, a gradient was introduced that may not align with Hotels.com brand standards.
Prompt dialog
Dark theme
Neutral theme
Brand
Observation:
Brand styling was applied.
Brand fluency was not demonstrated.
Token substitution ≠ brand interpretation.
Accessibility and Contrast Test
At this point, I could concede that v0 leads when it comes to creating engaging visual treatments, but decided that I may as well look at how well Loveable could support Accessibility requirements. For example, I noted right away the dark treatment introduced insufficient color contrast — dark text on a dark background.
The issue was corrected only after explicitly prompting for WCAG compliance, after a first round of WCAG review.
WCAG prompt and output 1
WCAG specific prompt and output
Screen Reader test
Using Safari + VoiceOver, I reviewed landmarks, headings, and form controls.
A few things stood out:
“Check-in Essentials” was designated as a Region; appropriate given its time-sensitive importance.
Two buttons in the Form Controls have no label other than “Button”, which is a WCAG violation: an assistive tech user won’t know what the purpose of the button is until it’s interacted with.
Accordion sections received H3s, which is reasonable. I would have introduced an H2 at the section title to broaden hierarchy and avoid structural depth creep.
Landmarks/Regions menu
Headings menu
Form controls menu
Annotations
Once I had a feel for the structure and labels for the page, I decided to see if Loveable could create annotated views. As a UX designer, it’s my job to provide annotations to my engineering team, so any program going forward has to have this capability, whether it’s for creating or reviewing or both.
I prompted the agent to generate numbered, color-coded annotations separating:
Heading structure (purple)
Screen reader roles and states (green)
Annotation labels mapped accurately to the UI and reduced friction when reviewing structure against presentation.
What I observed:
The heading structure list does not follow the visual order of the page. “Check-in Essentials” appears before the name of the stay — likely because the module was designated as a Region. While technically defensible, this could be confusing for someone less familiar with semantic hierarchy.
The “What’s nearby” accordion lacks an explicit status (closed/false), unlike the accordions within Check-in Essentials.
The “Things to do near your stay” carousel cards are not treated as interactive elements and lack heading structure. While not technically incorrect, it weakens clarity from an assistive tech perspective and obscures content type and intent.
I was pleased that Loveable accurately mapped annotation labels to their corresponding UI elements. It made reviewing structure alongside presentation significantly more efficient.
Heading structure annotations
Screen Reader for Form Controls annotations
Agent Capability Spectrum
Across experiments, agent capability is not uniform. Some struggle with structural fidelity. Others execute patterns reliably but require explicit direction. At this point, v0 has consistently demonstrated stronger visual reasoning and more refined output — though it remains unclear whether that reflects a durable advantage or a temporary lead in model tuning.
The gap between agents is significant.
Was this a successful experiment?
Yes and no. In some areas, expectations were low — and the agent exceeded them. In others, expectations were higher — and the output fell short. What became clear is that capability is not binary.
What surprised me?
The speed at which competent execution is becoming normalized — and how consistently discernment remains absent.
Several agents can now generate structurally sound, technically viable screens in seconds.
But none independently protected hierarchy, accessibility, or brand nuance without prompting.
What did I not expect?
Stepping back, I did not expect how misleading the term “AI is correct” can be. Agents can generate outputs that look complete and technically sound. But correctness is often surface-level.
What would I have done differently?
At this stage, I would move beyond screens and lean into task and process flows.
Interfaces are artifacts.
Flows expose cognition.
Testing how agents model decision paths, state changes, and user intent would likely reveal more about their true reasoning limits than visual reconstruction alone.