Screen → Ambient
Design systems assume a rectangle. A viewport. Pixels. Every token, every component, every layout primitive is built for a visual surface — something with width, height, a coordinate system, and a rendering engine.
AI doesn’t live on a rectangle.
The Shift
It lives in voice assistants where there are no pixels. In smartwatch notifications where there’s barely a screen. In agent-to-agent handoffs where there’s no human-facing interface at all. In background processes that surface a notification, run a workflow, and disappear. In car dashboards, AR overlays, ambient displays, and whatever comes after phones. The AI is everywhere, and “everywhere” includes a lot of places that don’t have a <div>.
This isn’t just “good cross-platform design.” Cross-platform means the same visual system renders on phone, tablet, and desktop. Ambient means the AI started a task on voice in your car, remembered the state, and surfaced the result as a card on your desktop ten minutes later. That’s a behavioral handoff across modalities, not a responsive layout adapting to screen size. The design system needs to encode the behavior — not just the appearance — of that transition.
If you think this is all going to be voice, think about how frustrating automated support calls already are. “Press 1 for billing. Press 2 for technical support. I’m sorry, I didn’t understand that.” We’ve had voice interfaces for decades and they’re still hostile to most people. Ambient doesn’t mean voice-first. It means the AI meets you wherever you are — voice, screen, notification, haptic, or no interface at all — and the experience is coherent across all of them. That requires behavioral tokens: what does “high confidence” sound like? What does “processing” feel like as a haptic pattern? How does a background agent communicate “I did something on your behalf” without a screen? A design system that only works with pixels is a design system that only works in the past.
And here’s a layer that makes this even harder: behavior is cultural. The way people expect to be addressed, the formality of a notification, what “urgent” sounds like, how much explanation is polite versus patronizing — all of this varies across languages, cultures, and contexts.
A voice assistant in Tokyo and a voice assistant in São Paulo aren’t just speaking different languages. They’re operating in different social contracts. Directness that’s efficient in one culture is rude in another. Ambient AI that crosses surfaces also crosses cultures. The behavioral token system doesn’t just need to work without pixels — it needs to work differently in different human contexts. We’re a long way from knowing how to encode that. But any design system aiming at ambient reach will have to face it.
Where Systems Stand Today
Both Fluent and Material have cross-platform token systems — CSS custom properties, Kotlin, Flutter, Dart. That’s a necessary foundation. Fluent powers web, Windows, iOS, Android, and Teams, giving it real multi-surface reach. Material’s Dynamic Color adapts to platform context. But in both cases, the tokens only cover visual properties. There are no motion tokens that translate to non-visual surfaces, no behavioral tokens for how the AI communicates across modalities. The multi-surface reach is skin-deep.
What Pushes a Score Up
If the token system only covers visual properties (color, type, spacing), you’re at a 3–4 even with cross-platform reach. A 7 means behavioral tokens — how the AI acts, not just how it looks — that translate across surfaces including ones without pixels.
Where this is going. This page is a working summary — the shift, the current state, the scoring rubric. The full deep dive expands each section with code-level evidence, specific component proposals, and mockups. Trust Expression is the first dimension getting the full treatment; the rest follow as they earn it.
If you’re building against this shift — or you see something the summary is missing — write back. The scorecard is debatable by design.