Design tokens are the most machine-legible part of a design system β but most token architectures were designed exclusively for human workflows. Re-architecting token hierarchies with LLM interpretation in mind unlocks a new class of AI capability: systems that can reason about design decisions, not just execute them.
The goal was to re-architect token hierarchies so that LLMs could reliably interpret design intent β not just token values β enabling AI tools to reason about visual relationships, generate contextually correct UI, and validate design decisions against system rules.
This built directly on the existing three-tier token architecture (core β semantic β component) but extended it with a new layer of machine-readable context.
Re-examined the existing three-tier token architecture to identify where LLMs were failing to interpret intent correctly β focusing on ambiguous names, missing relationships, and undocumented constraints.
Designed a lightweight annotation schema for semantic tokens β adding purpose, usage context, and constraint fields without disrupting existing token workflows.
Built a machine-readable token manifest in structured JSON, covering all tokens with values, tier, category, intent, and relationship data.
Ran iterative tests measuring how accurately LLMs could apply tokens in generated UI with and without the new manifest and annotations β tracking error rates and token hallucination.
Integrated the manifest into AI tooling workflows and published guidance on FreeStyle for teams using AI assistants with the design system.
Structured intent annotations and the token manifest significantly reduced cases where AI tools invented token names or defaulted to hardcoded values.
LLMs could reason about visual relationships and design decisions β not just apply values β enabling more sophisticated AI-assisted UI generation.
The token manifest gave AI tools a reliable, structured source of truth β queryable at generation time for any token's purpose, value, or constraints.
Token-to-component mappings ensured AI-generated code referenced the right tokens in the right contexts, maintaining system integrity.