Model
Data Visualization extends the Design Tokens v2 model to represent analytical meaning.
It follows the same architectural principles as the system while introducing a controlled extension for data-specific semantics.
Architecture
core.foundation → semantic.foundation
core.dataviz → semantic.dataviz
patterns/specs → consume semantic only
Core
Core tokens define raw, themeable values.
core.foundationcontains global primitives such as colors, spacing, typography, borders, radii, and motioncore.datavizcontains non-color encoding primitives specific to analytical visualization:- mark shapes, fill patterns, stroke dash arrays, and analytical opacity values
- Analytical colors are sourced from
core.colors.*— no separate dataviz color palette is needed
Core tokens must remain value-only and context-free.
Semantic
Semantic tokens define meaning.
semantic.foundationdefines UI semanticssemantic.datavizdefines analytical semantics
Semantic tokens:
- reference core tokens only
- express stable analytical roles
- form the public API of the category
Components and patterns must consume semantic tokens exclusively.
Patterns and Specifications
Patterns and specifications define how tokens are applied.
They are responsible for:
- chart type selection
- multi-view composition
- legend and labeling strategies
- tooltip behavior
- interaction design
- geospatial rendering details
These concerns are intentionally out of scope for tokens.
Semantic Boundary
Data Visualization tokens encode analytical meaning, not implementation.
Included
- categorical identity
- ordered magnitude
- midpoint comparison
- analytical references
- contextual states
- data status
- non-color encodings for redundancy
- geospatial overlay semantics
Excluded
- chart-specific configuration
- statistical methods
- rendering logic
- layout and composition rules
- visualization library behavior
- full map styling systems
Core Extension Rules
Data Visualization introduces new core tokens only when all of the following are true:
- the problem is unique to analytical visualization
- the concept is stable across multiple chart types and domains
- the value can be defined independently of context
This results in a minimal set of new primitives:
- encoding primitives for shape, pattern, stroke, and analytical opacity
Analytical opacity is distinct from foundation opacity. Foundation opacity is used for interface layering and interaction states. Data Visualization opacity is used as an encoding channel for analytical meaning.
All other needs must reuse existing foundation tokens.
Semantic Design Rules
Semantic tokens must follow these constraints.
1. Role-based naming
Tokens express analytical roles, not visual properties.
Examples:
seriesscale.sequentialscale.divergingreferencestatestatus
2. No chart-specific semantics
Tokens must not encode chart types or components.
Invalid examples:
bar.primaryline.highlightmap.region.fill
3. No library coupling
Tokens must remain independent from rendering technologies.
They cannot reference:
- specific chart libraries
- map providers
- rendering engines
4. Composability
Complex meaning must be expressed through composition across families.
Examples:
- forecast = color + stroke + optional opacity
- uncertainty = color + opacity + pattern
No single token should attempt to encode complex analytical meaning alone.
Geospatial Contract
Geospatial support follows an overlay-first approach.
Geography does not introduce a parallel semantic language for color or encoding. It defines the contextual contract for analytical overlays on spatial surfaces.
Geospatial layers
- Context: geographic background that supports orientation
- Overlay: analytical data rendered on top of geography
- State: spatial interaction such as focus and selection
Geospatial semantic tokens
dataviz.geo.context.muted
dataviz.geo.context.boundary
dataviz.geo.context.label
dataviz.geo.state.selection
dataviz.geo.state.focus
Geospatial rules
- geospatial overlays use
dataviz.color.*for analytical color meaning - geospatial overlays use
dataviz.encoding.*for non-color reinforcement dataviz.geo.*defines only contextual spatial semanticsdataviz.geo.*must not introduce a parallel color or encoding language
What geospatial semantics govern
- contextual reduction behind overlays
- supportive boundaries that preserve spatial reading
- contextual labels that preserve orientation
- explicit spatial focus and selection states
What geospatial semantics do not govern
- basemap design
- projection
- tiling
- zoom and generalization algorithms
- label placement systems
- provider-specific map style behavior
Validation Expectations
Analytical
- sequential scales must preserve perceptual order
- diverging scales must center around a meaningful midpoint
- series tokens must remain within bounded sets
- status tokens must clearly differentiate absence of data
Accessibility
- meaning must not rely on color alone
- encodings must provide redundancy when required
- critical graphical elements must remain perceptible
Geospatial
- overlays must remain legible against supported geographic context
- context reduction must not compete with the primary analytical layer
- spatial focus and selection must remain distinguishable from the base state
Summary
The Data Visualization model defines a minimal and strict semantic layer for analytical meaning.
It ensures that:
- tokens remain stable and reusable
- meaning is separated from implementation
- geospatial overlays reuse the same analytical semantics as other visualizations
- visualization systems can scale without semantic drift
By limiting scope and enforcing clear boundaries, the model provides a robust foundation for data-driven interfaces.