Add a proper public-facing documentation site using Mintlify with 19 MDX pages covering getting started, auto mode, commands, configuration, and all user-facing features. Move internal/SDK documentation (Pi SDK, TUI, context & hooks, research notes, ADRs) to docs-internal/ since they should not be part of the public documentation. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2 KiB
Encoding Taste & Aesthetics
The honest frontier: This is where all four models are most candid about current limitations.
What CAN Be Automated
| Technique | Description |
|---|---|
| Reference-based extraction | "Feels like Linear" → extract concrete attributes: spacing ratios, animation timing curves, color relationships, typography |
| Style specification | Convert extracted attributes to verifiable parameters: "transitions 150-200ms ease-out, 8px grid spacing, specific contrast ratios" |
| Automated verification | Lighthouse scores, visual regression tests, accessibility audits, performance budgets, design system linting |
| Visual comparison | Render output, compare against reference screenshots using vision-capable models |
| A/B comparison | Show two versions, human picks which "feels better" — faster than absolute judgment |
What CANNOT Be Automated
The gestalt — the overall feeling, emotional response, sense of quality emerging from a thousand small interacting decisions. Does this feel premium? Fast? Trustworthy? These are fundamentally subjective.
The Optimal Strategy
Narrow the gap by converting as much "taste" as possible into concrete, verifiable specifications upfront:
- Not "use nice spacing" → "16px between sections, 8px between related elements, 4px between tightly coupled elements"
- Exact animation timing curves, color values with contrast ratios, typography weights and sizes
Then reserve human review for the remaining subjective layer with structured, specific questions:
"Does the density feel right? Does the transition timing feel snappy enough? Does the empty state feel intentional or broken?"
The Emerging Frontier
Vision-capable models for aesthetic evaluation — render output, capture screenshot, compare against references on specific visual dimensions. Imperfect but improving rapidly. Grok reports ~80-85% of taste can be automated this way; the remaining 15% stays human-only.