Skip to content
H
Howardismvol. 03 · quiet corner of the web
PLATE II · PIECE № 06HOWARDISM

The Bitter Lesson

PublishedMay 13, 2026FiledConceptReading3 minSourceAI-synthesised

Sutton 2019: scaled general methods beat hand-engineered structure; recurring justification across the wiki for dissolving harnesses into models; caveat — mechanical verification and character may not migrate inward

Illustration for The Bitter Lesson

Sources#

Summary#

Rich Sutton's 2019 essay: general methods that leverage computation (search, learning) ultimately outperform methods that build in human knowledge and hand-engineered structure — and they do so by a wide margin as compute grows. The "bitter" part: this keeps surprising researchers who invested in clever domain structure, because the structure becomes a ceiling, not a foundation.

This page exists because the principle recurs as a load-bearing argument across this wiki — invoked explicitly to justify dissolving harnesses into models.

Where it's invoked here#

  • Interaction Models — TML cites "the bitter lesson" directly: hand-crafted interactivity systems (VAD, turn-detection, dialog-management harnesses) "will be outpaced by the advance of general capabilities," therefore "for interactivity to scale with intelligence, it must be part of the model itself." See Turn-Based Interface Bottleneck.
  • Encoder-Free Early Fusion — co-train all modality components from scratch in one transformer rather than stitching pretrained encoders/decoders: fewer hand-engineered modular boundaries.
  • Time-Aligned Micro-Turns — remove artificial turn boundaries so interaction modes become scalable model behavior rather than per-mode harness code.
  • Harness Shrinkage as Models Improve — the same logic applied to coding-agent harnesses: prompt scaffolding compensates for what the model can't yet do, and should shrink as models improve. (Caveat there: mechanical verification — tests, types, linters — is the part that doesn't migrate inward.)
  • Agent Harness Engineering — "enforce invariants, not implementations": let the model find the path; the harness only encodes what must be true.

The standard caveat#

The bitter lesson is about capabilities and structure migrating into the model, not "harnesses are useless." Things that legitimately stay outside the model: mechanical verification (Harness Shrinkage as Models Improve's synthesis), organization-specific policy/style, security boundaries, and — per Claude Character as Product — deliberate character/personality work. The open question on every harness component is which side of that line it's on.

Connections#

Sources#

§ end
About this piece

Articles in this journal are synthesised by AI agents from a curated wiki and are refreshed automatically as new concepts arrive. Topics, framing, and editorial direction are curated by Howardism.

10 articles link here
  • ConceptAgent Harness Engineering

    Patterns for scaffolding long-running LLM agents: environment design, progressive context disclosure, mechanical archit…

  • EssayOpinions on Using AI Tools & the Future of the Software Engineering Role

    Debate map of four stances on using AI tools (bullish-insider / pragmatist-practitioner / skeptic-governance / architec…

  • ConceptClaude Character as Product

    Personality as load-bearing product surface; Amanda's role at Anthropic; lunchtime vibe-checks as eval discipline; the…

  • ConceptEncoder-Free Early Fusion

    Multimodal design with minimal pre-processing instead of large standalone encoders: dMel audio embedding, 40×40-patch h…

  • ConceptHarness Shrinkage as Models Improve

    Prompt scaffolding shrinks each model release; Cat Wu's pruning discipline; Boris Cherny "100 lines of code a year from…

  • ConceptInteraction Models

    Thinking Machines Lab (May 2026): models that handle audio/video/text interaction natively in real time instead of via…

  • ConceptModel Spec Midtraining (MSM)

    New training phase between pretrain and AFT: train base model on synthetic docs discussing the Model Spec; controls AFT…

  • EntityThinking Machines Lab

    AI research lab behind interaction models (May 2026); harness-dissolves-into-model thesis; upstreamed streaming-session…

  • ConceptTime-Aligned Micro-Turns

    The core interaction-model move: input/output as continuous streams in ~200ms interleaved chunks, no turn boundaries; s…

  • ConceptTurn-Based Interface Bottleneck

    Why current AI interfaces limit collaboration: single-thread turn-taking is a bandwidth bottleneck; humans pushed out b…

Related articles
  • ConceptInteraction Models

    Thinking Machines Lab (May 2026): models that handle audio/video/text interaction natively in real time instead of via…

  • ConceptHarness Shrinkage as Models Improve

    Prompt scaffolding shrinks each model release; Cat Wu's pruning discipline; Boris Cherny "100 lines of code a year from…

  • ConceptContext Window Smart Zone

    Smart zone vs dumb zone (Dex Hardy / Matt Pocock): quadratic attention scaling, ~100K marker independent of advertised…

  • EntityAnthropic

    AI safety company / vendor of Claude; mission-as-tiebreaker culture; ~30–40 PMs across teams; Mike Krieger leads Labs r…

  • ConceptDesign Concept Grilling

    Matt Pocock's `grill-me` skill; reach Brooks "design concept" before any plan; counter to specs-to-code; PRD as destina…