Opacity is a scale threshold, not a class property

Type: kb/types/note.md · Status: seedling · Tags: learning-theory

Opacity isn't binary — any artifact class becomes practically opaque at sufficient scale. What distinguishes classes is the scale at which they cross that threshold. Distributed representations cross almost immediately: meaning is smeared across many weights jointly, so even a modestly-sized network resists per-unit inspection. Truly tiny networks can be read by hand (mechanistic interpretability on toy models does this), but the threshold is very low. Localized artifacts stay readable at much larger aggregates — per-unit readability plus search, diffing, and modular revision let you work with scales that would be hopeless in a weight matrix of comparable size.

"Readable" and "opaque" are convenient labels for where a class sits relative to current tooling, not intrinsic properties of the bytes. A prose KB past a certain size becomes practically opaque too; retrieval and filtering exist because unaided reading runs out. The right question is how much scale a class's structural affordances buy you before inspection breaks down.


Relevant Notes: