Opacity is a scale threshold, not a class property

Type: kb/types/note.md · Status: seedling · Tags: learning-theory

Opacity isn't binary. Any representational form (how an operative part is encoded and consumed) becomes practically opaque at sufficient scale. What distinguishes forms is the scale at which they cross that threshold. Distributed-parametric representations cross almost immediately: meaning is smeared across many weights or dense dimensions jointly, so even a modestly-sized network resists per-unit inspection. Truly tiny networks can be read by hand (mechanistic interpretability on toy models does this), but the threshold is very low. Localized artifacts stay readable at much larger aggregates — per-unit readability plus search, diffing, and modular revision let you work with scales that would be hopeless in a weight matrix of comparable size.

"Readable" and "opaque" are convenient labels for where a form sits relative to current tooling, not intrinsic properties of the bytes. A prose KB past a certain size becomes practically opaque too; retrieval and filtering exist because unaided reading runs out. The right question is how much scale a form's structural affordances buy you before inspection breaks down.


Relevant Notes: