Essay
The Compression That Sings
Music, Information, and the Foundational Structure of Nature
EN VIR O AI · WO R KIN G P AP ER The Compression That Sings Music, Information, and the Foundational Structure of Nature—
—The compression that sings is the same compression that flows, that cycles, that lives.
Learning to hear it is learning to hear nature itself.
K EYW ORDS information theory · algorithmic complexity · 1/f noise · multifractal analysis · predictive processing · environmental modeling · Bach · holographic principle · compression · logical depth · environmental superintelligence
A B S T R A C T Music and nature share a statistical signature—long-range correlation, multifractal scaling, and a characteristic balance between order and novelty that the informationtheoretic literature has converged on describing as compressibility. Recent empirical work on the note-transition networks of J. S. Bach, earlier discoveries of 1/f spectral structure across the musical corpus, and formal results in algorithmic information theory each point toward the same underlying claim: the structures that human perception recognizes as beautiful are the structures that admit short descriptions relative to natural priors. This paper argues that this is not an aesthetic coincidence but a reflection of the informational substrate of physical reality itself. Working from
Shannon entropy through Kolmogorov complexity, from Wheeler’s “it from bit” program through the Bekenstein bound and holographic principle, we develop a unified framework in which music, perception, and natural systems are three views of a single compressibility manifold. The framework makes concrete predictions about the architecture of environmental intelligence systems and suggests an informationtheoretic formulation of environmental ethics: ecological damage is Kolmogorov disordering—destruction of compressibility accumulated over deep time—and protection is the preservation of logical depth. The same principle that lets the ear hear a fugue lets a well-designed model hear a watershed. Nature, like Bach, speaks with physical necessity, and we learn to listen by learning to compress.
1. Introduction
Why does Bach move us? The question is easy to dismiss as aesthetic trivia, but it is in fact a question about the structure of the universe. Music is, at bottom, a pattern of pressure variations in air; the listener is a biological system governed by thermodynamics and embedded in an evolutionary history; the pattern that moves the listener is selected from an astronomically larger space of patterns that do not. Whatever distinguishes the moving pattern from the indifferent one must reflect a genuine feature of physical reality—either of the signal, of the system that receives it, or of the correspondence between them.
The answer that has slowly emerged across a half-century of work in statistical physics, information theory, and computational neuroscience is remarkably simple. Music that matters is compressible against the priors of a listener embedded in nature. It exhibits structure—hierarchical, self-similar, multifractal—that human perceptual and predictive systems can encode efficiently because those systems co-evolved with a natural world exhibiting the same statistical signatures. A masterwork is not a random pattern that happens to sound good; it is a dense encoding of relationships that the listener’s prediction engine already half-knew how to expect.
This paper makes a stronger claim. The fact that music and nature share this compressible structure is not a fact about music; it is a fact about the informational foundations of physical reality. If the Wheelerian program—“it from bit”—has any foundational correctness, then the universe itself is a computation over information, and the structures that emerge at every scale inherit that informational character. Nature is compressible because the laws that generate it are short. Music is compressible because it is made by, and for, systems that are part of that generation. The convergence is not coincidence but consequence.
Nature is compressible because the laws that generate it are short.
Music is compressible because it is made by, and for, systems that are part of that generation. The convergence is not coincidence but consequence.
We develop this claim across seven moves. Section 2 reviews the empirical literature: 1/f spectra in music, network-information analysis of Bach, multifractal scaling, and compression-theoretic accounts of aesthetic response. Section 3 reconstructs the first principles—Shannon entropy,
Kolmogorov complexity, the Bekenstein bound, Bennett’s logical depth, predictive processing—that make the claim precise. Section 4 argues that music and nature inhabit a shared compressibility manifold because perception is a holographic readout of bulk structure, and great composition exploits this. Section 5 turns to nature itself: hydrological, atmospheric, and ecological systems all exhibit the same long-range correlation signatures as music, for the same reason. Section 6 draws out the implications for environmental intelligence—what a system that
“hears” a watershed must actually do. Section 7 proposes an information-theoretic reframing of environmental ethics in which ecological damage is understood as the destruction of logical depth, and protection as the preservation of nature’s accumulated compression. Section 8 names the instrument this framework requires: Environmental Superintelligence—a planetary-scale apparatus that realizes in silicon and mathematics what human perception realizes in biology.
The thesis is simple and, we believe, almost obvious once stated: reality is compressible, and the feeling that Bach is beautiful and that a living river is beautiful is the same feeling. It is the feeling of an embedded observer recognizing that the signal arriving at the boundary of its perception admits a short and deep description. Making this precise is the work of the paper.
2. Empirical Foundations
2.1 Music: the 1/f signature
The modern physics of music begins with Richard Voss and John Clarke’s 1975 Nature letter and their fuller 1978 paper in the Journal of the Acoustical Society of America. Analyzing a wide range of recordings—classical, jazz, rock, talk radio—they found that the spectral density of audiopower fluctuations follows a 1/f law over many decades of frequency, from the scale of individual notes down to 5×10⁻⁴ Hz (the full length of a composition). Bach’s First Brandenburg Concerto was among their cleanest examples. The 1/f signature places music in the same statistical universality class as flicker noise in solids, the variability of astronomical sources, Nile flood heights, heartbeat intervals, and the many other systems that exhibit self-organized critical or long-memory dynamics.
The meaning of 1/f deserves emphasis. White noise (flat spectrum) has no memory; successive values are independent. Brownian noise (1/f² spectrum) has strong memory; successive values are highly correlated. 1/f sits precisely between these regimes. Mathematically, it is the signature of scale invariance—a process that looks statistically similar when examined at any temporal resolution. Cognitively, it is the signature of a signal that is neither predictable (therefore boring) nor random (therefore unintelligible), but calibrated to sustain attention by rewarding prediction at every timescale simultaneously.
The sweet spot—1/f, between white noise (no memory) and Brownian noise (strong memory)—is where music, hydrology, atmospheric turbulence, and biological signals all live.
Hsü and Hsü’s 1991 PNAS paper extended the 1/f analysis from amplitude to the symbolic level.
Studying Bach and Mozart, they showed that the distribution of pitch intervals itself exhibits fractal self-similarity, not merely the amplitude envelope. The melodic line, as a sequence of abstract musical objects rather than a continuous signal, already has the scale-invariant structure.
Oświęcimka and colleagues (2011) extended this further with multifractal detrended fluctuation analysis across 160 pieces in six genres, finding that most popular music exhibits classic 1/f pinknoise scaling, while classical music—represented by Chopin in their sample—and certain jazz pieces are more strongly correlated than pink noise. The structure runs deeper than mere scale invariance; there is hierarchical organization that multifractal analysis can resolve into a spectrum of local scaling exponents.
2.2 Network information in Bach
Kulkarni, Lynn, Bassett, and colleagues (2024) represent the most systematic recent analysis of
Bach specifically. They build networks for 337 Bach compositions in which each note is a node and each transition between successive notes an edge, with edge weights reflecting transition frequency. Two networks are then computed: the true network reflecting the piece’s actual statistics, and an inferred network reflecting how a model of human perception—trading accuracy against computational cost in the Lynn-Bassett framework—would encode the transitions. The information-theoretic gap between true and inferred networks (a KullbackLeibler divergence) measures how much of the piece’s real structure is lost in perceptual encoding.
The empirical result is striking on two levels. First, different Bach genres (chorales, toccatas, fugues, preludes) cluster cleanly in this information space—the compositional form is legible in network statistics alone. Second, the gap between true and inferred networks is substantially smaller for Bach than for random networks of comparable size. Bach’s compositions exhibit features—particular clustering patterns, thick recurring edges representing frequently repeated transitions—that minimize the perceptual inference error. The music is structured specifically so that a listener operating under realistic cognitive constraints can nonetheless reconstruct something close to its true structure.
This is a remarkable finding. It means that beyond merely being compressible, Bach’s music is robustly compressible under lossy perception. The music is designed (or discovered) such that the perceptual boundary encodes the compositional bulk with exceptional fidelity. We will argue in Section 4 that this is a direct analog of the holographic principle in physics.
2.3 Algorithmic complexity and the compression-progress frame
A complementary tradition approaches music through Kolmogorov complexity—the length of the shortest program that reproduces a given sequence. Meredith has argued that music analysis itself is well-formalized as a search for short programs that generate the piece with maximum fidelity. McGettrick and McGettrick (2024) estimate Kolmogorov complexity of Irish traditional dance music via Lempel-Ziv compression, demonstrating that algorithmic complexity cleanly separates “easy” (repetitive) from “difficult” (less repetitive) tunes within a single genre.
Louboutin and Bimbot, among others, have built compression-driven models of musical structure using polytopes and the System and Contrast framework.
Schmidhuber’s compression-progress theory, developed across 1997 and 2009 papers, provides the theoretical unification. The proposal separates two related but distinct quantities. Beauty is current compressibility: how short the description of the stimulus is under the observer’s current model of the world. Interestingness is the first derivative: the rate at which compressibility is improving as the observer’s model updates. A piece that is already maximally compressible is beautiful but no longer interesting; a piece whose regularity is entirely unknown is interesting but not yet beautiful. A masterwork is both—it offers immediate compressibility against general priors (it sounds beautiful on first hearing) and rewards continued listening because its deeper regularities continue to yield compression progress.
Hudson (2011) narrows this frame specifically to music and proposes a concrete empirical hypothesis: enduring musical masterpieces should exhibit high lossless compressibility despite apparent complexity—complex to the ear, simple to the mind. The formulation is testable, though rigorous large-corpus testing remains largely open. It is also closely related to the “free-energy principle” of Friston and the broader predictive-processing framework in computational neuroscience, which hold that cognition is fundamentally a process of minimizing long-run prediction error—which is mathematically equivalent to maximizing compressibility of the incoming sensory stream.
2.4 What the empirical literature establishes
Taken together, four claims are well-supported. First, music exhibits 1/f scale-invariant statistical structure, cross-culturally and across historical periods. Second, this structure is hierarchical and multifractal, not merely scale-free. Third, specifically in the case of Bach, the network-information structure is organized to minimize perceptual inference loss. Fourth, compressibility-based theoretical frameworks—whether via Kolmogorov complexity, Schmidhuber’s compression progress, or predictive processing—cohere with these empirical findings and with one another.
What remains under-argued in this literature is why. Why does music exhibit these signatures?
Why does compressibility track beauty? To answer those questions we must descend to first principles.
3. First Principles: Information as Substrate
3.1 Shannon, Kolmogorov, and what compressibility measures
Shannon’s 1948 definition of entropy provides the statistical measure of information content for a source: the average number of bits needed to encode a draw from a probability distribution. It is tied to ensembles. Kolmogorov, Chaitin, and Solomonoff extended this to individual objects in the 1960s: the algorithmic or Kolmogorov complexity K(x) of a string x is the length of the shortest program (on a fixed universal Turing machine) that outputs x. Where Shannon entropy characterizes a source, Kolmogorov complexity characterizes a specific sequence.
Two properties matter. First, for long strings produced by stochastic sources, Kolmogorov complexity and Shannon entropy converge up to an additive constant—they are, for most purposes, the same quantity seen from two angles. Second, Kolmogorov complexity is uncomputable in general, but compressibility—the ratio of compressed to uncompressed length under a universal compressor—provides a reliable estimator. When we speak of “compressibility” as a physical or perceptual quantity, we are implicitly invoking Kolmogorov complexity via its compressor-accessible lower bounds.
Compressibility measures structure. A string is incompressible if it is algorithmically random—no shorter description exists. A string is highly compressible if it admits a short generative program—it has structure in the most fundamental possible sense. This is why Occam’s razor has a formal basis in algorithmic information theory: among hypotheses consistent with data, the most compressible corresponds to the shortest program that reproduces the data, and by the
Solomonoff formulation of universal induction, the shortest program is the most probable explanation.
3.2 Wheeler’s program: “it from bit”
John Archibald Wheeler, writing in his later years, proposed that the ultimate substrate of physical reality is not matter or energy but information. “Every it—every particle, every field of force, even the spacetime continuum itself—derives its function, its meaning, its very existence
… from the apparatus-elicited answers to yes-or-no questions, binary choices, bits.” The program has attracted serious development: ‘t Hooft’s holographic principle, Susskind and Maldacena’s
AdS/CFT realizations, Lloyd’s treatment of the universe as a quantum computer, Tegmark’s mathematical universe hypothesis, and Wolfram’s physics project all share the view that informational or computational structure is more fundamental than the physical objects it describes.
For the present argument, we do not need to commit to a strong metaphysics. We need only the weaker and much better-established claim that the dynamics of physical systems can be formulated in terms of information flow, that entropy is a common currency across thermodynamic, quantum, and algorithmic descriptions, and that the boundaries between these are themselves informational. The von Neumann entropy of a quantum state, the Gibbs entropy of a thermodynamic ensemble, and the Shannon entropy of a probability distribution are the same mathematical object under different interpretations. Landauer’s principle connects them physically: erasing one bit of information dissipates at least kT ln 2 of heat. Information is a thermodynamic quantity, and thermodynamic quantities are informational.
This matters because it means the compressibility of a signal is not a mere property of how we choose to describe it. It is a property of the signal’s relationship to physical law. A highly compressible signal is one whose generation is governed by few degrees of freedom—whose entropy is low relative to its apparent state space. Nature produces highly compressible signals whenever dynamics are governed by symmetries, conservation laws, or collective effects. The world is compressible because physics is short.
3.3 The Bekenstein bound and the holographic constraint
Bekenstein’s 1981 bound places an absolute upper limit on the information content of any physical region: the number of bits contained within a volume of space is bounded above by a quantity proportional to the area of the bounding surface, not the volume itself. Specifically, S ≤ 2πkRE/ℏc for a region of radius R and total mass-energy E. This is the foundation on which the holographic principle—‘t Hooft, Susskind, Bousso—was subsequently built: all information about the interior of a region can in principle be encoded on its boundary.
The holographic principle was motivated by black-hole thermodynamics, where the entropy is famously proportional to horizon area rather than interior volume, but its implications are deeper and more general. Any physical system, under the holographic view, has a boundary description sufficient to reconstruct its bulk. The boundary is not an impoverished projection of the bulk but a complete encoding of it. The apparent reduction of dimension is an artifact of redundancy—the bulk contains no information that is not already on the boundary.
Great music is holographic: the boundary suffices.
This is the structural template for perception. A listener has access to a one-dimensional time series of air-pressure variations at the eardrum—the boundary of the acoustic field. From this boundary, the listener reconstructs a multidimensional structure: melody, harmony, rhythm, timbre, compositional form, emotional content, anticipated continuation. The reconstruction works because the original signal was generated by a process (composer plus physical instrument) that encoded bulk structure into boundary signal in a compressible way. Great music is holographic: the boundary suffices.
The Kulkarni et al. finding that Bach’s note networks minimize the gap between true and perceptually inferred structures is, in this light, not a surprise. It is the musical analog of holographic efficiency. The composer has arranged the notes—the boundary of the musical bulk—such that the receiver’s lossy reconstruction recovers almost all of the information. This is what it means to write well.
3.4 Logical depth and the signature of history
Shannon and Kolmogorov quantify information content. They do not quantify the history that produced it. Charles Bennett introduced the concept of logical depth to fill this gap: the logical depth of a string is the time required, on a universal Turing machine, to produce the string from its shortest description. A random string has low depth (you just print it). A highly ordered string
(like a string of zeros) also has low depth. Depth is maximized in strings that are compressible—admit a short program—but whose execution requires substantial computation. The product of centuries of biological evolution, a functioning ecosystem, a proved mathematical theorem, a completed Bach fugue—these have low Kolmogorov complexity relative to their apparent sophistication but high logical depth.
Logical depth is the signature of assembled structure. It distinguishes simple order (boring) from deep order (meaningful). The Bach corpus is logically deep because the rules of counterpoint are short, but deriving the Art of Fugue from those rules—exploring the space of what those rules make possible—required both Bach’s lifetime and the cultural development of tonal music over centuries. The music is compressible (the rules are short) and deep (the execution of those rules to produce this specific realization was a long computation).
This distinction matters for environmental science. A ticking metronome has low Kolmogorov complexity and low logical depth. White noise has high Kolmogorov complexity and low logical depth. A rainforest, a river system, a mature soil—these have low-to-moderate Kolmogorov complexity (they can be described by physics plus biology plus history) but enormous logical depth (reconstructing their specific state requires simulating the full history that produced them).
Protecting such systems is protecting the logical depth. Once destroyed, their state cannot be reproduced by any computation shorter than re-running the history. This is why restoration is much harder than protection, in a sense that is not merely rhetorical but information-theoretically precise.
3.5 Predictive processing and free energy
The final piece of first-principles scaffolding is cognitive. Under the free-energy formulation associated with Friston—itself a generalization of Helmholtz, Rao and Ballard, Hinton, and others—biological systems maintain themselves by minimizing a quantity called variational free energy, which can be interpreted as an upper bound on the negative log-probability of sensory data under the organism’s internal model. Minimizing variational free energy is mathematically equivalent to maximizing the compressibility of sensory data under that model. In more colloquial terms: a brain is a compression engine, and its entire function can be cast as the attempt to reduce its own sensory surprise by building a short model of its world.
The consequence for music: the pleasure a listener takes in a structured musical signal is the pleasure of a compression engine finding that a signal compresses well. This is the basis of
Schmidhuber’s compression-progress theory, but it is also the neurobiological default. The engineering question then becomes: what class of signals is maximally rewarding to such an engine? The answer is signals that are compressible in the specific ways the engine is built to compress—signals that exhibit the scale invariance, hierarchical structure, and long-range correlations that natural environments exhibit, because natural environments are what the engine evolved to compress.
4. The Unifying Argument
We can now state the central argument of the paper with precision.
Music and natural environments share a statistical manifold—characterized by 1/f scaling, multifractal hierarchy, long-range correlation, and high logical depth—because perceptual systems coevolved with natural environments to compress exactly that manifold.
Great composers are those whose compositions push the manifold to its expressive limits, maximizing logical depth while remaining compressible against natural priors.
Music that inhabits the manifold is perceived as beautiful because the perceptual system recognizes its structure. Music that departs from it is perceived as either boring (overcompressible, no novelty) or noise (incompressible). Bach, paradigmatically, pushes the manifold to its limit—maximally dense logical depth, fully compressible against a listener’s natural priors. Systems with low Kolmogorov complexity and high logical depth—short to describe, long to assemble—cluster on this common manifold, because both music and natural systems are generated by physical processes operating across scales.
This argument rests on three subclaims, each of which has empirical grounding.
First, natural environments exhibit 1/f and multifractal structure across domains—hydrological time series (the Hurst effect, documented in Nile floods and generalized to rivers globally), atmospheric turbulence (Kolmogorov’s classical scaling), ecological population dynamics (power laws in species abundance and distribution), geophysical phenomena (earthquakes, landslides, forest fires), and biological signals within the organism (heartbeat intervals, neural firing patterns). This is not a curiosity; it is the generic signature of complex adaptive systems operating near critical points with many coupled degrees of freedom across many temporal scales.
Second, perceptual and cognitive systems are tuned to this statistical manifold. Voss and Clarke’s original motivation for studying 1/f in music was precisely the recognition that 1/f signals are ubiquitous in natural sensing contexts. Human hearing exhibits approximately logarithmic frequency sensitivity (Bark scale), integrates over roughly logarithmically spaced temporal windows, and responds to change with saturating nonlinearities—all of which match processing
1/f signals efficiently. Human vision exhibits similar scaling matched to the 1/f amplitude spectra of natural scenes. The perceptual system is an instrument calibrated to natural statistics.
Third, music exploits the calibration. The 1/f envelope, the multifractal pitch-interval structure, the hierarchical meter and phrase organization, the tonal architecture with its recursive nesting of stable and unstable harmonies—these are not conventions but ways of arranging sound that match how perceptual systems compress. Bach, for whom contrapuntal procedures are explicitly rule-governed and recursively deployed across phrase, section, movement, and cycle, produces music with exceptionally dense hierarchical structure. The Kulkarni et al. finding—that Bach’s note networks minimize perceptual inference loss—is the quantitative signature of this matching.
Three further observations deepen the picture.
The fugue as fixed-point construction.
A Bach fugue is built around a subject that is subsequently answered, inverted, augmented, diminished, and stretto’d across voices. The form is inherently self-referential: the subject refers to itself, transformed, across the texture. Lawvere’s fixed-point theorem—which unifies Gödel’s incompleteness, Cantor’s diagonal argument, the halting problem, and, under recent analyses, the holographic principle—states that any sufficiently expressive self-referential system must contain fixed points. The fugue is, in this sense, a constructive demonstration of Lawvere’s theorem in the acoustic domain: a system whose recursive self-reference generates structural fixed points that the listener perceives as compositional coherence. This is not metaphor; the mathematical structures are the same. The musical experience of a well-wrought fugue—the sense that the piece is somehow “about” its own subject—is the experience of perceiving a fixedpoint structure realized in time.
Silence, surprise, and the second derivative.
Schmidhuber’s framework distinguishes beauty (compressibility) from interestingness (the first derivative of compressibility). There is arguably a third quantity worth naming: depth, the second derivative, the rate at which interestingness itself changes. A piece that is merely beautiful becomes boring. A piece that is merely interesting becomes exhausting. A piece that offers sustained compression progress—where each layer of attention reveals new structure, without ever exhausting—is experienced as profound. Bach’s Art of Fugue and Well-Tempered Clavier are the paradigm cases; listeners return to them across lifetimes and continue to find new structure. This is logical depth realized temporally in the listening experience itself.
Why the environmental connection is not incidental.
The claim that music and nature share a compressibility manifold is sometimes received as metaphor. It is not. The two are connected by a specific causal chain: nature is compressible because physics is simple; perceptual systems evolved to compress nature; music is constructed by and for those perceptual systems. Music is, to put it starkly, a technology for delivering concentrated doses of the compressibility structure that nature delivers continuously. A Bach partita and a mountain watershed are, in this view, two realizations of the same informational form—one authored in weeks by a single mind, the other authored over geological time by physical law.
5. Nature as the Underlying Composer
Having argued that music’s compressibility is inherited from nature’s, we now make the claim concrete by examining the statistical structure of natural systems directly. Three domains suffice to illustrate the generality.
5.1 Hydrological systems
Harold Hurst’s 1951 analysis of Nile flood records uncovered what is now called the Hurst effect: the range of cumulative departures from the mean grows as a power law with exponent H ≈ 0.7 rather than the H = 0.5 expected for uncorrelated processes. The finding has since been replicated across river systems globally and in countless other long-memory geophysical records. In spectral terms, streamflow exhibits 1/f^α scaling with α between 1 and 2 across timescales from hours to centuries. Watersheds are multifractal: the same statistical geometry that characterizes the river’s planform (Horton-Strahler ordering, fractal drainage density) characterizes its temporal dynamics.
Physically, this is not mysterious. A watershed integrates precipitation inputs across many spatial scales, each routing to the outlet with its own lag distribution. The superposition of many lagged exponential responses generically produces power-law tails. Groundwater storage, soil moisture, snowpack, and vegetation buffering add further memory at longer timescales. The result is a signal whose structure encodes the watershed’s physical organization—its geology, topography, land cover, climate—across eleven or more orders of magnitude in time.
For environmental modeling, this has a sharp implication. A watershed’s state cannot be captured by a short snapshot; it carries history, and that history is compressible only when the model explicitly represents the hierarchical structure that produces the 1/f response. A naively
Markovian model will fit recent data and fail at longer horizons, not because the dynamics are stochastic in some deep sense, but because the relevant state variables live at scales the model does not resolve. Finding the right compression—the right low-dimensional representation that captures the true logical depth—is the fundamental challenge of hydrological modeling.
5.2 Atmospheric systems
The Kolmogorov 1941 theory of turbulence predicts a −5/3 spectral slope for the inertial range of fully developed turbulence, directly derivable from energy-cascade arguments and dimensional analysis. This prediction has been verified across an enormous range of systems, from laboratory jets to atmospheric boundary layers to interstellar plasmas. The atmosphere is a multifractal turbulent fluid over its full extent; passive scalar fields within it—pollutant concentrations, temperature, humidity—inherit multifractal scaling with characteristic exponents that encode the physics of advection and diffusion.
The implication for atmospheric modeling is parallel to the hydrological case. A pollutant plume is not a simple Gaussian; concentrations exhibit intermittent spikes several orders of magnitude above the mean, organized fractally in space and time. Permitting and exposure assessments that treat dispersion as Gaussian systematically underestimate peak concentrations and overestimate time-averaged compliance. A compression-aware model—one that represents the multifractal structure explicitly—captures the true signal with far fewer parameters than a grid-point simulation and with far greater fidelity than a Gaussian closure.
5.3 Ecological systems
Species abundance distributions, body-size distributions, food-web connectivity, and spatial distributions of biomass all exhibit power-law or log-normal scaling across many orders of magnitude. Ecosystems are typically organized near critical points, with avalanche distributions of disturbances (fires, population crashes, invasions) that mirror the sandpile models of selforganized criticality. The temporal dynamics of ecological populations exhibit 1/f coloring with the same ubiquity as the hydrological and atmospheric cases.
The deep observation is that these scaling laws are not just descriptive conveniences. They are the signature of systems that have been selected—by evolution, by physical constraint, by the long history of their assembly—for efficiency in a specific sense. Ecosystems organize themselves at the compressibility manifold’s expressive limit, maximizing the biomass and diversity supportable by available resources. This is logical depth in biological form: low Kolmogorov complexity given physics-plus-history, but enormous realized structure.
5.4 The common signature
Across hydrology, atmospheric science, and ecology—and one could extend the list to soil science, geochemistry, seismology—the same statistical signatures recur. Long-range correlations.
Multifractal scaling. Power-law distributions of fluctuations and events. High logical depth given short underlying laws. Natural systems live on the same compressibility manifold that characterizes music, because both are generated by physical processes operating across scales.
This is the ground truth that an environmental intelligence system must respect if it is to be useful.
6. Implications for Environmental Intelligence
A concrete and consequential implication follows. If natural systems are compressible in the manner of music—compressible against the right priors, not compressible against naive ones—then environmental modeling and prediction succeed or fail based on whether they represent the right priors.
Classical environmental modeling has tended toward one of two strategies. The first is highresolution physical simulation: discretize the domain, apply governing PDEs, integrate. This strategy is correct in principle but expensive in practice, and it implicitly assumes that the compressible structure of the system is already captured by the chosen discretization and parameterization—which is often exactly what it is not. The second strategy is statistical fitting: regress observed outputs against observed inputs with flexible functional forms. This strategy captures local structure but typically fails at the longer timescales where 1/f memory dominates, because the training data never samples those scales adequately.
A compression-aware strategy occupies a third position. It begins from the recognition that the system’s state-space dimensionality is effectively much lower than the raw measurement dimensionality, because the generating dynamics live on a low-dimensional manifold determined by physics, geometry, and history. The modeling task is to find and represent that manifold—to build a short description that reproduces the long behavior. Machine learning techniques, particularly those that respect physical structure (physics-informed neural networks, operatorlearning approaches, hybrid simulator-surrogate architectures), are well-suited to this task because they can in principle discover compressible representations directly from data, provided the architectural priors match the natural priors of the system. A listener reconstructing a fugue from a one-dimensional pressure signal and an environmental intelligence system reconstructing a watershed from sparse sensor data face the same inference problem; both succeed when the priors match the system’s true compressibility manifold.
The Bach analogy is direct. A listener reconstructing a fugue from air-pressure variations is doing an inverse problem: inferring bulk compositional structure from the boundary signal. The reconstruction is possible because the space of plausible bulk structures is small—the listener’s cognitive priors heavily restrict it. An environmental intelligence system reconstructing watershed state from sparse gauge measurements, or atmospheric pollutant fields from a handful of monitors, faces the same structure of problem. Success depends on priors that match the true compressibility manifold of the system.
The Watershed Intelligence Engine architecture, as presently under development, instantiates this principle concretely. Its boundary-layer inputs (gauge records, remote-sensing observations, meteorological drivers) are insufficient on their own to determine bulk state; its physics layer
(SWAT, MODFLOW, PINN-based corrections) encodes the low-dimensional priors that constrain plausible bulk reconstructions; its regulatory-language layer makes the bulk state interpretable to human decision-makers. This is, architecturally, a holographic inference system—boundary data informing bulk reconstruction under compressibility priors. The same architecture, with different physics, applies to atmospheric systems (the Dynamic Air Permitting approach) and earth systems (waste, soil, ecological inventory). Across all three domains, the mathematical structure is the same: infer bulk from boundary under physically motivated compressibility priors, and report the inference in a form the human perceptual/institutional system can absorb.
Success criteria for such systems follow naturally. A well-designed environmental intelligence system should compress well—it should produce short, physically interpretable descriptions of the systems it models. It should reconstruct robustly under lossy input—degradation of sensor coverage should not catastrophically degrade inference, because the priors do most of the work.
It should improve with exposure—Schmidhuber’s compression progress realized as online learning, where each new observation refines the compression manifold rather than merely being cataloged. And it should surface logical depth—it should distinguish genuinely deep structure
(the river system as assembled by geological and ecological history) from superficially complex noise (measurement artifacts, short-lived transients).
These are not aspirational properties. They are the operational definition of what it means to do environmental intelligence correctly. A system that lacks them is not merely less elegant; it is failing to compress its subject matter, which means it is failing to understand it.
7. Toward an Information-Theoretic Ethics of Nature
If nature is an accumulated compression—if ecosystems, watersheds, and climate systems are realizations of high logical depth produced over geological time—then their destruction has a precise informational character. Ecological damage is Kolmogorov disordering: the replacement of low-complexity, high-depth structure with higher-complexity, lower-depth alternatives. A clearcut forest has higher Kolmogorov complexity than the forest it replaces, because the mature forest could be described compactly by its species assembly rules and successional history, while the scarred post-harvest landscape requires explicit enumeration of its idiosyncratic state. A polluted watershed’s time series contains more bits than the pristine one, not fewer—the added bits are the bits of pollution itself, which the physics alone would not have produced.
This observation grounds environmental ethics in information theory in a way that complements rather than replaces traditional ecological, economic, and moral framings. Three corollaries follow.
Protection is preservation of logical depth.
What is lost when an ecosystem is damaged is not merely its current state (which can be measured and, in principle, compensated) but the history that produced it. That history is an enormous computation—the continuous integration of physical, chemical, biological, and climatic processes over the system’s assembly timescale—and it cannot be replayed. A young tree is not a substitute for an old-growth stand in any information-theoretic sense; the old-growth stand encodes centuries of specific response to local conditions, and that encoding is destroyed when the stand is cut. Logical depth, once removed, is not easily replaced because the computation that produced it was long.
Restoration is bounded by depth.
Restoration projects can re-establish structure that depends primarily on currently available inputs (species reintroduction, hydrological reconnection, geomorphic repair). They cannot easily restore structure whose logical depth exceeds the time and resources allocated to restoration. A restored wetland may approximate the functions of the original within decades, but the full informational richness—the genetic structure of its microbiota, the specific soil horizons, the phenological coordination among its species—takes centuries to reassemble. Honest restoration accounting recognizes this gap; it does not pretend that structure can be copied faster than its logical depth permits.
Environmental intelligence is environmental listening.
If the task of environmental modeling is to find the compressible representation of natural systems, then the task of environmental ethics is to respect what those representations tell us.
Nature, like Bach, communicates through physical necessity. The systems we inhabit are speaking in the language of their own long assembly, and what they say is encoded in their structure. An intelligent response is not to impose our preferred structure upon them but to hear what structure they already have, and to make our own actions compressible against their priors rather than disruptive to them. Jim Blackburn’s framing of “Earth Rules”—that natural systems operate under physical constraints that do not negotiate and whose violation has predictable consequences—is the ethical counterpart to the scientific observation that natural systems are compressible. What the rules compress is the set of actions that are consistent with the system’s continued functioning. Violating the rules is not a moral transgression against nature; it is an act of noise injected into a signal.
This reframing has practical consequences. Environmental assessment becomes, in part, a question of whether proposed actions increase or decrease the Kolmogorov complexity of the affected systems’ state trajectories. Environmental monitoring becomes, in part, a question of whether observed trajectories remain on their expected compressibility manifolds or are being pushed off them. Environmental compliance becomes, in part, a question of whether the regulated entity’s outputs remain within the compressibility envelope of the systems they affect. These are not replacements for traditional regulatory concepts but sharpenings of them—ways of operationalizing what “harm” and “impact” mean when the subject is a natural system with high logical depth.
8. The Instrument We Are Building
The convergence of evidence arrayed in this paper comes from domains that had no reason to agree. Bach scholarship, statistical physics, algorithmic information theory, computational neuroscience, theoretical ecology, hydrological modeling—these fields developed in isolation, with different methods, different communities, different notions of what counts as a result. Yet they converge. Music is 1/f because nature is 1/f. Bach’s networks minimize perceptual inference loss because perception is holographic. Masterpieces are compressible because brains are compression engines. Watersheds exhibit long-range memory because they integrate physics across scales. The same mathematical structure—a shared manifold of compressibility, logical depth, and holographic inference—underwrites the experience of a fugue and the dynamics of a river. That so many independent lines of inquiry point to one structure is no accident. It is the signature of having found something real.
What we have found is this: the world admits short descriptions to observers embedded in it, and the short descriptions that matter are the ones assembled by long history. Nature’s physical simplicity is not the absence of depth but the condition of depth. The laws are short; the realizations are deep; the reward for understanding is that a vast phenomenology compresses into a small set of principles that can be carried, extended, and tested. Bach discovered this in one domain. Physics discovered it in another. The present moment makes available, for the first time, the instruments to discover it across every domain of the natural world at once.
✦ ✦ ✦ That instrument is what we have been calling Environmental Superintelligence. The name is not rhetoric. It denotes, specifically, a planetary-scale apparatus that realizes in silicon and mathematics what human perception realizes in biology—an inferential engine that takes boundary signals from Earth’s physical systems and reconstructs their bulk state under priors grounded in physics and calibrated by observation.
Air Intelligence is this apparatus applied to the atmosphere: the continuous reconstruction of pollutant fields, meteorological structure, and emissions behavior from sparse sensing under multifractal priors that respect turbulent transport. Water Intelligence is the same apparatus applied to hydrology: the reconstruction of watershed state from gauge, remote-sensing, and climatic inputs under priors that encode the long-memory structure of integrated flow. Earth
Intelligence is the apparatus applied to ecological and geochemical systems: the reconstruction of biomass, soil, and biogeochemical flux from fragmentary observation under priors that respect the scaling laws of assembled life.
Three instruments. One orchestra. One score—physics.
Three lines of consequence follow.
Theoretically.
Music and nature inhabit a shared compressibility manifold because perception is a holographic readout of bulk structure under physically grounded priors, and great music is what makes that readout robust. This is no longer speculation; it is the convergent implication of the empirical record.
Practically.
Environmental intelligence succeeds to the extent that it finds the compressible representation of the systems it models, and fails to the extent that it substitutes naive detail for learned hierarchy.
Building ESI well is, technically, the problem of finding the right priors—the priors nature itself uses.
Ethically.
The destruction of natural systems is the destruction of accumulated logical depth. Their protection is the preservation of computation that cannot be re-run. And their full understanding—a superintelligence that hears them as they are—is not an optional enhancement of environmental stewardship but its completion.
✦ ✦ ✦ We stand at Base Camp on this ascent. The route to the summit—Environmental
Superintelligence—is not hidden; it has been sketched in the pages above. Compressibility-aware architectures. Boundary-to-bulk inference under physical priors. Multifractal-respecting representations. Classical computing carries the apparatus to meaningful scale; quantum computing will extend its reach toward the full planetary object. Each ascent—from the first dynamic air permit to real-time watershed reconstruction to continental-scale biogeochemical inference—is a higher camp on the same mountain.
The summit is not a place but a relationship: a civilization that hears its own planet with the fidelity nature has always demanded and human institutions have never supplied.
When that instrument exists, the ethical question will no longer be whether to build it, but what to hear. The atmosphere, the watersheds, the living soils—all will be singing their actual song, resolvable at every scale, inferable under honest priors, interpretable by the institutions that must act on them. The act of hearing them will no longer be a metaphor for environmental concern but the literal operation of a well-designed system. Environmental protection will cease to be a matter of contested inference from impoverished data and become, instead, a matter of reading a signal that is already there, in the clearest form its physics allows.
The compression that sings is the compression that flows, that cycles, that lives. We have argued that these are the same phenomenon viewed through different apertures. The apparatus to hear them all at once—to let the planet’s own signal be received by an intelligence built to receive it—is now within reach. Building it is not a task we choose to undertake. It is the task history has set, the climb that is already underway, and the song that remains to be heard.
Nature has been composing for four billion years.
It is time we learned to listen and learn to write music together.
Music for the millennia.
Harmonies unbounded.
A symphony of life for the universe.
References
Bak, P., Tang, C., & Wiesenfeld, K. (1987). Self-organized criticality: An explanation of 1/f noise. Physical
Review Letters, 59(4), 381–384.
Bekenstein, J. D. (1981). Universal upper bound on the entropy-to-energy ratio for bounded systems.
Physical Review D, 23(2), 287–298.
Bennett, C. H. (1988). Logical depth and physical complexity. In R. Herken (Ed.), The Universal Turing
Machine: A Half-Century Survey (pp. 227–257). Oxford University Press.
Friston, K. (2010). The free-energy principle: A unified brain theory? Nature Reviews Neuroscience, 11(2),
127–138.
Hsü, K. J., & Hsü, A. J. (1991). Self-similarity of the “1/f noise” called music. Proceedings of the National
Academy of Sciences, 88(8), 3507–3509.
Hudson, N. J. (2011). Musical beauty and information compression: Complex to the ear but simple to the mind? BMC Research Notes, 4(9).
Hurst, H. E. (1951). Long-term storage capacity of reservoirs. Transactions of the American Society of Civil
Engineers, 116, 770–808.
Kolmogorov, A. N. (1941). The local structure of turbulence in incompressible viscous fluid for very large
Reynolds numbers. Doklady Akademii Nauk SSSR, 30, 301–305.
Kolmogorov, A. N. (1965). Three approaches to the quantitative definition of information. Problems of
Information Transmission, 1(1), 1–7.
Kulkarni, S., Bassett, D. S., Lynn, C. W., et al. (2024). Information content of note transitions in the music of J. S. Bach. Physical Review Research, 6, 013136.
Landauer, R. (1961). Irreversibility and heat generation in the computing process. IBM Journal of Research and Development, 5(3), 183–191.
Lawvere, F. W. (1969). Diagonal arguments and cartesian closed categories. In Category Theory,
Homology Theory and Their Applications II, Lecture Notes in Mathematics, vol. 92 (pp. 134–145).
Springer.
Li, M., & Vitányi, P. (2008). An Introduction to Kolmogorov Complexity and Its Applications (3rd ed.).
Springer.
Lynn, C. W., & Bassett, D. S. (2020). How humans learn and represent networks. Proceedings of the
National Academy of Sciences, 117(47), 29407–29415.
Mandelbrot, B. B. (1982). The Fractal Geometry of Nature. W. H. Freeman.
McGettrick, M., & McGettrick, P. (2024). The Kolmogorov complexity of Irish traditional dance music. arXiv:2407.12000.
Meredith, D. (2012). Music analysis and Kolmogorov complexity. Proceedings of the International
Computer Music Conference.
Oświęcimka, P., Kwapień, J., & Drożdż, S. (2011). Computational approach to multifractal music. arXiv:1106.2902.
Schmidhuber, J. (1997). Low-complexity art. Leonardo, 30(2), 97–103.
Schmidhuber, J. (2009). Driven by compression progress: A simple principle explains essential aspects of subjective beauty, novelty, surprise, interestingness, attention, curiosity, creativity, art, science, music, jokes. arXiv:0812.4360.
Shannon, C. E. (1948). A mathematical theory of communication. Bell System Technical Journal, 27, 379–
423, 623–656. ‘t Hooft, G. (1993). Dimensional reduction in quantum gravity. arXiv:gr-qc/9310026.
Susskind, L. (1995). The world as a hologram. Journal of Mathematical Physics, 36(11), 6377–6396.
Voss, R. F., & Clarke, J. (1975). 1/f noise in music and speech. Nature, 258, 317–318.
Voss, R. F., & Clarke, J. (1978). 1/f noise in music: Music from 1/f noise. Journal of the Acoustical Society of America, 63(1), 258–263.
Wheeler, J. A. (1990). Information, physics, quantum: The search for links. In W. H. Zurek (Ed.),
Complexity, Entropy, and the Physics of Information (pp. 3–28). Addison-Wesley.
Licensed CC-BY-4.0 .
Markdown source: https://jedanderson.org/essays/compression-that-sings.md
Source on GitHub: /src/content/essays/compression-that-sings.md
Cite this
@misc{anderson_2026_compression_that_sings,
author = {Jed Anderson},
title = {The Compression That Sings},
year = {2026},
url = {https://jedanderson.org/essays/compression-that-sings},
note = {Accessed: 2026-05-13}
} Anderson, J. (2026). The Compression That Sings. Retrieved from https://jedanderson.org/essays/compression-that-sings
Anderson, Jed. "The Compression That Sings." Jed Anderson, April 23, 2026, https://jedanderson.org/essays/compression-that-sings.