Information Engines
Abstract
The relationship between physical systems and intelligence has long fascinated researchers in computer science and physics. This talk explores fundamental connections between thermodynamic systems and intelligent decision-making through the lens of free energy principles.
We examine how concepts from statistical mechanics - particularly the relationship between total energy, free energy, and entropy - might provide novel insights into the nature of intelligence and learning. By drawing parallels between physical systems and information processing, we consider how measurement and observation can be viewed as processes that modify available energy. The discussion encompasses how model approximations and uncertainties might be understood through thermodynamic analogies, and explores the implications of treating intelligence as an energy-efficient state-change process.
While these connections remain speculative, they offer intriguing perspectives for discussing the fundamental nature of intelligence and learning systems. The talk aims to stimulate discussion about these potential relationships rather than present definitive conclusions.
Szilard Intelligence
The objective is to provide a mathematical definition of intelligence that grounds our intelligence in the physical sciences. Previous efforts at mathematically defining universal intelligence have reviewed the literature to distil the facets of intelligence, taking a descriptive approach to modeling intelligence.
Our definition is inspired by statistical mechanics descriptions of the universe and the simple idea that an intelligent decision is one that achieves a desired state change within the universe with the minimal use of resource. From a physical perspective there are two principal resources we can consider:
- Time
- Energy
In practice energy is conserved, so the resource of interest is free energy - the energy available to us to do work.
This work builds on ideas from Leo Szilard’s 1929 paper “On the decrease of entropy in a thermodynamic system by the intervention of intelligent beings” which established fundamental connections between information and thermodynamics.
Energy and Intelligence
In the physical universe, any form of intelligence - whether biological or artificial - must operate within the constraints of fundamental physical laws. The two most critical resources that bound intelligent behavior are:
- Time: The duration required to achieve a desired state change
- Free Energy: The portion of total energy available to perform useful work
While total energy in a closed system is conserved (first law of thermodynamics), free energy represents the energy available for useful work. This distinction is crucial because:
- Not all energy can be converted to useful work
- Some energy is always lost to entropy (second law of thermodynamics)
- Intelligent systems must optimize their use of available free energy
This framework allows us to quantify intelligence in terms of how efficiently a system uses these fundamental resources to achieve its objectives.
The connection between intelligence and thermodynamics was first explored in Leo Szilard’s seminal 1929 paper, which showed that the act of measurement and decision-making by an intelligent being has an associated energy cost. This established a fundamental link between information processing (a key aspect of intelligence) and physical resources.
Information and Physical Systems
The connection between information and physical systems was first formalized by Claude Shannon, but has deep roots in statistical mechanics and thermodynamics. This relationship becomes particularly important when we consider intelligent systems that must process information to make decisions.
Simulation Spaces and Domain Mapping
In the Szilard framework, intelligent systems operate by mapping between different spaces. Two fundamental spaces are particularly important:
- The Null Space: This represents states or transformations that are physically impossible or prohibited by the laws of nature.
- The Domain Space: This represents the set of all possible valid states or transformations available to the system.
The concept of separability is crucial in this framework. An intelligent system must be able to:
- Distinguish between null and domain spaces
- Efficiently identify valid state transformations
- Minimize resource expenditure in mapping between states
The quality of this separation and mapping is measured by domain fidelity - how accurately and efficiently the system can represent and navigate its operational domain.
Domain fidelity can be mathematically expressed as:
\(F_D = \frac{\text{Correct Domain Mappings}}{\text{Total Attempted Mappings}}\)
This measure relates directly to the SIQ, as higher domain fidelity generally correlates with more efficient resource usage. A system that frequently attempts invalid state transformations or fails to recognize impossible states will have lower efficiency and thus a lower SIQ.
This framework has practical implications for: * Designing AI systems that efficiently learn domain constraints * Optimizing resource allocation in intelligent systems * Developing better error detection and correction mechanisms * Understanding the limitations and capabilities of different types of intelligence
The separation between null and domain spaces also helps explain why some problems are inherently harder than others - they may require more complex domain mapping or have less clear separation between possible and impossible states.
Parameters and Domain Fidelity
To make the Szilard framework practical, we need measurable parameters. Three key parameters characterize an intelligent system’s performance:
- Resource Efficiency (η): How efficiently the system uses energy and time
- Domain Coverage (C): The proportion of valid domain space the system can access
- Mapping Accuracy (M): The precision of state transformations
Domain Fidelity (\(F_D\)) can be decomposed into its constituent parameters:
\(F_D = η × C × M\)
Where: * η (Resource Efficiency) = \(\frac{\text{Minimum Theoretical Resources}}{\text{Actual Resources Used}}\) * C (Domain Coverage) = \(\frac{\text{Accessible Valid States}}{\text{Total Valid States}}\) * M (Mapping Accuracy) = \(\frac{\text{Successful Transformations}}{\text{Attempted Transformations}}\)
Each parameter is bounded between 0 and 1, with 1 representing perfect performance.
This parameterization provides practical tools for:
- Evaluating System Performance:
- Identify bottlenecks in intelligence
- Compare different systems objectively
- Track improvements over time
- Optimization Strategies:
- Target specific parameters for improvement
- Balance trade-offs between parameters
- Allocate development resources effectively
- Design Decisions:
- Choose appropriate architectures
- Set realistic performance targets
- Define success metrics
The relationship between these parameters and the overall Szilard Intelligence Quotient (SIQ) is:
\(SIQ = F_D × \text{System Specific Factors}\)
This relationship acknowledges that while domain fidelity is crucial, other system-specific factors may influence overall intelligence. These might include: * Environmental constraints * Problem-specific requirements * Implementation limitations
Embodiment and Communication
A critical aspect of intelligence is how it is embodied in physical systems. This embodiment creates fundamental constraints on both computation and communication.
Embodiment Factors
bits/min | billions | 2,000 |
billion calculations/s |
~100 | a billion |
embodiment | 20 minutes | 5 billion years |
Figure: Embodiment factors are the ratio between our ability to compute and our ability to communicate. Relative to the machine we are also locked in. In the table we represent embodiment as the length of time it would take to communicate one second’s worth of computation. For computers it is a matter of minutes, but for a human, it is a matter of thousands of millions of years. See also “Living Together: Mind and Machine Intelligence” Lawrence (2017)
There is a fundamental limit placed on our intelligence based on our ability to communicate. Claude Shannon founded the field of information theory. The clever part of this theory is it allows us to separate our measurement of information from what the information pertains to.1
Shannon measured information in bits. One bit of information is the amount of information I pass to you when I give you the result of a coin toss. Shannon was also interested in the amount of information in the English language. He estimated that on average a word in the English language contains 12 bits of information.
Given typical speaking rates, that gives us an estimate of our ability to communicate of around 100 bits per second (Reed and Durlach, 1998). Computers on the other hand can communicate much more rapidly. Current wired network speeds are around a billion bits per second, ten million times faster.
When it comes to compute though, our best estimates indicate our computers are slower. A typical modern computer can process make around 100 billion floating-point operations per second, each floating-point operation involves a 64 bit number. So the computer is processing around 6,400 billion bits per second.
It’s difficult to get similar estimates for humans, but by some estimates the amount of compute we would require to simulate a human brain is equivalent to that in the UK’s fastest computer (Ananthanarayanan et al., 2009), the MET office machine in Exeter, which in 2018 ranked as the 11th fastest computer in the world. That machine simulates the world’s weather each morning, and then simulates the world’s climate in the afternoon. It is a 16-petaflop machine, processing around 1,000 trillion bits per second.
See Lawrence (2024) embodiment factor p. 13, 29, 35, 79, 87, 105, 197, 216-217, 249, 269, 353, 369.
Practical Implications
These theoretical connections between thermodynamics and intelligence have practical implications for how we design and understand intelligent systems. The efficiency with which a system uses energy and manages state transformations may provide a fundamental measure of intelligence.
Thanks!
For more information on these subjects and more you might want to check the following resources.
- book: The Atomic Human
- twitter: @lawrennd
- podcast: The Talking Machines
- newspaper: Guardian Profile Page
- blog: http://inverseprobability.com
References
the challenge of understanding what information pertains to is known as knowledge representation.↩︎