# Proof we are not living in a simulation

### If quantum biology is a thing, reality “computes” more efficiently than any computer.

**Introduction**

To some of us it is obvious, something we know in our hearts beyond doubt. There is nothing more real to me than my own consciousness. It is an example of what Immanuel Kant called a synthetic *a priori*. We may not know what our individual consciousness is, but we know unshakably that it is not something generated by computation.

Others might cite the results of experiments in parapsychology. If we have premonitions of the future, or if we can know what our distant friends are thinking, then clearly our minds are doing things that computers cannot do. This argument, however, cuts both ways. If our universe is a computer simulation it doesn’t have to be programmed in a way that conforms consistently to causal laws. If the programmer knows what is happening elsewhere (or elsewhen), he might leak that information into some of the simulated brains.

It is common among scientists to think in terms of physical realism, i.e., the hypothesis that physical reality is the only reality. It follows that consciousness is an epiphenomenon generated by the physical brain. If the brain is a computing engine, then it is reasonable to conclude that our multiple consciousnesses may be generated by a much larger computing engine.

To other scientists it is not at all obvious that physical realism is the right way to go. A competing world-view was articulated prominently by William James, rooted in the dualism of Descartes. It is that consciousness is a separate phenomenon, independent of physical reality, and the brain is not the generator but the transducer of consciousness. The function of the brain is to connect the non-physical world of thought to the physical realm, downloading intent and uploading sensation. From this perspective, the idea that we are living in a simulation is nonsensical.

Even for those who believe in physical realism, and whose subjective experience doesn’t support the synthetic *a priori*, there is a cogent argument from quantum physics that our universe is not a simulation.

**Summary**

We know that a simulation of the gross aspects of a macroscopic system can run much faster than the system itself. But this is not true of quantum systems. The complexity of computations in classical physics grows linearly with the number of particles, but the complexity of quantum mechanical computations grows exponentially. Solving the quantum wave equation for a system as simple as one molecule of H2O1 is far beyond the capacity of any conceivable digital computer. To the extent that our world depends on details at the quantum level—for example, if you believe at all that quantum biology is a real science—it cannot be simulated with any reasonably sized computer2.

We are so used to the idea that “theory is faster and more convenient than experiment” that we may not even recognize this as an assumption. Sometimes, it’s true.

If we can predict an orbit with a formula, it’s much, much faster and more convenient than watching a moon or a planet and tabulating its trajectory. For the sun and one planet, Newton’s gravitational equations can be solved to give the formula for an ellipse, and the planet’s position can be predicted with moderate accuracy using just pencil and paper.

For more complicated problems, three or more bodies, there is no formula. But the orbit can still be computed second-by-second. We use the fields now to predict the motion for the next second, then update the positions of all the bodies and update the fields. Then use the new positions to compute fields, and the new fields to compute the motions of all bodies. Rinse and repeat, second by second. This technique, called numerical integration, was known to Newton 300 years ago. It was tedious to do this calculation by hand, but not nearly as tedious as watching the planets and waiting for them to move. Today, a computer program can do numerical integration in a flash. Big supercomputer programs run for weeks at a time to calculate models of the entire universe with moderate precision.

It’s our expectation that — even if computer programs are complex and a bit slow — they are much, much faster than watching the galaxies or even the planets move. Of course, theory is always faster and more convenient than experiment. Except when it’s not.

**Quantum calculations are complex on a scale that leaves classical systems in the dust**

The equations of quantum mechanics are notoriously complex. But theoreticians routinely solve them for interactions of two particles. In classical mechanics, the three-body system is chaotic, meaning that it never settles into repetitive orbital motion. The quantum three-body problem doesn’t have this problem; there is a stationary ground state solution to the Schrödinger equation. But a different problem arises: the three-body problem requires enormously more computation compared to the two-body problem. The equation for the wave function of a Helium atom (two electrons and a nucleus) is just barely within the range of modern computing power.

And with each additional particle, solving the wave equation requires exponentially more computing power. In classical mechanics, it takes twice as long to calculate 4 planets as to calculate 2 plantes. But in quantum mechanics, it takes a billion billion times as long to calculate 4 electrons, compared to 2. Why is this? Quantum equations aren’t about particles — they are about configurations. The configuration for one particle lives in a 3-dimensional space .But the configuration for two particles lives in a 6 dimensional space, 3 for each particle. When we get to four electrons, we need a 12-dimensional space, and each dimension must be divided into roughly 1,000 pixels to get good resolution. So adding each new particles * multiplies *the difficulty of the problem by 1,000 * 1,000 * 1,000 = 1 billion.

A small molecule might have a few dozen electrons, and it would take a computer much larger than the universe to calculate it.

**You have heard that quantum mechanics makes very precise predictions**

Quantum physics has a reputation for making exact predictions that have been verified by experiment. * All the exact predictions of QM are based on two-particle systems*. For example, all the computations of high energy physics that are used to predict what happens inside particle accelerators are based on two particles colliding. Larger systems are solved only with a gross approximation, usually the assumption that individual electrons are not entangled with one another. We can calculate the hydrogen atom exactly (one electron, one proton), and the helium atom (three particles) with a stretch of our exact methods, but every larger atom is based on approximations and assumptions.

The theory of bonding that underlies our understanding of molecules is rooted in calculations that are based on the (quantum) hydrogen atom. We pretend that electrons don’t interact with one another, and then we can solve the equations. Supplemented by detailed empirical measurement, this becomes a reasonably useful approximation.

Solid state physics routinely calculates wave functions for clouds of huge numbers of particles, but the calculation is based on the assumption that each electron sees only the lattice of positive nuclei and not the sea of other electrons. It works pretty well to the extent that the other electrons are spread out and thus pushing equally from all directions. We can do better by taking the electron cloud calculated in this way and using it to calculate a corrected wave function. This is called perturbation theory, and it can be iterated indefinitely, each approximation used to produce a better next approximation. When this is done, the process converges on a solution, but it is not the exact solution that would be obtained if the interaction of every electron with every other were fully accounted.

**A quantum system “calculates” its own state far more efficiently than a computer can**

We start to see that nature is one huge quantum computer, performing parallel calculations far faster and more efficiently than we know how to solve the same system using our equations. This is an extraordinary fact. We may have exactly the right equation, but we can’t solve it exactly, so we look to nature’s matter computer to give us exact answers. This is a case where experiment is more efficient than theory. The experiments are faster, more convenient, and easier. In fact, solving the equations exactly is “impossible” with the any of the computing tools that we know how to build.

Every quantum system of more than three particles “simulates itself” far more efficiently than we can devise a computer program to simulate it.

For those of us seeking to understand physics of the universe, this is the ultimate cosmic joke. We have equations that we believe (with good reason) to be the basis of all physical interactions at “low energy”, where by “low” we mean up to and including energies that are reached at the center of the sun. And yet, we cannot know for sure, because we don’t know what predictions these equations make, that is, we cannot solve the equations.

As far as we know, the universe is working out the consequences of these equations in the most efficient possible way, far more efficient than any supercomputer we can program. In fact, the universe is working out answers far more efficiently than a hypothetical digital computer the size of the universe.

**Simulating a universe: Is detail at the quantum level really necessary?**

Maybe the approximations (independent electrons) on which chemistry is based are good enough to compute bulk behaviors of materials. Maybe we don’t have to compute the detailed behavior of every particle in the sun to know its average brightness. Maybe the solar flares and sunspots can be approximated well enough with random number generators.

Maybe.

But living systems seem to depend on quantum mechanics at the level of single molecules. There are molecular machines — single molecules that operate with extraordinary intelligence that we don’t understand. It seems a sure bet that these have evolved to exploit the full potential of quantum physics.

DNA polymerase is a single molecule that crawls along a single strand of DNA and pulls the right nucleotide base out of the surrounding plasma to pair with the nucleotide that it finds to be next along the chain.

Homologous DNA repair is the process of filling in a gap in a section of a chromosome that is severely damaged. There are molecules that seek out the sister chromosome for reference, using pairs of chromosomes as a backup system. The whole process of finding the corresponding area of the sister chromosome, reading the information, and using it to repair the break is accomplished by evolved molecular machines.

The Golgi apparatus, present in every eukaryotic cell, tags individual molecules with a destination address and guides each molecule to the place it is needed.

Photosynthesis is the most efficient conversion of light energy to chemical energy known to man. It is accomplished by the molecule called chlorophyl, using quantum tricks.

The mysterious, intelligent behavior of single biomolecules is likely mediated by the quantum properties of water, which support structures in response to molecular and surface interactions that extend ~100 micrometers from the source. This is a scale commensurate with the size of typical eukaryotic cells.

Brains are not static computation engines. The brain adapts to the way in which it is used, growing new capacity in areas where past experience has taught it to anticipate future demands. How this works is not understood, but it is easy to believe this ability invokes mechanisms at least as sophisticated as DNA repair.

A brain is not a neural network. Neural networks use pseudo-random number generators to determine the firing of each neuron. The behavior of real neurons may look to us to be partially random, simply because we don’t understand them; but we should not be so arrogant as to assume the detailed molecular interactions within each neuron are adequately modeled by randomness. Neural networks generally are structured with a pyramid shape, summarizing a large amount of data with a single output. Neural networks for artificial general intelligence have not to my knowledge been realized.

Stuart Kauffman has inspired and interpreted research suggesting that quantum criticality in biomolecules is an evolved adaptation. His idea is that, while human-designed machines are large enough to make quantum fluctuations irrelevant, evolution-designed molecular machines seem to be consistently poised on a quantum knife edge, as if nature * wanted* “quantum randomness” to determine macroscopic outcomes. There is a hint here that what we regard as “randomness” is not really random in biological systems. This is not physically unreasonable. Heisenberg’s “randomness” depends on the absence of long-range correlations, and Kauffman argues that biological systems are evolved to use large-scale quantum entanglement that we don’t understand, so the behavior appears random to us.

**QED**

To simulate in detail the quantum behavior of a single small molecule would require a computer larger than the universe.

The behavior of the universe can only be simulated (with a computer smaller than the universe) if we think that detailed quantum behavior is irrelevant, and can be replaced with pseudo-random algorithms. It is possible, but by no means certain, that non-living systems can be adequately simulated in a way that ignores quantum details. Life, however, seems to depend on large-scale quantum entanglement in ways that we are only beginning to understand. To the extent that living systems are evolved to exploit quantum entanglement, living systems cannot be simulated with a computer smaller or faster than the system itself.

Why won’t Substack let me use subscripts?

By “reasonably sized” I mean anything less than, for example, a googleplex times the size of our universe.

Doesn't Arthur Clarke's quote, "Any sufficiently advanced technology is indistinguishable from magic," that I believe you used in another post, imply that the type of "computer" needed to simulate the universe is completely possible?

It makes sense that the universe would be the most efficient system for "calculating" itself as nature tends not to be wasteful. I would only mention the existence of analog computing. It doesn't change the conclusion of your analysis-- analog computing can provide, at best, several orders of magnitude of increased computational performance, not enough to grow exponentially with the number of particles being simulated, but enough to suggest that digital-quantum hybrid computing will (or maybe already is) occupying a niche analogous to digital-analog hybrids in the 1950's and 60's. Perhaps digital-quantum hybrid computers could more easily solve some of the multi-particle quantum interactions hinted at by the DNA and photosynthesis processes you've mentioned.

https://en.wikipedia.org/wiki/Analog_computer