Beyond Expectations: Cosmic Anomalies

Scientific models are built to describe the universe as accurately as possible. Yet, modern cosmology contains a number of well‑documented observations that do not fully align with current theoretical expectations. These anomalies are not failures — they are opportunities. They highlight areas where our understanding may be incomplete, and where new logic, structure, or physical principles might be needed.

This page presents such anomalies in a clear and neutral way. The goal is not to dismiss the standard model, but to acknowledge the places where data consistently deviates from prediction, and to encourage open, sober exploration of the underlying mechanisms that may be missing.

Toca S2 Flare Analysis

ToCA S2 Flare Analysis: Geometric Latency and the 1.7x Residual Signal

Observations of the star S2 during its pericenter passage around Sagittarius A* have been modelled with high precision using General Relativity. However, while the overall orbital motion is well-captured, a rigorous residual audit reveals a consistent excess noise factor of approximately 1.7x (Reduced \bm{\chi^2 \approx 1.50 - 1.65}). This systematic deviation persists across independent datasets from Do et al. (2019), GRAVITY (2020), and GRAVITY (2024).

The pattern of these residuals suggests a specific physical mechanism: Geometric Latency. In the ToCA framework, the cosmic substrate (FCC-lattice) has a finite processing speed. As S2 reaches velocities of \bm{7,600 \text{ km/s}} (\bm{0.026c}) in a high-tension environment, the substrate’s geometric configuration cannot update instantaneously.

Key Findings:

• The 1.7x Excess: After full relativistic corrections, the data still exhibits ~70% more variance than predicted by a smooth spacetime model.

• Geometric Latency (\bm{\tau}): Reverse engineering of the residuals points to a substrate processing delay of \bm{\tau \approx 15–37} minutes.

• Phase Shift: This latency creates a detectable phase shift between the star's predicted position and the geometry it experiences—a unique signature of the underlying hardware.

Highlighting these deviations serves to document where the data reveals the N-shift update rate of the universe. This is not a contradiction of established theory, but a measurement of the substrate’s operational limits.

Image credit: X-ray: NASA/UMass/D.Wang et al., IR: NASA/STScI

Why Dust Cannot Be the Primary Explanation

Dust is an important factor in many astrophysical observations, but it cannot account for the specific anomalies discussed here. If dust were the dominant cause of the observed deviations, we would expect uniform and cumulative effects across all long‑duration space‑based instruments — including satellites, deep‑space probes, and telescopes such as Hubble.

However, these systems do not exhibit the kind of progressive “sandblasting” or signal degradation that a dust‑driven mechanism would imply. Their optical surfaces, detectors, and calibration curves remain stable over time within expected tolerances. The selective nature of the anomalies therefore suggests that dust is not the primary driver, and that additional physical factors must be considered.

The Diode Paradox

In conventional descriptions, a diode on a circuit board is said to “emit photons” or “accelerate electromagnetic energy” into free space. But the diode itself does not contain enough stored energy to accelerate anything to the speed of light. No physical component on the board is capable of imparting c to a particle or wavefront.

A more consistent interpretation is that the diode does not accelerate light at all. Instead, it triggers a discrete state‑change in the underlying structure 

a local snap in the lattice  and the resulting information update propagates through the substrate at the characteristic speed c. The diode provides the boundary condition; the substrate provides the propagation.

This resolves the apparent paradox: the diode does not launch a photon and push it to light‑speed. It simply initiates a transition in a medium that already supports propagation at c.

 

Perpetual Motion and Latency

If the universe had no latency — if changes could occur instantaneously — all dynamics would complete in a single update. There would be no process, no evolution, no unfolding, no time. A universe without latency would effectively be a perpetual‑motion machine: energy could circulate without loss, and systems could regulate themselves without delay.

In ToCA, time is not a separate dimension but a consequence of regulation within a medium. When there is a substrate, field, or connective medium through which all changes must propagate, several properties arise automatically:

  • latency (no update is instantaneous)

  • a maximum propagation speed (light)

  • time dilation (time slows in “denser” regions of the substrate)

  • the impossibility of perpetual motion (regulation always requires energy and time)

Accept a substrate — and the entire structure falls into place.

 

Why Early Physicists Missed the Digital Layer

For most of scientific history, physicists had no concept of digital systems. The idea of discrete updates, clock cycles, or step‑based computation simply did not exist. Their world was mechanical, and their intuition was shaped by gears, levers, pulleys, and continuous motion.

Ironically, many of the machines they built — especially early perpetual‑motion devices — were digital in structure: systems of discrete teeth, steps, and mechanical states. But because the output looked smooth and analog, they interpreted the entire system as analog.

They saw the flow, not the steps.

This is why the possibility of a digital substrate never entered classical physics. The tools, the language, and the mental models weren’t there yet.

Human biology reinforces this bias. Our nervous system can only experience the world as continuous. If we were exposed to the raw “n‑step” update rate of a digital substrate, the result would be overload — we are not built to perceive discrete updates directly. Biology smooths everything into analog experience.

But the existence of analog experience does not mean the underlying substrate is analog. It only means we are calibrated to perceive it that way.

Just as early engineers overlooked the digital nature of their own gear systems, we overlook the possibility that the universe itself may operate on discrete regulatory steps beneath the smooth flow of time.

Leonardo da Vinci and the Human Search for Perpetual Motion

Leonardo da Vinci (1452–1519) was a pioneer of science and engineering whose mechanical imagination was centuries ahead of its time. Among his many explorations, he sketched early concepts for perpetual‑motion machines — devices imagined to run forever without any external energy input.

Like many thinkers before and after him, Leonardo was fascinated by the idea of endless motion. The dream of a “perpetuum mobile” has followed humanity for millennia, appearing in ancient manuscripts, medieval engineering, Renaissance notebooks, and even modern thought experiments. It reflects a deep human desire to understand — and perhaps overcome — the limits of nature.

But Leonardo’s genius was not just in imagining such machines. Through careful study of friction, materials, and mechanical constraints, he ultimately demonstrated why true perpetual motion is impossible. His notebooks show a clear understanding that energy is always lost to friction and resistance, and that no mechanism can sustain motion indefinitely.

In this way, Leonardo represents both sides of the human story: the eternal search for perfect motion, and the scientific clarity that reveals why nature forbids it.

 

The Brain, Biology, and the Digital Substrate

Human biology is fundamentally analog. Our nervous system processes the world through continuous signals, gradients, delays, and thresholds. Every sensation, every emotion, every decision is shaped by biological latency — the time it takes for signals to propagate, integrate, and stabilize.

Yet the world we build — from computers to communication networks — runs on a digital substrate. It updates in discrete steps, at speeds far beyond anything biology can directly experience.

This creates a deep asymmetry:

  • Biology must experience reality as a smooth, continuous flow.

  • The underlying substrate of the universe may update in discrete steps.

  • We cannot perceive those steps directly — we would burn out if we did.

If the brain were exposed to the raw “n‑step” update rate of a digital substrate, the result would be catastrophic: sensory overload, loss of coherence, and breakdown of regulation. Our biology is calibrated to filter, smooth, and integrate — to turn discrete events into continuous experience.

This does not mean the discrete steps aren’t there. It means we are not built to perceive them.

Just as a movie appears continuous even though it is made of frames, consciousness appears continuous even though it is built on neural spikes and regulatory cycles.

Latency protects us. Continuity is a biological necessity. And the analog experience of time is the interface layer between the organism and the substrate it inhabits.