Skip to content

Can You Spell Science without AI? - DTNS Hangout

New research shows AI models are evolving to process visual data like human neurology, even perceiving motion in static optical illusions. A DTNS Hangout with Dr. Kiki Sanford & Andy Beach dives into this study, bridging biological perception and computer vision, and questioning AI's interpreta

Table of Contents

New research indicates that deep neural networks trained to mimic human visual prediction can be deceived by optical illusions, suggesting that artificial intelligence systems are evolving to process visual data in ways that strikingly resemble human neurology. A recent discussion on the DTNS Hangout featuring neurophysiologist Dr. Kiki Sanford and media technology expert Andy Beach highlighted a study from Japan’s National Institute for Basic Biology, where an AI model perceived motion in static images known as "rotating snakes." This development bridges a significant gap between biological perception and computer vision, raising critical questions about how machines interpret reality and the future computational power required to simulate true cognitive function.

Key Points

  • Biological Mimicry: Japanese researchers found that "PredNet," a deep neural network designed to predict visual input, perceives motion in static optical illusions, validating theories that AI vision is beginning to mirror human neural processing.
  • The Focus Gap: While AI experiences the illusion of motion, it currently lacks the human ability to fixate on a specific point to halt the perceived movement, highlighting a divergence in attentional control.
  • Data Efficiency: Expert analysis suggests that computer vision models do not require 4K resolution to function effectively, with 1080p or even 720p often providing sufficient data for identification and tracking.
  • Quantum Hypotheses: Emerging research into the "Necker Cube" illusion suggests human perception may involve quantum tunneling-like processes, a complexity that classical computing models struggle to replicate without massive code.

Bridging the Gap Between Biological and Computer Vision

The convergence of biological science and artificial intelligence has reached a new milestone with the application of PredNet (Predictive Neural Network). Researchers at the National Institute for Basic Biology in Japan trained this deep neural network to operate similarly to the human visual system, which relies heavily on prediction. In human cognition, the brain predicts incoming visual data and compares it against actual light input to reconcile reality. By replicating this process, the AI model was successfully "tricked" by the "rotating snakes" illusion, perceiving movement where there was none.

However, the study revealed a distinct limitation in current AI capabilities regarding attentional focus. Dr. Kiki Sanford, host of This Week in Science, explained the nuance of how the human brain processes these images versus the machine model.

"If you focus on one circle, it'll stop moving because you're focusing on it. And because a deep neural network has no way of focusing, it can't do that either... The visual system is not just optic nerve into your brain and then visual cortex. There’s the spatial activation of different sensory rods and cones... it sums at every level to become a more complex perception."

This suggests that while AI can mimic the "passive" perception of motion through predictive algorithms, it lacks the active, top-down attentional control that allows humans to override the illusion. The implication is that for AI to truly "see" like a human, it must move beyond simple input processing and develop mechanisms for active scanning and focal adjustment.

Optimizing Data for Machine Perception

As neural networks become more sophisticated, the hardware and data requirements to run them are under scrutiny. Contrary to the consumer trend toward 8K resolution, expert analysis suggests that effective computer vision models prioritize distinct patterns over raw pixel count. In practical applications, such as identifying athletes on a moving field, the "resolution ceiling" for AI is surprisingly low.

Andy Beach, a media technology strategist, noted that handing over excessive data to a model can be computationally wasteful.

"I haven't seen a case where the more richer data is somehow some additional benefit that it gets above and beyond. It just sort of uses what it needs and then it kind of ignores the rest... 1080p is actually totally fine. Even 720 will get you a lot of the way there unless it's just a really bad camera."

This distinction is vital for the development of real-time processing in autonomous systems. By understanding that AI—much like the human eye—relies on "good enough" resolution combined with complex processing, developers can optimize bandwidth and storage without sacrificing model accuracy.

Quantum Computing and the Future of Cognitive AI

The discussion extended beyond classical computing into the realm of quantum mechanics. Recent studies involving the "Necker Cube"—a wireframe cube that visually toggles between two orientations—indicate that human perception switches states in a manner analogous to quantum tunneling. Researchers at Charles Sturt University in Australia simulated this effect in a model, which was able to replicate the human timing of perceptual switching.

This raises the hypothesis that true Artificial General Intelligence (AGI) may require quantum computing to effectively simulate the synchronization of brain waves and perception. Current Large Language Models (LLMs) and vision models are essentially separate "centers" of a digital brain. Combining them into a cohesive, reasoning entity might necessitate hardware that operates on quantum probabilities rather than binary logic.

"You can just tell a quantum computer like 'hey, act like you're entangled' and it's like 'great, I am.' Whereas a classical computer is like 'great, give me five million lines of code and I can pretend I'm entangled.' ... It feels like until we hit true quantum compute, that isn't a thing we can even really tackle."

Beyond the technical architecture of AI, the volume of data generated by these systems presents a challenge for human consumption. The concept of a "news diet" has become increasingly relevant as algorithms—often driven by AI—curate information feeds that can lead to sampling bias and perceptual distortion. Much like a neural network requires clean data to avoid hallucinations, human cognition requires a balanced intake of information to maintain an objective worldview.

The rapid evolution of media technology, from Usenet groups to algorithmic social feeds, necessitates a higher degree of media literacy. Experts argue that triangulating sources and understanding the "backend" of how news is delivered is essential for navigating the modern digital ecosystem without succumbing to information overload.

Moving forward, the integration of distinct AI modules—vision, language, and reasoning—will likely depend on breakthroughs in computing architecture. As researchers continue to map the parallels between biological neural networks and synthetic models, the industry moves closer to creating systems that do not just process data, but perceive the world with human-like nuance.

Latest

The creator of Clawd: "I ship code I don't read"

The creator of Clawd: "I ship code I don't read"

Peter Steinberger, creator of Clawd, merges 600 commits daily using a fleet of AI agents. In this deep dive, discover how he challenges engineering norms by shipping code he doesn't read, treating PRs as "Prompt Requests," and replacing manual review with autonomous loops.

Members Public
The Clawdbot Craze | The Brainstorm EP 117

The Clawdbot Craze | The Brainstorm EP 117

The AI landscape is shifting to autonomous agents, led by the viral "Claudebot." As developers unlock persistent memory, OpenAI refines ad models, and Tesla hits new milestones, software intelligence meets real-world utility. Tune into The Brainstorm EP 117.

Members Public