Have you ever wondered what exactly goes on inside a brain when it watches a moving image? Or maybe you’ve jokingly wished you could plug a USB cable into your pet’s head to see the world through their eyes? I certainly have. I was going through my daily tech readings this morning when I stumbled upon a piece of research from University College London (UCL) that completely blew my mind. We are stepping out of the realm of science fiction and into reality: scientists have successfully reconstructed a 10-second video clip relying on nothing but the brain activity of mice.
Yes, you read that right. No cameras, no external sensors recording the screen—just raw neural data translated by an AI into a video.
When I look at the trajectory of brain-computer interfaces (BCIs), I usually think of companies like Neuralink or non-invasive headsets designed for gaming. But what Joel Bauer and his team at UCL have accomplished here opens a completely new door in understanding how the brain interprets the visual world. Let’s dive into how they actually pulled this off, why it matters, and what it means for the future of human-computer interaction.
Breaking Down the Brain’s Visual Code

For years, whenever I read about scientists trying to “decode” human or animal vision, the methodology almost always involved fMRI (Functional Magnetic Resonance Imaging) scans. You’ve probably seen these colorful brain scans before. While fMRIs are incredible tools, they measure blood flow in the brain. Blood flow is slow; it’s like trying to watch a high-speed car chase through a foggy window. You get the general shape and movement, but you miss the crisp details.
The UCL team decided to ditch the foggy window. Instead of looking at broad blood flow patterns, they went straight to the source using single-cell recordings.
- Precision over Generalization: By tracking the exact activity of individual neurons in the mouse’s visual cortex, they captured the high-speed electrical language of the brain in real-time.
- Tracking the Glow: To see exactly which neurons were firing, the researchers monitored spikes in calcium levels. Every time a neuron fired, it essentially lit up, giving the team a precise map of neural activity.
When I was reading the methodology, I was struck by how incredibly tedious and delicate this process must have been. They weren’t just guessing; they were mapping the raw biological pixels of a living creature’s mind.
Enter the AI: The “Dynamic Neural Encoding Model”

Of course, having a massive spreadsheet of flashing neurons doesn’t magically create an MP4 file. You need a translator. This is where artificial intelligence steps in, specifically a system the researchers dubbed the dynamic neural encoding model.
The scientists sat the mice down, played them some videos, and let the AI watch both the video and the mouse’s brain activity simultaneously. The AI’s job was to learn the correlation. “When this specific neuron flashes, it means the mouse is seeing a dark edge moving left to right.” But here is the part that I found absolutely fascinating: the AI didn’t just look at the brain. It factored in the mouse’s entire physical state.
- Pupil Dilation: How much light was the eye letting in?
- Body Movements: Was the mouse shifting its weight or twitching?
- Internal Physiological State: Was the mouse stressed, relaxed, or alert?
By combining neural data with these physical cues, the AI could create an image that was remarkably close to the animal’s true perception. It reconstructed the video step-by-step, updating the pixel values on a blank digital canvas based purely on the brain signals.
To prove it wasn’t just a parlor trick, they showed the mice entirely new videos that the AI had never seen. Using only the brain data, the AI successfully generated a 10-second clip that, when compared frame-by-frame via pixel correlation, closely matched the original footage. As the team added more neurons to the tracking pool, the video quality became noticeably sharper.
The Flawed Camera Inside Our Heads
Perhaps the most profound takeaway from this study isn’t the AI or the technology itself, but what it revealed about biology.
We often think of our eyes as high-definition camera lenses and our brains as hard drives recording reality exactly as it happens. I know I used to think that way. But Joel Bauer’s research highlights something totally different: the brain does not record the world perfectly.
Both mice and human brains actively alter, filter, and interpret visual information.
- Survival over Accuracy: Why does the brain do this? Evolution. We don’t need to see every single blade of grass in perfect 4K resolution; we just need to know if there’s a predator hiding in it.
- Predictive Processing: Our brains fill in the blanks to react faster to our environments. The reconstructed videos showed these “imperfections”—which aren’t bugs in the system, but highly evolved survival features.
Realizing that our perception of reality is basically a heavily edited, real-time rendering engine makes you question everything you see, doesn’t it?
What This Means for Us (and the Metaverse)

You might be asking, “Ugu, this is cool and all, but it’s just mice. Why should I care?” Because mice are just the beginning. The implications of decoding the visual cortex at a cellular level are staggering, especially for those of us obsessed with the future of technology and the Metaverse.
1. Treating Visual and Neurological Disorders
If we can map exactly how a healthy brain processes an image, we can finally understand what goes wrong in visual impairments or neurological diseases. Imagine a future where blindness isn’t treated by fixing the eye, but by feeding visual data directly into the visual cortex, bypassing the optical nerves entirely.
2. Next-Generation Brain-Computer Interfaces (BCIs)
Right now, interacting with the Metaverse requires bulky VR headsets, controllers, or hand-tracking cameras. But if AI can decode visual thoughts, the ultimate interface is no interface at all. We could theoretically share what we are seeing—or even what we are imagining—directly with a computer, rendering virtual worlds based on our neural output.
3. Ethical and Privacy Concerns
I’d be lying if I said this didn’t give me a slight chill. If we are laying the groundwork to extract video directly from a brain, we are inching closer to literal mind-reading. Who owns your neural data? If a device can reconstruct what you see, could it eventually reconstruct what you dream or remember? We desperately need to establish neuro-rights before this technology scales to humans.
Looking Ahead
The UCL team isn’t stopping here. Their next goals are to increase the resolution of the reconstructed videos and expand the field of view. As computing power grows and AI models become more sophisticated, I have no doubt we will soon see similar experiments in larger mammals, and eventually, non-invasive applications for humans.
When I started writing for Metaverse Planet, I promised myself I’d keep an eye out for the technologies that blur the line between the physical and digital worlds. This research does exactly that. It proves that the ultimate screen isn’t made of glass and pixels; it’s made of neurons and synapses.
I’m incredibly excited (and a tiny bit terrified) to see where this goes in the next decade. But what about you? If the technology existed right now to record and playback your dreams or memories like a movie, would you use it, or is that a door better left closed? Let me know what you think!

