In an era flooded with information, you might expect clarity to improve. Instead, as Roger Highfield argues in this thought-provoking Royal Society lecture, we are witnessing a paradox:
“We’ve got more signal—yet less clarity. More science communication—yet less confidence.”
This isn’t just a communication failure. It’s something far deeper—rooted in how our brains construct reality itself.
๐ The Paradox of Trust in Science
The lecture opens with striking statistics. Public trust in science remains relatively high—82% believe scientists contribute positively to society. Yet confidence in scientific information has declined:
- Belief that science information is “generally true”: 50% → 40% (2019–2025)
- People feeling informed about science: 51% → 43%
- Strong trust in science: 53% → 34%
This creates a troubling contradiction:
More access to science, but less confidence in it.
Highfield’s central question emerges:
๐ Why does increased exposure not translate into increased trust?
๐ง Reality as a “Controlled Hallucination”
To answer this, the lecture pivots inward—into the brain.
Drawing on ideas from Anil Seth, Highfield introduces a radical idea:
“Your perception of reality is a controlled hallucination.”
Rather than passively receiving information, the brain:
- Predicts what it expects to see
- Updates those predictions with sensory input
- Constructs a narrative
Reality, then, is not simply observed—it is actively built.
Even more striking:
“What we know as reality is when we all agree on our hallucinations.”
๐บ The Triangle That Thinks (But Doesn’t)
One of the lecture’s most memorable examples comes from a 1944 animation experiment. Participants watched simple geometric shapes moving around.
And yet…
People described “conflict,” “romance,” even “aggression” in the triangles.
A triangle becomes “angry.” Another becomes “territorial.”
Why? Because the brain:
- Detects motion without clear cause
- Infers intention
- Constructs a story
This tendency is evolutionarily useful—better to assume agency than miss a threat. But it also makes us vulnerable to false narratives.
๐ The Dress That Broke the Internet
Highfield revisits the viral phenomenon of “the dress”—blue/black vs white/gold.
The key insight:
- People saw different realities from the same data
Why?
Because their brains made different assumptions about lighting:
- Blue sky → subtract blue → white/gold
- Artificial light → subtract yellow → blue/black
This wasn’t disagreement. It was different perception itself.
๐งฉ Pattern-Seeking Gone Wrong
Humans are wired to find patterns—even when none exist. This shows up in:
- Seeing faces in clouds (pareidolia)
- Hearing words in noise
- Detecting “hidden truths” in random events
And crucially:
People who believe conspiracy theories are more likely to see illusory patterns.
This insight reframes misinformation—not just as ignorance, but as overactive pattern detection.
⚽ Tribal Brains: Beliefs as Identity
Beliefs aren’t just about truth—they signal belonging.
A striking experiment:
- Manchester United fans helped injured people more if they wore a United shirt than a Liverpool FC shirt.
The implication?
๐ We don’t just believe things because they’re true.
๐ We believe them because they align with our group.
This extends to science:
- Trust varies by political identity
- Influenced by leaders more than evidence
๐งช The Confirmation Bias Trap
Highfield demonstrates a classic cognitive bias with a simple puzzle:
Given the sequence: 2, 4, 8
People assume the rule is doubling.
But the actual rule?
“Each number must simply be larger than the previous one.”
The mistake:
- People test confirming examples (16, 32, 64)
- They rarely test disconfirming ones (3, 5, 7)
This is the essence of confirmation bias:
๐ We seek evidence that proves us right, not wrong.
๐ฌ Scientists Are Not Immune
In a refreshingly honest moment, Highfield turns the lens on science itself:
- P-hacking
- Cherry-picking results
- “Publish or perish” pressures
These contribute to the reproducibility crisis.
Science, he argues, works not because scientists are perfect—but because:
“The scientific method is designed to correct our collective irrationality.”
๐ค AI: Amplifying Our Weaknesses
The lecture’s most urgent section examines AI.
Highfield warns of a dangerous convergence:
- Human brains → optimized for survival, not truth
- AI systems → optimized for plausibility and engagement
The result?
“A perfect storm.”
Key dangers:
- AI hallucinations (“confabulations”)
- Fake studies, fake experts, fake images
- Error rates up to 73% in some summarization tasks
One chilling example:
- A fake disease (“Bixonia”) was invented
- AI systems later treated it as real
๐ฌ The “Forbidden Planet” Metaphor
Drawing from Forbidden Planet, Highfield offers a powerful analogy:
A machine amplifies a scientist’s unconscious mind—creating destructive illusions.
Similarly today:
- AI reflects and amplifies our biases
- Social media reinforces extreme beliefs
๐ The Social Media Effect
Modern platforms accelerate misinformation by:
- Connecting like-minded individuals instantly
- Reinforcing confirmation bias
- Nudging users toward more extreme views
“No matter how strange your belief, you can find a community that confirms it within minutes.”
๐งญ So What Can Be Done?
Highfield doesn’t end in pessimism. Instead, he outlines practical solutions:
1. Better Narratives (Not Just More Facts)
“Humans can resist bad stories, but only by encountering better ones.”
Science must:
- Tell compelling stories
- Without sacrificing rigor
2. “Pre-bunking” Misinformation
The “Bad News Game” trains users to create fake news themselves.
Result:
- Builds “cognitive antibodies”
- Helps recognize manipulation
3. Reforming Science Itself
Example: Pre-registration
- Before: 57% of studies reported positive effects
- After: only 8%
Less exciting—but more reliable science.
4. Shift to “Interpretation Literacy”
Instead of just teaching facts, teach:
- Uncertainty
- Probability
- Cognitive bias
“Audiences don’t lack data—they lack tools to evaluate narratives.”
5. Embrace Uncertainty
A key cultural shift:
๐ Science is not about certainty
๐ It is about managing uncertainty
๐ฌ Science as a Habit of Mind
Highfield’s most powerful message comes near the end:
“Science is not a set of facts—it’s a habit of mind.”
And the Royal Society’s motto captures it perfectly:
“Nullius in verba” — Take nobody’s word for it.
๐งฉ Final Reflection: The Real Problem Is Us
Perhaps the most sobering insight:
“The problem is not that we tell stories. The problem is when our stories tell us how to think.”
Misinformation isn’t just about bad actors or faulty technology.
It’s about:
- Our brains
- Our biases
- Our need to belong
⭐ Verdict: A Lecture That Lingers
This is not a comfortable lecture—but it’s an essential one.
What makes it powerful:
- Blends neuroscience, psychology, and AI
- Uses vivid examples (triangles, dresses, fake diseases)
- Turns critique inward, including toward science
What makes it memorable:
-
Its central inversion:
๐ The battle for reality is not “out there”
๐ It is inside our heads
๐ง Takeaway
In a world of infinite information and increasingly persuasive machines:
๐ Science matters not because it gives us answers
๐ But because it helps us question our own thinking
And that may be the only reliable path back to reality.
See the full video here:
No comments:
Post a Comment