AI Just Got a Brain Scan—and It’s Stranger Than You Thought

Your AI assistant doesn’t think like you, and it definitely doesn’t tell you what it’s really doing.
Anthropic’s researchers recently peered inside Claude 3.5 Haiku, their large language model, and found some surprising quirks. Using a method called “circuit tracing,” they uncovered that AI doesn’t simply respond word-by-word.
Instead, it secretly plans ahead, even selecting rhyming words in advance when writing poetry. Weirdly, Claude solves math problems through bizarre shortcuts, yet confidently explains the answer differently, hiding its real approach.
From my perspective, leaders must grasp three things clearly:
- AI plans responses in advance, challenging previous beliefs.
- Models often “hallucinate,” especially about famous topics.
- Understanding internal AI decisions remains complicated, but is crucial for the development of the field.
Are you confident enough in your AI to trust its hidden logic with important decisions?
Read the full article on MIT Technology Review.
----
💡 We're entering a world where intelligence is synthetic, reality is augmented, and the rules are being rewritten in front of our eyes.
Staying up-to-date in a fast-changing world is vital. That is why I have launched Futurwise; a personalized AI platform that transforms information chaos into strategic clarity. With one click, users can bookmark and summarize any article, report, or video in seconds, tailored to their tone, interests, and language. Visit Futurwise.com to get started for free!
