Back to The Alarming Rise of Stupidity Amplified
Chapter 2

Understanding Intelligence in the Age of AI

Audio Narration
Coming Soon

Ask a hundred people to define intelligence, and you'll get a hundred different answers. Is it the ability to solve complex mathematical problems? The capacity to adapt to new situations? Creativity? Emotional awareness? The truth is that we've spent centuries trying to define intelligence, and the arrival of artificial intelligence has only made the question more complicated—and more urgent.

Understanding what intelligence actually is becomes critical when we're building systems designed to amplify it. Because if we don't know what we're amplifying, how can we distinguish between enhancing human cognition and merely accelerating our mistakes? This chapter explores the nature of intelligence, how we measure it, what AI reveals about our assumptions, and why getting this right matters more than ever.

The IQ Trap

For over a century, we've tried to quantify intelligence with a single number: the Intelligence Quotient, or IQ. Developed in the early 1900s, IQ tests measure certain cognitive abilities—pattern recognition, mathematical reasoning, verbal comprehension, spatial awareness. And they do predict certain outcomes: academic success, problem-solving ability, even life expectancy to some degree.

But here's what IQ doesn't measure: wisdom, creativity, emotional intelligence, practical judgment, ethical reasoning, or the ability to distinguish truth from falsehood. You can have a genius-level IQ and still believe in conspiracy theories. You can ace every cognitive test and still make catastrophically poor decisions. You can be brilliant in your field and spectacularly ignorant about everything else.

The arrival of AI has exposed this limitation dramatically. Machines can now exceed human performance on virtually every traditional intelligence test, yet they lack common sense. They can solve complex mathematical equations but can't understand why it's inappropriate to discuss funeral arrangements at a wedding. They can beat world champions at chess and Go but don't understand the concept of "winning" or why anyone would care about a game.

"We built machines that can pass intelligence tests, only to discover that passing tests isn't the same as being intelligent."

Multiple Intelligences

Psychologist Howard Gardner proposed the theory of "multiple intelligences" in 1983, arguing that intelligence isn't a single trait but a collection of different capacities: linguistic, logical-mathematical, spatial, musical, bodily-kinesthetic, interpersonal, intrapersonal, and naturalistic. While this framework has critics, it captures an important truth: intelligence is multifaceted.

AI excels at some of these intelligences while completely lacking others. It can demonstrate extraordinary logical-mathematical intelligence, processing calculations that would take humans lifetimes. It can show remarkable linguistic intelligence, generating coherent and even creative text. But ask an AI about interpersonal dynamics, emotional nuance, or ethical dilemmas, and you quickly hit its limitations.

More importantly, AI has no integrated intelligence—no ability to combine these different forms of intelligence into a unified understanding of the world. A human child knows that pushing someone will make them fall, that falling hurts, that hurting people is wrong, and that their parents will be upset if they do wrong. An AI can understand each of these concepts separately but doesn't naturally integrate them into intuitive behavior.

Crystallized vs. Fluid Intelligence

Psychologists distinguish between two types of intelligence: crystallized intelligence—the accumulated knowledge and skills we acquire over time—and fluid intelligence—our ability to think logically and solve new problems independent of acquired knowledge.

AI has virtually unlimited crystallized intelligence. It can be trained on the sum total of human knowledge, accessing information instantly that would take humans years to learn. Need to know about medieval European history, quantum physics, and French cooking? AI has that knowledge available immediately.

But fluid intelligence—the ability to genuinely reason about novel situations, to apply common sense, to navigate ambiguity and uncertainty—remains challenging for AI. Humans excel at this. We can walk into a completely new situation and make reasonable judgments based on general principles, emotional cues, and contextual understanding. AI, despite all its knowledge, often struggles with basic common sense in novel contexts.

Intelligence vs. Understanding

Here's a crucial distinction that AI has forced us to confront: the difference between intelligence and understanding. An AI can produce remarkably intelligent-seeming responses without actually understanding anything. It can complete your sentences, write essays, explain complex topics—all while being, in a fundamental sense, unconscious.

This is what philosopher John Searle illustrated with his "Chinese Room" thought experiment. Imagine a person in a room with a book of instructions for manipulating Chinese symbols. They receive Chinese text, follow the instructions to produce appropriate Chinese responses, and appear to understand Chinese—yet they don't actually understand a word. They're performing symbol manipulation without comprehension.

Current AI systems are, in many ways, sophisticated Chinese Rooms. They process patterns, predict likely continuations, generate appropriate responses—without genuine understanding. And this matters tremendously when we're relying on these systems to amplify human intelligence, because they can amplify performance without amplifying wisdom.

"The most dangerous machine is one intelligent enough to be convincing, but not wise enough to be careful."

The Turing Test and Its Limits

Alan Turing proposed a test for machine intelligence in 1950: if a machine can converse with a human such that the human can't tell they're talking to a machine, then the machine should be considered intelligent. For decades, this seemed like a reasonable benchmark.

But we've now effectively passed the Turing Test with systems like GPT-4, and we're discovering that conversational ability isn't the same as intelligence. A system can generate human-like text without understanding meaning. It can produce explanations without grasping concepts. It can seem intelligent without being wise.

The Turing Test measures appearance of intelligence, not intelligence itself. And in an age where AI can convincingly fake expertise, where deepfakes can reproduce anyone's voice and face, where generated content is indistinguishable from human-created content, the ability to fake intelligence has become nearly as powerful as intelligence itself.

Artificial Intelligence vs. Artificial Stupidity

If AI can amplify intelligence, it can also amplify stupidity. And here's where things get interesting: AI doesn't distinguish between the two. It's a pattern-matching and pattern-generating system. Feed it intelligent patterns, and it produces intelligent outputs. Feed it stupid patterns, and it produces stupid outputs. Feed it biased data, and it amplifies those biases.

This is why AI trained on internet data can be simultaneously impressive and appalling. The internet contains humanity's knowledge and humanity's ignorance, our wisdom and our foolishness, our creativity and our prejudices. An AI trained on this data learns all of it indiscriminately.

Early chatbots learned racist language from Twitter. Image generators learned to associate "CEO" with men and "nurse" with women. Content recommendation algorithms learned to maximize engagement by promoting outrage and conspiracy theories. The AI wasn't malfunctioning—it was working exactly as designed, amplifying the patterns it found in its training data, stupid and intelligent alike.

Collective Intelligence and Collective Stupidity

Here's another dimension we must consider: collective intelligence. Humans individually are smart, but we become vastly more capable when we collaborate, share knowledge, build on each other's work. Science, technology, civilization itself—these are achievements of collective intelligence.

But we also exhibit collective stupidity: moral panics, mass delusions, financial bubbles, conspiracy movements. When misinformation spreads through a population, when confirmation bias becomes self-reinforcing in communities, when groupthink overrides individual skepticism, collective intelligence becomes collective stupidity.

AI amplifies both. It can facilitate collaboration, knowledge sharing, and problem-solving at unprecedented scales. But it can also accelerate the spread of misinformation, create echo chambers, and enable conspiracy theories to find and reinforce each other instantly across the globe. The tools we built to enhance collective intelligence are equally effective at amplifying collective stupidity.

What We're Really Amplifying

So what is intelligence in the age of AI? It's not just raw cognitive ability or accumulated knowledge. True intelligence in this era requires:

Critical thinking - The ability to evaluate information, identify biases, distinguish correlation from causation, and update beliefs based on evidence.

Epistemic humility - Recognizing the limits of one's knowledge, understanding when to defer to experts, and being willing to say "I don't know."

Contextual wisdom - Knowing not just what is true, but when information is relevant, how to apply it appropriately, and what its implications might be.

Ethical reasoning - Understanding not just what we can do with technology, but what we should do, and being able to navigate moral complexity.

These are precisely the forms of intelligence that AI struggles with, and they're becoming more critical as AI becomes more powerful. Because when you have tools that can generate any content, prove any claim, and create any reality you want, the only thing standing between wisdom and catastrophe is human judgment.

Understanding intelligence in the age of AI isn't just an academic exercise. It's essential for recognizing what we need to cultivate in ourselves and what we can safely delegate to machines. Because the alternative—treating AI as an intelligence we can blindly trust—is the fastest path to amplifying our stupidity rather than our wisdom.