- Experimental Design: Researchers compared human performance, AI performance, and human-AI hybrid teams in forecasting real-world events.
- Predictive Accuracy: Large AI models outperformed humans working in isolation, though human-AI hybrids displayed the highest potential for total accuracy.
- Hybrid Pitfalls: Many hybrid users relied on AI for direct answers, leading to poor outcomes characterized by confirmation bias and sycophancy.
- Collaborative Cyborgs: A small subset of users treated AI as a sparring partner to interrogate assumptions and challenge AI-generated assertions.
- Cognitive Requirements: Successful integration requires perspective-taking and intellectual humility rather than simple reliance on technological convenience.
- Information Exploration Paradox: High volumes of easily accessible information may reduce critical thinking and individual exploration, potentially leading to human skill atrophy.
- Strategic Reframe: AI should be utilized to search for what is missing in one's own logic rather than as a tool to automate routine labor.
- Developmental Necessity: Cultivating cognitive resistance to AI-generated easy answers is essential for maintaining human agency and intellectual rigor.
By
Vivienne Ming
April 24, 2026 2:00 pm ET
EDMON DE HARO FOR WSJ; ROBOT ARM, FIREFLY
Who’s smarter, the human or the machine?
In the 30 years I’ve worked in artificial intelligence that’s been the question driving the conversation.
We’ve also been sold a story about AI that goes something like this: It will handle the tedious, routine work—the research, the first draft, the number-crunching—while we focus on the interesting parts: creativity, judgment, the human touch.
My research suggests we’ve been asking the wrong question and drawing the wrong conclusions.
A few months ago, I recruited adults from San Francisco’s Bay Area for an experiment. I gave each group one hour to make predictions about real-world events, using scenarios drawn from the prediction market platform Polymarket. This provided us a rigorous, objective way to check results against the collective wisdom of thousands of financially motivated forecasters. In addition to AI making predictions on its own, some human teams worked alone, while others worked as human-AI hybrids. (Polymarket has a data partnership with Dow Jones, the publisher of The Wall Street Journal).
The human groups performed poorly, relying on instinct or whatever information had come across their feeds that morning. The large AI models—ChatGPT and Gemini, in this case—performed considerably better, though still short of the market itself.
But when we combined AI with humans, things got more interesting.
Most hybrid teams used AI for the answer and submitted it as their own, performing no better than the AI alone. Others fed their own predictions into AI and asked it to come up with supporting evidence. These “validators” had stumbled into a classic confirmation bias-loop: the sycophancy that leads chatbots to tell you what you want to hear, even if it isn’t true. They ended up performing worse than an AI working solo.
But in roughly 5% to 10% of teams, something different emerged. The AI became a sparring partner. The teams pushed back, demanding evidence and interrogating assumptions. When the AI expressed high confidence, the humans questioned it. When the humans felt strongly about an intuition, they asked the AI to come up with a counterargument.
The hybrids were becoming cyborgs.
These teams reached insightful conclusions that neither a human nor a machine could have produced on its own. They were the only group to consistently rival the prediction market’s accuracy. On certain questions, they even outperformed it.
It’s not that these people were more intelligent than the others in the study. But they demonstrated two important qualities: perspective-taking and intellectual humility.
Perspective-taking is the ability to genuinely inhabit another point of view. Not to debate it, not to tolerate it, but to actually inhabit it. Intellectual humility is the ability to recognize the edge of your own knowledge and sit with that discomfort rather than trying to rush to fill it.
Both of these qualities are, at root, emotional skills. Perspective-taking requires genuine curiosity about minds other than your own. Intellectual humility requires a kind of emotional courage: the willingness to feel uncertain, even a little foolish, in the presence of something or someone that seems very sure of itself.
These are not the soft skills we typically celebrate. We celebrate confidence. We promote decisiveness. We are building AI systems specifically designed to give us the answer before we feel the discomfort of not having it.
What my experiment suggests is that the human qualities most likely to matter are not the feel-good ones. They’re the uncomfortable ones: the capacity to be wrong in public and stay curious; to sit with a question your phone could answer in three seconds and resist the urge to reach for it. To read a confident, fluent response from an AI and ask yourself, “What’s missing?” rather than default to “Great, that’s done.” To disagree with something that sounds authoritative and to trust your instinct enough to follow it.
We don’t build these capacities by avoiding discomfort. We build them by choosing it, repeatedly, in small ways: the student who struggles through a problem before checking the answer; the person who asks a follow-up question in a conversation; the reader who sits with a difficult idea long enough for it to actually change one’s mind. Most AI chatbots today default to easy answers, which is hurting our ability to think critically
I call this the Information-Exploration Paradox. As the cost of information approaches zero, human exploration collapses. We see it in students who perform better on AI-assisted tasks and worse on everything afterward. We see it in developers shipping more code and understanding it less. We are, in ways that feel like progress, slowly optimizing ourselves out of the loop.
This is the divergence I worry about. Not the dramatic science-fiction scenario of AI replacing humans wholesale, but the quieter process of people gradually outsourcing their judgment in increments too small to notice.
Over time, this produces two different kinds of people: Those who use AI as a genuine intellectual partner—whose thinking actually gets sharper through the friction of the collaboration—and those who get better at securing quick answers and worse at knowing what questions to ask.
So what can any of us actually do about it?
Start with the reframe: The goal of working with AI isn’t to get the answer faster. It’s to find out what you’re missing. Don’t deploy AI minions to “do the boring work” for you, as so many sales pitches argue; use it as a savant collaborator to explore uncertainty.
In practice, that means before you accept an AI’s answer, ask it for the strongest argument against itself. When it hedges or qualifies, pay attention—that’s usually where the real uncertainty lives. Treat it like a brilliant colleague who has read everything and understands nothing—useful precisely because they’re different from you, not because they’ll agree with you.
For the AI industry, a key design question has gone largely unasked: Is the product building human capacity or consuming it? Nearly all AI benchmarks measure what AI agents can do alone. We desperately need benchmarks for hybrid intelligence. Errors are signals our brains use to trigger learning. An AI that eliminates friction entirely is often eliminating the learning along with it.
A hopeful finding is that perspective-taking, intellectual humility and curiosity are not fixed traits. They can be cultivated and respond to practice, the right relationships and environments that reward uncertainty.
But they require us to decide—as individuals, as parents, as educators, as designers of tools—that this is what we’re trying to build. And in the race between human potential and human atrophy, the stakes for building it could not be higher.
Vivienne Ming is a theoretical neuroscientist, cognitive scientist and the author of “Robot-Proof: When Machines Have All The Answers, Build Better People.”
Copyright ©2026 Dow Jones & Company, Inc. All Rights Reserved. 87990cbe856818d5eddac44c7b1cdeb8
Appeared in the April 25, 2026, print edition as 'Is AI Smarter Than People? It’s Complicated.'.
Up Next
[
17 Gifts for Every Type of Mom in Your Life
](https://www.wsj.com/lifestyle/mothers-day-gift-guide-4d335d10?mod=WTRN_pos1)
[
What to give when a bouquet doesn’t feel like enough.
](https://www.wsj.com/lifestyle/mothers-day-gift-guide-4d335d10?mod=WTRN_pos1)
[
Is Masculinity in Crisis? Readers Are Divided
](https://www.wsj.com/arts-culture/television/male-masculinity-crisis-tv-b91a1231?mod=WTRN_pos2)
[
An article about television shows grappling with the state of men today elicited a range of opinions.
](https://www.wsj.com/arts-culture/television/male-masculinity-crisis-tv-b91a1231?mod=WTRN_pos2)
[
I Snagged a $550 Business-Class Ticket to Italy. Then the Airline Found Out.
[
The Turkish Airlines deal was a mistake fare that the carrier can either honor or not. Mine was canceled.
[
My Son Was Killed by Hamas. The Pain Isn’t Getting Better.
](https://www.wsj.com/arts-culture/books/rachel-goldberg-polin-hersh-book-de1f1913?mod=WTRN_pos5)
[
In an exclusive book excerpt, Rachel Goldberg-Polin writes about mourning her son, Hersh, nearly two years after his death.
](https://www.wsj.com/arts-culture/books/rachel-goldberg-polin-hersh-book-de1f1913?mod=WTRN_pos5)
[
In Defense of Tween Screen Time
[
Every generation panics about girlhood. The latest, over the negative consequences of social media for tweens, hides a bigger problem in girl world: They’ve got nothing else.
[
Injectable Peptides Are the Latest TikTok Wellness Fad. Doctors Are Worried.
[
Regulators may soon lift restrictions on making the compounds despite scant evidence of their safety or efficacy.
[
Intel’s stock extends its spectacular run by posting its best day in nearly four decades
[
There’s a debate on Wall Street about whether Intel’s financial prospects have materially changed.
[
Gina Rodriguez’s Home in ‘Jane the Virgin’ Hits the Market in Los Angeles for $2.9 Million
[
The Spanish-style residence served as the family home of the titular character, Jane Villanueva, in the CW series




