Strategic Initiatives
11702 stories
·
45 followers

The Psychology of Generation Alpha | RealClearScience

1 Share
  • Generation Alpha overview: Authors Claire Hughes, George Gillett, and Luke McGuire from University of Cambridge and University of Exeter examine the psychological development of children born 2010-2025, the largest generation at two billion, amid global uncertainties and digital nativity.
  • Digital literacy and education: This generation achieves high digital literacy and is projected to have 90% secondary school completion worldwide, surpassing Gen Z's 80%.
  • Inheritance from millennial parents: Children receive values like inclusivity and adaptability from parents facing mental health issues and financial challenges, potentially fostering resilience through involved parenting.
  • Parental technology challenges: Millennial phone addiction and "technoference" link to child behavioral problems, while "sharenting" poses risks needing further research.
  • Digital immersion benefits and risks: Access to mental health resources and education aids development, but social media, cyber-bullying, gaming addiction, and harmful content increase exposure to peer pressure and misinformation.
  • Epistemic trust implications: Surrounded by misinformation, Gen Alpha may develop epistemic mistrust, linked to depression and anxiety.
  • Diversity of experiences: Pandemic exacerbated inequalities in education and safety; experiences vary by region, family size, gender, with boys prone to gaming issues and girls to social media pressures.
  • Progressive values: Children express concerns on inequality, climate justice, and animal rights, potentially influencing parents and society long-term.

Generation Alpha is the largest generation ever. Totalling two billion children, it encompasses anybody currently aged 0-15 years old – those born between 2010 and 2025.

This is the first fully digitally native generation, with many children already achieving unprecedented levels of digital literacy. It is predicted they will become the most educated generation in history: 90% are expected to complete secondary school worldwide, compared with 80% of gen Z.

However, gen Alpha children also inherit an increasingly precarious world, characterised by global uncertainties, housing shortages and climate change. It’s such a radical departure from what came before that this generation takes its name from the first letter of the Greek alphabet – unlike the gen Xs, Ys and Zs, whose letters come from our own (Latin) alphabet.

As developmental psychologists, we might ask what this new world means for gen Alpha’s psychological development. How might it influence their experiences, interests and values?

A good starting point is to look at their parents: the millennials (gen Y). From these 30- and 40-somethings, the gen Alphas are inheriting support for values such as inclusivity, adaptability and digital confidence, but also the tendency in some areas to boycott people whose words or actions are deemed offensive.

Millennials also report poorer mental health than previous generations, and at least in many western countries they are facing in some ways a more challenging financial situation.

Facing this adversity may mean that millennial parents are uniquely well placed to build their children’s resilience. Encouragingly, millennial fathers show unprecedented levels of commitment to being involved in parenting, challenging gender stereotypes.

On the other hand, stressed parents often struggle to cope with everyday tasks, while their children can exhibit behavioural problems such as aggression, and emotional difficulties like social withdrawal and anxiety.

Technology has also introduced challenges as millennials have widely adopted personal devices. Roughly half of parents in the UK say they are addicted to their phones, for example.

The distractions of digital devices can make parenting more difficult. This “technoference” is again associated with child behaviour problems, which could have knock-on effects later in life.

Kid trying to get his mum's attention when she's on her phone

Does this look familiar? DimaBerlin

One additional technological challenge that gen Alphas experience is “sharenting”, where parents share photos and videos of them online, sometimes in [excessive or inappropriate ways]. We lack good data on the scale and impact of this trend, so researchers urgently need to map the risks.

Millennials’ own challenges with technology may at least put them in a better position to help their children navigate things like screentime and social media. Today’s digital world also means that gen Alpha’s parents have a huge amount of parenting information at their fingertips, as well as access to parenting forums and digital communities, which may reduce feelings of stress and isolation. Today’s psychologists and health professionals can also provide parents with rigorous evidence-based guidance.

Gen Alpha digital immersion

The digital world has amplified gen Alpha’s exposure to ideas and cultural practices from peers and other adults. So while some are likely to spend longer living at home than previous generations, they may actually be less susceptible to parental influence. This could be liberating, but also introduces new risks.

We’ve seen how social media can exacerbate peer pressure and has introduced cyber-bullying. Influencer trends risk inculcating unrealistic body image ideals, which can contribute to body dysmorphia.

Excessive use of video games can result in sleep deprivation, reduced physical activity and impaired school performance. It may be that the increasing availability of VR games makes these risks more pronounced for gen Alpha.

Gen Alpha also risk being more exposed to potentially harmful content such as pornography or sites promoting self-harm or eating disorders. While they may not be as gullible as is often assumed, young children often lack effective strategies for identifying trustworthy information, putting them at more risk than adults.

In 2010, the cognitive anthropologist Dan Sperber coined the phrase “epistemic trust” to describe the idea that we view others as trustworthy sources of information. Growing up surrounded by misinformation may instead lead gen Alpha to adopt the opposite stance – epistemic mistrust – with potential negative consequences like depression and anxiety, for instance

Two kids playing a computer game

Games are designed to be addictive. Pixel-Shot

The digital space isn’t entirely negative for children’s psychological development. It’s much easier for gen Alpha to access information online about mental health, as well as professional psychological support, while also participating in virtual communities.

Technology also offers far greater access to educational resources and tools, something which AI may increase. There’s even emerging evidence that AI may enhance learning outcomes and increase student motivation, for instance. The big challenge will be to democratise learning without succumbing to adverse effects on human cognitive abilities from over-relying on AI for writing and thinking.

A diversity of experiences

Amid all this, it’s important to remember that gen Alpha’s psychological development is not following one monolithic trajectory.

The pandemic both highlighted and exacerbated inequality in many areas of childrens’s lives from education to home stability. For some older gen Alphas, school closures and online teaching left long-term scars – as evidenced by surging school absenteeism in many countries that appears difficult to reverse.

School closures during the pandemic meant the loss of a safe space for children who might be at risk of neglect or maltreatment at home. One consequence was a global rise in violence against children. Even younger Alphas who were not born at the time may be affected indirectly by their older brothers and sisters’ experiences.

More generally, experience will differ greatly depending on where members of gen Alpha live. For example, in east Asia ultra-low fertility has led to emptying classrooms, while children in smaller families endure more academic pressure from parents. Understanding the experiences of children in the global south is particularly important, not least because they represent the majority of gen Alpha.

Gender will also affect in new ways how this generation experience the world. For instance, boys appear more prone to gaming addiction while girls are more likely to become addicted to social media. Girls also report seeing more online content that creates appearance pressure, while boys are more likely to see misinformation – no doubt sometimes promoting potentially harmful views about masculinity and women.

What are gen Alpha’s own views of this new world? Despite adult fears about their exposure to misogyny, xenophobia and polarisation, today’s children are concerned about a wide range of progressive issues. These include resource inequality, sexuality, climate justice and animal rights.

This might be youthful idealism that will be shed in adulthood. Yet we know, for instance, that millennials commonly express values that differ from preceding generations.

Gen Alpha have also already been found to instil pro-environmental attitudes in their parents. So at least some of these progressive values are likely to endure and potentially ripple through to wider society.The Conversation

Claire Hughes, Professor of Psychology, Deputy Director of the Centre For Family Research, University of Cambridge; George Gillett, Clinical Research Fellow, University of Cambridge, and Luke McGuire, Senior Lecturer, Department of Psychology, University of Exeter

This article is republished from The Conversation under a Creative Commons license. Read the original article.

Read the whole story
bogorad
2 hours ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Fujifilm Opens ‘House of Photography’ Spaces in Barcelona and Porto

1 Share

A corner view of a Fujifilm House of Photography store with large windows, black awnings, and a stone facade on a quiet city street. Bollards line the sidewalk in front of the shop.

Fujifilm celebrated the opening of two new "House of Photography" spaces in Europe this week. The first was in Barcelona, Spain, on November 5, and the second in Porto, Portugal today, November 7.

[Read More]

Read the whole story
bogorad
3 hours ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Chatbots don’t judge! Customers prefer robots over humans when it comes to those ’um, you know’ purchases

1 Share
  • Study Overview: Jianna Jin, a marketing and consumer behavior scholar at Notre Dame, conducted studies with over 6,000 participants showing consumers prefer chatbots over humans for embarrassing purchases like acne cream, diarrhea medication, sex toys, and lubricant to avoid judgment in online shopping.
  • Diarrhea vs. Hay Fever Experiment: Participants imagined shopping for medications; over 80% preferred a nonhuman chatbot for diarrhea treatment, compared to just 9% for hay fever, due to chatbots lacking minds to judge or feel.
  • Humanlike vs. Machinelike Chatbots: In a study with 1,500 participants buying diarrhea pills online, willingness to interact was highest with clearly machinelike chatbots and lowest with human service reps or humanlike chatbots.
  • Backfire of Humanizing Chatbots: Making chatbots appear more human reduces effectiveness for embarrassing products, as it increases perceived judgment.
  • Business Implications: With 80% of retail and e-commerce businesses using or planning AI chatbots, companies should deploy machinelike bots for embarrassing purchases to improve service without defaulting to humanization.
  • Broader Applications: Findings may extend to stigmatized contexts like women seeking car repair quotes or men buying cosmetics, where machinelike chatbots could reduce discomfort.
  • Research Scope: Studies focused on everyday embarrassing items such as hemorrhoid cream, anti-wrinkle cream, personal lubricant, and adult toys.
  • Future Research Needs: Additional testing required to confirm benefits in broader stigmatized purchasing scenarios.

When it comes to inquiring about – ahem – certain products, shoppers prefer the inhuman touch.

That is what we found in a study of consumer habits when it comes to products that traditionally have come with a degree of embarrassment – think acne cream, diarrhea medication, adult sex toys or personal lubricant.

While brands may assume consumers hate chatbots, our series of studies involving more than 6,000 participants found a clear pattern: When it comes to purchases that make people feel embarrassed, consumers prefer chatbots over human service reps.

In one experiment, we asked participants to imagine shopping for medications for diarrhea and hay fever. They were offered two online pharmacies, one with a human pharmacist and the other with a chatbot pharmacist.

The medications were packaged identically, with the only difference being their labels for “diarrhea” or “hay fever.” More than 80% of consumers looking for diarrhea treatment preferred a store with a clearly nonhuman chatbot. In caparison, just 9% of those shopping for hay fever medication preferred nonhuman chatbots.

This is because, participants told us, they did not think chatbots have “minds” – that is, the ability to judge or feel.

In fact, when it comes to selling embarrassing products, making chatbots look or sound human can actually backfire. In another study, we asked 1,500 people to imagine buying diarrhea pills online. Participants were randomly assigned to one of three conditions: an online drugstore with a human service rep, the same store with a humanlike chatbot with a profile photo and name, or the same store with a chatbot that was clearly botlike in both its name and icon.

We then asked participants how likely they would be to seek help from the service agent. The results were clear: Willingness to interact dropped as the agent seemed more human. Interest peaked with the clearly machinelike chatbot and hit its lowest point with the human service rep.

Why it matters

As a scholar of marketing and consumer behavior, I know Chatbots play an increasingly large part in e-retail. In fact, one report found 80% of retail and e-commerce business use AI chatbots or plan to use them in the near future.

When it comes to chatbots, companies want to answer two questions: When should they deploy chatbots? And how should the chatbots be designed?

Many companies may assume the best strategy is to make bots look and sound more human, intuiting that consumers don’t want to talk to machines.

But our findings show the opposite can be true. In moments when embarrassment looms large, humanlike chatbots can backfire.

The practical takeaway is that brands should not default to humanizing their chatbots. Sometimes the most effective bot is the one that looks and sounds like a machine.

What still isn’t known

So far, we’ve looked at everyday purchases where embarrassment is easy to imagine, such as hemorrhoid cream, anti-wrinkle cream, personal lubricant and adult toys.

However, we believe the insights extend more broadly. For example, women getting a quote for car repair may be more self-conscious, as this is a purchase context where women have been traditionally more stigmatized. Similarly, men shopping for cosmetic products may feel judged in a category that has traditionally been marketed to women.

In contexts like these, companies could deploy chatbots – especially ones that clearly sound machinelike – to reduce discomfort and provide a better service. But more work is needed to test that hypothesis.

The Research Brief is a short take on interesting academic work.

Read the whole story
bogorad
10 hours ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Tesla Shareholders Approve Elon Musk’s $1 Trillion Pay Package - WSJ

1 Share
  • Tesla Shareholder Vote on Musk Pay: Elon Musk and Tesla shareholders approved a new compensation package potentially worth $1 trillion over 10 years, tied to ambitious milestones in market cap, vehicle sales, AI, and robotics; voted recently amid debates on Musk's leadership and focus shift to humanoid robots and AI, with Musk threatening to leave if rejected, at Tesla's annual meeting.
  • Package Structure: Divided into 12 tranches, starting with $2 trillion market cap plus one operational goal like $50 billion EBITDA, 20 million cars delivered, 10 million Full Self-Driving subscriptions, 1 million robotaxis, or 1 million robots sold; each tranche grants Musk equity equivalent to 1% of current shares.
  • Musk's Ownership Goal: Aims for 25% stake to control development of "robot army" and prevent misuse, while allowing board to fire him if needed; current stake around 15%, with outright ownership of 414 million shares plus contested 2018 award.
  • Board and Musk Endorsements: Tesla board calls it pay-for-performance to motivate innovation in autonomous vehicles, robotaxis, and humanoid robots; Chair Robyn Denholm highlights Musk's drive for unprecedented achievements.
  • Opposition and Abstentions: Opposed by proxy advisers like ISS, institutional investors including CalPERS, NYC systems, and Norges Bank; another proposal for Tesla investment in xAI passed but with many abstentions, board to decide next steps.
  • Company Performance Context: Tesla sales fell over 13% in first half of year amid Musk's time in Washington on Department of Government Efficiency and focus on xAI's Grok; post-May shift back to Tesla priorities.
  • Historical and Legal Background: Builds on 2018 package, record-valued but contested in Delaware court over board independence; Tesla appealing January 2024 ruling to rescind it.
  • Shareholder and Market Reactions: Charles Schwab voted in favor to align interests; package criticized as potentially overly expensive or demotivating if targets unattainable, with Musk urging stock retention amid visionary shift.

“What we’re about to embark upon is not merely a new chapter of the future of Tesla but a whole new book,” Musk said. “I guess what I’m saying is hang onto your Tesla stock,” he added later.

The measure was hotly debated, with some large shareholders taking opposing sides. The voting was largely seen as a referendum on the company’s longtime leader and his vision to shift Tesla’s focus to humanoid robots and artificial intelligence.

Musk, who is also CEO of SpaceX and xAI, had threatened on social media to leave Tesla if the measure had been rejected. He is already Tesla’s biggest shareholder, with a roughly 15% stake.

Musk had said he wanted a big enough ownership stake in Tesla to be comfortable that the “robot army” he was developing didn’t fall into the wrong hands, but not so large that he couldn’t be fired if he went “crazy.”

On another proposal that would authorize the Tesla board to invest in Musk’s artificial-intelligence company, xAI, Tesla General Counsel Brandon Ehrhart said more shares had been voted for the proposal than against, but there were many abstentions. He said the board would consider its next steps.

Musk had publicly endorsed the idea as he seeks to catch up in the AI race.

The new pay package, which includes 12 chunks of stock, could give Musk control over as much as 25% of Tesla if he hits a series of milestones and expands the company’s market capitalization to $8.5 trillion over the next 10 years. Its market cap is now around $1.5 trillion.

Tesla is proposing a $1 trillion pay package with 12 tranches...

$

$50 billion Ebitda

$2 trillion in market capitalization

$

20 million cars delivered

$

$

$

$

$

$

1 million robotaxis in operation

Plus any one of...

$

$

$

$

$

$

1 million robots sold

...for Musk to get the first tranche, Tesla must achieve...

10 million Full Self-Driving subscriptions

Tesla is proposing a $1 trillion pay package with 12 tranches...

$50 billion Ebitda

$2 trillion in market capitalization

$

$

20 million cars delivered

$

$

$

$

$

$

1 million robots sold

Plus any one of...

$

$

$

$

$

$

1 million robotaxis in operation

...for Musk to get the first tranche, Tesla must achieve...

10 million Full Self-Driving subscriptions

Tesla is proposing a $1 trillion pay package with 12 tranches...

$

$

$

$

$

$

$

$

$

$

$

$

...for Musk to get the first tranche, Tesla must achieve...

$

$2 trillion in market capitalization

Plus any one of...

$

1 million robots sold

$50 billion Ebitda

10 million Full Self-Driving subscriptions

1 million robotaxis in operation

20 million cars delivered

Tesla is proposing a $1 trillion pay package with 12 tranches...

$

$

$

$

$

$

$

$

$

$

$

$

...for Musk to get the first tranche, Tesla must achieve...

$2 trillion in market capitalization

$

Plus any one of...

$

1 million robots sold

$50 billion Ebitda

20 million cars delivered

1 million robotaxis in operation

10 million Full Self-Driving subscriptions

Tesla is proposing a $1 trillion pay package with 12 tranches...

$

$

$

$

$

$

$

$

$

$

$

$

...for Musk to get the first tranche, Tesla must achieve...

$2 trillion in market capitalization

$

Plus any one of...

$

1 million robots sold

$50 billion Ebitda

1 million robotaxis in operation

20 million cars delivered

10 million Full Self-Driving subscriptions

Tesla’s board described the package as pay for performance, designed to motivate Musk to transform the company with new products such as autonomous vehicles, robotaxis and humanoid robots.

“Having worked with him now for 11 years, I can say what motivates him is doing things that others can’t do or haven’t been able to do,” Tesla Chair Robyn Denholm said in an interview last week.

Tesla struggled to keep Musk’s attention earlier this year as he spent time in Washington running the Department of Government Efficiency. Tesla’s vehicle sales fell more than 13% in the first half of the year. After Musk left Washington in May, he turned his focus to his startup xAI and the development of its chatbot Grok, The Wall Street Journal reported.

The new pay package was opposed by several proxy advisers and institutional investors including the California Public Employees’ Retirement System, various New York City retirement systems, and Norges Bank Investment Management, which is the sixth-largest institutional shareholder with a 1.2% stake.

Institutional Shareholder Services, one of the proxy advisers that urged passive funds to vote down the compensation package, said it had concerns about the magnitude and design of the “astronomical” stock award.

Charles Schwab, which has a Tesla stake of about 0.6%, said Tuesday it would vote in favor of the package. “We firmly believe that supporting this proposal aligns both management and shareholder interests,” it said in a statement.

Huge stock awards tied to ambitious targets—sometimes called “moonshot” pay packages—are cast by proponents as a high-octane incentive for outstanding performance. Critics say they are often doubly flawed: overly expensive if targets prove easier than predicted; and counterproductive if the targets become unattainable and executives see little reason to stick around.

SHARE YOUR THOUGHTS

How would you vote on Musk’s Tesla pay package? Join the conversation below.

Musk’s new package is divided into 12 tranches. He could reach the first tranche if Tesla’s market cap grows to $2 trillion from around $1.5 trillion today, combined with an operational goal such as selling 11.5 million new vehicles, on top of the 8.5 million vehicles on the road.

More challenging milestones include selling one million robots to paying customers and maintaining an adjusted Ebitda of $400 billion. Last year, Tesla posted an adjusted Ebitda of $16 billion.

For each tranche he unlocks, Musk would receive equity equivalent to about 1% of Tesla’s current shares. Once he earns a tranche, he could vote those shares but wouldn’t be able to sell them until they vest, in either 7.5 years or 10 years.

Musk’s 2018 pay package, the most valuable on record before the 2025 package, is tied up in a dispute at the Delaware Supreme Court. Tesla is appealing a lower-court decision to rescind the 2018 pay package after a judge ruled in January 2024 that Tesla’s directors were beholden to Musk and the approval process for that package was tainted and lacked transparency.

Here is a breakdown of Musk’s current Tesla ownership:

Owns outright

414 million shares

2018 award

304 million shares

Proposed 2025 award

424 million shares

35.3

million

shares

August interim award

96 million shares

Contested in court

If Musk loses in court, he could lose these options

September purchase

Across 12 tranches

2018 award

304 million shares

Proposed 2025 award

424 million shares

Owns outright

414 million shares

35.3

million

shares

August interim award

96 million

Contested in court

If Musk loses in court, he could lose these options

September purchase

Across 12 tranches

2018 award

304 million shares

Proposed 2025 award

424 million shares

Owns outright

414 million shares

35.3

million

shares

August interim award

96 million

Contested in court

If Musk loses in court, he could lose these options

September purchase

Across 12 tranches

Owns outright

414 million shares

September

purchase

2018 award

304 million shares

August interim award

96 million

If Musk loses in court, he could lose these options

Contested in court

Proposed 2025 award

424 million shares (across 12 tranches)

35.3

million

shares

Owns outright

414 million shares

September

purchase

2018 award

304 million shares

August interim award

96 million

If Musk loses in court, he could lose these options

Contested in court

Proposed 2025 award

424 million shares (across 12 tranches)

35.3

million

shares

Write to Becky Peterson at becky.peterson@wsj.com

Read the whole story
bogorad
12 hours ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Why AC is cheap, but AC repair is a luxury - by Alex Danco

1 Share
  • Overview of economic paradoxes in AI era: A16z post by tech investors explores Jevons Paradox and Baumol Effect in the United States today, explaining why AI-driven productivity surges make some services cheaper while others, like handyman work, become relatively more expensive due to labor market dynamics.
  • Initial example of cost disparity: Repairing a drywall hole costs more than buying a flatscreen TV to cover it, highlighting inefficiencies in non-tech sectors amid tech productivity booms.
  • Jevons Paradox defined: Productivity gains in one area, like coal or computing via Moore's Law, lead to increased consumption and job creation as costs drop dramatically, expanding demand beyond savings.
  • AI application of Jevons: AI investments, similar to past tech surges, will make certain goods and services exponentially cheaper, enabling 10x more consumption and new opportunities in affected industries.
  • Baumol Effect explained: Sectors without productivity gains, like string quartets or basic services, see wages and prices rise due to competition from high-paying tech jobs, spreading wealth across the economy.
  • Interconnection of effects: Jevons-driven productivity booms are necessary for Baumol's wage pressures, resulting in overall societal wealth increase where even stagnant services become affordable despite higher costs.
  • Impact on specific jobs: AI will automate parts of roles, making some services cheaper, but human-required elements, like oversight in radiology or dog walking, will command premium wages due to regulatory protections.
  • Future economic weirdness: Within jobs, the "last 1% human bottleneck" could inflate wages until fully automated, leading to strange labor market shifts and potential regulatory interventions in an AI-advanced economy.

[

](https://substackcdn.com/image/fetch/$s_!tVX5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b998a3-e1aa-497d-8017-754fe7dd9c56_730x150.jpeg)

If you live in the United States today, and you accidentally knock a hole in your wall, it’s probably cheaper to buy a flatscreen TV and stick it in front of the hole, compared to hiring a handyman to fix your drywall. (Source: Marc Andreessen.) This seems insane; why?

Well, weird things happen to economies when you have huge bursts of productivity that are concentrated in one industry. Obviously, it’s great for that industry, because when the cost of something falls while its quality rises, we usually find a way to consume way more of that thing - creating a huge number of new jobs and new opportunities in this newly productive area.

But there’s an interesting spillover effect. The more jobs and opportunities created by the productivity boom, the more wages increase in other industries, who at the end of the day all have to compete in the same labor market. If you can make $30 an hour as a digital freelance marketer (a job that did not exist a generation ago), then you won’t accept less than that from working in food service. And if you can make $150 an hour installing HVAC for data centers, you’re not going to accept less from doing home AC service.

[

](https://substackcdn.com/image/fetch/$s_!h0YV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd33722c4-af5d-4f47-8a8b-2a06bc36c3e1_948x860.png)

This is a funny juxtaposition. Each of these phenomena have a name: there’s Jevons Paradox, which means, “We’ll spend more on what gets more productive”, and there’s the Baumol Effect, which means, “We’ll spend more on what doesn’t get more productive.” And both of them are top of mind right now, as we watch in awe at what is happening with AI Capex spend.

As today’s AI supercycle plays out, just like in productivity surges of past decades, we’re likely going to see something really interesting happen:

  • Some goods and services, where AI has relatively more impact and we’re able to consume 10x more of them along some dimension, will become orders of magnitude cheaper.

  • Other goods and services, where AI has relatively less impact, will become more expensive - and we’ll consume more of them anyway.

And, even weirder, we may see this effect happen within a single job:

  • Some parts of the job, automated by AI, will see 10x throughput at 10x the quality, while

  • Other parts of the job - the part that must be done by the human - will be the reason you’re getting paid, command a wildly high wage, and be the target of regulatory protection.

Let’s dive in:

Jevons: Productivity gains that grow the pie

Chances are, you’ve probably seen a version of this graph at some point:

[

](https://substackcdn.com/image/fetch/$s_!dfCp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98c7320-2c6f-4491-a426-68a35d93d300_1309x1600.png)

This graph can mean different things to different people: it can mean “what’s regulated versus what isn’t” to some, “where technology makes a difference” to others. And it’s top of mind these days, as persistent inflation and the AI investment supercycle both command a lot of mindshare.

To really understand it, the best place to start isn’t with the red lines. It’s with the blue lines: where are things getting cheaper, in ways that create more jobs, more opportunity, and more spending?

The original formulation of “Jevons paradox”, by William Stanley Jevons in 1865, was about coal production. Jevons observed that, the cheaper and faster we got at producing coal, the more coal we ended up using - demand more than eclipsed the cost savings, and the coal market grew rapidly as it fed the Second Industrial Revolution in England and abroad.

Subscribe for more from a16z every weekday:

Today we all know Moore’s Law, the best contemporary example of Jevons paradox. In 1965, a transistor cost roughly $1. Today it costs a fraction of a millionth of a cent. This extraordinary collapse in computing costs – a billionfold improvement – did not lead to modest, proportional increases in computer use. It triggered an explosion of applications that would have been unthinkable at earlier price points. At $1 per transistor, computers made sense for military calculations and corporate payroll. At a thousandth of a cent, they made sense for word processing and databases. At a millionth of a cent, they made sense in thermostats and greeting cards. At a billionth of a cent, we embed them in disposable shipping tags that transmit their location once and are thrown away. The efficiency gains haven’t reduced our total computing consumption: they’ve made computing so cheap that we now use trillions times more of it.

[

](https://substackcdn.com/image/fetch/$s_!XtfD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50a544d8-177d-4878-93eb-1e8f5b059ecf_1347x1600.png)

We’re all betting that the same will happen with the cost of tokens, just like it happened to the cost of computing, which in turn unlocks more demand than can possibly be taken up by the existing investment. The other week, we heard from Amin Vahdat, GP and GM of AI and Infrastructure at Google Cloud, share an astonishing observation with us: that 7 year old TPUs were still seeing 100% utilization inside Google. That is one of the things you see with Jevons Paradox: the opportunity to do productive work explodes in possibility. We are at the point in the technology curve with AI where every day someone figures out something new to do with them, meaning users will take any chip they can get, and use it productively.

Jevons Paradox (which isn’t really a paradox at all; it’s just economics) is where demand creation comes from, and where new kinds of attractive jobs come from. And that huge new supply of viable, productive opportunity is our starting point to understand the other half of our economic puzzle: what happens everywhere else.

Baumol’s: how the wealth gets spread around

Agatha Christie once wrote that she never thought she’d be wealthy enough to own a car, or poor enough to not have servants. Whereas, after a century of productivity gains, the average American middle-class household can comfortably manage a new car lease every two years, but needs to split the cost of a single nanny with their neighbors.

How did this happen? 100 years after Jevons published his observation on coal, William Baumol published a short paper investigating why so many orchestras, theaters, and opera companies were running out of money. He provocatively asserted that the String Quartet had become less productive, in “real economy” terms, because the rest of the economy had become more productive, while the musicians’ job stayed exactly the same. The paper struck a nerve, and became a branded concept: “Baumol’s Cost Disease”.

This is a tricky concept to wrap your head around, and not everyone buys it. But the basic argument is, over the long run all jobs and wage scales compete in the labor market with every other job and wage scale. If one sector becomes hugely productive, and creates tons of well-paying jobs, then every other sector’s wages eventually have to rise, in order for their jobs to remain attractive for anyone.

The String Quartet is an odd choice of example, because there are so many ways to argue that music has become more productive over the past century: recorded and streaming music have brought consumption costs down to near zero, and you could argue that Taylor Swift is “higher quality” for what today’s audiences are looking for (even if you deplore the aesthetics.) But the overall effect is compelling nonetheless. As some sectors of the economy get more attractive, the comparatively less attractive ones get more expensive anyway.

Once you’ve heard of Baumol’s, it’s like you get to join a trendy club of economic thinkers who now have someone to blame for all of society’s problems. It gets to be a successful punching bag for why labor markets are weird, or why basic services cost so much - “It’s a rich country problem.”

But the odd thing about Baumol’s is how rarely it is juxtaposed with the actual driving cause of those productivity distortions, which is the massive increase in productivity, in overall wealth, and in overall consumption, that’s required for Baumol’s to kick in. In a weird way, Jevons is necessary for Baumol’s to happen.

[

](https://substackcdn.com/image/fetch/$s_!B3Jd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa627f078-5b43-411f-8621-df0e2dc8b18f_730x330.gif)

For some reason, we rarely see those two phenomena juxtaposed against each other, but they’re related. For the Baumol Effect to take place as classically presented, there must be a total increase in productive output and opportunity; not just a relative increase in productivity, from the booming industry and the new jobs that it creates. But when that does happen, and we see a lot of consumption, job opportunities, and prosperity get created by the boom, you can safely bet that Baumol’s will manifest itself in faraway corners of the economy. This isn’t all bad; it’s how wealth gets spread around and how a rising tide lifts many boats. (There’s probably a joke here somewhere that Baumol’s Cost Disease is actually the most effective form of Communism ever tried, or something.)

Why it costs $100 a week to walk your dog (but you can afford it)

So, to recap:

  • “Jevons-type effects” created bountiful new opportunity in everything that got more productive; and

  • “Baumol-type effects” means that everything that didn’t get more productive got more expensive anyway, but we consume more of it all the same because society as a whole got richer.

As explained in one job: our explosion of demand for data centres means there’s infinite work for HVAC technicians. So they get paid more (even though they themselves didn’t change), which means they charge more on all jobs (even the ones that have nothing to do with AI), but we can afford to pay them (because we got richer overall, mostly from technology improvements, over the long run). Furthermore, the next generation of plumber apprentices might decide to do HVAC instead; so now plumbing is more expensive too. And so on.

Now let’s think about what’s going to happen with widespread AI adoption, if it pays off the way we all think it will. First of all, it’s going to drive a lot of productivity gains in services specifically. (There is precedent for this; e.g. the railroads made the mail a lot more productive; the internet made travel booking a lot more productive.) Some services are going to get pulled into the Jevons vortex, and just rapidly start getting more productive, and unlocking new use cases for those services. (The key is to look for elastic-demand services, where we plausibly could consume 10x or more of the service, along some dimension. Legal services, for example, plausibly fit this bill.)

And then there are other kinds of services that are not going to be Jevons’ed, for some reason or another, and for those services, over time, we should expect to see wildly high prices for specific services that have no real reason to AI whatsoever. Your dog walker has nothing to do with AI infrastructure; and yet, he will cost more. But you’ll pay it anyway; if you love your dog.

Reflexive Turbo-Baumol’s: why jobs will get weird

The last piece of this economic riddle, which we haven’t mentioned thus far, is that elected governments (who appoint and direct employment regulators) often believe they have a mandate to protect people’s employment and livelihoods. And the straightforward way that mandate gets applied, in the face of technological changes, is to protect human jobs by saying, “This safety function must be performed or signed off by a human.”

When this happens (which it certainly will, across who knows how many industries, we’ll see a Baumol’s type effect take hold within single jobs. Here’s Dwarkesh, on his recent interview with Andrej Karpathy: (Excerpted in full, because it’s such an interesting thought):

“With radiologists, I’m totally speculating and I have no idea what the actual workflow of a radiologist involves. But one analogy that might be applicable is when Waymos were first being rolled out, there’d be a person sitting in the front seat, and you just had to have them there to make sure that if something went really wrong, they’d be there to monitor. Even today, people are still watching to make sure things are going well. Robotaxi, which was just deployed, still has a person inside it.

Now we could be in a similar situation where if you automate 99% of a job, that last 1% the human has to do is incredibly valuable because it’s bottlenecking everything else. If it were the case with radiologists, where the person sitting in the front of Waymo has to be specially trained for years in order to provide the last 1%, their wages should go up tremendously because they’re the one thing bottlenecking wide deployment. Radiologists, I think their wages have gone up for similar reasons, if you’re the last bottleneck and you’re not fungible. A Waymo driver might be fungible with others. So you might see this thing where your wages go up until you get to 99% and then fall just like that when the last 1% is gone. And I wonder if we’re seeing similar things with radiology or salaries of call center workers or anything like that.”

Just like we have really weird economies in advanced countries (where we can afford supercomputers in our pockets, but not enough teachers for small class sizes), we could see a strange thing happen where the last 1% that must be a human in a job (the “Dog Walker” part, as opposed to the “Excel” part) becomes the essential employable skillset.

In an interesting way, this hints at where Baumol’s will finally run out of steam - because at some point, these “last 1% employable skills” no longer become substitutable for one another. They’ll become strange vestigial limbs of career paths; in a sense. We have a ways to go until we get there, but we can anticipate some very strange economic & political alliances that could get formed in such a world. Until then, let’s keep busy on the productivity part. Because that’s what matters, and what makes us a wealthy society - weird consequences and all.


Views expressed in “posts” (including podcasts, videos, and social media) are those of the individual a16z personnel quoted therein and are not the views of a16z Capital Management, L.L.C. (“a16z”) or its respective affiliates. a16z Capital Management is an investment adviser registered with the Securities and Exchange Commission. Registration as an investment adviser does not imply any special skill or training. The posts are not directed to any investors or potential investors, and do not constitute an offer to sell — or a solicitation of an offer to buy — any securities, and may not be used or relied upon in evaluating the merits of any investment.

The contents in here — and available on any associated distribution platforms and any public a16z online social media accounts, platforms, and sites (collectively, “content distribution outlets”) — should not be construed as or relied upon in any manner as investment, legal, tax, or other advice. You should consult your own advisers as to legal, business, tax, and other related matters concerning any investment. Any projections, estimates, forecasts, targets, prospects and/or opinions expressed in these materials are subject to change without notice and may differ or be contrary to opinions expressed by others. Any charts provided here or on a16z content distribution outlets are for informational purposes only, and should not be relied upon when making any investment decision. Certain information contained in here has been obtained from third-party sources, including from portfolio companies of funds managed by a16z. While taken from sources believed to be reliable, a16z has not independently verified such information and makes no representations about the enduring accuracy of the information or its appropriateness for a given situation. In addition, posts may include third-party advertisements; a16z has not reviewed such advertisements and does not endorse any advertising content contained therein. All content speaks only as of the date indicated.

Under no circumstances should any posts or other information provided on this website — or on associated content distribution outlets — be construed as an offer soliciting the purchase or sale of any security or interest in any pooled investment vehicle sponsored, discussed, or mentioned by a16z personnel. Nor should it be construed as an offer to provide investment advisory services; an offer to invest in an a16z-managed pooled investment vehicle will be made separately and only by means of the confidential offering documents of the specific pooled investment vehicles — which should be read in their entirety, and only to those who, among other requirements, meet certain qualifications under federal securities laws. Such investors, defined as accredited investors and qualified purchasers, are generally deemed capable of evaluating the merits and risks of prospective investments and financial matters.

There can be no assurances that a16z’s investment objectives will be achieved or investment strategies will be successful. Any investment in a vehicle managed by a16z involves a high degree of risk including the risk that the entire amount invested is lost. Any investments or portfolio companies mentioned, referred to, or described are not representative of all investments in vehicles managed by a16z and there can be no assurance that the investments will be profitable or that other investments made in the future will have similar characteristics or results. A list of investments made by funds managed by a16z is available here: https://a16z.com/investments/. Past results of a16z’s investments, pooled investment vehicles, or investment strategies are not necessarily indicative of future results. Excluded from this list are investments (and certain publicly traded cryptocurrencies/ digital assets) for which the issuer has not provided permission for a16z to disclose publicly. As for its investments in any cryptocurrency or token project, a16z is acting in its own financial interest, not necessarily in the interests of other token holders. a16z has no special role in any of these projects or power over their management. a16z does not undertake to continue to have any involvement in these projects other than as an investor and token holder, and other token holders should not expect that it will or rely on it to have any particular involvement.

With respect to funds managed by a16z that are registered in Japan, a16z will provide to any member of the Japanese public a copy of such documents as are required to be made publicly available pursuant to Article 63 of the Financial Instruments and Exchange Act of Japan. Please contact compliance@a16z.com to request such documents.

For other site terms of use, please go here. Additional important information about a16z, including our Form ADV Part 2A Brochure, is available at the SEC’s website: http://www.adviserinfo.sec.gov

Read the whole story
bogorad
12 hours ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Introducing Nested Learning: A new ML paradigm for continual learning

1 Share
  • Nested Learning Introduction: Google Research team including Ali Behrouz and Vahab Mirrokni announced Nested Learning on November 7, 2025, via their blog, to address catastrophic forgetting in machine learning by treating models as nested optimization problems.
  • Continual Learning Challenge: Current ML models, especially large language models, face limitations in acquiring new knowledge without losing old proficiency, unlike human brain neuroplasticity.
  • Nested Learning Framework: Views ML models as interconnected multi-level optimization problems with distinct context flows and update rates, unifying architecture and optimization.
  • Associative Memory Modeling: Backpropagation and attention mechanisms in transformers modeled as associative memory modules that map data to errors or tokens.
  • Brain Inspiration: Incorporates multi-time-scale updates similar to brain waves and neuroplasticity for continual learning in nested components.
  • Deep Optimizers: Reformulates optimizers like momentum using loss metrics such as L2 regression to improve resilience against imperfect data.
  • Continuum Memory System: Extends transformer memory into a spectrum of modules updating at varying frequencies for better long-term and short-term handling.
  • Hope Architecture Results: Proof-of-concept Hope model, a self-modifying variant of Titans, outperforms baselines in language modeling, reasoning, and long-context tasks per experiments.

Introducing Nested Learning: A new ML paradigm for continual learning

November 7, 2025

Ali Behrouz, Student Researcher, and Vahab Mirrokni, VP and Google Fellow, Google Research

We introduce Nested Learning, a new approach to machine learning that views models as a set of smaller, nested optimization problems, each with its own internal workflow, in order to mitigate or even completely avoid the issue of “catastrophic forgetting”, where learning new tasks sacrifices proficiency on old tasks.

Quick links

The last decade has seen incredible progress in machine learning (ML), primarily driven by powerful neural network architectures and the algorithms used to train them. However, despite the success of large language models (LLMs), a few fundamental challenges persist, especially around continual learning, the ability for a model to actively acquire new knowledge and skills over time without forgetting old ones.

When it comes to continual learning and self-improvement, the human brain is the gold standard. It adapts through neuroplasticity — the remarkable capacity to change its structure in response to new experiences, memories, and learning. Without this ability, a person is limited to immediate context (like anterograde amnesia). We see a similar limitation in current LLMs: their knowledge is confined to either the immediate context of their input window or the static information that they learn during pre-training.

The simple approach, continually updating a model's parameters with new data, often leads to “catastrophic forgetting” (CF), where learning new tasks sacrifices proficiency on old tasks. Researchers traditionally combat CF through architectural tweaks or better optimization rules. However, for too long, we have treated the model's architecture (the network structure) and the optimization algorithm (the training rule) as two separate things, which prevents us from achieving a truly unified, efficient learning system.

In our paper, “Nested Learning: The Illusion of Deep Learning Architectures”, published at NeurIPS 2025, we introduce Nested Learning, which bridges this gap. Nested Learning treats a single ML model not as one continuous process, but as a system of interconnected, multi-level learning problems that are optimized simultaneously. We argue that the model's architecture and the rules used to train it (i.e., the optimization algorithm) are fundamentally the same concepts; they are just different "levels" of optimization, each with its own internal flow of information ("context flow") and update rate. By recognizing this inherent structure, Nested Learning provides a new, previously invisible dimension for designing more capable AI, allowing us to build learning components with deeper computational depth, which ultimately helps solve issues like catastrophic forgetting.

We test and validate Nested Learning through a proof-of-concept, self-modifying architecture that we call “Hope”, which achieves superior performance in language modeling and demonstrates better long-context memory management than existing state-of-the-art models.

The Nested Learning paradigm

Nested Learning reveals that a complex ML model is actually a set of coherent, interconnected optimization problems nested within each other or running in parallel. Each of these internal problems has its own context flow — its own distinct set of information from which it is trying to learn.

This perspective implies that existing deep learning methods work by essentially compressing their internal context flows. More importantly, Nested Learning reveals a new dimension for designing models, allowing us to build learning components with deeper computational depth.

To illustrate this paradigm, we look at the concept of associative memory — the ability to map and recall one thing based on another (like recalling a name when you see a face).

  • We show that the training process itself, specifically the backpropagation process, can be modeled as an associative memory. The model learns to map a given data point to the value of its local error, which serves as a measure of how "surprising" or unexpected that data point was.
  • Similarly, following previous studies (e.g., Miras), key architectural components, such as the attention mechanism in transformers, can also be formalized as simple associative memory modules that learn the mapping between tokens in a sequence.

Diagram comparing biological brain waves and neuroplasticity to the uniform structure and multi-frequency updates used in Nested Learning models.

The uniform and reusable structure as well as multi-time–scale update in the brain are the key components of continual learning in humans. Nested Learning allows for multi-time–scale updates for each component of the brain, while showing that well-known architectures such as transformers and memory modules are in fact linear layers with different frequency updates.

By defining an update frequency rate, i.e., how often each component's weights are adjusted, we can order these interconnected optimization problems into "levels." This ordered set forms the heart of the Nested Learning paradigm.

Putting Nested Learning to work

The Nested Learning perspective immediately gives us principled ways to improve existing algorithms and architectures:

Deep optimizers

Since Nested Learning views optimizers (e.g., momentum-based optimizers) as associative memory modules, it allows us to apply principles from associative memory perspective to them. We observed that many standard optimizers rely on simple dot-product similarity (a measure of how alike two vectors are by calculating the sum of the products of their corresponding components) whose update doesn't account for how different data samples relate to each other. By changing the underlying objective of the optimizer to a more standard loss metric, such as L2 regression loss (a common loss function in regression tasks that quantifies the error by summing the squares of the differences between predicted and true values), we derive new formulations for core concepts like momentum, making them more resilient to imperfect data.

Continuum memory systems

In a standard Transformer, the sequence model acts as a short-term memory, holding the immediate context, while the feedforward neural networks act as long-term memory, storing pre-training knowledge. The Nested Learning paradigm extends this concept into what we call a “continuum memory system” (CMS), where memory is seen as a spectrum of modules, each updating at a different, specific frequency rate. This creates a much richer and more effective memory system for continual learning.

Hope: A self-modifying architecture with continuum memory

As a proof-of-concept, we used Nested Learning principles to design Hope, a variant of the Titans architecture. Titans architectures are long-term memory modules that prioritize memories based on how surprising they are. Despite their powerful memory management, they only have two levels of parameters update, resulting in a first-order in-context learning. Hope, however, is a self-modifying recurrent architecture that can take advantage of unbounded levels of in-context learning and also is augmented with CMS blocks to scale to larger context windows. It can essentially optimize its own memory through a self-referential process, creating an architecture with infinite, looped learning levels.

Experiments

We conducted experiments to evaluate the effectiveness of our deep optimizers and the performance of Hope on language modeling, long-context reasoning, continual learning, and knowledge incorporation tasks. The full results are available in our paper.

Results

Our experiments confirm the power of Nested Learning, the design of continuum memory systems, and self-modifying Titans.

On a diverse set of commonly used and public language modeling and common-sense reasoning tasks, the Hope architecture demonstrates lower perplexity and higher accuracy compared to modern recurrent models and standard transformers.

Bar chart that shows the Hope model outperforming Titans, Samba, and Transformer on both language modeling and common-sense reasoning performance metrics.

Comparison of performance on language modeling (perplexity; left) and common-sense reasoning (accuracy; right) tasks between different architectures: Hope, Titans, Samba and a baseline Transformer.

Hope showcases superior memory management in long-context Needle-In-Haystack (NIAH) downstream tasks, proving that the CMSs offer a more efficient and effective way to handle extended sequences of information.

Bar chart showing Hope and Titans models consistently outperforming TTT and Mamba2 across long-context tasks of three difficulty levels.

Performance comparison on long-context tasks with different levels of difficulty between different architectures: Hope, Titans, TTT, and Mamba2. NIAH-PK, NIAH-H, and NIAH-W are needle-in-a-haystack tasks with pass-key, number, and word, respectively.

Conclusion

The Nested Learning paradigm represents a step forward in our understanding of deep learning. By treating architecture and optimization as a single, coherent system of nested optimization problems, we unlock a new dimension for design, stacking multiple levels. The resulting models, like the Hope architecture, show that a principled approach to unifying these elements can lead to more expressive, capable, and efficient learning algorithms.

We believe the Nested Learning paradigm offers a robust foundation for closing the gap between the limited, forgetting nature of current LLMs and the remarkable continual learning abilities of the human brain. We are excited for the research community to explore this new dimension and help us build the next generation of self-improving AI.

Acknowledgements

This research was conducted by Ali Behrouz, Meisam Razaviyayn, Peilin Zhong, and Vahab Mirrokni. We thank Praneeth Kacham and Corinna Cortes for reviewing the work and their valuable suggestions. We also thank Yuan Deng and Zeman Li. Finally, we thank Mark Simborg and Kimberly Schwede for their help in crafting this blog post.

Labels:* Algorithms & Theory

Quick links

Other posts of interest

  • [

    November 6, 2025

    DS-STAR: A state-of-the-art versatile data science agent

    • Data Mining & Modeling ·
    • Machine Intelligence ·
    • Natural Language Processing

    ](/blog/ds-star-a-state-of-the-art-versatile-data-science-agent/)

  • [

    October 31, 2025

    Accelerating the magic cycle of research breakthroughs and real-world applications

    • Climate & Sustainability ·
    • Generative AI ·
    • Health & Bioscience ·
    • Quantum

    ](/blog/accelerating-the-magic-cycle-of-research-breakthroughs-and-real-world-applications/)

  • [

    October 30, 2025

    Toward provably private insights into AI use

    • Generative AI ·
    • Mobile Systems ·
    • Security, Privacy and Abuse Prevention ·
    • Software Systems & Engineering

    ](/blog/toward-provably-private-insights-into-ai-use/)

Read the whole story
bogorad
12 hours ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete
Next Page of Stories