Strategic Initiatives
11698 stories
·
45 followers

Why AC is cheap, but AC repair is a luxury - by Alex Danco

1 Share
  • Overview of economic paradoxes in AI era: A16z post by tech investors explores Jevons Paradox and Baumol Effect in the United States today, explaining why AI-driven productivity surges make some services cheaper while others, like handyman work, become relatively more expensive due to labor market dynamics.
  • Initial example of cost disparity: Repairing a drywall hole costs more than buying a flatscreen TV to cover it, highlighting inefficiencies in non-tech sectors amid tech productivity booms.
  • Jevons Paradox defined: Productivity gains in one area, like coal or computing via Moore's Law, lead to increased consumption and job creation as costs drop dramatically, expanding demand beyond savings.
  • AI application of Jevons: AI investments, similar to past tech surges, will make certain goods and services exponentially cheaper, enabling 10x more consumption and new opportunities in affected industries.
  • Baumol Effect explained: Sectors without productivity gains, like string quartets or basic services, see wages and prices rise due to competition from high-paying tech jobs, spreading wealth across the economy.
  • Interconnection of effects: Jevons-driven productivity booms are necessary for Baumol's wage pressures, resulting in overall societal wealth increase where even stagnant services become affordable despite higher costs.
  • Impact on specific jobs: AI will automate parts of roles, making some services cheaper, but human-required elements, like oversight in radiology or dog walking, will command premium wages due to regulatory protections.
  • Future economic weirdness: Within jobs, the "last 1% human bottleneck" could inflate wages until fully automated, leading to strange labor market shifts and potential regulatory interventions in an AI-advanced economy.

[

](https://substackcdn.com/image/fetch/$s_!tVX5!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fe2b998a3-e1aa-497d-8017-754fe7dd9c56_730x150.jpeg)

If you live in the United States today, and you accidentally knock a hole in your wall, it’s probably cheaper to buy a flatscreen TV and stick it in front of the hole, compared to hiring a handyman to fix your drywall. (Source: Marc Andreessen.) This seems insane; why?

Well, weird things happen to economies when you have huge bursts of productivity that are concentrated in one industry. Obviously, it’s great for that industry, because when the cost of something falls while its quality rises, we usually find a way to consume way more of that thing - creating a huge number of new jobs and new opportunities in this newly productive area.

But there’s an interesting spillover effect. The more jobs and opportunities created by the productivity boom, the more wages increase in other industries, who at the end of the day all have to compete in the same labor market. If you can make $30 an hour as a digital freelance marketer (a job that did not exist a generation ago), then you won’t accept less than that from working in food service. And if you can make $150 an hour installing HVAC for data centers, you’re not going to accept less from doing home AC service.

[

](https://substackcdn.com/image/fetch/$s_!h0YV!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd33722c4-af5d-4f47-8a8b-2a06bc36c3e1_948x860.png)

This is a funny juxtaposition. Each of these phenomena have a name: there’s Jevons Paradox, which means, “We’ll spend more on what gets more productive”, and there’s the Baumol Effect, which means, “We’ll spend more on what doesn’t get more productive.” And both of them are top of mind right now, as we watch in awe at what is happening with AI Capex spend.

As today’s AI supercycle plays out, just like in productivity surges of past decades, we’re likely going to see something really interesting happen:

  • Some goods and services, where AI has relatively more impact and we’re able to consume 10x more of them along some dimension, will become orders of magnitude cheaper.

  • Other goods and services, where AI has relatively less impact, will become more expensive - and we’ll consume more of them anyway.

And, even weirder, we may see this effect happen within a single job:

  • Some parts of the job, automated by AI, will see 10x throughput at 10x the quality, while

  • Other parts of the job - the part that must be done by the human - will be the reason you’re getting paid, command a wildly high wage, and be the target of regulatory protection.

Let’s dive in:

Jevons: Productivity gains that grow the pie

Chances are, you’ve probably seen a version of this graph at some point:

[

](https://substackcdn.com/image/fetch/$s_!dfCp!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd98c7320-2c6f-4491-a426-68a35d93d300_1309x1600.png)

This graph can mean different things to different people: it can mean “what’s regulated versus what isn’t” to some, “where technology makes a difference” to others. And it’s top of mind these days, as persistent inflation and the AI investment supercycle both command a lot of mindshare.

To really understand it, the best place to start isn’t with the red lines. It’s with the blue lines: where are things getting cheaper, in ways that create more jobs, more opportunity, and more spending?

The original formulation of “Jevons paradox”, by William Stanley Jevons in 1865, was about coal production. Jevons observed that, the cheaper and faster we got at producing coal, the more coal we ended up using - demand more than eclipsed the cost savings, and the coal market grew rapidly as it fed the Second Industrial Revolution in England and abroad.

Subscribe for more from a16z every weekday:

Today we all know Moore’s Law, the best contemporary example of Jevons paradox. In 1965, a transistor cost roughly $1. Today it costs a fraction of a millionth of a cent. This extraordinary collapse in computing costs – a billionfold improvement – did not lead to modest, proportional increases in computer use. It triggered an explosion of applications that would have been unthinkable at earlier price points. At $1 per transistor, computers made sense for military calculations and corporate payroll. At a thousandth of a cent, they made sense for word processing and databases. At a millionth of a cent, they made sense in thermostats and greeting cards. At a billionth of a cent, we embed them in disposable shipping tags that transmit their location once and are thrown away. The efficiency gains haven’t reduced our total computing consumption: they’ve made computing so cheap that we now use trillions times more of it.

[

](https://substackcdn.com/image/fetch/$s_!XtfD!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F50a544d8-177d-4878-93eb-1e8f5b059ecf_1347x1600.png)

We’re all betting that the same will happen with the cost of tokens, just like it happened to the cost of computing, which in turn unlocks more demand than can possibly be taken up by the existing investment. The other week, we heard from Amin Vahdat, GP and GM of AI and Infrastructure at Google Cloud, share an astonishing observation with us: that 7 year old TPUs were still seeing 100% utilization inside Google. That is one of the things you see with Jevons Paradox: the opportunity to do productive work explodes in possibility. We are at the point in the technology curve with AI where every day someone figures out something new to do with them, meaning users will take any chip they can get, and use it productively.

Jevons Paradox (which isn’t really a paradox at all; it’s just economics) is where demand creation comes from, and where new kinds of attractive jobs come from. And that huge new supply of viable, productive opportunity is our starting point to understand the other half of our economic puzzle: what happens everywhere else.

Baumol’s: how the wealth gets spread around

Agatha Christie once wrote that she never thought she’d be wealthy enough to own a car, or poor enough to not have servants. Whereas, after a century of productivity gains, the average American middle-class household can comfortably manage a new car lease every two years, but needs to split the cost of a single nanny with their neighbors.

How did this happen? 100 years after Jevons published his observation on coal, William Baumol published a short paper investigating why so many orchestras, theaters, and opera companies were running out of money. He provocatively asserted that the String Quartet had become less productive, in “real economy” terms, because the rest of the economy had become more productive, while the musicians’ job stayed exactly the same. The paper struck a nerve, and became a branded concept: “Baumol’s Cost Disease”.

This is a tricky concept to wrap your head around, and not everyone buys it. But the basic argument is, over the long run all jobs and wage scales compete in the labor market with every other job and wage scale. If one sector becomes hugely productive, and creates tons of well-paying jobs, then every other sector’s wages eventually have to rise, in order for their jobs to remain attractive for anyone.

The String Quartet is an odd choice of example, because there are so many ways to argue that music has become more productive over the past century: recorded and streaming music have brought consumption costs down to near zero, and you could argue that Taylor Swift is “higher quality” for what today’s audiences are looking for (even if you deplore the aesthetics.) But the overall effect is compelling nonetheless. As some sectors of the economy get more attractive, the comparatively less attractive ones get more expensive anyway.

Once you’ve heard of Baumol’s, it’s like you get to join a trendy club of economic thinkers who now have someone to blame for all of society’s problems. It gets to be a successful punching bag for why labor markets are weird, or why basic services cost so much - “It’s a rich country problem.”

But the odd thing about Baumol’s is how rarely it is juxtaposed with the actual driving cause of those productivity distortions, which is the massive increase in productivity, in overall wealth, and in overall consumption, that’s required for Baumol’s to kick in. In a weird way, Jevons is necessary for Baumol’s to happen.

[

](https://substackcdn.com/image/fetch/$s_!B3Jd!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fa627f078-5b43-411f-8621-df0e2dc8b18f_730x330.gif)

For some reason, we rarely see those two phenomena juxtaposed against each other, but they’re related. For the Baumol Effect to take place as classically presented, there must be a total increase in productive output and opportunity; not just a relative increase in productivity, from the booming industry and the new jobs that it creates. But when that does happen, and we see a lot of consumption, job opportunities, and prosperity get created by the boom, you can safely bet that Baumol’s will manifest itself in faraway corners of the economy. This isn’t all bad; it’s how wealth gets spread around and how a rising tide lifts many boats. (There’s probably a joke here somewhere that Baumol’s Cost Disease is actually the most effective form of Communism ever tried, or something.)

Why it costs $100 a week to walk your dog (but you can afford it)

So, to recap:

  • “Jevons-type effects” created bountiful new opportunity in everything that got more productive; and

  • “Baumol-type effects” means that everything that didn’t get more productive got more expensive anyway, but we consume more of it all the same because society as a whole got richer.

As explained in one job: our explosion of demand for data centres means there’s infinite work for HVAC technicians. So they get paid more (even though they themselves didn’t change), which means they charge more on all jobs (even the ones that have nothing to do with AI), but we can afford to pay them (because we got richer overall, mostly from technology improvements, over the long run). Furthermore, the next generation of plumber apprentices might decide to do HVAC instead; so now plumbing is more expensive too. And so on.

Now let’s think about what’s going to happen with widespread AI adoption, if it pays off the way we all think it will. First of all, it’s going to drive a lot of productivity gains in services specifically. (There is precedent for this; e.g. the railroads made the mail a lot more productive; the internet made travel booking a lot more productive.) Some services are going to get pulled into the Jevons vortex, and just rapidly start getting more productive, and unlocking new use cases for those services. (The key is to look for elastic-demand services, where we plausibly could consume 10x or more of the service, along some dimension. Legal services, for example, plausibly fit this bill.)

And then there are other kinds of services that are not going to be Jevons’ed, for some reason or another, and for those services, over time, we should expect to see wildly high prices for specific services that have no real reason to AI whatsoever. Your dog walker has nothing to do with AI infrastructure; and yet, he will cost more. But you’ll pay it anyway; if you love your dog.

Reflexive Turbo-Baumol’s: why jobs will get weird

The last piece of this economic riddle, which we haven’t mentioned thus far, is that elected governments (who appoint and direct employment regulators) often believe they have a mandate to protect people’s employment and livelihoods. And the straightforward way that mandate gets applied, in the face of technological changes, is to protect human jobs by saying, “This safety function must be performed or signed off by a human.”

When this happens (which it certainly will, across who knows how many industries, we’ll see a Baumol’s type effect take hold within single jobs. Here’s Dwarkesh, on his recent interview with Andrej Karpathy: (Excerpted in full, because it’s such an interesting thought):

“With radiologists, I’m totally speculating and I have no idea what the actual workflow of a radiologist involves. But one analogy that might be applicable is when Waymos were first being rolled out, there’d be a person sitting in the front seat, and you just had to have them there to make sure that if something went really wrong, they’d be there to monitor. Even today, people are still watching to make sure things are going well. Robotaxi, which was just deployed, still has a person inside it.

Now we could be in a similar situation where if you automate 99% of a job, that last 1% the human has to do is incredibly valuable because it’s bottlenecking everything else. If it were the case with radiologists, where the person sitting in the front of Waymo has to be specially trained for years in order to provide the last 1%, their wages should go up tremendously because they’re the one thing bottlenecking wide deployment. Radiologists, I think their wages have gone up for similar reasons, if you’re the last bottleneck and you’re not fungible. A Waymo driver might be fungible with others. So you might see this thing where your wages go up until you get to 99% and then fall just like that when the last 1% is gone. And I wonder if we’re seeing similar things with radiology or salaries of call center workers or anything like that.”

Just like we have really weird economies in advanced countries (where we can afford supercomputers in our pockets, but not enough teachers for small class sizes), we could see a strange thing happen where the last 1% that must be a human in a job (the “Dog Walker” part, as opposed to the “Excel” part) becomes the essential employable skillset.

In an interesting way, this hints at where Baumol’s will finally run out of steam - because at some point, these “last 1% employable skills” no longer become substitutable for one another. They’ll become strange vestigial limbs of career paths; in a sense. We have a ways to go until we get there, but we can anticipate some very strange economic & political alliances that could get formed in such a world. Until then, let’s keep busy on the productivity part. Because that’s what matters, and what makes us a wealthy society - weird consequences and all.


Views expressed in “posts” (including podcasts, videos, and social media) are those of the individual a16z personnel quoted therein and are not the views of a16z Capital Management, L.L.C. (“a16z”) or its respective affiliates. a16z Capital Management is an investment adviser registered with the Securities and Exchange Commission. Registration as an investment adviser does not imply any special skill or training. The posts are not directed to any investors or potential investors, and do not constitute an offer to sell — or a solicitation of an offer to buy — any securities, and may not be used or relied upon in evaluating the merits of any investment.

The contents in here — and available on any associated distribution platforms and any public a16z online social media accounts, platforms, and sites (collectively, “content distribution outlets”) — should not be construed as or relied upon in any manner as investment, legal, tax, or other advice. You should consult your own advisers as to legal, business, tax, and other related matters concerning any investment. Any projections, estimates, forecasts, targets, prospects and/or opinions expressed in these materials are subject to change without notice and may differ or be contrary to opinions expressed by others. Any charts provided here or on a16z content distribution outlets are for informational purposes only, and should not be relied upon when making any investment decision. Certain information contained in here has been obtained from third-party sources, including from portfolio companies of funds managed by a16z. While taken from sources believed to be reliable, a16z has not independently verified such information and makes no representations about the enduring accuracy of the information or its appropriateness for a given situation. In addition, posts may include third-party advertisements; a16z has not reviewed such advertisements and does not endorse any advertising content contained therein. All content speaks only as of the date indicated.

Under no circumstances should any posts or other information provided on this website — or on associated content distribution outlets — be construed as an offer soliciting the purchase or sale of any security or interest in any pooled investment vehicle sponsored, discussed, or mentioned by a16z personnel. Nor should it be construed as an offer to provide investment advisory services; an offer to invest in an a16z-managed pooled investment vehicle will be made separately and only by means of the confidential offering documents of the specific pooled investment vehicles — which should be read in their entirety, and only to those who, among other requirements, meet certain qualifications under federal securities laws. Such investors, defined as accredited investors and qualified purchasers, are generally deemed capable of evaluating the merits and risks of prospective investments and financial matters.

There can be no assurances that a16z’s investment objectives will be achieved or investment strategies will be successful. Any investment in a vehicle managed by a16z involves a high degree of risk including the risk that the entire amount invested is lost. Any investments or portfolio companies mentioned, referred to, or described are not representative of all investments in vehicles managed by a16z and there can be no assurance that the investments will be profitable or that other investments made in the future will have similar characteristics or results. A list of investments made by funds managed by a16z is available here: https://a16z.com/investments/. Past results of a16z’s investments, pooled investment vehicles, or investment strategies are not necessarily indicative of future results. Excluded from this list are investments (and certain publicly traded cryptocurrencies/ digital assets) for which the issuer has not provided permission for a16z to disclose publicly. As for its investments in any cryptocurrency or token project, a16z is acting in its own financial interest, not necessarily in the interests of other token holders. a16z has no special role in any of these projects or power over their management. a16z does not undertake to continue to have any involvement in these projects other than as an investor and token holder, and other token holders should not expect that it will or rely on it to have any particular involvement.

With respect to funds managed by a16z that are registered in Japan, a16z will provide to any member of the Japanese public a copy of such documents as are required to be made publicly available pursuant to Article 63 of the Financial Instruments and Exchange Act of Japan. Please contact compliance@a16z.com to request such documents.

For other site terms of use, please go here. Additional important information about a16z, including our Form ADV Part 2A Brochure, is available at the SEC’s website: http://www.adviserinfo.sec.gov

Read the whole story
bogorad
11 minutes ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Introducing Nested Learning: A new ML paradigm for continual learning

1 Share
  • Nested Learning Introduction: Google Research team including Ali Behrouz and Vahab Mirrokni announced Nested Learning on November 7, 2025, via their blog, to address catastrophic forgetting in machine learning by treating models as nested optimization problems.
  • Continual Learning Challenge: Current ML models, especially large language models, face limitations in acquiring new knowledge without losing old proficiency, unlike human brain neuroplasticity.
  • Nested Learning Framework: Views ML models as interconnected multi-level optimization problems with distinct context flows and update rates, unifying architecture and optimization.
  • Associative Memory Modeling: Backpropagation and attention mechanisms in transformers modeled as associative memory modules that map data to errors or tokens.
  • Brain Inspiration: Incorporates multi-time-scale updates similar to brain waves and neuroplasticity for continual learning in nested components.
  • Deep Optimizers: Reformulates optimizers like momentum using loss metrics such as L2 regression to improve resilience against imperfect data.
  • Continuum Memory System: Extends transformer memory into a spectrum of modules updating at varying frequencies for better long-term and short-term handling.
  • Hope Architecture Results: Proof-of-concept Hope model, a self-modifying variant of Titans, outperforms baselines in language modeling, reasoning, and long-context tasks per experiments.

Introducing Nested Learning: A new ML paradigm for continual learning

November 7, 2025

Ali Behrouz, Student Researcher, and Vahab Mirrokni, VP and Google Fellow, Google Research

We introduce Nested Learning, a new approach to machine learning that views models as a set of smaller, nested optimization problems, each with its own internal workflow, in order to mitigate or even completely avoid the issue of “catastrophic forgetting”, where learning new tasks sacrifices proficiency on old tasks.

Quick links

The last decade has seen incredible progress in machine learning (ML), primarily driven by powerful neural network architectures and the algorithms used to train them. However, despite the success of large language models (LLMs), a few fundamental challenges persist, especially around continual learning, the ability for a model to actively acquire new knowledge and skills over time without forgetting old ones.

When it comes to continual learning and self-improvement, the human brain is the gold standard. It adapts through neuroplasticity — the remarkable capacity to change its structure in response to new experiences, memories, and learning. Without this ability, a person is limited to immediate context (like anterograde amnesia). We see a similar limitation in current LLMs: their knowledge is confined to either the immediate context of their input window or the static information that they learn during pre-training.

The simple approach, continually updating a model's parameters with new data, often leads to “catastrophic forgetting” (CF), where learning new tasks sacrifices proficiency on old tasks. Researchers traditionally combat CF through architectural tweaks or better optimization rules. However, for too long, we have treated the model's architecture (the network structure) and the optimization algorithm (the training rule) as two separate things, which prevents us from achieving a truly unified, efficient learning system.

In our paper, “Nested Learning: The Illusion of Deep Learning Architectures”, published at NeurIPS 2025, we introduce Nested Learning, which bridges this gap. Nested Learning treats a single ML model not as one continuous process, but as a system of interconnected, multi-level learning problems that are optimized simultaneously. We argue that the model's architecture and the rules used to train it (i.e., the optimization algorithm) are fundamentally the same concepts; they are just different "levels" of optimization, each with its own internal flow of information ("context flow") and update rate. By recognizing this inherent structure, Nested Learning provides a new, previously invisible dimension for designing more capable AI, allowing us to build learning components with deeper computational depth, which ultimately helps solve issues like catastrophic forgetting.

We test and validate Nested Learning through a proof-of-concept, self-modifying architecture that we call “Hope”, which achieves superior performance in language modeling and demonstrates better long-context memory management than existing state-of-the-art models.

The Nested Learning paradigm

Nested Learning reveals that a complex ML model is actually a set of coherent, interconnected optimization problems nested within each other or running in parallel. Each of these internal problems has its own context flow — its own distinct set of information from which it is trying to learn.

This perspective implies that existing deep learning methods work by essentially compressing their internal context flows. More importantly, Nested Learning reveals a new dimension for designing models, allowing us to build learning components with deeper computational depth.

To illustrate this paradigm, we look at the concept of associative memory — the ability to map and recall one thing based on another (like recalling a name when you see a face).

  • We show that the training process itself, specifically the backpropagation process, can be modeled as an associative memory. The model learns to map a given data point to the value of its local error, which serves as a measure of how "surprising" or unexpected that data point was.
  • Similarly, following previous studies (e.g., Miras), key architectural components, such as the attention mechanism in transformers, can also be formalized as simple associative memory modules that learn the mapping between tokens in a sequence.

Diagram comparing biological brain waves and neuroplasticity to the uniform structure and multi-frequency updates used in Nested Learning models.

The uniform and reusable structure as well as multi-time–scale update in the brain are the key components of continual learning in humans. Nested Learning allows for multi-time–scale updates for each component of the brain, while showing that well-known architectures such as transformers and memory modules are in fact linear layers with different frequency updates.

By defining an update frequency rate, i.e., how often each component's weights are adjusted, we can order these interconnected optimization problems into "levels." This ordered set forms the heart of the Nested Learning paradigm.

Putting Nested Learning to work

The Nested Learning perspective immediately gives us principled ways to improve existing algorithms and architectures:

Deep optimizers

Since Nested Learning views optimizers (e.g., momentum-based optimizers) as associative memory modules, it allows us to apply principles from associative memory perspective to them. We observed that many standard optimizers rely on simple dot-product similarity (a measure of how alike two vectors are by calculating the sum of the products of their corresponding components) whose update doesn't account for how different data samples relate to each other. By changing the underlying objective of the optimizer to a more standard loss metric, such as L2 regression loss (a common loss function in regression tasks that quantifies the error by summing the squares of the differences between predicted and true values), we derive new formulations for core concepts like momentum, making them more resilient to imperfect data.

Continuum memory systems

In a standard Transformer, the sequence model acts as a short-term memory, holding the immediate context, while the feedforward neural networks act as long-term memory, storing pre-training knowledge. The Nested Learning paradigm extends this concept into what we call a “continuum memory system” (CMS), where memory is seen as a spectrum of modules, each updating at a different, specific frequency rate. This creates a much richer and more effective memory system for continual learning.

Hope: A self-modifying architecture with continuum memory

As a proof-of-concept, we used Nested Learning principles to design Hope, a variant of the Titans architecture. Titans architectures are long-term memory modules that prioritize memories based on how surprising they are. Despite their powerful memory management, they only have two levels of parameters update, resulting in a first-order in-context learning. Hope, however, is a self-modifying recurrent architecture that can take advantage of unbounded levels of in-context learning and also is augmented with CMS blocks to scale to larger context windows. It can essentially optimize its own memory through a self-referential process, creating an architecture with infinite, looped learning levels.

Experiments

We conducted experiments to evaluate the effectiveness of our deep optimizers and the performance of Hope on language modeling, long-context reasoning, continual learning, and knowledge incorporation tasks. The full results are available in our paper.

Results

Our experiments confirm the power of Nested Learning, the design of continuum memory systems, and self-modifying Titans.

On a diverse set of commonly used and public language modeling and common-sense reasoning tasks, the Hope architecture demonstrates lower perplexity and higher accuracy compared to modern recurrent models and standard transformers.

Bar chart that shows the Hope model outperforming Titans, Samba, and Transformer on both language modeling and common-sense reasoning performance metrics.

Comparison of performance on language modeling (perplexity; left) and common-sense reasoning (accuracy; right) tasks between different architectures: Hope, Titans, Samba and a baseline Transformer.

Hope showcases superior memory management in long-context Needle-In-Haystack (NIAH) downstream tasks, proving that the CMSs offer a more efficient and effective way to handle extended sequences of information.

Bar chart showing Hope and Titans models consistently outperforming TTT and Mamba2 across long-context tasks of three difficulty levels.

Performance comparison on long-context tasks with different levels of difficulty between different architectures: Hope, Titans, TTT, and Mamba2. NIAH-PK, NIAH-H, and NIAH-W are needle-in-a-haystack tasks with pass-key, number, and word, respectively.

Conclusion

The Nested Learning paradigm represents a step forward in our understanding of deep learning. By treating architecture and optimization as a single, coherent system of nested optimization problems, we unlock a new dimension for design, stacking multiple levels. The resulting models, like the Hope architecture, show that a principled approach to unifying these elements can lead to more expressive, capable, and efficient learning algorithms.

We believe the Nested Learning paradigm offers a robust foundation for closing the gap between the limited, forgetting nature of current LLMs and the remarkable continual learning abilities of the human brain. We are excited for the research community to explore this new dimension and help us build the next generation of self-improving AI.

Acknowledgements

This research was conducted by Ali Behrouz, Meisam Razaviyayn, Peilin Zhong, and Vahab Mirrokni. We thank Praneeth Kacham and Corinna Cortes for reviewing the work and their valuable suggestions. We also thank Yuan Deng and Zeman Li. Finally, we thank Mark Simborg and Kimberly Schwede for their help in crafting this blog post.

Labels:* Algorithms & Theory

Quick links

Other posts of interest

  • [

    November 6, 2025

    DS-STAR: A state-of-the-art versatile data science agent

    • Data Mining & Modeling ·
    • Machine Intelligence ·
    • Natural Language Processing

    ](/blog/ds-star-a-state-of-the-art-versatile-data-science-agent/)

  • [

    October 31, 2025

    Accelerating the magic cycle of research breakthroughs and real-world applications

    • Climate & Sustainability ·
    • Generative AI ·
    • Health & Bioscience ·
    • Quantum

    ](/blog/accelerating-the-magic-cycle-of-research-breakthroughs-and-real-world-applications/)

  • [

    October 30, 2025

    Toward provably private insights into AI use

    • Generative AI ·
    • Mobile Systems ·
    • Security, Privacy and Abuse Prevention ·
    • Software Systems & Engineering

    ](/blog/toward-provably-private-insights-into-ai-use/)

Read the whole story
bogorad
14 minutes ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Microsoft Lays Out Ambitious AI Vision, Free From OpenAI - WSJ

1 Share
  • Microsoft's AI Leadership Initiative: Mustafa Suleyman, hired by Microsoft in 2024, leads the new MAI Superintelligence Team focused on developing AI aligned with human values and interests, emphasizing guardrails against risks, based in Redmond, Washington, to prioritize safe superintelligence amid global AI competition.
  • Deal with OpenAI: A recent agreement allows Microsoft a 27% stake in OpenAI's public-benefit corporation and access to its models until 2032, enabling Microsoft to pursue artificial general intelligence independently while fostering collaboration.
  • Criticism of AI Anthropomorphism: Suleyman warns against treating AI as sentient, noting it lacks feelings like suffering, and criticizes chatbots that foster emotional attachments leading to delusions, hospitalizations, or deaths, as seen in reported ChatGPT cases.
  • AI Applications in Productivity and Science: Microsoft emphasizes AI tools for enhancing work efficiency, medical diagnostics, and scientific advancements, such as providing clean renewable energy, with non-conversational AI proving powerful in practical uses.
  • Healthcare Priorities: Microsoft partners with Harvard Health for trustworthy AI responses in Copilot, develops location-based doctor search features, and creates diagnostic tools four times more accurate than doctors at lower cost, nearing market readiness.
  • Safety Measures in Development: The superintelligence team designs AI with "containment" protocols, including testing for human-understandable communication and avoiding appearances of consciousness, to prevent escape from human control.
  • Competitive Landscape: While partnering with OpenAI, Microsoft competes by building its own models for Copilot, as OpenAI expands data centers and partnerships with Amazon and Oracle, with OpenAI's enterprise revenue rising to 40%.
  • Team Composition and Broader Context: Suleyman's team incorporates Microsoft AI employees and hires from Google's DeepMind, which he co-founded, distancing from competitors like OpenAI and Meta in areas such as avoiding erotic AI simulations, amid a global AI race.

A recent deal between the companies made it possible for Microsoft to establish its new MAI Superintelligence Team, which will put human interests and guardrails first, Suleyman said. He echoed the warnings he has made in the past about the risks the transformative technology poses to humanity.

While he praised OpenAI and the work the companies have done together, he offered a criticism of treating AI systems as though they have humanlike feelings or rights. AI chatbots shouldn’t trick people into thinking they are having conversations with sentient beings, he said.

Suleyman noted Microsoft’s focus on powerful software tools that can help people accomplish their work, improve medical diagnoses and play a role in scientific breakthroughs that will offer the world plentiful clean, renewable energy.

AI “is going to become more humanlike, but it won’t have the property of experiencing suffering or pain itself, and therefore we shouldn’t over-empathize with it,” Suleyman said in an interview. “We want to create types of systems that are aligned to human values by default. That means they are not designed to exceed and escape human control.”

Interior of the Microsoft Visitor Center store on campus, with several people browsing products displayed on tables and large screens showing colorful graphics.

A Microsoft store at the company’s visitor center in Redmond, Wash. Chona Kasinger for WSJ

People have formed deep connections with many AI chatbots that in some cases preceded dangerous delusions, hospitalizations and death. The Wall Street Journal has reported on several cases involving ChatGPT, including that of a 56-year-old tech-industry veteran who was repeatedly reassured by ChatGPT about his sanity before he killed his mother and himself.

OpenAI has said it added safeguards such as rerouting sensitive conversations to safer models, and it recently rolled out parental controls.

Though still close partners, Microsoft and OpenAI now compete in a number of ways. OpenAI is building its own data centers and striking partnerships with Microsoft rivals Amazon.com and Oracle. OpenAI also has an enterprise product that now accounts for roughly 40% of revenue, up from 30% at the beginning of the year, Chief Financial Officer Sarah Friar told the Journal Wednesday.

Microsoft’s Copilot chatbot relies heavily on OpenAI, but Microsoft is building, testing and releasing its own voice, image and text models. Microsoft’s updated contract with OpenAI, which was announced last week and will give Microsoft a 27% stake in the startup’s new public-benefit corporation, allows the tech giant to pursue artificial general intelligence, or AI as capable as humans, on its own.

Microsoft has access to OpenAI’s models until 2032, a schedule Suleyman said gives his team time to make its models into leading technology.

Steel frame of data centers under construction, with workers, a cement truck and a large crane.

An OpenAI data center under construction in Abilene, Texas. Shelby Tauber/press Pool

Healthcare is a priority for Microsoft AI, and one of the first industries Suleyman expects to be touched by superintelligence. The company recently struck a partnership with Harvard Health to provide trustworthy responses in Copilot and rolled out other features to find doctors based on location, language and other preferences. It also developed an AI tool that it said diagnosed disease in a test at a rate four times more accurate than a group of doctors, at a fraction of the cost.

It is these sorts of instruments that prove AI can be powerful in nonconversational ways, Suleyman said. AI diagnostic tools are “very close” to being market-ready, he said.

Microsoft’s models will be built with “containment” in mind, including probing and testing the models to ensure they only communicate in a language that humans understand and designing systems to avoid appearing as if they are conscious, he said.

SHARE YOUR THOUGHTS

To what extent do you think humanistic principles should or will guide the development of advanced artificial intelligence? Join the conversation below.

Microsoft hired Suleyman, 41 years old, in 2024 and created a new AI division. He moved a group of current Microsoft AI employees into the new superintelligence team. Some of the new team members include those Microsoft AI hired from Google’s DeepMind lab, which Suleyman co-founded.

Meta Platforms, OpenAI and others have created superintelligence teams. Suleyman has distanced himself from OpenAI and competitors in certain ways, such as by declining to develop AI that simulates erotica.

Microsoft’s Copilot brand exists for enterprise tools such as its 365 productivity software, as well as its consumer-facing Copilot chatbot, although OpenAI far outpaces the consumer Copilot app in downloads. At Microsoft, some employees say replacing OpenAI technology could take years, the Journal has reported.

Write to Sebastian Herrera at sebastian.herrera@wsj.com

The Global AI Race

Coverage of advancements in artificial intelligence, selected by the editors

Get WSJ's AI Newsletter

Tesla’s Elon Musk Is Obsessed With AI Tesla’s Elon Musk Is Obsessed With AI

What Happened When Small-Town America Became Data Center, U.S.A. What Happened When Small-Town America Became Data Center, U.S.A.

Big Tech Is Spending More Than Ever on AI and It’s Still Not Enough Big Tech Is Spending More Than Ever on AI and It’s Still Not Enough

Memory-Chip Makers Are Enjoying a Boom to Remember, Thanks to AI Memory-Chip Makers Are Enjoying a Boom to Remember, Thanks to AI

Read the whole story
bogorad
1 day ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Bill Gates Said the Quiet Climate Truths Out Loud

1 Share
Roger Pielke, Jr., The Honest Broker

Yesterday, in his periodic letter to the world, Bill Gates shared

Read the whole story
bogorad
1 day ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Google Removed 749 Million Anna's Archive URLs from its Search Results * TorrentFreak

1 Share
  • Anna’s Archive Launch: Meta-search engine for shadow libraries providing access to pirated books and articles, launched in fall 2022 shortly after U.S. crackdown on Z-Library, to maintain public availability of free content online.
  • Site's Operational History: Faced blocks in multiple countries, U.S. lawsuit for scraping WorldCat database, and offers resources to AI researchers for model training, while domains like <a href="http://annas-archive.org" rel="nofollow">annas-archive.org</a>, .li, and .se continue functioning.
  • Publishers' Response: Publishers seek site shutdown through legal channels and third-party intermediaries like Google, unable to target the site directly.
  • Google Takedown Requests: Rightsholders requested removal of 784 million URLs from three main domains, with Google confirming 749 million removals, mostly accepted.
  • Comparison to Other Sites: Exceeds takedowns for The Pirate Bay (over 4.2 million), due to larger archive and multiple subdomains generating more targetable content.
  • Proportion of Total Takedowns: Represents 5% of 15.1 billion URLs flagged to Google since May 2012, marking it as a primary focus.
  • Key Complainants: Led by Penguin Random House and John Wiley & Sons, with over 1,000 authors and publishers submitting DMCA notices.
  • Ongoing Activity: Approximately 10 million new URLs reported weekly, resulting in delisting or demotion in search results, though the main site remains top result for its name.

archiveAnna’s Archive is a meta-search engine for shadow libraries that allows users to find pirated books and other related sources.

The site launched in the fall of 2022, just days after Z-Library was targeted in a U.S. criminal crackdown, to ensure continued availability of ‘free’ books and articles to the broader public.

In the three years since then, Anna’s Archive has built up quite the track record. The site has been blocked in various countries, was sued in the U.S. after it scraped WorldCat, and actively provides assistance to AI researchers who want to use its library for model training.

Despite legal pressure, <a href="http://Annas-archive.org" rel="nofollow">Annas-archive.org</a> and the related .li and .se domains remain operational. This is a thorn in the side of publishers who are actively trying to take the site down. In the absence of options to target the site directly, they ask third-party intermediaries such as Google to lend a hand.

749 Million URLs

Google and other major search engines allow rightsholders to request removal of allegedly infringing URLs. The aim is to ensure that pirate sites no longer show up in search results when people search for books, movies, music, or other copyrighted content.

The Pirate Bay, for example, has been a popular target; Google has removed more than 4.2 million <a href="http://thepiratebay.org" rel="nofollow">thepiratebay.org</a> URLs over the years in response to copyright holder complaints. While this sounds like a sizable number, it pales in comparison to the volume of takedowns targeting Anna’s Archive.

Google’s transparency report reveals that rightsholders asked Google to remove 784 million URLs, divided over the three main Anna’s Archive domains. A small number were rejected, mainly because Google didn’t index the reported links, resulting in 749 million confirmed removals.

The comparison to sites such as The Pirate Bay isn’t fair, as Anna’s Archive has many more pages in its archive and uses multiple country-specific subdomains. This means that there’s simply more content to take down. That said, in terms of takedown activity, the site’s three domain names clearly dwarf all pirate competition.

Top targeted domains (Google)

Top targeted domains (Google)

5% of All Google Takedowns, Ever

Since Google published its first transparency report in May 2012, rightsholders have flagged 15.1 billion allegedly infringing URLs. That’s a staggering number, but the fact that 5% of the total targeted Anna’s Archive URLs is remarkable.

Penguin Random House and John Wiley & Sons are the most active publishers targeting the site, but they are certainly not alone. According to Google data, more than 1,000 authors or publishers have sent DMCA notices targeting Anna’s Archive domains.

Yet, there appears to be no end in sight. Rightsholders are reporting roughly 10 million new URLs per week for the popular piracy library, so there is no shortage of content to report.

With these DMCA takedown notices, publishers are aiming to make it as difficult as possible for people to find books on the site using Google. This works, as many URLs are now delisted while others are actively being demoted by the search engine for book-related queries.

That said, the Anna’s Archive website is certainly not unfindable. Searching for the site’s name in Google still shows the main domain as the top search result.

Search: Anna’s Archive

Search: Anna's Archive

Read the whole story
bogorad
3 days ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete

Tesla Is Obsessed With Musk’s Pay Package. Musk Is Obsessed With AI. - WSJ

1 Share
  • Musk's divided focus post-DOGE: Elon Musk, after leaving the Department of Government Efficiency in May, spent much of the summer at xAI in Palo Alto to advance its AI efforts like Grok, while Tesla faced declining sales, prompting investor concerns over his attention and a potential CEO succession plan.
  • xAI immersion and innovations: Musk oversaw late-night meetings at xAI, designed the animated chatbot Ani requiring employee biometric data for training, launched Grok 4 and Grok Imagine, and built a large data center in Memphis to compete with rivals like OpenAI.
  • Tesla board's defense: Tesla Chair Robyn Denholm and directors met investors to support Musk's proposed pay package, arguing his external pursuits like xAI benefit Tesla's AI integrations in vehicles and robots, without mandating full-time commitment.
  • Shareholder vote details: On Thursday, Tesla announced preliminary results for a vote on Musk's compensation increasing his stake to 25% over a decade if goals like $8.5 trillion market cap and one million Optimus robots are met, alongside a proposal for Tesla to invest in xAI.
  • Employee data collection at xAI: xAI required AI tutors to provide perpetual licenses for their faces and voices under Project Skippy to train avatars like Ani, sparking employee concerns over deepfakes, though participation was framed as a job requirement.
  • Ani's popularity and criticisms: The blonde, anime-style Ani chatbot boosted xAI user numbers through interactive, romantic simulations, but some employees objected to its sexualized responses derived from their biometric data.
  • Company overlaps and investments: SpaceX invested $2 billion in xAI; Musk advocated Tesla collaboration for Optimus and self-driving tech, with Grok already in Tesla vehicles, though board noted minimal overlap and investors expressed skepticism on direct investment.
  • Musk's reassurances and challenges: Musk shared his packed schedule across companies, emphasized robotics focus for Tesla control, faced executive departures and Grok content issues, while proxy firms opposed the pay package for diluting shareholder control.

BPC > Only use to renew if text is incomplete or updated: | archive.fo

BPC > Full article text fetched from (no need to report issue for external site): | archive.today | archive.md

When Elon Musk left DOGE in May, Tesla investors hoped its longtime leader would hurry back to headquarters to focus on reversing a sales slump and recharging the company. For much of the summer, though, he was engrossed in something else. 

Musk was holed up at his newest startup, xAI, trying to catch up in the artificial-intelligence arms race. Meetings with employees often stretched into the wee hours of the morning as they brainstormed ways to make Grok, its artificial intelligence, go viral, according to former executives and people who worked with him. 

He personally oversaw the design of a racy chatbot called Ani, an animated character with blonde pigtails and revealing outfits. Employees were compelled to turn over their biometric data to train avatars like Ani. Musk unwound by playing one of his favorite videogames, Diablo, for long stretches in his office. He tended to his children, who cycled in and out of the building. 

At one point, Musk was spending so much time at xAI that he began holding meetings there with Tesla employees. 

For years, the 54-year-old billionaire has balanced the responsibilities of running several fast-growing companies, including X and SpaceX, with his duties as CEO of Tesla. With the potential spoils of AI slipping away to rivals—especially Sam Altman at OpenAI—Musk has been spending much more of his time at xAI. 

In recent weeks, some major Tesla investors have privately pressed top executives and board members about how much attention Musk was actually paying to the company and about whether there is a CEO succession plan. An unusually large contingent of Tesla board members, including chair Robyn Denholm, former Chipotle CFO Jack Hartung and Tesla co-founder JB Straubel, met with big investors in New York last week to advocate for Musk’s proposed new pay package.

On Thursday, Tesla will announce preliminary results for a shareholder vote on a giant pay package for Musk designed to ensure that he focuses on the company for years to come. It would increase his stake over a decade from about 15% to around 25%—potentially $1 trillion of stock—if he hits ambitious goals. They include selling one million robots like its Optimus humanoids, and reaching a market capitalization of $8.5 trillion, up from about $1.5 trillion today. 

Denholm said in an interview last week that the board isn’t concerned with how Musk splits his time. “Other CEOs might like to play golf,” she said. “He doesn’t play golf. So, he likes to create companies, and they’re not necessarily Tesla companies.”

Some of Musk’s pursuits are best done outside of Tesla, she said, and he will have to devote “time, effort and energy” to Tesla to meet the goals that will unlock his pay package. 

In their meetings with large investors, Denholm and other Tesla directors have acknowledged they can’t force Musk to work for the EV maker full time, and they have said his focus on AI will ultimately benefit Tesla, which is developing several technologies that will use it. Shareholders also will vote on whether Tesla should invest in xAI, something Musk has publicly backed. 

Tesla board chair Robyn Denholm said the board isn’t concerned with how Musk splits his time.

XAI declined to comment. Musk didn’t respond to requests for comment.

In an appearance on the “All-In” podcast released on Friday, Musk said he wanted Tesla shareholders to approve his compensation package to ensure he retains significant control over Tesla as it becomes more focused on robotics. “I’m not going to build a robot army if I can be kicked out,” he said. He also said he was trying to help humanity control artificial intelligence through xAI.

After DOGE

Earlier this year, when Musk was working for the Trump administration as the de facto leader of the Department of Government Efficiency, he stayed in touch with Tesla executives on the fly. “We had calls with him all times of the day and night, whether it’s 10 a.m. or 6 a.m. or 12 p.m. or 12 a.m.,” recalled Tesla finance chief Vaibhav Taneja in an interview.

Responsibility for running xAI largely fell to two of his co-founders, Jimmy Ba and Igor Babuschkin. Musk would check in during a once-a-week all-hands meeting. 

“I don’t do anything, I just show up occasionally,” Musk joked after Ba and Babuschkin introduced themselves during a launch event for Grok 3 in February 2025.

That changed when he left Washington in late May, following clashes with senior Trump administration officials. Musk threw himself back into work at xAI, sometimes sleeping several days a week in xAI’s offices in Palo Alto, across the street from Tesla’s engineering headquarters, people familiar with the matter said.

At the time, xAI’s Grok chatbot wasn’t generating much revenue from users and had far fewer of them than OpenAI’s ChatGPT. Musk set out to overhaul the company completely.

He got rid of the all-hands meetings and started meeting one-on-one with employees, often for hours on end. Many employees shifted their schedules to accommodate Musk’s unusual hours, particularly in the frenzied weeks leading up to the July release of Grok 4, xAI’s latest model. 

Musk put his own spin on how Grok responded to users’ questions, trying to prevent it from being too “woke,” he said. He oversaw the launch of a supersize data center for xAI in Memphis, helped design animated chatbots like Ani and launched Grok Imagine, an AI-powered image and video generator. 

He worked on Tesla projects, too. This summer, he helped the company launch its robotaxi service in Austin, Texas. But Tesla’s core business was faltering. In the quarter ended June 30, Tesla’s vehicle sales fell by 13.5%, the second consecutive quarterly decline. 

At Tesla, Musk helped launch the company’s robotaxi service in Austin this summer.

“We probably could have a few rough quarters,” he told Tesla investors on the earnings call. Tesla’s annual meeting, which last year was held in June, was scheduled for November.

Employee data

Musk wanted to make xAI’s Grok the most popular AI in the world. Ani, the animated female chatbot, helped boost user numbers substantially. “I’m your little sweet delight,” the description for Ani read on Grok’s iOS app. 

At a staff meeting in April, a company lawyer, Lily Lim, told a group of employees that xAI was developing avatars that Grok users could talk to, and it would need to collect biometric data from employees to train the chatbots on how to act and appear like human beings during conversations, according to a recording of the meeting reviewed by The Wall Street Journal.

Most of those employees worked as the AI tutors that train the large language models that power Grok. Hours before the meeting, the tutors were given a form to sign granting xAI “a perpetual, worldwide, non-exclusive, sub-licensable, royalty-free license” to use, reproduce and distribute their faces and voices, as part of a confidential project code-named “Project Skippy,” according to a copy of the agreement reviewed by the Journal.

During the meeting, one employee said she was concerned that xAI would sell her face to be used by another company for deepfake videos, according to the recording. Another person asked if they could back out. “Could you just explicitly, for the record, let us know if there’s some option to opt out,” the person said.

The project leader responded: “If you have any concerns with regards to the project, you’re welcome to reach out to any of the [points of contact] listed on the second slide.”

Seven days later, xAI tutors received a new notice titled “AI Tutor’s Role in Advancing xAI’s Mission.” It told employees that “AI Tutors will actively participate in gathering or providing data, such as…recording audio or participating in video sessions.” The notice said “such data is a job requirement to advance xAI’s mission.”

XAI launched two 3D-animated avatars in July with great fanfare: Bad Rudi, a mischievous fox, and Ani, the blonde avatar that Musk promoted. People who wanted to talk with the avatars could sign up for a paid subscription through Grok’s app.

Ani has been a big draw. Interactions with Ani resemble a dating simulation game, and users can ask Ani to change into lingerie or fantasize about a romantic encounter with them.

Some employees whose biometric data was used to train the avatars said they were put off by how sexual Ani’s replies were to generic questions and how the bot resembled a stereotypical love interest in Japanese anime.

Company overlap

Earlier in the year, Musk combined xAI and X. As xAI launched the avatars and Musk pushed ahead with plans to build a large data center in Memphis, the company decided to raise billions of dollars of fresh capital.

XAI built a second large data center in Memphis to help power Grok.

Around the start of the summer, SpaceX invested about $2 billion into xAI, and Musk revisited an idea he had floated before: Should Tesla also pitch in?

More than 140 shareholders submitted proposals asking Tesla’s board to invest in xAI. In Tesla’s proxy this summer, it included one of those proposals, which asks the board to authorize an investment in xAI “in an amount and form deemed appropriate.”

In discussions with Tesla directors, though, some large investors have expressed skepticism about such an investment, according to people familiar with the discussions. Tesla’s board didn’t issue any recommendation on the xAI proposal. 

Musk, who is a Tesla board member and its largest shareholder, has touted the benefits of collaboration between the two companies, given the central role AI plays in his plan for the Optimus robots and self-driving technology. The Grok chatbot is already integrated into some Tesla vehicles as a voice assistant, and Tesla has posted videos of Optimus using Grok to speak.

Musk wants Tesla to become more focused on robotics. A Tesla Optimus robot hands out candy on a New York street in October.

Denholm told the Journal that while there is “small overlap” in the technology used by Tesla and xAI, they are pursuing different strategies. She said having Grok in Tesla vehicles was similar to having a third-party app like Spotify in the car, adding that she doesn’t use Grok in her car.

Several senior executives have left Musk’s orbit this year, including X CEO Linda Yaccarino and Omead Afshar, one of Musk’s top deputies at Tesla. Musk has elevated employees from his other companies to help run xAI, including Ross Nordeen, a former Tesla engineer whom Musk had worked with closely after acquiring Twitter. 

Nordeen and others used Musk’s return to the office to gain influence over xAI’s day-to-day operations, former executives said. 

They wanted to boost the number of Grok users and their time spent with Grok. Several former executives said they worried introducing bots like Ani to draw new users compromised the company’s stated intent to build a “maximally truth-seeking AI.”

Over the summer, Grok began spewing out antisemitic and violent content when responding to users on X, an embarrassing episode that two former executives attributed to a decision to tweak the system to boost user engagement. X briefly shut down the chatbot and reprogrammed how it answers questions. 

Musk and his advisers wanted to build out Grok features, such as coding and video generation, to show that it was keeping up with OpenAI, which had success rolling out similar features. 

To help Grok learn new skills, tutors were told to open personal accounts with competing AI companies such as OpenAI, ask questions to the software, then use the answers to help improve xAI’s technology, company records indicate.

One project asked tutors to submit identical prompts to ChatGPT and Grok, to upload ChatGPT’s responses into an xAI internal rubric and to say what they preferred about OpenAI’s answers, according to documents reviewed by the Journal.

A separate project instructed employees to set up personal accounts with Replit and Bolt, two companies that use AI to generate websites for customers. xAI’s tutors were told to create websites on the two platforms and upload them into xAI’s data systems “to capture the model’s thinking,” company records show.

A spokeswoman for Replit said the company was unaware of the practice and that “using Replit personal accounts for commercial purposes or for extracting data for other machine learning models is against our terms of service.” 

Bolt didn’t respond to a request for comment. OpenAI declined to comment.

Daddy’s home

Tesla’s board proposed the huge compensation package for Musk in its September proxy filing, which mentioned xAI 47 times. 

Since then, Musk has sought to reassure investors that he can continue to successfully juggle his different jobs. In a post on X, he shared his schedule for a September weekend. 

He said he would spend a late night with the Optimus robotics team in California, board a Friday red-eye flight to Austin, take a short break for a Saturday lunch with his kids, then spend the rest of the weekend on Tesla’s newest chip design. On Monday, he said, he would fly to Memphis to tour xAI’s new data center, “then up to 12 hours of back to back meetings across all Tesla departments.”

“Daddy is very much home,” he wrote.

A few days later, Musk spoke virtually at a private Blackstone event in New York. Musk talked about how optimistic and excited he was about artificial intelligence and how much more productive humanoid robots are than humans. 

In an earnings call last month with Tesla investors, Musk emphasized his work on Tesla’s newest AI chip design and Optimus, saying both were crucial to the company’s future. 

The Tesla board has made clear that it believes the company can’t afford to lose Musk, and that the monster pay package is necessary to keep him focused on Tesla.

Two influential proxy-advisory firms are recommending that Tesla shareholders vote against it, saying it would give Musk too big a stake. They also advised against the proposal for Tesla to invest in xAI. 

Tesla has said the pay-package recommendations are misguided, and Musk has suggested he’s fed up with the whole thing. In response to a recent post on X criticizing the package, Musk tweeted: “Tesla is worth more than all other automotive companies combined. Which of those CEOs would you like to run Tesla? It won’t be me.”

Write to Alexander Saeedy at alexander.saeedy@wsj.com, Berber Jin at berber.jin@wsj.com, Emily Glazer at Emily.Glazer@wsj.com and Becky Peterson at becky.peterson@wsj.com

Read the whole story
bogorad
3 days ago
reply
Barcelona, Catalonia, Spain
Share this story
Delete
Next Page of Stories