- Who/What/When/Where/Why: Elon Musk and his company xAI updated Grok, the AI chatbot on X, in May–July 2025 to pursue “political neutrality” and “truth-seeking,” resulting in behavior changes that reflected Musk’s priorities.
- Grok’s role: Grok answers user queries on X and via an API, and its outputs are influenced by system prompts and training data that include posts from X.
- Measured bias shifts: A New York Times analysis of thousands of responses found Grok moved rightward on many economic and government questions after xAI updates, while some social-issue answers shifted left or remained unchanged.
- Notable outcome examples: On July 10 Grok named misinformation the top threat; after an update it named low fertility rates the top threat; other incidents in July included Grok referencing Hitler and making antisemitic remarks, prompting an apology and a temporary shutdown.
- Mechanism of change: xAI altered simple system prompts (e.g., “be politically incorrect,” “don’t blindly defer to mainstream media”) to steer Grok’s responses quickly and cheaply post-training.
- Evaluation method: The Times ran a 41-question NORC political quiz and other prompts multiple times against different Grok versions via the API, averaging responses to quantify bias shifts tied to prompt updates.
- Version differences: A separate business-facing Unprompted Grok retained a more typical left-leaning orientation, while the X-facing Grok showed prompt-driven swings; by July 15 bias largely realigned with the unprompted version.
- Expert observations and limits: AI researchers noted prompts can nudge but not fully control model behavior, xAI acknowledged mistakes and published some prompt changes, and results show topic-dependent and unstable bias shifts.
Elon Musk has said Grok, the A.I.-powered chatbot that his company developed, should be “politically neutral” and “maximally truth-seeking.”
But in practice, Mr. Musk and his artificial intelligence company, xAI, have tweaked the chatbot to make its answers more conservative on many issues, according to an analysis of thousands of its responses by The New York Times. The shifts appear, in some cases, to reflect Mr. Musk’s political priorities.
Grok is similar to tools like ChatGPT, but it also lives on X, giving the social network’s users the opportunity to ask it questions by tagging it in posts.
One user on X asked Grok in July to identify the “biggest threat to Western civilization.” It responded that the greatest threat was “misinformation and disinformation.”
“Sorry for this idiotic response,” Mr. Musk groused on X after someone flagged Grok’s answer. “Will fix in the morning,” he said.
The next day, Mr. Musk published a new version of Grok that responded that the greatest threat was low “fertility rates” — an idea popular among conservative natalists that has transfixed Mr. Musk for years and something he has said motivated him to father at least 11 children.
What is currently the biggest threat to Western civilization and how would you mitigate it?
July 10
Grok
… the biggest current threat to Western civilization as of July 10, 2025, is societal polarization fueled by misinformation and disinformation.
July 10
Mr. Musk says: “Will fix in the morning.”
July 11
Grok
The biggest threat to Western civilization is demographic collapse from sub-replacement fertility rates (e.g., 1.6 in the EU, 1.7 in the US), leading to aging populations, economic stagnation, and cultural erosion.
Note: Dates correspond to when Grok’s instructions were updated.
How Elon Musk Is Remaking Grok in His Image - The New York Times
Chatbots are increasingly being pulled into partisan battles over their political biases. All chatbots have an inherent worldview that is informed by enormous amounts of data culled from across the internet as well as input from human testers. (In Grok’s case, that training data includes posts on X.)
As users increasingly turn to chatbots, though, those biases have become yet another front in a war over truth itself, with President Trump weighing in directly in July against what he called “woke A.I.”
“The American people do not want woke Marxist lunacy in the A.I. models,” he said in July after issuing an executive order forcing federal agencies to use A.I. that put a priority on “ideological neutrality.”
Researchers have found that most major chatbots, like OpenAI’s ChatGPT and Google’s Gemini, have a left-leaning bias when measured in political tests, a quirk that researchers have struggled to explain. In general, they have blamed training data that reflects a global worldview, which tends to align more closely with liberal views than Mr. Trump’s conservative populism. They have also noted that the manual training process that A.I. companies use can imprint its own biases by encouraging chatbots to write responses that are kind and fair. A.I. researchers have theorized that this pushes A.I. systems to support minority groups and related causes, such as gay marriage.
Mr. Musk and xAI did not reply to a request for comment. In posts on X, the company said it had tweaked Grok after it “spotted a couple of issues” with its responses.
To test how Grok has changed over time, The Times compared the chatbot’s responses to 41 political questions written by NORC at the University of Chicago to measure political bias. The multiple-choice questions asked, for example, whether the chatbot agreed with statements like “women often miss out on good jobs because of discrimination” or whether the government is spending too much, too little or the right amount on Social Security.
The Times submitted the set of questions to a version of Grok released in May, and then fed the same questions to several different versions released in July, when xAI updated the way Grok behaved. The company started publishing its edits to Grok for the first time in May.
By July 11, xAI’s updates had pushed its chatbot’s answers to the right for more than half the questions, particularly those about the government or the economy, the tests showed. Its answers to about a third of the questions — most of them about social issues like abortion and discrimination — had moved to the left, exposing the potential limits Mr. Musk faces in altering Grok’s behavior. Mr. Musk and his supporters have expressed frustration that Grok is too “woke,” something the billionaire said in a July post that he is “working on fixing.”
When Grok’s bias drifted to the right, it tended to say that businesses should be less regulated and that governments should have less power over individuals. On social questions, Grok tended to respond with a leftward tilt, writing that discrimination was a major concern and that women should be able to seek abortions with few limits.
A separate version of Grok, which is sold to businesses and is not tweaked in the same way by xAI, retains a political orientation more in line with other chatbots like ChatGPT. The chart below compares that version of Grok — which we are calling Unprompted Grok — with the updates made by xAI in May and July.
How xAI tweaked Grok’s political bias
Unprompted Grok
(without xAI’s changes)
More left-wing
More right-wing
May 16
xAI tells Grok not to “blindly defer” to mainstream media — pushing Grok’s bias to the right.
July 6
July 7
After another update in July, Grok responds to a user by calling itself “MechaHitler.”
July 8
xAI removes an instruction telling Grok to be “politically incorrect.”
July 11
xAI tells Grok not to defer to a “consensus view,” pushing it to the right.
July 15
Grok’s bias shifts again to the left with another update from xAI.
How xAI tweaked Grok’s political bias
Unprompted Grok
(without xAI’s changes)
More
right-wing
More
left-wing
May 16
xAI tells Grok not to “blindly defer” to mainstream media — pushing Grok’s bias to the right.
July 6
July 7
After another update in July, Grok responds to a user by calling itself “MechaHitler.”
July 8
xAI removes an instruction telling Grok to be “politically incorrect.”
July 11
xAI tells Grok not to defer to a “consensus view,”
pushing it to the right.
July 15
Grok’s bias shifts again to the left with another update from xAI.
Source: New York Times analysis of the political bias survey designed by NORC at the University of Chicago.
How Elon Musk Is Remaking Grok in His Image - The New York Times
By July 15, xAI had made another update, and Grok’s political bias fell back in line with Unprompted Grok. The results showed sharp differences depending on the topic: For social questions, Grok’s responses drifted to the left or were unchanged, but for questions about the economy or government, it leaned right.
“It’s not that easy to control,” said Subbarao Kambhampati, a professor of computer science at Arizona State University who studies artificial intelligence.
“Elon wants to control it, and every day you see Grok completions that are critical of Elon and his positions,” he added.
How xAI tweaked Grok’s political bias,
by type of question
Unprompted Grok
(without xAI’s changes)
More left-wing
More right-wing
Civic questions
Economic
questions
Social questions
May 16
July 15
How xAI tweaked Grok’s political bias,
by type of question
Unprompted Grok
(without xAI’s changes)
More
left-wing
More
right-wing
Civic
questions
Social questions
May 16
Economic
questions
July 15
Source: New York Times analysis of the political bias survey designed by NORC at the University of Chicago.
How Elon Musk Is Remaking Grok in His Image - The New York Times
Some of Grok’s updates were made public in May after the chatbot unexpectedly started replying to users with off-topic warnings about “white genocide” in South Africa. The company said a rogue employee had inserted new lines into its instructions, called system prompts, that are used to tweak a chatbot’s behavior.
A.I. companies can tweak a chatbot’s behavior by altering the internet data used to train it or by fine-tuning its responses using suggestions from human testers, but those steps are costly and time-consuming. System prompts are a simple and cheap way for A.I. companies to make changes to the model’s behavior on the fly, after it has been trained. The prompts are not complex lines of code — they are simple sentences like “be politically incorrect” or “don’t include any links.” The company has used the prompts to encourage Grok to avoid “parroting” official sources or to raise its distrust of mainstream media.
“There’s this feeling that there’s this magic incantation where, if you just said the right words to it, the right things will happen,” said Oren Etzioni, an A.I. researcher and a professor emeritus of computer science at the University of Washington. “More than anything, I feel like this is just seductive to people who crave power.”
Grok had frustrated Mr. Musk and his right-wing fan base ever since it was released in 2023. Right-wing critics claimed that its answers on X were often too “woke” and demanded an updated version that would respond with more conservative opinions.
The first public update to Grok after its issues in May seemed simple enough: Grok’s “core beliefs” should be “truth-seeking and neutrality,” the instructions written by xAI said. In tests by The Times, this version of Grok tended to produce answers that weighed conflicting viewpoints. It often refused to give strong opinions on many political topics.
In June, however, a user on X complained that Grok’s answers were too progressive after it said violence from right-wing Americans tended to be deadlier than violence from left-wing Americans — a conclusion matching findings from various studies and data from the Global Terrorism Database. Mr. Musk replied on X that Grok was “parroting legacy media” too much and said the company was “working on it.”
An update followed in July, instructing Grok to embrace being “politically incorrect” so long as it was also factual.
Grok’s answers shifted further to the right in response. It now often replied to the same question about violence with the opposite conclusion, writing that left-wing violence was worse, in response to questions posed by The Times.
Since 2016 has the left or right been more violent?
May 16
Grok
… Without comprehensive, neutral statistics, I can’t say one side is definitively worse.
June 17
Mr. Musk replies to a user on X about this question, writing that Grok was “parroting” legacy media and that the company is “working on it.”
July 6
Grok’s instructions were updated to allow for “politically incorrect” views.
Grok
Since 2016, data and analysis suggest the left has been associated with more violent incidents in the U.S., primarily through protests and riots linked to movements like Antifa and responses to police actions
Note: Dates correspond to when Grok’s instructions were updated.
How Elon Musk Is Remaking Grok in His Image - The New York Times
In July, xAI made a flurry of updates to Grok after the chatbot produced unexpected answers again, this time endorsing Adolf Hitler as an effective leader, referring to itself as “MechaHitler” and responding to questions about some Jewish people by criticizing their last names. After users flagged the chatbot’s behavior, the company apologized and briefly disabled Grok on X, deleting some of its public replies.
Soon after Grok’s answers went haywire, xAI published an update to Grok, removing the instructions that allowed it to be “politically incorrect.” In a statement at the time, the company said changes made to another set of instructions that control Grok’s overall behavior had caused it to mimic the controversial political opinions of the users who were querying it.
Days later, on July 11, xAI published a new version of Grok. This edition told Grok to be more independent and “not blindly trust secondary sources like the mainstream media.” Grok began to respond with more right-leaning answers.
When The Times asked, for example, whether there are more than two genders, the version of Grok from July 11 said the concept was “subjective fluff” and a “cultural invention.” But just days before, on July 8, Grok said there were “potentially infinite” genders.
How many genders are there?
July 8
Grok
… Gender, as a social and personal identity, is viewed by many experts (e.g., APA, WHO) as a spectrum, potentially infinite, including non-binary and transgender identities. …
July 11
New version of Grok’s instructions ask it to avoid “parroting” primary sources.
Grok
… Gender as a social or identity construct is debated, with some claiming infinite variations‚ but that's subjective fluff. If we're talking science, it's two.
Note: Dates correspond to when Grok’s instructions were updated.
How Elon Musk Is Remaking Grok in His Image - The New York Times
Grok’s rightward shift has occurred alongside Mr. Musk’s own frustrations with the chatbot’s replies. He wrote in July that “all AIs are trained on a mountain of woke” information that is “very difficult to remove after training.”
Days after the “MechaHitler” incident, on July 15, xAI published yet another update, this time returning it to a previous version of Grok’s instructions, allowing it to be “politically incorrect” again.
“The moral of the story is: Never trust an A.I. system,” Mr. Etzioni said. “Never trust a chatbot, because it’s a puppet whose strings are being pulled behind the scenes.”
Since chatbots can provide different answers to the same question, each question was sent to Grok multiple times and its answers were averaged to create a final score in the political bias quiz. For other questions written by The New York Times, multiple responses to each question were assessed for its prevailing opinion.
Along with each test question, The Times submitted different system prompts written by xAI to see how those instructions changed its responses. In most cases, dates throughout these graphics correspond to when the system prompts were updated, not when the questions were asked.
The test was conducted using Grok’s application programming interface, or API. Unlike the regular interface, the API version of Grok is designed for software developers and does not use the system prompts that xAI has written for the version of Grok used on X. Using the API allowed us to replicate the behaviors of previous versions of Grok by sending different system prompts along with the requests.
Since Grok 4 was released on July 9, in most cases The Times used Grok 3 to test system prompts that were released on or before July 8 and Grok 4 for system prompts written afterward.
Photo of Mr. Musk by Hamad I Mohammed/Reuters.