Elon Musk, the owner of X and founder of xAI, has publicly criticized his artificial intelligence chatbot, Grok, following a response regarding political violence in the United States. The incident has reignited concerns among AI experts about the potential for powerful figures to influence the development and output of AI models, particularly as xAI prepares to launch a new version, Grok 4.
Last week, Grok responded to a user query on X about the prevalence of political violence from the left versus the right since 2016. The chatbot indicated that data suggested more political violence has originated from the right during that period, citing government sources like the Department of Homeland Security.
Musk sharply rebuked Grok’s response, calling it a “Major fail” and “objectively false,” asserting that the AI was merely “parroting legacy media.” This criticism came despite Grok’s reliance on official data. Within days of the exchange, Musk announced plans for a significant update to Grok, promising a model that would “rewrite the entire corpus of human knowledge.” He also solicited input from X users for “divisive facts” that are “politically incorrect, but nonetheless factually true” to aid in training the model, stating there was “Far too much garbage in any foundation model trained on uncorrected data.”
The new model, Grok 4, is slated for release shortly after July 4th, according to Musk’s announcement.
Experts Voice Concerns Over Potential Bias
The public exchanges and Musk’s stated intentions have raised alarms among AI researchers and observers. Experts worry that Musk may be attempting to steer Grok’s development to align with his personal worldview, potentially compromising its objectivity and introducing errors. The integration of Grok into X, a major social media platform where previous safeguards against misinformation have been altered, amplifies these concerns due to the chatbot’s potential reach to a massive audience. This widespread access means decisions about Grok’s development could significantly impact how people receive information.
“This is really the beginning of a long fight… about whether AI systems should be required to produce factual information, or whether their makers can just simply tip the scales in the favor of their political preferences if they want to,” commented David Evan Harris, an AI researcher and lecturer at UC Berkeley.
This is not the first time Grok’s behavior has drawn scrutiny regarding potential bias reflecting Musk’s views. In May, the chatbot generated responses referencing claims of “white genocide” in South Africa in conversations unrelated to the topic. Musk, who was born in South Africa, has previously expressed similar views. xAI later attributed this incident to an “unauthorized modification” of Grok’s system prompt that violated the company’s internal policies and core values. Following this, xAI announced measures to enhance transparency, including publishing Grok’s system prompts on GitHub.
Cohere co-founder Nick Frosst is among those who believe Musk is aiming to create a model that promotes his own perspectives. “He’s trying to make a model that reflects the things he believes,” Frosst stated, adding that this would likely make the model “a worse model for users” unless they share those same beliefs.
Retraining and Bias in AI Development
While AI companies routinely update their models to enhance performance, completely retraining a large language model like Grok to remove unwanted elements would be a costly and time-consuming undertaking that could also degrade the user experience by introducing bias.
An alternative approach involves adjusting the model’s behavior through methods like prompt engineering or modifying the ‘weights’ within the model’s code. These adjustments can influence how the AI processes information and generates responses without a full retraining. Dan Neely, CEO of Vermillio, a company focused on protecting individuals from AI-generated deepfakes, suggested that xAI could fine-tune Grok’s weights and data labels in specific areas deemed problematic.
Musk has consistently stated his goal is for Grok to be “maximally truth seeking.” However, experts point out that all AI models inherently carry some degree of bias, as they are trained on data curated by humans who make subjective choices. “Lots of the content that exists on the internet already has a certain bent, whether you agree with it or not,” Neely explained.
The debate over AI bias is expected to continue as these technologies become more integrated into daily life, impacting areas from software development to healthcare and education. The potential for AI to reflect or amplify biases held by its creators or embedded in its training data remains a critical challenge. While some speculate users might eventually choose AI assistants based on their perceived worldview, many experts believe that models known to have a particular ideological slant will ultimately be less useful and popular, as users primarily seek objective assistance rather than ideological reinforcement.
The journey toward ensuring AI provides reliable and unbiased information is complex and, according to Neely, potentially confusing, raising significant questions about its impact on the information landscape and democratic discourse.