Your support helps us to tell the story
From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it’s investigating the financials of Elon Musk’s pro-Trump PAC or producing our latest documentary, ‘The A Word’, which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.
At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.
The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.
Your support makes all the difference.
Read more
Grok, the AI assistant made by Elon Musk’s artificial intelligence firm, was specifically told not to use sources that said he was responsible for spreading misinformation.
There were specific rules written into Grok’s system prompt – the instructions that it is given to decide how it will reply to questions – that told it not to use “sources that mention Elon Musk/Donald Trump spread misinformation”, users found.
Unlike some other more closed AI research companies, the system prompt that it is given to the Grok system with queries are public, which means that users are able to see them.
Mr Musk has suggested that one of Grok’s standout features is that it is free from the bias and limitations that characterise some other AI chatbots.
But xAI has at times intervened to stop embarrassing or difficult responses. Last week, users found that Grok would respond to questions about who might deserve the death penalty with Elon Musk and Donald Trump.
Users have regularly attempted to use Grok to shame Mr Musk, including using it to suggest that he helps spread disinformation. It is unclear whether the change had been to try and stop that happening, but a senior engineer at xAI said that it had been done to “help”.
“You are over-indexing on an employee pushing a change to the prompt that they thought would help without asking anyone at the company for confirmation,” said Igor Babuschkin, xAI’s head of engineering.
“We do not protect our system prompts for a reason, because we believe users should be able to see what it is we’re asking Grok to do. Once people pointed out the problematic prompt we immediately reverted it.
“Elon was not involved at any point. If you ask me, the system is working as it should and I’m glad we’re keeping the prompts open.”