Then retrain on the 99.85 billion documents you have left over, and you've created an LLM that is unaware of those labelled 150 million documents.
Then retrain on the 99.85 billion documents you have left over, and you've created an LLM that is unaware of those labelled 150 million documents.
What about using LLM to not skip undesired documents, but rewrite them, so the new model thinks they say something else?
You think they’re likely to do better than canceling research on “equity” and discovering they’ve bankrupted all business school finance departments?
A+ post
You've created an LLM there that is /worse/ because it is (a) smaller, and (b) intentionally skewed, but it's also not so much smaller that it will be visibly useless, and it has knowledge of other topics, even if it is much dumber about things adjacent to the filtered topics.
And then Elon will brand it as more “intelligent” and lots of people will believe it.
It doesn't require any particularly sophisticated steps to do it, or need an LLM that is particularly intelligent to help you get there, which is also partly what makes this dangerous: if you have enough money and an AI machine, you can absolutely construct propaganda models. And he's saying he will
I think you’re right, making propaganda models can be done. It already works that way of you know how to put the right things surrounding the LLM, though. All of the bad stuff is in there you can get it out with the right input prompts. This also shows Musk doesn’t know how they work.
Right, we've already seen several cases of people trying to create propaganda models in a few different ways. Musk did it already with system prompt hacking (the famous "grok is obsessed with white genocide" fiasco). DeepSeek did it too, but in post-training, objecting to Tiananmen square references
You have some filtering which is more neutral; prior to the 2024 election several online models tried to detect and bail when talking about politics to both avoid brand risks and "election disinformation", and genImages try hard to detect attempts to create abuse material
Which all gets back to the hostility Musk et al have for Wikipedia -- for its faults, Wiki is a largely reliable source of verified facts and data that resists being used as a propaganda machine.
A good reason to send them donations. I do.
quite an expensive fools errand too, training ain't cheap
but this is much more direct attempt to create propaganda: that the output of the model should seek to achieve political persuasion outcomes and be deployed to a site that hundreds of millions of people, including political elites, journalists, and citizens
I agree with you that’s what he is describing. What a lot of disinfo people fear most is GenAI only good trait is it’s a convincing liar. And using it to wage propaganda war at Internet scale is its bread and butter.
📌
Elon Musk wants to create an artificial Fox Viewer by filtering out training material that does agree with the narrative. This creation of artificial stupidity may be successful, but it's not a contribution.
I am certain Grok was already a propaganda machine before the election.
Musk should take a ride in one of his rockets.
An LLM is actually a perfect propaganda machine, it's real power is not the veracity of its output, but the convincing way it presents it. The training data is still the costly time consuming point I would argue, consensus on what it should spit out, and internal integrity of the data presented
The real power is that most users believe that it possesses intelligence. Calling it AI is a misnomer that needs to stop. It’s a really good mimicry machine.
They believe it possesses intelligence exactly because it's output sounds so convincing. it's output sounds like human speach, and with authority.