It really strains credulity to say that a Musk-owned ai model that answers controversial questions by looking up what his Twitter profile says was completely out of the blue. Unless they are able to somehow show this wasn't built into the training process I don't see anyone taking this model seriously for its intended use, besides maybe the sycophants who badly need to a summary of Elon Musk's tweets.
So in that sense, Grok and Gemini aren't that far apart, just the other side of the extreme.
Apparently it's very hard to create an AI that behaves balanced. Not too woke, and not too racist.
Well, it's hard to build things we don't even understand ourselves, especially about highly subjective topics. What is "woke" for one person is "basic humanity" for another, and "extremism" for yet another person, and same goes for most things.
If the model can output subjective text, then the model will be biased in some way I think.
Musk said "stop making it sound woke" after re-training it and changing the fine tuning dataset, it was still sounding woke. After he fired a bunch more researchers, I suspect they thought "why not make it search what musk thinks?" boom it passes the woke test now.
Thats not an emergent behaviour, that's almost certainly deliberate. If someone manages to extract the prompt, you'll get conformation.