It really strains credulity to say that a Musk-owned ai model that answers controversial questions by looking up what his Twitter profile says was completely out of the blue. Unless they are able to somehow show this wasn't built into the training process I don't see anyone taking this model seriously for its intended use, besides maybe the sycophants who badly need to a summary of Elon Musk's tweets.
Musk said "stop making it sound woke" after re-training it and changing the fine tuning dataset, it was still sounding woke. After he fired a bunch more researchers, I suspect they thought "why not make it search what musk thinks?" boom it passes the woke test now.
Thats not an emergent behaviour, that's almost certainly deliberate. If someone manages to extract the prompt, you'll get conformation.