huge. the activation scan, which looks for which nodes change the most when prompted with the words "Golden Gate Bridge" and later an image of the same bridge, is eerily reminiscent of a brain scan under similar prompts...
replies(2):
One such example: The Internal State of an LLM Knows When It's Lying (https://arxiv.org/abs/2304.13734)
Searching phrases like "llm interpretability" and "llm activation analysis" uncover more
(drop-out was found to increase resilience in models because they had to encode information in the weights differently, i.e. could not rely on single neuron (at the limit))
When you look at a specific input, you can look to see what gets activated or not. Orthogonal but related ideas for inspecting the activations to see effects