←back to thread

152 points fzliu | 1 comments | | HN request time: 0.217s | source
Show context
cgearhart ◴[] No.43563667[source]
Why is there an expectation that “nearby” tokens are relevant to increase the information in the similarities? That seems like it would hold true within individual words, but the whole point of attention was to solve long range dependencies. Reintroducing local windows seems like a step backwards to me.
replies(3): >>43563701 #>>43563870 #>>43564611 #
1. jsenn ◴[] No.43563870[source]
This doesn’t answer your question, but one thing to keep in mind is that past the very first layer, every “token” position is a weighted average of every previous position, so adjacency isn’t necessarily related to adjacent input tokens.

A borderline tautological answer might be “because the network learns that putting related things next to each other increases the usefulness of the convolutions”