←back to thread

176 points nxa | 1 comments | | HN request time: 0.84s | source

I've been playing with embeddings and wanted to try out what results the embedding layer will produce based on just word-by-word input and addition / subtraction, beyond what many videos / papers mention (like the obvious king-man+woman=queen). So I built something that doesn't just give the first answer, but ranks the matches based on distance / cosine symmetry. I polished it a bit so that others can try it out, too.

For now, I only have nouns (and some proper nouns) in the dataset, and pick the most common interpretation among the homographs. Also, it's case sensitive.

Show context
__MatrixMan__ ◴[] No.43989961[source]
Here's a challenge: find something to subtract from "hammer" which does not result in a word that has "gun" as a substring. I've been unsuccessful so far.
replies(8): >>43990011 #>>43990015 #>>43990016 #>>43990019 #>>43990027 #>>43990335 #>>43990495 #>>43995910 #
mrastro ◴[] No.43990019[source]
The word "gun" itself seems to work. Package this as a game and you've got a pretty fun game on your hands :)
replies(1): >>43991561 #
1. __MatrixMan__ ◴[] No.43991561[source]
Doh why didn't I think of that