←back to thread

176 points nxa | 3 comments | | HN request time: 0.781s | source

I've been playing with embeddings and wanted to try out what results the embedding layer will produce based on just word-by-word input and addition / subtraction, beyond what many videos / papers mention (like the obvious king-man+woman=queen). So I built something that doesn't just give the first answer, but ranks the matches based on distance / cosine symmetry. I polished it a bit so that others can try it out, too.

For now, I only have nouns (and some proper nouns) in the dataset, and pick the most common interpretation among the homographs. Also, it's case sensitive.

1. matallo ◴[] No.43989584[source]
uncle + aunt = great-uncle (91%)

great idea, but I find the results unamusing

replies(1): >>43989784 #
2. HWR_14 ◴[] No.43989784[source]
Your aunt's uncle is your great-uncle. It's more correct than your intuition.
replies(1): >>43989911 #
3. matallo ◴[] No.43989911[source]
I asked ChatGPT (after posting my comment) and this is the response. "Uncle + Aunt = Great-Uncle is incorrect. A great-uncle is the brother of your grandparent."