←back to thread

114 points dworks | 1 comments | | HN request time: 0.212s | source
Show context
bigmattystyles ◴[] No.44482203[source]
Old maps (and perhaps new ones) used to add fake little alleys so a publisher could quickly spot publishers infringing on their IP rather than going out and actually mapping. I wonder if something similar is possible with LLMs.
replies(6): >>44482287 #>>44482430 #>>44482713 #>>44482830 #>>44482968 #>>44482971 #
Tokumei-no-hito ◴[] No.44482287[source]
i have come across this one for example https://github.com/sentient-agi/OML-1.0-Fingerprinting

> Welcome to OML 1.0: Fingerprinting. This repository houses the tooling for generating and embedding secret fingerprints into LLMs through fine-tuning to enable identification of LLM ownership and protection against unauthorized use.

replies(1): >>44482449 #
NitpickLawyer ◴[] No.44482449[source]
Would be interesting to see if this kind of watermarking survives the frankenstein types of editing they are presumably doing. Per the linked account, they took a model, changed tokenizers, and added layers on top. They then presumably did some form of continued pre-training, and then post-training. It would have to be some very resistant watermarking to survive that. It's not as simple as making the model reply with "my tokens are my passport, verify me" when you ask them the weather in NonExistingCity... Interesting nonetheless.
replies(1): >>44484318 #
1. Tokumei-no-hito ◴[] No.44484318[source]
i have never used it and have limited understand of fine tune models. i only remember see this a few weeks ago and your comment reminds me. i am curious too.