Huh. That’s an interesting additional risk. I don’t think it is what the original commenter meant, because they were talking about catching cheaters. But it is interesting to think about…
I dunno. There generally isn’t super high security around preprint papers (lots of people just toss their own up on arxiv, after all). But, yeah, it is something that you’ve been asked to look after for somebody, which is quite important to them, so it should probably be taken pretty seriously…
I dunno. The extent to which, and the timelines for, the big proprietary LLMs to feed their prompts back into the training set, are hard to know. So, hard to guess whether this is a serious vector for leaks (and in the absence of evidence it is best to be prudent with this sort of thing and not do it). Actually, I wonder if there’s an opening for a journal to provide a review-helper LLM assistant. That way the journal could mark their LLM content however they want, and everything can be clearly spelled out in the terms and conditions.