Most active commenters
  • smartmic(3)

←back to thread

427 points JumpCrisscross | 11 comments | | HN request time: 0.804s | source | bottom
Show context
mrweasel ◴[] No.41901883[source]
The part that annoys me is that students apparently have no right to be told why the AI flagged their work. For any process where an computer is allowed to judge people, where should be a rule in place that demands that the algorithm be able explains EXACTLY why it flagged this person.

Now this would effectively kill off the current AI powered solution, because they have no way of explaining, or even understanding, why a paper may be plagiarized or not, but I'm okay with that.

replies(8): >>41902108 #>>41902131 #>>41902463 #>>41902522 #>>41902919 #>>41905044 #>>41905842 #>>41907688 #
1. smartmic ◴[] No.41902522[source]
I agree with you, but I would go further and turn the tables. An AI should simply not be allowed to evaluate people, in any context whatsoever. For the simple reason that it has been proven not to work (and will also never).

Anyone interested to learn more about it, I recommend the recent book "AI Snake Oil" from Arvind Narayanan and Sayash Kapoor [1]. It is a critical but nuanced book and helps to see the whole AI hype a little more clearly.

[1] https://press.princeton.edu/books/hardcover/9780691249131/ai....

replies(2): >>41902634 #>>41903001 #
2. fullstackchris ◴[] No.41902634[source]
I'm definitely no AI hypster, but saying anything will "never" work over an infinite timeline is a big statement... do you have grounds why some sort of AI system could one day "never" work at evaluating some metric about someone? Seems we have reliable systems already doing that in some areas (facial recognition at airport boarding, for example)
replies(2): >>41902909 #>>41908929 #
3. smartmic ◴[] No.41902909[source]
Okay, let me try to be more precise. By "evaluate", I mean using an AI to make predictions about human behavior, either retrospectively (as is the case here in trying to make an accusation of cheating) or prospectively (i.e. automating criminal justice). Even if you could collect all the parameters (features?) that make up a human being, there is the randomness in humans and in nature in general, which simply destroys any ultimate prediction machine. Not to mention the edge cases we wander into. You can try to measure and average a human being, and you will get a certain accuracy well above 50%, but you will never cross the threshold of such high accuracy that a human being should be measured against, especially in life-deciding questions like career decisions or any social matters.

Reliable systems in some areas? - Absolutely, and yes, even facial recognition. I agree, it works very well, but that is a different issue as it does not reveal or try to guess anything about the inner person. There are other problems that arise from the fact that it works so well (surveillance, etc.), but I did not mean that part of the equation.

replies(1): >>41903126 #
4. raincole ◴[] No.41903001[source]
Statistical models (which "AI" is) have been used to evaluate people's outputs since forever.

Examples: Spam detection, copyrighted material detection, etc.

replies(1): >>41903876 #
5. _heimdall ◴[] No.41903126{3}[source]
This feels like an argument bigger than AI evaluations. All points you raised could very well be issues with humans evaluating other humans to attempt to predict future outcomes.
replies(1): >>41904190 #
6. freilanzer ◴[] No.41903876[source]
But not in cheating or grades, etc. Spam filters are completely different from this.
replies(2): >>41904125 #>>41905424 #
7. baby_souffle ◴[] No.41904125{3}[source]
> But not in cheating or grades, etc. Spam filters are completely different from this.

Really? A spammer is trying to ace a test where my attention is the prize. I don't really see a huge difference between a student/diploma and a spammer/my attention.

Education tech companies have been playing with ML and similar tech that is "AI adjacent" for decades. If you went to school in the US any time after computers entered the class room, you probably had some exposure to a machine generated/scored test. That data was used to tailor lessons to pupil interest/goals/state curricula. Good software also gave instructor feedback about where each student/cohort is struggling or not.

LLMs are just an evolution of tech that's been pretty well integrated into academic life for a while now. Was anything in academia prepared for this evolution? No. But banning it outright isn't going to work

8. smartmic ◴[] No.41904190{4}[source]
They are not wrong. And the art of predicting future outcomes proves to be difficult and fraught with failure. But human evaluation of other humans is more like an open level field to me. A human is accountable for what he or she says or predicts about others, subject to interrogation or social or legal consequences. Not so easy with AI, because it steps out of all these areas - at least many actors using AI do not seem to stay responsible and take on all these mistakes.
replies(1): >>41904505 #
9. _heimdall ◴[] No.41904505{5}[source]
In my experience, we're really bad at holding humans accountable for their predictions too. That may even be a good thing, but I'm less confident that we would be holding LLMs less accountable for their predictions than humans.
10. gs17 ◴[] No.41905424{3}[source]
> But not in cheating or grades

I had both, over a decade ago in high school. Plagiarism detection is the original AI detection, although they usually told you specifically what you were accused of stealing from. A computer-based English course I took over the summer used automated grading to decide if what you wrote was good enough (IIRC they did have a human look over it at some point).

11. PeterisP ◴[] No.41908929[source]
There's the dichotomy of an irresistible force meeting an immovable object - only one of these is possible.

Either there can be an undefeatable AI detector, or an undetectable AI writer, both can't exist in the same universe. And my assumption is that with sufficient advances there could be a fully human-equivalent AI that is not distinguishable from a human in any way, so in that sense being able to detect it will actually never work.