←back to thread

395 points pseudolus | 9 comments | | HN request time: 0s | source | bottom
1. SamBam ◴[] No.43633756[source]
I feel like Anthropic has an incentive to minimize how much students use LLMs to write their papers for them.

In the article, I guess this would be buried in

> Students also frequently used Claude to provide technical explanations or solutions for academic assignments (33.5%)—working with AI to debug and fix errors in coding assignments, implement programming algorithms and data structures, and explain or solve mathematical problems.

"Write my essay" would be considered a "solution for academic assignment," but by only referring to it obliquely in that paragraph they don't really tell us the prevalence of it.

(I also wonder if students are smart, and may keep outright usage of LLMs to complete assignments on a separate, non-university account, not trusting that Anthropic will keep their conversations private from the university if asked.)

replies(3): >>43634021 #>>43634024 #>>43642992 #
2. radioactivist ◴[] No.43634021[source]
Most of their categories have straightforward interpretations in terms of students using the tool to cheat. They don't seem to want to/care to analyze that further and determine which are really cheating and which are more productive uses.

I think that's a bit telling on their motivations (esp. given their recent large institutional deals with universities).

replies(1): >>43634130 #
3. vunderba ◴[] No.43634024[source]
Exactly. There's a big difference between a student having a back-and-forth dialogue with Claude around "the extent to which feudalism was one of the causes of the French Revolution.", versus another student using their smartphone to take a snapshot of the actual homework assignment, pasting it into Claude and calling it a day.
replies(1): >>43634222 #
4. SamBam ◴[] No.43634130[source]
Indeed. I called out the second-top category, but you could look at the top category as well:

> We found that students primarily use Claude to create and improve educational content across disciplines (39.3% of conversations). This often entailed designing practice questions, editing essays, or summarizing academic material.

Sure, throwing a paragraph of an essay at Claude and asking it to turn it into a 3-page essay could have been categorized as "editing" the essay.

And it seems pretty naked the way they lump "editing an essay" in with "designing practice questions," which are clearly very different uses, even in the most generous interpretation.

I'm not saying that the vast majority of students do use AI to cheat, but I do want to say that, if they did, you could probably write this exact same article and tell no lies, and simply sweep all the cheating under titles like "create and improve educational content."

5. PeterStuer ◴[] No.43634222[source]
From what I could observe, the latter is endemic amongst high school students. And don't kid yourself. For many it is just a step up from copy/pasting the first Google result.

They never could be arsed to learn how to input their assignments into Wolfram Alpha. It was always the ux/ui effort that held them back.

replies(1): >>43642468 #
6. chii ◴[] No.43642468{3}[source]
THe question is would those students have done any better or worse if there hadn't been LLM for them to "copy" off?

In other words, is the school certificaftion meant to distinguish those who genuinely learnt, or was it merely meant to signal (and thus, those who used to copy pre-llm are going to do the same, and thus reach the same level of certification regardless of whether they learnt or not)?

7. ignoramous ◴[] No.43642992[source]
> feel like Anthropic has an incentive to minimize how much students use LLMs to write their papers for them

You're right.

Quite incredibly, they also do the opposite, in that they hype-up / inflate the capability of their LLMs. For instance, they've categorised "summarisation" as "high-order thinking" ("Create", per Bloom's Taxonomy). It patently isn't. Comical they'd not only think so, but also publicly blog about it.

replies(1): >>43644703 #
8. xpe ◴[] No.43644703[source]
> Bloom's taxonomy is a framework for categorizing educational goals, developed by a committee of educators chaired by Benjamin Bloom in 1956. ... In 2001, this taxonomy was revised, renaming and reordering the levels as Remember, Understand, Apply, Analyze, Evaluate, and Create. This domain focuses on intellectual skills and the development of critical thinking and problem-solving abilities. - Wikipedia

This context is important: this taxonomy did not emerge from artificial intelligence nor cognitive science. So its levels are unlikely to map to how ML/AI people assess the difficulty of various categories of tasks.

Generative models are, by design, fast (and often pretty good) at generation (creation), but this isn't the same standard that Bloom had in mind with his "creation" category. Bloom's taxonomy might be better described as a hierarchy: proper creation draws upon all the layers below it: understanding, application, analysis, and evaluation.

replies(1): >>43644912 #
9. xpe ◴[] No.43644912{3}[source]
Here is one key take-away, phrased as a question: when a student uses an LLM for "creation", are underlying aspects (understanding, application, analysis, and evaluation) part of the learning process?