Most active commenters

    ←back to thread

    395 points pseudolus | 13 comments | | HN request time: 1.215s | source | bottom
    Show context
    dtnewman ◴[] No.43633873[source]
    > A common question is: “how much are students using AI to cheat?” That’s hard to answer, especially as we don’t know the specific educational context where each of Claude’s responses is being used.

    I built a popular product that helps teachers with this problem.

    Yes, it's "hard to answer", but let's be honest... it's a very very widespread problem. I've talked to hundreds of teachers about this and it's a ubiquitous issue. For many students, it's literally "let me paste the assignment into ChatGPT and see what it spits out, change a few words and submit that".

    I think the issue is that it's so tempting to lean on AI. I remember long nights struggling to implement complex data structures in CS classes. I'd work on something for an hour before I'd have an epiphany and figure out what was wrong. But that struggling was ultimately necessary to really learn the concepts. With AI, I can simply copy/paste my code and say "hey, what's wrong with this code?" and it'll often spot it (nevermind the fact that I can just ask ChatGPT "create a b-tree in C" and it'll do it). That's amazing in a sense, but also hurts the learning process.

    replies(34): >>43633957 #>>43634006 #>>43634053 #>>43634075 #>>43634251 #>>43634294 #>>43634327 #>>43634339 #>>43634343 #>>43634407 #>>43634559 #>>43634566 #>>43634616 #>>43634842 #>>43635388 #>>43635498 #>>43635830 #>>43636831 #>>43638149 #>>43638980 #>>43639096 #>>43639628 #>>43639904 #>>43640528 #>>43640853 #>>43642243 #>>43642367 #>>43643255 #>>43645561 #>>43645638 #>>43646665 #>>43646725 #>>43647078 #>>43654777 #
    1. 0xffff2 ◴[] No.43634842[source]
    >For many students, it's literally "let me paste the assignment into ChatGPT and see what it spits out, change a few words and submit that".

    Does that actually work? I'm long past having easy access to college programming assignments, but based on my limited interaction with ChatGPT I would be absolutely shocked if it produced output that was even coherent, much less working code given such an approach.

    replies(5): >>43634915 #>>43634958 #>>43635406 #>>43639090 #>>43644616 #
    2. StefanBatory ◴[] No.43634915[source]
    I have some subjects, at Masters - that are solvable by one prompt. One.

    Quality of CS/Software Engineering programs vary that much.

    replies(1): >>43640607 #
    3. bongodongobob ◴[] No.43634958[source]
    Why are you asking? Go try it. And yes, depending on the task, it does.
    replies(1): >>43635747 #
    4. rufus_foreman ◴[] No.43635406[source]
    >> Does that actually work?

    Sure. Works in my IDE. "Create a linked list implementation, use that implementation in a method to reverse a linked list and write example code to demonstrate usage".

    Working code in a few seconds.

    I'm very glad I didn't have access to anything like that when I was doing my CS degree.

    replies(1): >>43639248 #
    5. 0xffff2 ◴[] No.43635747[source]
    As I said, I'm not a student, so I don't have access to a homework assignment to paste in. Ironically I have pretty much everything I ever submitted for my undergrad, but it seems like I absolutely never archived the assignments for some reason.
    replies(1): >>43639240 #
    6. izacus ◴[] No.43639090[source]
    It doesn't matter who coherent the output is - the students will paste it anyway, then fail the assignment (and you need to deal with grading it) and then complain to parents and school board that you're incompetent because you're failing the majority of the class.

    Your post is based in a misguided idea that students actually care about some basic quality of their work.

    7. bongodongobob ◴[] No.43639240{3}[source]
    I was able to get ~80% one shots on Advent of Code with 4o up to about day 12 iirc.
    8. krull10 ◴[] No.43639248[source]
    Yeah, and forget about giving skeleton code to students they should fill in; using an AI can quite frequently completely ace a typical undergraduate level assignment. I actually feel bad for people teaching programming courses, as the only real assessment one can now do is in-class testing without computers, but that is a strange way to test students’ ability to write and develop code to solve certain classes of problems…
    replies(2): >>43639556 #>>43641555 #
    9. fn-mote ◴[] No.43639556{3}[source]
    Hopefully someone is thinking about adapting the assessments. Asking questions that focus on a big picture understanding instead of details on those in-class tests.
    10. porridgeraisin ◴[] No.43640607[source]
    Yeah. On the other hand, "implement boruvkas MST algorithm in cuda such that only the while(numcomponents > 1) loop runs on the CPU, and everything else runs in the gpu. Memcpy everything onto the gpu first and only transfer back the count each iteration/keep it in pinned memory"

    It never gets it right, even after many reattempts in cursor. And even if it gets it right, it doesn't do the parallelization effectively enough - it's a hard problem to parallelize.

    11. i_am_proteus ◴[] No.43641555{3}[source]
    Why do the in-class testing without computers?

    We use an airgapped lab (it has LAN and a local git server for submissions, no WAN) to give coding assessments. It works.

    replies(1): >>43654904 #
    12. currymj ◴[] No.43644616[source]
    since late 2024/early 2025 it now is the case, especially with a reasoning model like Sonnet 3.7, DeepSeek-r1, o3, Gemini 2.5, etc., and especially if you upload the textbook, slides, etc alongside the homework to be cheated on.

    most normal-difficulty undergraduate assignments are now doable reliably by AI with little to no human oversight. this includes both programming and mathematical problem sets.

    for harder problem sets that require some insight, or very unstructured larger-scale programming projects, it wouldn't work so reliably.

    but easier homework assignments serve a valid purpose to check understanding, and now they are no longer viable.

    13. chipsrafferty ◴[] No.43654904{4}[source]
    At my college, we did in-class testing with psuedocode, because we were being tested on concepts, not specific programming languages or syntax.