←back to thread

38 points dxdxdt | 2 comments | | HN request time: 0.578s | source
Show context
forgotpwd16 ◴[] No.46253035[source]
>After reviewing the core defrag logic myself, I've come to a conclusion that it's AI slop.

Will call it a human slop. AI may've given them some code but they certainly haven't use it fully. I uploaded the defrag.c in ChatGPT asking to review on performance/correctness/safety and pointed the sames issues as you (alongside bunch of others but not interested at the moment to review them).

replies(2): >>46253490 #>>46260702 #
dxdxdt ◴[] No.46260702[source]
I did the same. Was genuinely curious. Didn't get much from it. I'm still confused.

The code base is huge for an LLM to handle, perhaps it was generated over multiple prompts idk. Not sure if someone can train a model on the kernel code or exfatprogs and generate the code. I doubt someone with such expertise would even go through the process when they can just write the code themselves which is much easier.

replies(1): >>46262383 #
1. forgotpwd16 ◴[] No.46262383[source]
Multiple prompts are mandatory for anything non-trivial or/and larger in scope. That said, the exfatprogs repo is ~60k tokens (in 8k LOC) and Linux's exfat driver* is ~40k tokens (in 6k LOC). So directly relevant code is ~100k tokens (in 14k LOC). Not that extensive.

>Not sure if someone can train a model on the kernel code or exfatprogs and generate the code.

They can certainly finetune such a model. Not a crazy idea, just computationally expensive. (But less expensive than training from scratch.)

*Of course Linux driver also uses many includes so if consider those alongside linked code the number goes significantly up.

replies(1): >>46262665 #
2. dxdxdt ◴[] No.46262665[source]
> just computationally expensive. (But less expensive than training from scratch.)

Model training requires GPUs w/ 1kW TDP. I can shit out code on noodles and red bulls. Not sure about the quality, but still way less energy :)

Jokes aside, the defrag program probably was a slob to some extent.