←back to thread

423 points sohkamyung | 1 comments | | HN request time: 0s | source
Show context
roguecoder ◴[] No.45670387[source]
I am curious if LLMs evangelists understand how off-putting it is when they knee-jerk rationalize how badly these tools are performing. It makes it seem like it isn't about technological capabilities: it is about a religious belief that "competence" is too much to ask of either them or their software tools.
replies(7): >>45670776 #>>45670799 #>>45670830 #>>45671500 #>>45671741 #>>45672916 #>>45673109 #
senordevnyc ◴[] No.45670799[source]
I'm curious if LLM skeptics bother to click through and read the details on a study like this, or if they just reflexively upvote it because it confirms their priors.

This is a hit piece by a media brand that's either feeling threatened or is just incompetent. Or both.

replies(1): >>45671281 #
smt88 ◴[] No.45671281[source]
Whether a hitpiece or not, it rhymes with my experience and provides receipts. Can you provide yours?
replies(1): >>45671607 #
1. lyu07282 ◴[] No.45671607[source]
Because yours is anecdotal evidence, a study like this should have a higher bar than that and be useful to support your experience, but it doesn't do that. It doesn't even say what exact models they evaluated ffs