←back to thread

Tim Bray on Grokipedia

(www.tbray.org)
175 points Bogdanp | 2 comments | | HN request time: 0.445s | source
Show context
tptacek ◴[] No.45777117[source]
Why give it oxygen?
replies(6): >>45777142 #>>45777160 #>>45777311 #>>45777327 #>>45777329 #>>45777411 #
meowface ◴[] No.45777160[source]
To play devil's advocate: Grok has historically actually been one of the biggest debunkers of right-wing misinformation and conspiracy theories on Twitter, contrary to popular conception. Elon keeps trying to tweak its system prompt to make it less effective at that, but Grokipedia was worth an initial look from me out of curiosity. It took me 10 seconds to realize it was ideologically-motivated garbage and significantly more right-biased than Wikipedia is left-biased.

(Unfortunately, Reply-Grok may have been successfully partially lobotomized for the long term, now. At the time of writing, if you ask grok.com about the 2020 election it says Biden won and Trump's fraud claims are not substantiated and have no merit. If you @grok in a tweet it now says Trump's claims of fraud have significant merit, when previously it did not. Over the past few days I've seen it place way too much charity in right-wing framings in other instances, as well.)

replies(4): >>45777225 #>>45777240 #>>45777294 #>>45777386 #
pstuart ◴[] No.45777225[source]
The problem of debunking right-wing misinformation is that it doesn't seem to matter. The consumers of that misinformation want it and those of us who think it's bad for society already know that its garbage.

It feels like we've reached Peak Stupidity but it's clear it can (and likely will) get much worse with AI videos.

replies(6): >>45777259 #>>45777318 #>>45777362 #>>45777581 #>>45778665 #>>45784645 #
bawolff ◴[] No.45777318[source]
I think there is a problem sometimes that "debunkers" are often more interested in scoring points with secondary audiences (i.e. people who already agree with them) than actually convincing the people who believe the misinformation.

Most people who believe bullshit were convinced by something. It might not have been fully rational but there is usually a kernel of something there that triggered that belief. They also probably have heard at least the surface level version of the oppising argument at some point before. Too many debunkers just reiterate the surface argument without engaging with whatever is convincing their opponent. Then when it doesn't land they complain their opponent is brainwashed. Which sometimes might even be true, but sometimes their argument just misses the point of why their opponent believes what they do.

replies(3): >>45777407 #>>45777444 #>>45777613 #
1. meowface ◴[] No.45777613[source]
This is very, very true. The best debunkers avoid being hostile and make the other side feel like they're being heard and that their feelings and fears are being validated. And they do it in a way that feels honest and not condescending and patronizing (like talking to a child). They make frequent (sincere) concessions and hedges and find as much common ground as they can.

Although he's more populist-left and I'm more establishment-liberal (and so I might find him a bit overly conciliatory with certain conspiracy theorists), Andrew Callaghan of Channel 5/All Gas No Brakes demonstrates a good example of this in the first few minutes of this video: https://youtu.be/QU6S3Cbpk-k?t=38

replies(1): >>45778252 #
2. pstuart ◴[] No.45778252[source]
I'm a fan of Andrew and am impressed by how he's evolved from documenting stupid kids to actually reporting on issues of interest.

I agree that one catches more flies with honey rather than vinegar, but many times it doesn't matter what you say or how you say it -- they're gonna stick to their guns. A prime example of this is in Jordan Klepper interviews where he asks Trump supporters how they feel about something horrible that Biden did, to which they express their indignation; then he reveals that it was actually Trump and they dismiss it because it "doesn't matter".