Most active commenters
  • chrismorgan(3)
  • nashashmi(3)

←back to thread

443 points wg0 | 46 comments | | HN request time: 0.434s | source | bottom
1. chrismorgan ◴[] No.45899143[source]
The current title (“Pakistani newspaper mistakenly prints AI prompt with the article”) isn’t correct, it wasn’t the prompt that was printed, but trailing chatbot fluff:

> If you want, I can also create an even snappier “front-page style” version with punchy one-line stats and a bold, infographic-ready layout—perfect for maximum reader impact. Do you want me to do that next?

The article in question is titled “Auto sales rev up in October” and is an exceedingly dry slab of statistic-laden prose, of the sort that LLMs love to err in (though there’s no indication of whether they have or not), and for which alternative (non-prose) presentations can be drastically better. Honestly, if the entire thing came from “here’s tabular data, select insights and churn out prose”… I can understand not wanting to do such drudgework.

replies(9): >>45899255 #>>45899348 #>>45899636 #>>45899711 #>>45899852 #>>45900787 #>>45902114 #>>45903466 #>>45904945 #
2. wg0 ◴[] No.45899255[source]
Thank you, yes that's accurate and I am not sure if article itself is accurate. Don't think so it would have no incorrect stats.

By "AI prompt" I mean "prompted by AI"

Edit: Note about prompt's nature.

replies(1): >>45899554 #
3. layer8 ◴[] No.45899348[source]
The AI is prompting the human here, so the title isn't strictly wrong. ;)
replies(2): >>45900301 #>>45902047 #
4. nashashmi ◴[] No.45899554[source]
It might be better to mention “Dawn newspaper” instead of “Pakistani newspaper”.
replies(2): >>45900269 #>>45903127 #
5. kleene_op ◴[] No.45899636[source]
I guess in the end the journalist didn't feel necessary to impact his readers with punchy one line stats and bold infographic-ready layouts, considering he opted for the first draft.
6. michaelbuckbee ◴[] No.45899711[source]
For years, both the financial and sports news sides of things have generated increasingly templated "articles", this just feels like the latest iteration.
replies(8): >>45899930 #>>45900236 #>>45900320 #>>45900682 #>>45901528 #>>45901990 #>>45902025 #>>45902084 #
7. righthand ◴[] No.45899852[source]
I think AI-Prompt is synonymous with the chat before an LLM prints the intended garbage.
replies(1): >>45902019 #
8. reaperducer ◴[] No.45899930[source]
Legitimate news organizations announce their use of A.I.

I believe the New York Times weather page is automated, but that started before the current "A.I." hype wave.

And I think the A.P. uses LLMs for some of its sports coverage.

9. dredmorbius ◴[] No.45900236[source]
This dates back to at least the late 1990s for financial reports. A friend demoed such a system to me at that time.

Much statistically-based news (finance, business reports, weather, sport, disasters, astronomical events) are heavily formulaic and can at least in large part or initial report be automated, which speeds information dissemination.

Of course, it's also possible to distribute raw data tables, charts, or maps, which ... mainstream news organisations seem phenomenally averse to doing. Even "better" business-heavy publications (FT, Economist, Bloomberg, WSJ) do so quite sparingly.

A few days ago I was looking at a Reuters report on a strategic chokepoint north of the Philippines which it and the US are looking toward to help contain possible Chinese naval operations. Lots of pictures of various equipment, landscapes, and people. Zero maps. Am disappoint.

replies(2): >>45900466 #>>45900813 #
10. nomdep ◴[] No.45900269{3}[source]
Only Pakistanis knew from where the Dawn newspaper is, so the current title is more informative
replies(1): >>45901165 #
11. dwringer ◴[] No.45900301[source]
Gemini has been doing this to me for the past few weeks at the end of basically every single response now, and it often seems to result in the subsequent responses getting off track and lower quality as all these extra tangets start polluting the context. Not to mention how distracting it is as it throws off the reply I was already halfway in the middle of composing by the time I read it.
replies(4): >>45901512 #>>45901950 #>>45901979 #>>45903775 #
12. cantor_S_drug ◴[] No.45900320[source]
Not just that we know from heavy reddit posters that they have branching universe templates for all eventualities, so that they are "ready" whatever the outcome.
13. jrjeksjd8d ◴[] No.45900466{3}[source]
At least in the case of Bloomberg they would like you to pay for that raw data. That's their bread and butter.
replies(1): >>45900674 #
14. dredmorbius ◴[] No.45900674{4}[source]
True.

But there's the approach the Economist takes. For many decades, it's relied on a three-legged revenue model: subscriptions, advertising, and bespoke consulting and research through the Economist Intelligence Unit (EIU). My understanding is that revenues are split roughly evenly amongst these, and that they tend to even out cash-flow throughout economic cycles (advertising is famously pro-cyclical, subscriptions and analysis somewhat less so).

To that extent, the graphs and maps the Economist actually does include in its articles (as well as many of its "special reports") are both teasers and loss-leader marketing for EIU services. I believe that many of the special reports arise out of EIU research.

<https://www.eiu.com/n/>

<https://en.wikipedia.org/wiki/Economist_Intelligence_Unit>

15. jerf ◴[] No.45900682[source]
A non-"AI" template is probably getting filled in with numbers straight from some relevant source. AI may produce something more conversational today but as someone else observed, this is a high-hallucination point for them. Even if they get one statistic right they're pretty inclined to start making up statistics that weren't provided to them at all if they sound good.
16. abdullahkhalids ◴[] No.45900787[source]
The newspaper in question is Pakistan's English language "newspaper of record", which has wide readership.

For some reason, they rarely ever add any graphs or tables to financial articles, which I have never understood. Their readership is all college educated. One time I read an Op-Ed, where the author wrote something like: If you go to this gov webpage, and take the data and put it on excel, and plot this thing vs that thing, you will see X trend.

Why would they not just take the excel graph, clean it up and put it in their article?

replies(2): >>45900936 #>>45902307 #
17. RobotToaster ◴[] No.45900813{3}[source]
Obviously the solution is to use AI to extract the raw data from their AI generated fluff.

It's like the opposite of compression.

replies(1): >>45901999 #
18. IAmBroom ◴[] No.45900936[source]
Because it was BS opinion, dressed in scientifical sounding clothing?
19. rat9988 ◴[] No.45901165{4}[source]
It doesn't matter much in which country it is located actually. It also provides less information.
replies(1): >>45902017 #
20. layer8 ◴[] No.45901512{3}[source]
Occasionally I find it helpful, but it would be good to have the option to remove it from the context.
replies(1): >>45902066 #
21. the__alchemist ◴[] No.45901528[source]
Both categories are and have-been bottom-feeder copy, and have been prior to the prevalence of LLMs.
22. Razengan ◴[] No.45901950{3}[source]
I think AI should present those continuation prompts as dynamic buttons, like "Summarize", "Yes, explain more" etc. based on the AI's last message, like the NPC conversation dialogs in some RPGs
replies(1): >>45902554 #
23. butlike ◴[] No.45901979{3}[source]
Why do you respond to its prompting? It's a machine
replies(1): >>45902261 #
24. mlmonkey ◴[] No.45901990[source]
In the mid-late naughts, there used to be a content farm called "Associated Content". They would get daily lists of top searched terms from various search engines (Yahoo, Dogpile, Altavista, etc. etc.) and for each search term, pay an English major to write a 2-page fluff article. Regardless of what the topic was, they churned out articles by the bushel. Then they place ads on these articles and sat back and watched the dollars roll in.
25. mh- ◴[] No.45901999{4}[source]
>It's like the opposite of compression.

I believe the word is depression, which seems apt when thinking of the idea of people using AI to make content longer and then the readers all using AI to make it shorter again.

26. mh- ◴[] No.45902017{5}[source]
How does including the country of origin provide less information than omitting it?
replies(1): >>45903727 #
27. chrismorgan ◴[] No.45902019[source]
The prompt is the chat before it prints the intended garbage. This is the engagement bait the LLM appends after the intended garbage.
28. shagie ◴[] No.45902025[source]
https://www.npr.org/sections/money/2015/05/20/406484294/an-n...

    ...
    The rules for the race: Both contenders waited for Denny's, the diner company, to come out with an earnings report. Once that was released, the stopwatch started. Both wrote a short radio story and get graded on speed and style.
https://www.wired.com/story/wordsmith-robot-journalist-downl... https://archive.ph/gSdmb

And this has been going on for a while... https://en.wikipedia.org/wiki/Automated_journalism

    StatSheet, an online platform covering college basketball, runs entirely on an automated program. In 2006, Thomson Reuters announced their switch to automation to generate financial news stories on its online news platform. Reuters used a tool called Tracer. An algorithm called Quakebot published a story about a 2014 California earthquake on The Los Angeles Times website within three minutes after the shaking had stopped.
Sports and financial are the two easiest to do since they both work from well structured numeric statistics.
replies(1): >>45902252 #
29. chrismorgan ◴[] No.45902047[source]
I have decided to call it engagement bait.
30. drivers99 ◴[] No.45902066{4}[source]
You can if you script the request yourself, or you could have a front end that lets you cut out those paragraphs from the conversation. I only say that because yesterday I followed this guide: https://fly.io/blog/everyone-write-an-agent/ except I had to figure out how to do it with Gemini API instead. The context is always just (essentially) a list of strings (or "parts" anyway, doesn't have to be strings) that you pass back to the model so you can make the context whatever you like. It shouldn't be too hard to make a frontend that lets you edit the context, and fairly easy to mock up if you just put the request in a script that you add to.
31. ◴[] No.45902084[source]
32. xhkkffbf ◴[] No.45902114[source]
Do we know it was an AI? I realize that it rings with a sycophantic tone that the AIs love to use, but I've worked with some humans who speak the same way. AIs didn't invent brownnosing.
33. simonw ◴[] No.45902252{3}[source]
I like Quakebot as an example of how to do this kind of thing ethically and with integrity: https://www.latimes.com/people/quakebot

> Quakebot is a software application developed by the Los Angeles Times to report the latest earthquakes as fast as possible. The computer program reviews earthquake notices from the U.S. Geological Survey and, if they meet certain criteria, automatically generates a draft article. The newsroom is alerted and, if a Times editor determines the post is newsworthy, the report is published.

replies(1): >>45903365 #
34. dwringer ◴[] No.45902261{4}[source]
Because if I don't, it tends to misinterpret the next thing I say because it reads that as an answer to the question it just asked me.
replies(1): >>45902478 #
35. pseudony ◴[] No.45902307[source]
Maybe the model just wasn’t multi-modal back then ;)
36. catlifeonmars ◴[] No.45902478{5}[source]
Try one-shotting. Rather than a continuous conversation, refine your initial prompt and restart.
37. xnorswap ◴[] No.45902554{4}[source]
Claude code already does this, it'll present a series of questions with pre-set answers, and the opportunity to answer "custom: <free text>"
38. wg0 ◴[] No.45903127{3}[source]
Nobody outside Pakistan knows Dawn even though it is the newspaper that was founded by Muhammad Ali Jinnah (considered founding father of the nation) and one of the largest and most prestigious as well.
replies(1): >>45903676 #
39. mschuster91 ◴[] No.45903365{4}[source]
> The computer program reviews earthquake notices from the U.S. Geological Survey

Probably a service that is provided to the general public for free, similar to NOAA and weather data - so chances are rather high it ends up on the chopping block or for-money only.

40. lawlessone ◴[] No.45903466[source]
>” and is an exceedingly dry slab of statistic-laden prose

Thats the kinda thing i'd be worried AI would say make up a stat in, something really boring that most people aren't going to follow up on to verify.

41. nashashmi ◴[] No.45903676{4}[source]
It is like the NYT for the country. But the relevant detail here is the printing of the prompt in a nationally recognized newspaper. The brand, as local as it maybe, still provides more context than some random newspaper in a foreign country would.

And I have ran into Dawn newspaper on google news frontpage several times, usually on entertainment stuff.

42. nashashmi ◴[] No.45903727{6}[source]
Does omitting the brand provide more information? Newspaper brand is more relevant than the country.
replies(1): >>45904330 #
43. lubujackson ◴[] No.45903775{3}[source]
Add "Complete this request as a single task and do not ask any follow-up questions." Or some variation of that. They keep screwing with default behavior, but you can explicitly direct the LLM to override it.
replies(1): >>45905986 #
44. ◴[] No.45904330{7}[source]
45. SoftTalker ◴[] No.45904945[source]
> it wasn’t the prompt that was printed, but trailing chatbot fluff

I've seen that sort of thing copy/pasted in several emails at work, usually ones that are announcing something on a staff email list.

Sort of a givaway that the email isn't very important.

46. astrange ◴[] No.45905986{4}[source]
That doesn't help GPT-5; it /really/ wants to suggest follow-ups and ignored me telling it not to.