←back to thread

172 points marban | 1 comments | | HN request time: 0s | source
Show context
bearjaws ◴[] No.40052158[source]
The focus on TOPS seems a bit out of line with reality for LLMs. TOPs doesn't matter for LLMs if your memory bandwidth can't keep up. Since it doesn't have quad channel memory mentioned I guess it's still dual channel?

Even top of the line DDR5 is around 128GB/s vs a M1 at 400GB/s.

At the end of the day, it still seems like AI in consumer chips is chasing a buzzword, what is the killer feature?

On mobile there are image processing benefits and voice to text, translation... but on desktop those are no where near common use cases.

replies(3): >>40052204 #>>40052260 #>>40052353 #
futureshock ◴[] No.40052353[source]
Upscaling for gaming or video.

Local context aware search

Offline Speech to text and TTS

Offline generation of clip art or stock images for document editing

Offline LLM that can work with your documents as context and access application and OS APIs

Improved enemy AI in gaming

Webcam effects like background removal or filters.

Audio upscaling and interpolation like for bad video call connections.

replies(2): >>40052491 #>>40052497 #
bearjaws ◴[] No.40052491[source]
> Upscaling for gaming or video.

Already exists on all three major GPU manufacturers, and it definitely makes sense as a GPU workload.

> Local context aware search

You don't need an AI processor to do this, Windows search used to work better and had even less compute resources to work with.

> Offline Speech to text and TTS

See my point about not a very common use case for desktops & laptops vs cell phones.

> Offline LLM that can work with your documents as context and access application and OS APIs

Maybe for some sort of background task or only using really small models <13B parameters. Anything real time is going to run at 1-2t/s with a large model.

Small models are pretty terrible though, I doubt people want even more incorrect information and hallucinations.

> Improved enemy AI in gaming

See Ageia PhysX

> Webcam effects like background removal or filters.

We already have this without NPUs.

> Audio upscaling and interpolation like for bad video call connections.

I could see this, or noise cancellation.

replies(3): >>40052653 #>>40052700 #>>40052949 #
1. futureshock ◴[] No.40052949[source]
> Upscaling for gaming or video. > Already exists on all three major GPU manufacturers, and it definitely makes sense as a GPU workload. These AMD chips are APUs that are often the only GPU, not every user will have a dedicated GPU.

> Local context aware search > You don't need an AI processor to do this, Windows search used to work better and had even less compute resources to work with. You could still improve it with increased natural language understanding instead of simple keyword. “Give me all documents about dogs” instead of searching for each breed as a keyword.

> Offline Speech to text and TTS > See my point about not a very common use case for desktops & laptops vs cell phones. Maybe not for you but accessibility is a key feature for many users. You think blind users should suffer through bad TTS?

> Offline LLM that can work with your documents as context and access application and OS APIs > Maybe for some sort of background task or only using really small models <13B parameters. Anything real time is going to run at 1-2t/s with a large model. > Small models are pretty terrible though, I doubt people want even more incorrect information and hallucinations. Small model have been improving and better capabilities in consumer chips will allow larger models to run faster.

> Improved enemy AI in gaming > See Ageia PhysX Surely you’re not suggesting that enemy AI is solved problem in gaming?

> Webcam effects like background removal or filters. > We already have this without NPUs. Sure but it could go from obvious and distracting to seamless and convincing.

> Audio upscaling and interpolation like for bad video call connections. I could see this, or noise cancellation.