←back to thread

172 points marban | 1 comments | | HN request time: 0s | source
Show context
bearjaws ◴[] No.40052158[source]
The focus on TOPS seems a bit out of line with reality for LLMs. TOPs doesn't matter for LLMs if your memory bandwidth can't keep up. Since it doesn't have quad channel memory mentioned I guess it's still dual channel?

Even top of the line DDR5 is around 128GB/s vs a M1 at 400GB/s.

At the end of the day, it still seems like AI in consumer chips is chasing a buzzword, what is the killer feature?

On mobile there are image processing benefits and voice to text, translation... but on desktop those are no where near common use cases.

replies(3): >>40052204 #>>40052260 #>>40052353 #
VHRanger ◴[] No.40052260[source]
The killer feature is presumably inference at the edge, but I don't see that being used on desktop much at all right now.

Especially since most desktop applications people use are web apps. Of the native apps people use that leverage this sort of stuff, almost all are GPU accelerated already (eg. image and video editing AI tools)

replies(1): >>40052360 #
jzig ◴[] No.40052360{3}[source]
What does “at the edge” mean here?
replies(4): >>40052515 #>>40052529 #>>40052531 #>>40052991 #
1. PeterSmit ◴[] No.40052529{4}[source]
Not in the cloud.