Most active commenters
  • boomskats(6)
  • eightysixfour(3)
  • almostgotcaught(3)

←back to thread

486 points dbreunig | 12 comments | | HN request time: 0.002s | source | bottom
Show context
eightysixfour ◴[] No.41863546[source]
I thought the purpose of these things was not to be fast, but to be able to run small models with very little power usage? I have a newer AMD laptop with an NPU, and my power usage doesn't change using the video effects that supposedly run on it, but goes up when using the nvidia studio effects.

It seems like the NPUs are for very optimized models that do small tasks, like eye contact, background blur, autocorrect models, transcription, and OCR. In particular, on Windows, I assumed they were running the full screen OCR (and maybe embeddings for search) for the rewind feature.

replies(7): >>41863632 #>>41863779 #>>41863821 #>>41863886 #>>41864628 #>>41864828 #>>41869772 #
boomskats ◴[] No.41863779[source]
That's especially true because yours is a Xilinx FPGA. The one that they just attached to the latest gen mobile ryzens is 5x more capable too.

AMD are doing some fantastic work at the moment, they just don't seem to be shouting about it. This one is particularly interesting https://lore.kernel.org/lkml/DM6PR12MB3993D5ECA50B27682AEBE1...

edit: not an FPGA. TIL. :'(

replies(5): >>41863852 #>>41863876 #>>41864048 #>>41864435 #>>41865733 #
errantspark ◴[] No.41863852[source]
Wait sorry back up a bit here. I can buy a laptop that has a daughter FPGA in it? Does it have GPIO??? Are we seriously building hardware worth buying again in 2024? Do you have a link?
replies(2): >>41863959 #>>41864293 #
1. eightysixfour ◴[] No.41863959[source]
It isn't as fun as you think - they are setup for specific use cases and quite small. Here's a link to the software page: https://ryzenai.docs.amd.com/en/latest/index.html

The teeny-tiny "NPU," which is actually an FPGA, is 10 TOPS.

Edit: I've been corrected, not an FPGA, just an IP block from Xilinx.

replies(2): >>41864036 #>>41864062 #
2. wtallis ◴[] No.41864036[source]
It's not a FPGA. It's an NPU IP block from the Xilinx side of the company. It was presumably originally developed to be run on a Xilinx FPGA, but that doesn't mean AMD did the stupid thing and actually fabbed a FPGA fabric instead of properly synthesizing the design for their laptop ASIC. Xilinx involvement does not automatically mean it's an FPGA.
replies(2): >>41864064 #>>41864111 #
3. boomskats ◴[] No.41864062[source]
Yes, the one on the ryzen 7000 chips like the 7840u isn't massive, but that's the last gen model. The one they've just released with the HX370 chip is estimated at 50 TOPS, which is better than Qualcomm's ARM flagship that this post is about. It's a fivefold improvement in a single generation, it's pretty exciting.

A̵n̵d̵ ̵i̵t̵'̵s̵ ̵a̵n̵ ̵F̵P̵G̵A̵ It's not an FPGA

replies(1): >>41864248 #
4. eightysixfour ◴[] No.41864064[source]
Thanks for the correction, edited.
5. boomskats ◴[] No.41864111[source]
Do you have any more reading on this? How come the XDNA drivers depend on Xilinx' XRT runtime?
replies(2): >>41864232 #>>41864296 #
6. almostgotcaught ◴[] No.41864232{3}[source]
because XRT has a plugin architecture: XRT<-shim plugin<-kernel driver. The shims register themselves with XRT. The XDNA driver repo houses both the shim and the kernel driver.
replies(1): >>41864611 #
7. almostgotcaught ◴[] No.41864248[source]
> And it's an FPGA.

nope it's not.

replies(1): >>41864925 #
8. wtallis ◴[] No.41864296{3}[source]
It would be surprising and strange if AMD didn't reuse the software framework they've already built for doing AI when that IP block is instantiated on an FPGA fabric rather than hardened in an ASIC.
replies(1): >>41864630 #
9. boomskats ◴[] No.41864611{4}[source]
Thanks, that makes sense.
10. boomskats ◴[] No.41864630{4}[source]
Well, I'm irrationally disappointed, but thanks. Appreciate the correction.
11. boomskats ◴[] No.41864925{3}[source]
I've just ordered myself a jump to conclusions mat.
replies(1): >>41865072 #
12. almostgotcaught ◴[] No.41865072{4}[source]
Lol during grad school my advisor would frequently cut me off and try to jump to a conclusion, while I was explaining something technical often enough he was wrong. So I did really buy him one (off eBay or something). He wasn't pleased.