> Upscaling for gaming or video.
Already exists on all three major GPU manufacturers, and it definitely makes sense as a GPU workload.
> Local context aware search
You don't need an AI processor to do this, Windows search used to work better and had even less compute resources to work with.
> Offline Speech to text and TTS
See my point about not a very common use case for desktops & laptops vs cell phones.
> Offline LLM that can work with your documents as context and access application and OS APIs
Maybe for some sort of background task or only using really small models <13B parameters. Anything real time is going to run at 1-2t/s with a large model.
Small models are pretty terrible though, I doubt people want even more incorrect information and hallucinations.
> Improved enemy AI in gaming
See Ageia PhysX
> Webcam effects like background removal or filters.
We already have this without NPUs.
> Audio upscaling and interpolation like for bad video call connections.
I could see this, or noise cancellation.