It all depends on the GPU. If it’s something integrated in the CPU it will probably not so better, if it’s a 2000$ dedicated GPU with 48GB of VRAM is will be very powerful for Neural Net computing. NPUs are most often implemented as small, low-power, embedded solutions. Their goal is not to compete with data centers or workstations, it’s to enable some basic “AI” features on portable devices. E.g: “smart” camera with object recognition to give you alerts.
It all depends on the GPU. If it’s something integrated in the CPU it will probably not so better, if it’s a 2000$ dedicated GPU with 48GB of VRAM is will be very powerful for Neural Net computing. NPUs are most often implemented as small, low-power, embedded solutions. Their goal is not to compete with data centers or workstations, it’s to enable some basic “AI” features on portable devices. E.g: “smart” camera with object recognition to give you alerts.