Open Menu
AllLocalCommunitiesAbout
lotide
AllLocalCommunitiesAbout
Login

My new laptop chip has an 'AI' processor in it, and it's a complete waste of space

⁨180⁩ ⁨likes⁩

Submitted ⁨⁨1⁩ ⁨day⁩ ago⁩ by ⁨cm0002@piefed.world⁩ to ⁨technology@lemmy.zip⁩

https://www.pcgamer.com/hardware/my-new-laptop-chip-has-an-ai-processor-in-it-and-its-a-complete-waste-of-space/

source

Comments

Sort:hotnewtop
  • corroded@lemmy.world ⁨1⁩ ⁨day⁩ ago

    I have to wonder if NPUs are just going to eventually become a normal part of the instruction set.

    When SIMD was first becoming a thing, it was advertised as accelerating “multimedia,” as that was the hot buzzword of the 1990s. Now, SIMD instructions are used everywhere, any place there is a benefit from processing an array of values in parallel.

    I could see NPUs becoming the same. Developers start using NPU instructions, and the compiler can “NPU-ify” scalar code when it thinks it’s appropriate.

    NPUs are advertised for “AI,” but they’re really just a specialized math coprocessor. I don’t really see this as a bad thing to have. Surely there are plenty of other uses.

    source
    • olafurp@lemmy.world ⁨3⁩ ⁨hours⁩ ago

      It’s tricky to use in programming though for non neural network math, I can see it used in video compression and decompression or some very specialised video game math.

      Video game AI could be a big one though where difficulty would be AI based instead of just stat modifiers.

      source
    • Dudewitbow@lemmy.zip ⁨1⁩ ⁨day⁩ ago

      The problem that (local) ai has at the current moment is that its not just a single type of compute, and because of that, breaks usefulness in the pool of what you can do with it.

      on the Surface level, “AI” is a mixture of what is essentially FP16, FP8, and INT8 accelerators, and different implementations have been using different ones. NPUs are basically INT8 only, while GPU intensive ones are FP based, making them not inherently cross compatible.

      It forces devs to either think of the NPUs themselves with small things (e.g background blur with camera) as there isn’t any consumer level chip with a massive INT8 co processor except for the PS5 Pro (300 TOPS INT8, which compared to laptop cpus, have a 50 TOPs, so on a completely different league, PS5 Pro uses it to upscale)

      source
    • addie@feddit.uk ⁨1⁩ ⁨day⁩ ago

      SIMD is pretty simple really, but it’s been 30 years since it’s been a standard-ish feature in CPUs, and modern compilers are “just about able to sometimes” use SIMD if you’ve got a very simple loop with fixed endpoints that might use it. It’s one thing that you might fall back to writing assembly to use - the FFmpeg developers had an article not too long ago about getting a 10% speed improvement by writing all the SIMD by hand.

      Using an NPU means recognising algorithms that can be broken down into parallelizable, networkable steps with information passing between cells. Basically, you’re playing a game of TIS-100 with your code. It’s fragile and difficult, and there’s no chance that your compiler will do that automatically.

      Best thing to hope for is that some standard libraries can implement it, and then we can all benefit. It’s an okay tool for ‘jobs that can be broken down into separate cells that interact’, so some kinds of image processing, maybe things like liquid flow simulations. There’s a very small overlap between ‘things that are just algorithms that the main CPU would do better’ and ‘things that can be broken down into many many simple steps that a GPU would do better’ where an NPU really makes sense, tho.

      source
    • pftbest@sh.itjust.works ⁨19⁩ ⁨hours⁩ ago

      I agree, we should push more to get the open and standardized API for these accelerators, better drivers and better third party software support. As long as the manufacturers keep them locked and proprietary, we won’t be able to use them outside of niche copilot features no one wants anyway.

      source
  • wuffah@lemmy.world ⁨1⁩ ⁨day⁩ ago

    Microsoft, Google, and Apple are all quietly putting these and the software infrastructure in place in their operating systems to do on device classification of content: Windows Recall, Google Secure Core, and Apple Intelligence. These services are obsequiously marketed as being for your benefit, while are all privacy and surveillance nightmares. When the security breaking features of these systems are mentioned, each company touts some convoluted workaround to justify the tech.

    The real reason is to watch everything you do in your device, use the tensor hardware you may or may not know that you purchased to analyze the data locally, then export and sell that “anonymized” information to advertisers and the government. All while cryptographically tying the data to your device and the device to you for “security”. This enables mass surveillance, digital rights management, and target advertised previously unseen.

    Each of these providers already scans everything you upload to their services:

    Microsoft has been caught using your stored passwords to decrypt uploaded archives.

    Apple developed client side scanning for CSAM before backlash shut it down, and they already locally scan your photos with a machine learning classification algorithm whether you like it or not. You can’t turn it off.

    Google recently implemented local content scanning with SecureCore to protect you from unwanted content.

    I would rather saw off my own nuts with a rusty spork before willfully purchasing a device with one of these NPUs on-device. I fear that in the next 5-10 years, you won’t be able to avoid them.

    ** Do you trust the rising fascist regimes and their tech lackeys in America and the UK to use this information morally and responsibly?**

    Do you really believe that these features that no one asked for are and that you cannot disable are for your benefit?

    source
    • sturger@sh.itjust.works ⁨20⁩ ⁨hours⁩ ago

      They’re not selling chips. They’re selling stock.

      source
    • Clent@lemmy.dbzer0.com ⁨20⁩ ⁨hours⁩ ago

      Microsoft, Google, and Apple are all quietly integrating NPUs into their devices, and implementing the software infrastructure in their operating systems to do on-device classification of content:

      Too late. Apple has had an NPU in their phones since 2017. It’s been standard on flagship android devices since 2020 with various amd and intel processors starting around the same time.

      I’m not here to defend LLMs or AI feature but for a comment to start with such a misinformed assessment of the state of reality reminds me of someone spouting off about chemtrails and weather machines.

      At least NPUs actually exist.

      If someone wants to avoid this stuff they are going to need to pick an open source platform that does not use these processor features.

      source
      • wuffah@lemmy.world ⁨2⁩ ⁨hours⁩ ago

        You’re right, it appears that NPU hardware was introduced in the iPhone 8, in 2017, and I say this typing away from an Apple device. Sounds like I’ll be busy this weekend with that rusty spork.

        However, it’s the recent generation of NPUs that provide the processing power needed to run the expanded scanning services implemented by these OS’s. That’s why Apple Intelligence requires an iPhone 15, and Microsoft is hawking Recall laptops.

        While I admit the pins-and-string-on-a-corkboard tone, I don’t think we actually disagree on anything here. Eventually, an open source platform will become the only way to avoid this kind of hardware enabled surveillance.

        source
  • Mwa@thelemmy.club ⁨1⁩ ⁨day⁩ ago

    Am curious if NPUS can be used in Ollama or Local LLMS,if it can’t then its completely useless also useless if you don’t use AI at all.

    source
    • girsaysdoom@sh.itjust.works ⁨18⁩ ⁨hours⁩ ago

      This might partially answer your question: https://github.com/ollama/ollama/issues/5186.

      It looks like the answer is, it depends on what you want to run as some configs are partially supported but there’s no clear cut support yet?

      source
      • sheogorath@lemmy.world ⁨16⁩ ⁨hours⁩ ago

        I tried running some models on an Intel 155h NPU and the performance is actually worse than using the CPU directly for inference. However, it wins on power consumption front IIRC.

        source
  • irotsoma@lemmy.blahaj.zone ⁨23⁩ ⁨hours⁩ ago

    I mean, even if the NPU space can’t be replaced by more useful components easily or cheaply, just removing it is sure to save a small amount of power which equates to a possibly not so small amount of heat that needs to be dissipated, which takes not insignificant amounts of and/or requires slowing the system down. Additionally, the pathways likely could be placed to create less interference with each other and direct heat transfer which is likely to mean more stability overall.

    Of course without a comparable processor without the NPU to compare to, these are really difficult things to quantify, but are true of nearly all compact chips on power sensitive platforms.

    source
    • pftbest@sh.itjust.works ⁨19⁩ ⁨hours⁩ ago

      No, it will not save any power at all. The power is only consumed during switching, so when some module is properly clock gated, it will not consume any power. There are many parts of the chip that are dark, for example the full CPU core could be disabled for various reasons and it does not affect power consumption when it’s dark. Maybe you know the Steam Deck, it is a battery operated device with the best power efficiency in its class. But what people don’t know is that it has more than 20% of its chip area disabled, as it relates to stereoscopic displays, because the same exact chip is also used by some AR or VR goggles, I forgot the name.

      Also in general the modern chips are more limited by thermal rather than space. So realistically, even if you remove the NPU, you won’t be able to place anything high power there anyways, maybe you can put a couple hundred K of sram for cache in its place, but it won’t matter much in the end.

      source
  • TropicalDingdong@lemmy.world ⁨1⁩ ⁨day⁩ ago

    would he rip the cuda cores out of his Nvidia GPU?

    source
  • CixoUwU@lemmy.cixoelectronic.pl ⁨1⁩ ⁨day⁩ ago

    Yes, i agree and if it must run neural network it could do it on GPU, NPU is not necesary.

    source
    • Endmaker@ani.social ⁨1⁩ ⁨day⁩ ago

      Someone with the expertise should correct me if I am wrong; it’s been 4-5 years since I learnt about NPUs during my internship so I am very rusty:

      You don’t even need a GPU if all you want to do is to run (i.e. perform inference using) a neural network (abbreviating it to NN). Just a CPU would do if the NN is sufficiently lightweight. The GPU is only needed to speed up the training of NNs.

      The thing is, the CPU is a general-purpose processor, so it won’t be able run the NN optimally / as efficiently as possible. Imagine you want to do something that requires the NN and as a result, you can’t do anything else on your phone / laptop (it won’t be problem for desktops with GPUs though).

      Where NPU really shines is when there are performance constraints on the model: when it has to be fast, lightweight and memory efficient. Use cases include mobile computing and IoT.

      In fact, there’s news about live translation on Apple AirPod. I think this may be the perfect scenario for using NPUs - ideally housed within the earphones directly but if not, within a phone.

      Disclaimer: I am only familiar with NPUs in the context of “old-school” convolutional neural networks (boy, tech moves so quickly). I am not familiar with NPUs for transformers - and LLMs by extension - but I won’t be surprised if NPUs have been adapted to work with them.

      source
      • rumba@lemmy.zip ⁨1⁩ ⁨day⁩ ago

        I’m not exactly an expert either but I believe the NPUs were seeing in the wild here are more like efficiency cores for AI.

        Using the GPU would be faster, but have much larger energy consumption. They’re basically mathco processors that are good at matrix calculations.

        source
  • twinnie@feddit.uk ⁨1⁩ ⁨day⁩ ago

    I expect we’re eventually going to start seeing AI in more sensible places and these NPUs will prove useful. Hopefully soon the bubble will burst and we’ll stop seeing it crammed in everywhere, then it’ll start being used where it actually improves a product rather than wherever an LLM will fit.

    source
    • ohwhatfollyisman@lemmy.world ⁨1⁩ ⁨day⁩ ago

      i’m upvoting this comment from my internet enabled toaster.

      source
  • cmnybo@discuss.tchncs.de ⁨1⁩ ⁨day⁩ ago

    It would be nice if Frigate supported the NPU in AMD CPUs. Then it could be used for object detection with CCTV cameras.

    source
    • ch00f@lemmy.world ⁨1⁩ ⁨day⁩ ago

      I love how Frigate is like…the only thing you can use AI processors for. I got a Coral M.2 card and used it for Frigate for like a year and then got bored. Sure, I can have it ID birds, but there’s not much more practical use that I found.

      source
      • cmnybo@discuss.tchncs.de ⁨1⁩ ⁨day⁩ ago

        You can use object detection to send an alert if there is a person in your yard or if a car pulls into your driveway. It avoids the frequent false alerts you get with normal motion detection.

        source
        • -> View More Comments
  • yarr@feddit.nl ⁨1⁩ ⁨day⁩ ago

    From a PC manufacturer perspective, the important part about including an NPU is that you can slap an “AI” sticker on your marketing. This is regardless of whether it has any actual use cases or not.

    There’s definitely a large slice of “AI” features shipping now just to excite shareholders and serve no actual function to the user.

    source
  • Dudewitbow@lemmy.zip ⁨1⁩ ⁨day⁩ ago

    So, yes, it would be unrealistic to suggest AMD could rip out the NPU and easily put that space to good use in this current chip.

    a few amd chips did have the NPU removed in them, except they were mostly allocated to handheld gaming devices. the Ryzen Z1E for example.

    when redreshed, there were a handful of hawkpoint cpus that were NPU less (Ryzen 7 8745HS)

    Strix point does not have that dumped die cycle yet

    source
  • Gullible@sh.itjust.works ⁨1⁩ ⁨day⁩ ago

    I was excited when I learned that a new business laptop had a removable battery, a decent graphics card, and 1tb storage standard. I planned to buy it used for a fraction of its current price price, 2k usd new, once some doofus got bored of underusing their machine and decided to trade up. Then I saw the AI chip and my desire wavered. You think there will ever be workarounds to make use of this garbage? I really want that removable battery.

    source
    • Ashiette@lemmy.world ⁨1⁩ ⁨day⁩ ago

      That NPU is a math coprocessor. It can be very useful. It’s like a cuda core.

      source
      • AnyOldName3@lemmy.world ⁨1⁩ ⁨day⁩ ago

        A CUDA core is just a vector processor like every GPUs since the late 90s has been made of, but with a different name so it sounds special. It doesn’t just run CUDA, it runs everything else a GPU has traditionally been for, too, and that was stuff people were doing before CUDA was introduced. There are lots of tasks that require the same sequence of operations to be applied to groups of 32 numbers.

        An NPU is a much more specialised piece of hardware, and it’s only really neural network training and inference that it can help with. There aren’t many tasks that require one operation to be applied over and over to groups of hundreds of numbers. Most people aren’t finding that they’re spending lots of time waiting for neural network inference or draining their batteries doing neural network inference, so making it go faster and use less power isn’t a good use of their money compared to making their computer better at the things they do actually do.

        source
  • BigTrout75@lemmy.world ⁨1⁩ ⁨day⁩ ago

    Bring on the games that require AI

    source
    • Endmaker@ani.social ⁨1⁩ ⁨day⁩ ago

      I’m pretty sure most video games require AI already. I have difficulty naming even one that doesn’t use AI.

      Neural nets, on the other hand - I find it hard to imagine one running locally without impacting performance.

      source
      • Mika@sopuli.xyz ⁨1⁩ ⁨day⁩ ago

        Hard, but should definitely be possible. I’m waiting for someone to write a decent roguelike, like imagine the possibilities.

        source