They use HBM (High Bandwidth Memory). PCs, laptops and phones don’t use this type of RAM.

  • CaptainBasculin@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    126
    ·
    28 days ago

    Manufacturers have a certain amount of chips they can manufacture, let’s say they can manufacture 10 million chips per year. Normally they adjust for demand, like manufacture 7 million chips for consumers and 3 million for enterprise customers. Company A contacts them and says “We need 6 million chips for this year, here’s the money”, But the factories still can manufacture only 10 million chips in total, so they adjust their factories to manufacture more enterprise focused chips, decreasing the amount of chips manufactured used by consumers.

    • Hawke@lemmy.world
      link
      fedilink
      arrow-up
      45
      ·
      28 days ago

      I think it’s more like “we need 6 million chips for this year, and we’ll pay you eventually, maybe”.

      But yeah the outcome is the same.

  • coolie4@lemmy.world
    link
    fedilink
    arrow-up
    49
    ·
    28 days ago

    Without even getting into the electronics similarities, they use common raw materials and manufacturing facilities. Diverting resources to one lowers supply of the other, affecting costs.

  • Treczoks@lemmy.world
    link
    fedilink
    arrow-up
    42
    ·
    28 days ago

    Because the manufacturers use the facilities that once produced DDR5 RAM chips to produce HBM chips instead.

  • gdog05@lemmy.world
    link
    fedilink
    arrow-up
    27
    ·
    28 days ago

    The better question is, why are they doing all of this without an actual purchase contract?

    • kbal@fedia.io
      link
      fedilink
      arrow-up
      26
      ·
      28 days ago

      Because it’s AI, haven’t you heard? Does it make sense for the business? Who cares, it’s AI. Is it financially sustainable? Dude, it’s AI though. Will there be any customers for any of it? The AI says there will be. You’ve got to understand, this is AI we’re talking about. It’s the AI revolution that will transform the world. We’ve got to bet everything on the AI, or we’ll be left out of the AI future. I asked the AI and it was very clear about that.

      • Asafum@lemmy.world
        link
        fedilink
        arrow-up
        20
        ·
        28 days ago

        Hey @grok is this true?

        Grok: yes, I am very great. I am good at everything because I am AI. Also, fuck the Jews.

  • Appoxo@lemmy.dbzer0.com
    link
    fedilink
    arrow-up
    18
    ·
    27 days ago

    Production capacity.
    Somewhere I read that to produce one HBM module you could produce three regular DDR modules

  • brucethemoose@lemmy.world
    link
    fedilink
    arrow-up
    17
    ·
    28 days ago

    To add to what others said:

    LPDDRX is used in some inference hardware. The same stuff you find in laptops and smartphones.

    Also, the servers need a whole lot of regular CPU DIMMs since they’re still mostly EPYC/Xeon severs with 8 GPUs in each. And why are they “wasting” so much RAM on CPU RAM that isn’t really needed, you ask? Same reason as a lot of AI: it’s immediately accessible, already targeted by devs, and AI dev is way more conservative and wasteful than you’d think.

    Same for SSDs. Regular old servers (including AI servers) need it too. In a perfect world they’d use centralized storage for images/weights with near-“diskless” inference/training servers. Some AI servers do this, but most don’t.


    Basically, the waste is tremendous, for the same reason they use cheap gas generators on-site: it’s faster-to-market.

  • Robin@lemmy.world
    link
    fedilink
    English
    arrow-up
    17
    ·
    28 days ago

    Besides what others have already said, HBM is only used for the GPUs. These AI servers also use regular DDR5 chips, just with an extra EEC chip.

  • kbal@fedia.io
    link
    fedilink
    arrow-up
    12
    ·
    28 days ago

    The type of RAM that they use is different in that it takes up even more of all the things that would otherwise be used to produce the RAM that you use.

  • kmirl@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    28 days ago

    If RAM and GPUs were cheap people like us would be more likely to set up local LLMs to prevent our data from being productized by power-grabbing corporations.

      • kmirl@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        28 days ago

        Not claiming it’s the reason since it clearly isn’t, only that it will help drive traffic to commercial AI products.