“In 10 years, computers will be doing this a million times faster.” The head of Nvidia does not believe that there is a need to invest trillions of dollars in the production of chips for AI::Despite the fact that Nvidia is now almost the main beneficiary of the growing interest in AI, the head of the company, Jensen Huang, does not believe that

  • General_Effort@lemmy.world
    link
    fedilink
    English
    arrow-up
    75
    ·
    5 months ago

    Despite the fact that Nvidia is now almost the main beneficiary of the growing interest in AI, the head of the company, Jensen Huang, does not believe that additional trillions of dollars need to be invested in the industry.

    *Because of

    You heard it, guys. There’s no need to create competition to Nvidia’s chips. It’s perfectly fine if all the profits go to Nvidia, says Nvidia’s CEO.

  • BetaDoggo_@lemmy.world
    link
    fedilink
    English
    arrow-up
    29
    arrow-down
    1
    ·
    edit-2
    5 months ago

    This isn’t necessarily about just hardware. Current ML architectures and inference engines are far from being at peak efficiency. Just last year we saw 20x speedups for llm inference on some hardware. “a million times” is obviously hyperpole though.

    • NegativeInf@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      5 months ago

      Literally reading preprint papers daily on more efficient implementations of self attention approximations.

  • Esqplorer@lemmy.zip
    link
    fedilink
    English
    arrow-up
    23
    ·
    5 months ago

    He doesn’t want a new competitor. He’s just spouting whatever will make the line move up. It has nothing to do with his opinion.

  • rebelsimile@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    2
    ·
    5 months ago

    Honestly as someone who has watched the once-fanciful prefixes “giga” and “tera” enter common parlance, and saw kilobytes of RAM turn to gigabytes, it’s really hard for me to think what he’s saying is impossible.

    • Buddahriffic@lemmy.world
      link
      fedilink
      English
      arrow-up
      17
      arrow-down
      1
      ·
      5 months ago

      Even if he is accurate, specialist hardware will outperform generic hardware at what it is specialized for.

      I remember a story sometime in the 00s about PCs finally getting to the point where they were as fast as one of the WWII code breaking computers (or something like that). It wasn’t because we backtracked in computer speeds after WWII, but because even that ancient hardware was able to get good performance when it was purpose-built, but it couldn’t do anything else and likely would have required a lot of work to adjust to a different kind of cypher scheme, if it could be adapted at all.

      So GP compute might be a million times faster in a decade, but specialist AI chips might be a million times faster than that.

      A hardware neural net might be able to eliminate memory latency by giving each neuron fast resisters to handle all their memory needs. If it doesn’t need to change connections, each connection could be hard wired. A GPU wouldn’t have a chance at keeping up no matter how wide that memory bus gets or how many channels it gets split into. It might even use way less power (though with the elimination of memory latency, it could go fast enough to use way more, too).

    • TheGrandNagus@lemmy.world
      link
      fedilink
      English
      arrow-up
      14
      arrow-down
      1
      ·
      5 months ago

      Nobody is saying it won’t happen eventually. But a million times within the next decade, i.e. 4x better every year for 10 years?

      This generation isn’t better than last generation by even close to that. Nevermind doing 4x for 10 years straight.

      • fidodo@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 months ago

        He was probably not being literal with the number, but when you’re the head of a computer chip hardware company you should pick numbers carefully.

  • Buffalox@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    1
    ·
    edit-2
    5 months ago

    Sorry I have doubts, because that would require a factor 4x increase every year for 10 years! 4x^10 = 1,048,576x
    Considering they historically have had problems achieving just twice the speed per year, it does not seem likely.

      • Buffalox@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        5 months ago

        Why does that make a difference? Compute for AI is build on the progress for compute first for GPU then for data center. They are similar in nature.
        Yes they have exceeded 2x for AI for a while, but that has been achieved through exploding die size and cost, but even that won’t make a million times faster in 10 years possible, because they can’t increase die sizes any further.

        • ryannathans@aussie.zone
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 months ago

          Building an ASIC for purpose built computation is significantly faster than generic gpu compute cores. Like when ASICs were built for bitcoin mining/sha256 and a little 5 watt usb device could outperform the best GPUs

          • frezik@midwest.social
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            5 months ago

            It may be even more specialized than that. It might be a return to analog computers.

            Which isn’t going to work for Nvidia’s traditional products, either.

          • Buffalox@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            5 months ago

            The H200 is evolved from Nvidia GPU designs, and will be by far the most powerful AI component in existence when it arrives later this year, AI is now so complex, that it doesn’t really make sense to call it an ASIC or to use an ASIC for the purpose, and the cost is $40,000.- for a single H200 unit!!! So no not small 5 watt units, more like 100x that.
            If they could make small ASICS that did the same, they’d all do it. Nvidia AMD Intel Google Amazon Huawei etc. But it’s simply not an option.

            Edit:

            In principle the H200 AI/Compute system, is a giant cluster of tiny ASICS built onto one chip for massive parallel compute and greater speed.

        • fidodo@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          5 months ago

          There’s also software improvements to consider, there’s a lot of room for efficiency improvements.

      • Buffalox@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        5 months ago

        Yes, but usually we keep those 2 kinds of optimizations separate, only combining chip design and production process. Because if the software is optimized, the hardware isn’t really doing the same thing.
        So yes AI speed may increase more than just the hardware, but for the most sophisticated systems, the tasks will be more complex, which may again slow the software down.
        So I think they will never be able to achieve it even when considering software optimizations too. Just the latest Tesla cars boast about 4 times higher resolution cameras, that will require 4 times the processing power to process image recognition, which then will be more accurate, but relatively slower.
        We are not where we want to be, and the systems of the future will clearly be more complex, and on the software are more likely to be slower than faster.

  • Coreidan@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    4
    ·
    5 months ago

    Oh well. The world is going to burn anyway. Fuck this shit hole we call earth