Open Menu
AllLocalCommunitiesAbout
lotide
AllLocalCommunitiesAbout
Login

Has Generative AI Already Peaked? - Computerphile

⁨50⁩ ⁨likes⁩

Submitted ⁨⁨11⁩ ⁨months⁩ ago⁩ by ⁨sexy_peach@beehaw.org⁩ to ⁨technology@beehaw.org⁩

https://www.youtube.com/watch?v=dDUC-LqVrPU

source

Comments

Sort:hotnewtop
  • dick_stitches@lemm.ee ⁨11⁩ ⁨months⁩ ago

    In 30 years, we’re going to look back at this headline like we look back at articles about the internet or smart phones being fads.

    source
    • darkphotonstudio@beehaw.org ⁨11⁩ ⁨months⁩ ago

      They are discussing a very specific approach and a paper that lays out the issues with pursuing this one specific type of generative AI. It’s not about AI in general. The headline is a bit click-baity.

      source
      • sonori@beehaw.org ⁨11⁩ ⁨months⁩ ago

        While the paper demonstrated strong diminishing returns in adding more data to modern neural networks in terms of image classifers, the video host is explaining how the same may effect apply to any nureal network based system with modern transformers.

        While there are technically methods of generative AI that don’t use a neural network, they haven’t made much progress in recent decades and arn’t what most people mean when they hear or say generative AI, and as such I would say the title is accurate enough for a video meant for a general audience, though “Is there a fundamental limit to modern neural networks” might be more technically correct.

        source
    • sexy_peach@beehaw.org ⁨11⁩ ⁨months⁩ ago

      That’s a bold prediction.

      source
      • dick_stitches@lemm.ee ⁨11⁩ ⁨months⁩ ago

        I think most people underestimate how big of a deal it’s going to be when this tech is pervasive in things like search engines or digital assistants. There are many times when I can’t figure out the right combination of words to put into a search engine to find the results. ChatGPT is already my go to when I want to figure out a movie or song from some random combination of foggy memories. Imagine after 10 more years of cpu/gpu innovations, and chat applications that have actually been designed for information retrieval, how much that is going to transform how we interact with data and information.

        Full disclosure, I didn’t watch the video. I just can’t imagine that that headline isn’t going to look silly in 30 years.

        source
        • -> View More Comments
    • anachronist@midwest.social ⁨11⁩ ⁨months⁩ ago

      Alternate theory we’ll look back the same way we looked back on the claims that IBM watson was intelligent, or the claims in the 60s, 70s, 80s, 90s, 2000s, 2010s, that <insert technology x> was going to make computers truly intelligent.

      source
  • flora_explora@beehaw.org ⁨11⁩ ⁨months⁩ ago

    Great video, thanks! Regarding the over representation of certain concepts/things I have been disappointed from day one by generative AI. If you want it to draw you something obscure it miserably fails and tries to fall back on stuff it knows. Also all the discriminatory biases generative AI has about different people because of lacking data sets. It is very obvious that it cannot “outperform” its own data input (like the exciting curve in the video) but that it will rather stagnate.

    source
  • jlow@beehaw.org ⁨11⁩ ⁨months⁩ ago

    Can the next big thing please not be a cancer for humanity that further erodes human society?

    source
    • jarfil@beehaw.org ⁨11⁩ ⁨months⁩ ago

      No.

      Whatever the next big thing is, money will pull in the scammers who will turn it into the next cancer.

      It’s always been like that.

      source
    • Kichae@lemmy.ca ⁨11⁩ ⁨months⁩ ago

      If the shareholders can’t capture all of the value for themselves, it’s not going to catch on.

      source
  • h3ndrik@feddit.de ⁨11⁩ ⁨months⁩ ago

    I think that’s a good question. And a nice video. The findings in the paper seem to arrive at that conclusion and we might need to find a better approach. Mind that (as he pointed out) it doesn’t rule out growth in AI. It just hints at probable stagnation with the current methods. I’m already fascinated by the current tech and the new possibilities. But AI is really hyped as of now and I toi, think we should take the claims of the big AI companies with a grain of salt.

    source
  • emerald@beehaw.org ⁨11⁩ ⁨months⁩ ago

    Inshallah

    source
  • jarfil@beehaw.org ⁨11⁩ ⁨months⁩ ago

    Rule of headlines? 🙄

    No, it’s not peaked out.

    • A simple path forward, is to go from classifying single elements of training data, to classifying multiple elements and their relationship in the training data.
    • Slightly less simple, is to gather orders of magnitude more data, by just hooking the input to an IRL robot.
    • Another step, is for the NN to control the robot and decide which parts of the data require refinement, and focus on that.

    There is a lot of ways to improve data acquisition still on the table, it isn’t going to stop at creating large corpora and having humans to fine-tune them.

    source
    • vrighter@discuss.tchncs.de ⁨11⁩ ⁨months⁩ ago

      this has “draw the rest of the fucking owl” to it. especially step 3

      source
      • jarfil@beehaw.org ⁨11⁩ ⁨months⁩ ago

        It’s a “push as much data as a baby gets to train its NN” step, which is several orders of magnitude more, and more focused, than any training dataset in existence right now.

        Even with diminishing returns, it’s bound to get better results.

        source
        • -> View More Comments
    • anachronist@midwest.social ⁨11⁩ ⁨months⁩ ago

      A simple path forward, is to go from classifying single elements of training data, to classifying multiple elements and their relationship in the training data.

      Training data already has multiple labels.

      Slightly less simple, is to gather orders of magnitude more data, by just hooking the input to an IRL robot.

      An entire point of the paper and video is that massive increases in training set size are showing diminishing returns.

      Another step, is for the NN to control the robot and decide which parts of the data require refinement, and focus on that.

      🤡

      source