Open Menu
AllLocalCommunitiesAbout
lotide
AllLocalCommunitiesAbout
Login

OpenAI Insider Estimates 70 Percent Chance That AI Will Destroy or Catastrophically Harm Humanity

⁨100⁩ ⁨likes⁩

Submitted ⁨⁨11⁩ ⁨months⁩ ago⁩ by ⁨floofloof@lemmy.ca⁩ to ⁨technology@beehaw.org⁩

https://futurism.com/the-byte/openai-insider-70-percent-doom

source

Comments

Sort:hotnewtop
  • thingsiplay@beehaw.org ⁨11⁩ ⁨months⁩ ago

    How did he calculate the 70% chance? Without an explanation this opinion is as much important as a Reddit post. It’s just marketing fluff talk, so people talk about AI and in return a small percentage get converted into people interested into AI. Let’s call it clickbait talk.

    First he talks about high chance that humans get destroyed by AI. Follows with a prediction it would achieve AGI in 2027 (only 3 years from now). No. Just no. There is a loong way to get general intelligence. But isn’t he trying to sell you why AI is great? He follows with:

    “We’re proud of our track record providing the most capable and safest AI systems and believe in our scientific approach to addressing risk,”

    Ah yes, he does.

    source
    • LibertyLizard@slrpnk.net ⁨11⁩ ⁨months⁩ ago

      Insider from OpenAI PR department speaks out!

      source
    • joelfromaus@aussie.zone ⁨11⁩ ⁨months⁩ ago

      How did he calculate the 70% chance?

      Maybe they asked ChatGPT?

      source
      • MagicShel@programming.dev ⁨11⁩ ⁨months⁩ ago

        ChatGPT says 1-5%, but I told it to give me nothing but a percentage and it gave me a couple of paragraphs like a kid trying to distract from the answer by surrounding it with bullshit. I think it’s onto us…

        (I kid. I attribute no sentience or intelligence to ChatGPT.)

        source
    • eveninghere@beehaw.org ⁨11⁩ ⁨months⁩ ago

      This is a horoscope trick. They can always say AI destroyed humanity.

      Trump won in 2016 and there was Cambridge Analytica doing data analysis: AI technology destroyed humanity!

      Israel used AI-guided missiles to attack Gaza: AI destroyed humanity!

      Whatever. You can point at whatever catastrophe and there is always AI behind because already in 2014 AI is a basic technology.

      source
    • chicken@lemmy.dbzer0.com ⁨11⁩ ⁨months⁩ ago

      The person who predicted 70% chance of AI doom is Daniel Kokotajlo, who quit OpenAI because of it not taking this seriously enough. The quote you have there is a statement by OpenAI, not by Kokotajlo, this is all explicit in the article. The idea that this guy is motivated by trying to do marketing for OpenAI is just wrong, the article links to some of his extensive commentary where he is advocating for more government oversight specifically of OpenAI and other big companies instead of the favorable regulations that company is pushing for. The idea that his belief in existential risk is disingenuous also doesn’t make sense, it’s clear that he and other people concerned about this take it very seriously.

      source
  • millie@beehaw.org ⁨11⁩ ⁨months⁩ ago

    I think when people think of the danger of AI, they think of something like Skynet or the Matrix. It either hijacks technology or builds it itself and destroys everything.

    But what seems much more likely, given what we’ve seen already, is corporations pushing AI that they know isn’t really capable of what they say it is and everyone going along with it because of money and technological ignorance.

    This is much more likely, and you can already see the warning signs. Cars that run pedestrians over, search engines that tell people to eat glue, customer support AI that have no idea what they’re talking about, endless fake reviews and articles. It’s already hurt people, but so far only on a small scale.

    But the profitablity of pushing AI early, especially if you’re just pumping and dumping a company for quarterly profits, is massive. The more that gets normalized, the greater the chance one of them gets put in charge of something important, or becomes a barrier to something important.

    That’s what’s scary about it. It isn’t AI itself, it’s AI as a vector for corporate recklessness.

    source
    • fwygon@beehaw.org ⁨11⁩ ⁨months⁩ ago

      It isn’t AI itself, it’s AI as a vector for corporate recklessness.

      This. 1000% this. Many of Issac Asimov novels warned about this sort of thing too; as did any number of novels inspired by Asimov.

      It’s not that we didn’t provide the AI with rules. It’s not that the AI isn’t trying not to harm people. It’s that humans, being the clever little things we are, are far more adept at deceiving and tricking AI into saying things and using that to justify actions to gain benefit.

      …Understandably this is how that is being done. By selling AI that isn’t as intelligent as it is being trumpeted as. As long as these corporate shysters can organize a team to crap out a “Minimally Viable Product” they’re hailed as miracle workers and get paid fucking millions.

      Ideally all of this should violate the many, many laws of many, many civilized nations…but they’ve done some black magic with that too; by attacking and weakening laws and institutions that can hold them liable for this and even completely ripping out or neutering laws that could cause them to be held accountable by misusing their influence.

      source
    • localhost@beehaw.org ⁨11⁩ ⁨months⁩ ago

      I don’t think your assumption holds. Corporations are not, as a rule, incompetent - in fact, they tend to be really competent at squeezing profit out of anything. They are misaligned, which is much more dangerous.

      I think the more likely scenario is also more grim:

      AI actually does continue to advance and gets better and better displacing more and more jobs. It doesn’t happen instantly so barely anything gets done. Some half-assed regulations are attempted but predictably end up either not doing anything, postponing the inevitable by a small amount of time, or causing more damage than doing nothing would. Corporations grow in power, build their own autonomous armies, and exert pressure on governments to leave them unregulated. Eventually all resources are managed by and for few rich assholes, while the rest of the world tries to survive without angering them.
      If we’re unlucky, some of those corporations end up being managed by a maximizer AGI with no human supervision and then the Earth pretty much becomes an abstract game with a scoreboard, where money (or whatever is the equivalent) is the score.

      Limitations of human body act as an important balancing factor in keeping democracies from collapsing. No human can rule a nation alone - they need armies and workers. Intellectual work is especially important (unless you have some other source of income to outsource it), but it requires good living conditions to develop and sustain. Once intellectual work is automated, infrastructure like schools, roads, hospitals, housing cease to be important for the rulers - they can give those to the army as a reward and make the rest of the population do manual work. Then if manual work and policing through force become automated, there is no need even for those slivers of decency.
      Once a single human can rule a nation, there is enough rich psychopaths for one of them to attempt it.

      There are also other AI-related pitfalls that humanity may fall into in the meantime - automated terrorism (e.g. swarms of autonomous small drones with explosive charges using face recognition to target entire ideologies by tracking social media), misaligned AGI going rogue (e.g. the famous paperclip maximizer, although probably not exactly this scenario), collapse of the internet due to propaganda bots using next-gen generative AI… I’m sure there’s more.

      source
      • Juice@midwest.social ⁨11⁩ ⁨months⁩ ago

        Ai doesn’t get better. Its completely dependent on computing power. They are dumping all the power into it they can, and it sucks ass. The larger the dataset the more power it takes to search it all. Your imagination is infinite, computing power is not. you can’t keep throwing electricity at a problem. It was pushed out because there was a bunch of excess computing power after crypto crashed, or semi stabilized. Its an excuse to lay off a bunch of workers after covid who were gonna get laid off anyway. Managers were like sweet I’ll trim some excess employees and replace them with ai! Wrong. Its a grift. It might hang on for a while but policy experts are already looking at the amount of resources being thrown at it and getting weary. The technological ignorance you are responding to, that’s you. You don’t know how the economy works and you don’t know how ai works so you’re just believing all this roku’s basilisk nonsense out of an overactive imagination. Its not an insult lots of people are falling for it, ai companies are straight up lying, the media is stretching the truth of it to the point of breaking. But I’m telling you, don’t be a sucker. Until there’s a breakthrough that fixes the resource consumption issue by like orders of magnitude, I wouldn’t worry too much about Ellison’s AM becoming a reality

        source
        • -> View More Comments
    • 0x815@feddit.de ⁨11⁩ ⁨months⁩ ago

      Yes. We need human responsibility for everything what AI does. It’s not the technology that harms but human beings and those who profit from it.

      source
    • Ilandar@aussie.zone ⁨11⁩ ⁨months⁩ ago

      Yes, it’s very concerning and frustrating that more people don’t understand the risks posed by AI. It’s not about AI becoming sentient and destroying humanity, it’s about humanity using AI to destroy itself. I think this fundamental misunderstanding of the problem is the reason why you get so many of these dismissive “AI is just techbro hype” comments. So many people are genuinely clueless about a) how manipulative this technology already is and b) the rate at which it is advancing.

      source
    • newtraditionalists@mastodon.social ⁨11⁩ ⁨months⁩ ago

      @millie @floofloof this is so well articulated I can't stand it. I want to have it printed out and hand it to anyone who asks me anything about AI. Thank you for this!

      source
      • masterspace@lemmy.ca ⁨11⁩ ⁨months⁩ ago

        It’s not tho.

        source
    • coffeetest@beehaw.org ⁨11⁩ ⁨months⁩ ago

      Calling LLMs, “AI” is one of the most genius marketing moves I have ever seen. It’s also the reason for the problems you mention.

      I am guessing that a lot of people are just thinking, “Well AI is just not that smart… yet! It will learn more and get smarter and then, ah ha! Skynet!” It is a fundamental misunderstanding of what LLMs are doing. It may be a partial emulation of intelligence. Like humans, it uses its prior memory and experiences (data) to guess what an answer to a new question would look like. But unlike human intelligence, it doesn’t have any idea what it is saying, actually means.

      source
  • Retiring@lemmy.ml ⁨11⁩ ⁨months⁩ ago

    I feel this is all just a scam, trying to drive the value of AI stocks. Noone in the media seems to talk about the hallucination problem, the problem with limited data for new models (Habsburg-AI), the energy restrictions etc.

    It’s all uncritical believe that „AI“ will just become smart eventually. This technology is built upon a hype, it is nothing more than that. There are limitations, and they reached them.

    source
    • aStonedSanta@lemm.ee ⁨11⁩ ⁨months⁩ ago

      And these current LLMs aren’t just gonna find sentience for themselves. Sure they’ll pass a Turing test but they aren’t alive lol

      source
      • knokelmaat@beehaw.org ⁨11⁩ ⁨months⁩ ago

        I think the issue is not wether it’s sentient or not, it’s how much agency you give it to control stuff.

        Even before the AI craze this was an issue. Imagine if you were to create an automatic turret that kills living beings on sight, you would have to make sure you add a kill switch or you yourself wouldn’t be able to turn it off anymore without getting shot.

        The scary part is that the more complex and adaptive these systems become, the more difficult it can be to stop them once they are in autonomous mode. I think large language models are just another step in that complexity.

        An atomic bomb doesn’t pass a Turing test, but it’s a fucking scary thing nonetheless.

        source
    • lvxferre@mander.xyz ⁨11⁩ ⁨months⁩ ago

      Habsburg-AI? Do you have an idea on how much you made me laugh in real life with this expression??? It’s just… perfect! Model degeneration is a lot like what happened with the Habsburg family’s genetic pool.

      When it comes to hallucinations in general, I got another analogy: someone trying to use a screwdriver with nails, failing, and calling it a hallucination. In other words I don’t think that the models are misbehaving, they’re simply behaving as expected, and that any “improvement” in this regard is basically a band-aid being added to humans to a procedure that doesn’t yield a lot of useful outputs to begin with.

      source
      • Retiring@lemmy.ml ⁨11⁩ ⁨months⁩ ago

        I am glad you liked it. Can’t take the credit for this one though, I first heard it from Ed Zitron in his podcast „Better Offline“. Highly recommend.

        source
    • averyminya@beehaw.org ⁨11⁩ ⁨months⁩ ago

      Energy restrictions actually could be pretty easily worked around using analog converting methods. Otherwise I agree completely though, and what’s the point of using energy on useless tools. There’s so many great things that AI is and can be used for, but of course like anything exploitable whatever is “for the people” is some amalgamation of extracting our dollars.

      The funny part to me is that like mentioned “beautiful” AI cabins that are clearly fake – there’s this weird dichotomy of people just not caring/too ignorant to notice the poor details, but at the same time so many generative AI tools are specifically being used to remove imperfection during the editing process. And that in itself is something that’s too bad, I’m definitely guilty of aiming for “the perfect composition” but sometimes nature and timing forces your hand which makes the piece ephemeral in a unique way. Shadows are going to exist, background subjects are going to exist.

      The current state of marketed AI is selling the promise of perfection, something that’s been getting sold for years already. Just now it’s far easier to pump out scam material with these tools, something that gets easier with each advancement in these sorts of technologies, and now with more environmental harm than just a victim of a predator.

      It really sucks being an optimist sometimes.

      source
    • darkphotonstudio@beehaw.org ⁨11⁩ ⁨months⁩ ago

      It could be only hype. But I don’t entirely agree. Personally, I believe we are only a few years away from AGI. Will it come from OpenAI and LLMs? Maybe, but it will likely come from something completely different. Like it or not, we are within spitting distance of a true Artificial Intelligence, and it will shake the foundations of the world.

      source
  • lvxferre@mander.xyz ⁨11⁩ ⁨months⁩ ago

    May I be blunt? I estimate that 70% of all OpenAI are full of crap.

    What people are calling nowadays “AI” is not a magic solution for everything. It is not an existential threat either. The main risks that I see associated with it are:

    1. Assumptive people taking LLM output for granted, to disastrous outcomes. Think on “yes, you can safely mix bleach and ammonia” tier (note: made up example).
    2. Supply and demand. Generative models have awful output, but sometimes “awful” = “good enough”.
    3. Heavy increase in energy and resources consumption.

    None of those issues was created by machine “learning”, it’s just that it synergises with them.

    source
    • BarryZuckerkorn@beehaw.org ⁨11⁩ ⁨months⁩ ago

      Your scenario 1 is the actual danger. It’s not that AI will outsmart us and kill us. It’s that AI will trick us into trusting them with more responsibility than the AI can responsibly handle, to disastrous results.

      It could be small scale, low stakes stuff, like an AI designing a menu that humans blindly cook. Or it could be higher stakes stuff that actually does things like affect election results, crashes financial markets, causes a military to target the wrong house, etc. The danger has always been that humans will act on the information provided by a malfunctioning AI, not that AI and technology will be a closed loop with no humans involved.

      source
      • lvxferre@mander.xyz ⁨11⁩ ⁨months⁩ ago

        Yup, it is a real risk. But on a lighter side, it’s a risk that we [humanity] have been fighting against since forever - the possibility of some of us causing harm to the others not due to malice, but out of assumptiveness and similar character flaws.

        source
    • CanadaPlus@lemmy.sdf.org ⁨11⁩ ⁨months⁩ ago

      Not yet, anyway.

      source
      • lvxferre@mander.xyz ⁨11⁩ ⁨months⁩ ago

        I’m reading your comment as “[AI is] Not yet [an existential threat], anyway”. If that’s inaccurate, please clarify, OK?

        With that reading in mind: I don’t think that the current developments in machine “learning” lead towards the direction of some hypothetical system that would be an existential threat. The closest to that would be the subset of generative models, that looks like a tech dead end - sure, it might see some applications, but I don’t think that it’ll progress much past the current state.

        In other words I believe that the AI that would be an existential threat would be nothing like what’s being created and overhyped now.

        source
        • -> View More Comments
  • starman@programming.dev ⁨11⁩ ⁨months⁩ ago

    OpenAI Insider

    Ah, what a reliable and unbiased source

    source
  • django@discuss.tchncs.de ⁨11⁩ ⁨months⁩ ago

    The energy demand of AI will harm humanity, because we keep feeding it huge amounts of energy produced by burning fossile fuels.

    source
  • darkphotonstudio@beehaw.org ⁨11⁩ ⁨months⁩ ago

    If humanity is doomed, it will be our own stupid fault, not AI. And by us, I mean billionaires and their politicians.

    source
    • Kichae@lemmy.ca ⁨11⁩ ⁨months⁩ ago

      I think much of it comes from “futurologists” spending too much time smelling each others’ farts. These AI guys think so very much of themselves.

      source
      • Hazzia@discuss.tchncs.de ⁨11⁩ ⁨months⁩ ago

        It’s crazy how little experts like these think of humanity, or just underestimate our tollerance and adaptability to weird shit. People used to talk about how “if we ever learned UFOs were a real phenomena, there would be global mayhem!” because people’s world views would collapse and they’d riot, or whatever. After getting a few articles the past few years since that first NY Times article, I’ve basically not heard anyone really caring (who didn’t already seem to be into them before, anyway). Hell, we had a legitimate attempt to overthrow our own government, and the large majority of our population just kept on with their lives.

        The same AI experts 10 years ago would have thought the AI we have right now would have caused societal collapse.

        source
        • -> View More Comments
      • darkphotonstudio@beehaw.org ⁨11⁩ ⁨months⁩ ago

        Agreed, partially. However, the “techbros” in charge, for the most part, aren’t the researchers. There are futurologists who are real scientists and researchers. Dismissing them smacks of the anti-science knuckleheads ignoring warnings about the dangers of not wearing masks and getting vaccines during the pandemic.

        source
        • -> View More Comments
    • verdare@beehaw.org ⁨11⁩ ⁨months⁩ ago

      The only danger to humans is humans.

      I’m sorry, but this is a really dumb take that borders on climate change denial logic. A sufficiently large comet is an existential threat to humanity. You seem to have this optimistic view that humanity is invincible against any threat but itself, and I do not think that belief is justified.

      People are right to be very skeptical about OpenAI and “techbros.” But I fear this skepticism has turned into outright denial of the genuine risks posed by AGI.

      I find myself exhausted by this binary partitioning of discourse surrounding AI. Apparently you have to either be a cult member who worships the coming god of the singularity, or think that AI is either impossible or incapable of posing a serious threat.

      source
      • darkphotonstudio@beehaw.org ⁨11⁩ ⁨months⁩ ago

        You seem to have this optimistic view that humanity is invincible against any threat but itself

        I didn’t say that. You’re making assumptions. However, I don’t take AGI as a serious risk, not directly anyway. AGI is a big question mark at this time and hardly comparable to a giant comet or pandemic, of which we have experience or solid scientific evidence. Could it be a threat? Yeah. Do I personally think so? No. Our reaction to and exploitation of will likely do far more harm than any direct action by an AGI.

        source
  • technocrit@lemmy.dbzer0.com ⁨11⁩ ⁨months⁩ ago

    OpenAI Insider Investor Estimates 70 Percent Chance That AI Will Destroy or Catastrophically Harm Humanity

    source
  • Floey@lemm.ee ⁨11⁩ ⁨months⁩ ago

    This fear mongering is just beneficial to Altman. If his product is powerful enough to be a threat to humanity then it is also powerful enough to be capable of many useful things, things it has not proven itself to be capable of. Ironically spreading fear about its capabilities will likely raise investment, so if you actually are afraid of openai somehow arriving at agi that is dangerous then you should really be trying to convince people of its lack of real utility.

    source
    • tal@lemmy.today ⁨11⁩ ⁨months⁩ ago

      The guy complaining left the company:

      Fed up, Kokotajlo quit the firm in April, telling his team in an email that he had “lost confidence that OpenAI will behave responsibly” as it continues trying to build near-human-level AI.

      I don’t think that he stands to benefit.

      He also didn’t say that OpenAI was on the brink of having something like this either.

      source
      • Floey@lemm.ee ⁨11⁩ ⁨months⁩ ago

        I didn’t think it was a choreographed publicity stunt. I just know Altman has used AI fear in the past to keep people from asking rational questions like “What can this actually do?” He obviously stands to gain from people thinking they are on the verge of agi. And someone looking for a new job in the field also has to gain from it.

        As for the software thing, if it’s done by someone it won’t be openai and megacorporations following in its footsteps. They seem insistent at throwing more data (of diminishing quality) and more compute (an impractical amount) at the same style of models hoping they’ll reach some kind of tipping point.

        source
  • KevonLooney@lemm.ee ⁨11⁩ ⁨months⁩ ago

    I just realized something: since most people have no idea what AI is, it could easily be used to scam people. I think that will be it’s main function originally.

    Like the average person does not have access to real time stock data. You could make a fake AI program that pretends to be a trading algorithm and makes a ton of pretend money as the mark watches. The data would be 100% real and verifiable, just picked a few seconds after the the fact.

    Since most people care a lot about money, this will be some of the first widespread applications of real time AI. Just tricking people out of money.

    source
    • scrubbles@poptalk.scrubbles.tech ⁨11⁩ ⁨months⁩ ago

      Yeah I’ll admit I was freaked out at the beginning. So I learned about models, used them, and got familiar with them. Now I’m less freaked out and more “oh my god so many people are going to get scammed/tricked”.

      Go on Facebook and you’ll see it’s a good 50-70% AI garbage now. My favorite are “log cabin” and kitchen posts that are just images of them with blanket titles like “wish I lived here” with THOUSANDS of comments of people saying “YES” or “it’s so beautiful”. Of course it is it has no supports! The cabinets are held up by nothing! There are 9 kinds of lanterns and most are floating. Jesus people are not ready for it.

      source
      • frog@beehaw.org ⁨11⁩ ⁨months⁩ ago

        The “Willa Wonka Experience” event comes to mind. The images on the website were so obviously AI-generated, but people still coughed up £35 a ticket to take their kids to it, and were then angry that the “event” was an empty warehouse with a couple of plastic props and three actors trying to improvise because the script they’d been given was AI-generated gibberish. Straight up scam.

        source
  • 2xsaiko@discuss.tchncs.de ⁨11⁩ ⁨months⁩ ago

    I mean I give it a 100% chance if they keep going like this considering the enormous energy and water consumption, essentially slave labor to classify data for training because it’s such a huge amount that it would never be financially viable to fairly pay people, and end result which is to fill the internet with garbage.

    You really don’t need to be an insider to see that.

    source
    • Railison@aussie.zone ⁨11⁩ ⁨months⁩ ago

      When I think of AI ruining humanity, this is how I picture it

      source
  • qqq@programming.dev ⁨11⁩ ⁨months⁩ ago

    Wake me up when nixpkgs issues decline significantly from 5k+ due to AI.

    source
  • MayonnaiseArch@beehaw.org ⁨11⁩ ⁨months⁩ ago

    He was interviewed after his septum replacement surgery, got a brand new teflon one

    source
  • autotldr@lemmings.world [bot] ⁨11⁩ ⁨months⁩ ago

    🤖 I’m a bot that provides automatic summaries for articles:

    Click here to see the summary

    In an interview with The New York Times, former OpenAI governance researcher Daniel Kokotajlo accused the company of ignoring the monumental risks posed by artificial general intelligence (AGI) because its decision-makers are so enthralled with its possibilities. Kokotajlo’s spiciest claim to the newspaper, though, was that the chance AI will wreck humanity is around 70 percent — odds you wouldn’t accept for any major life event, but that OpenAI and its ilk are barreling ahead with anyway. The 31-year-old Kokotajlo told the NYT that after he joined OpenAI in 2022 and was asked to forecast the technology’s progress, he became convinced not only that the industry would achieve AGI by the year 2027, but that there was a great probability that it would catastrophically harm or even destroy humanity. Kokotajlo became so convinced that AI posed massive risks to humanity that eventually, he personally urged OpenAI CEO Sam Altman that the company needed to “pivot to safety” and spend more time implementing guardrails to reign in the technology rather than continue making it smarter. Fed up, Kokotajlo quit the firm in April, telling his team in an email that he had “lost confidence that OpenAI will behave responsibly” as it continues trying to build near-human-level AI. “We’re proud of our track record providing the most capable and safest AI systems and believe in our scientific approach to addressing risk,” the company said in a statement after the publication of this piece. — Saved 56% of original text.

    source