• TheGrandNagus@lemmy.world
    link
    fedilink
    English
    arrow-up
    24
    ·
    1 year ago

    This is just based on a survey of researchers (not even AI researchers specifically) where they were asked if they think it’ll happen or not and they gave anything from vague predictions to complete gut-feeling guesses back.

    It’s a worthless article. Not that I even needed to click to find that out.

    • themeatbridge@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      1 year ago

      83% of statisticians polled say you cannot predict future events from the statistical results of an opinion poll, which means you can still make those predictions 17% of the time. Maybe this is one of that 17%?

        • themeatbridge@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          ·
          1 year ago

          Bartlet: Sweden has a 100% literacy rate, Leo. 100%! How do they do that?

          McGarry: Well, maybe they don’t and they also can’t count.

  • Yer Ma@lemm.ee
    link
    fedilink
    English
    arrow-up
    8
    ·
    edit-2
    1 year ago

    +/-95%

    “AI experts’ predictions should not be seen as a reliable guide to objective truth”

  • hoot@lemmy.ca
    link
    fedilink
    English
    arrow-up
    9
    arrow-down
    2
    ·
    1 year ago

    “Researchers say”

    Researchers also say things they pulled out of their ass 90% of the time.

    Source: I researched it

    • GigglyBobble@kbin.social
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      “Researchers say” isn’t something researchers say though but journalists or “journalists” (bloggers etc).

  • drdiddlybadger@pawb.social
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    1
    ·
    1 year ago

    Guys who’s job depends on hype around their field continue to hype their field. News at 11.

    All that ad space just to leave out the whys.

  • HeavyDogFeet@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    1 year ago

    What are the chances that it won’t cause extinction, but will just make things a bit (or maybe a lot) worse for a lot of people?

    I don’t think any reasonable people are all that concerned with doomsday scenario as much as with the slow march towards a bullshit future.

  • originalucifer@moist.catsweat.com
    link
    fedilink
    arrow-up
    4
    arrow-down
    1
    ·
    1 year ago

    this is just a subset of the 100% chance humans will cause human extinction.

    ‘by ai’ is just one of very many options were looking into…

  • silverbax@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    1 year ago

    So, 95% chance that humans will cause human extinction.

    And humans created AI, so even if AI does in the human race, it will still have been humans.

    I guess if humans go extinct, it’s close to 100% due to humans.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    2
    ·
    1 year ago

    This is the best summary I could come up with:


    In the short term, researchers estimate AI will become significantly more advanced, able to create a Top 40 pop song and write an NYT bestseller before 2030.

    “While AI experts’ predictions should not be seen as a reliable guide to objective truth, they can provide one important piece of the puzzle,” said researchers from Berkeley and the University of Oxford who conducted the study in December.

    “Their familiarity with the technology and the dynamics of its past progress puts them in a good position to make educated guesses about the future of AI.”

    The study highlights the perceived danger around creating a powerful artificial intelligence from the world’s leading researchers.

    Anthropic has a constitution laid over its AI systems to ensure they act in alignment with our society’s rules.

    Experts anticipate that AI will be able to assemble LEGOs, translate newfound languages, and build video games before 2033.


    The original article contains 402 words, the summary contains 148 words. Saved 63%. I’m a bot and I’m open source!

  • LainOfTheWired@lemy.lol
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    1 year ago

    Didn’t a bunch of “researchers” think the world was going to end in 2012 for some reason?

    • ripcord@kbin.social
      link
      fedilink
      arrow-up
      5
      ·
      1 year ago

      No?

      I mean, if you stretch those airquotes pretty hard, but not sure why you’d compare those people to this case

  • Hal-5700X@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    You have nothing to be afraid of, Dave.

    Can it happen, maybe. Will it happen. Beats me, man.

    • WHYAREWEALLCAPS@kbin.social
      link
      fedilink
      arrow-up
      2
      ·
      1 year ago

      The more I think about it, the more I suspect it’ll be completely by accident. Some AI designed drug will pass all trials, get approved, be used for years or decades only to find out that some bit of it kind of acts like a generic prion that affects all life. Oh, and that bit also passes right through you, too, so by the time they figure it out, the pseudo-prion is already out there in the wild, infecting fish and other aquatic creatures. And before long the ecosystems of the world’s oceans and lakes collapse. Meanwhile land animals also start dying off due to their drinking supply being polluted by it, so a full scale ecological collapse begins. As the pseudoprion sticks around indefinitely, every attempt by nature to evolve new life ends because of it. Eventually it gets buried by all the detritus of time and new life does once again rise only to have humanity’s ticking time bomb waiting in the ground for something to dig it up and start the whole cycle all over again…