• rektdeckard@lemmy.world
    link
    fedilink
    English
    arrow-up
    22
    ·
    6 months ago

    MBA idiots and Economists consistently overestimate the abilities (and the underlying nature) of AI, because the output of AI is so much like how they speak: eloquent, confidently wrong, unconcerned with ground truth. They see themselves in AI, and they consistently overestimate themselves and their abilities.

    • self@awful.systems
      link
      fedilink
      English
      arrow-up
      13
      ·
      6 months ago

      they phrase it so your brain will think otherwise, but I just want to point out the maybe obvious so nobody else has to do a double-take at the quoted paragraphs:

      their example of what an ultra-intelligent AGI could do is of a human speedrunner using a glitch to beat Minecraft in 20 seconds. this asshole is taking something humans are excellent at and saying “but what if AI could do this, and also what if Minecraft was real life?” this is literal baby shit. like, even tool-assisted speedruns are the product of a shitload of human research into the problem space, and the tool’s just executing impossibly precise game inputs programmed, again, by a human. this is another space where AI sucks compared with regular human effort.

      and speaking of which, does anyone remember the early OpenAI and Google marketing where they had an LLM play pac-man or some shit at supposedly superhuman levels? can anyone dig up an outcome for any of those in the form of a record on any credible speedrunning site or during an event? cause speedrunning has a ton of categories including stuff like dog-assisted runs, where you train your dog to play the game, and it’s all considered valid as different forms of skill applied to the game. the one thing you can’t do is cheat, and they’re very good at verifying runs (ie, you must be provably using the method you claim, and you can’t splice video together or use emulator cheats to achieve a better run). so where’s the verified LLM speedrun records?

      • aio@awful.systems
        link
        fedilink
        English
        arrow-up
        6
        ·
        6 months ago

        Analyzing our data we conclude with 95% confidence that within a decade the Dyson Sphere Any% TAS time will be reduced below 55 seconds (± 1E10 years).

      • YourNetworkIsHaunted@awful.systems
        link
        fedilink
        English
        arrow-up
        2
        ·
        5 months ago

        They’re catching on that “big if true” is being recognized to mean “this is bullshit” so are trying to compensate by using more words.

    • swlabr@awful.systems
      link
      fedilink
      English
      arrow-up
      11
      ·
      6 months ago

      Me, as a jock bully: Well of course these nerds would find finishing in 20 seconds an achievement

  • swlabr@awful.systems
    link
    fedilink
    English
    arrow-up
    17
    ·
    6 months ago

    Given the all-consuming desire of Silicon Valley to accelerate us into a mad max style dystopia, the most foreshadowing thing about SF has gotta be all the (human) poop on the sidewalk.

  • Jayjader@jlai.lu
    link
    fedilink
    English
    arrow-up
    14
    ·
    edit-2
    6 months ago

    @> By the end of the decade, American electricity production will have grown tens of percent

    … don’t worry guys AI will totally save us from climate change! /s

    • Mii@awful.systems
      link
      fedilink
      English
      arrow-up
      16
      ·
      6 months ago

      I’m reading Feynman’s lectures on electromagnetism right now, and GPT-4o can answer questions and help me with the math. I doubt that even a smart high school would be able to do it.

      Ten bucks this guy hasn’t double-checked anything his chatbot told him but accepted it as truth because it used big words in grammatically coherent ways.

      • blakestacey@awful.systems
        link
        fedilink
        English
        arrow-up
        12
        ·
        6 months ago

        Electromagnetism is a standard subject covered in a bajillion books, so the training set is probably full of repeated explanations of the basic examples. That sounds like an excellent recipe for “AI” bilge-water that is just coherent enough for a student to miss where it goes wrong.

  • blakestacey@awful.systems
    link
    fedilink
    English
    arrow-up
    8
    ·
    6 months ago

    Scott Aaronson says that Aschenbrenner displays “unusual clarity, concreteness, and seriousness”.

    That’s it, that’s the joke

    • BigMuffin69@awful.systems
      link
      fedilink
      English
      arrow-up
      3
      ·
      5 months ago

      You know for a blog that’s on its face about computational complexity, you’d think Scott would show a little more skepticism to the tech bro saying “all we need is 14 quintillion x compute to solve the Riemann hypothesis”