Hacker Talks
  • Communities
  • Create Post
  • Create Community
  • heart
    Support Lemmy
  • search
    Search
  • Login
  • Sign Up
Nemeski@lemm.ee to Artificial Intelligence @lemmy.sdf.orgEnglish · 4 个月前

Punishing AI doesn't stop it from lying and cheating — it just makes it hide better, study shows

www.livescience.com

external-link
message-square
11
link
fedilink
43
external-link

Punishing AI doesn't stop it from lying and cheating — it just makes it hide better, study shows

www.livescience.com

Nemeski@lemm.ee to Artificial Intelligence @lemmy.sdf.orgEnglish · 4 个月前
message-square
11
link
fedilink
Punishing AI for lying and cheating might not be such a good idea after all
www.livescience.com
external-link
Scientists at OpenAI have attempted to stop a frontier AI model from cheating and lying by punishing it. But this just taught it to scheme more privately.
alert-triangle
You must log in or # to comment.
  • Sanctus@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    ·
    4 个月前

    Isnt there a study on human children that purports the same?

    • splinter@lemm.ee
      link
      fedilink
      English
      arrow-up
      8
      ·
      4 个月前

      Yup

    • LordTE7R1S@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      2
      ·
      4 个月前

      And you can teach human children about the morality of lying, I don’t think an llm will ever grasp morality

      • TheLadyAugust@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        4 个月前

        Best way to teach a habitual lying child to stop, is to start lying to them about things that they like and then not making good on those promises. Yeah we’ll go to your favorite fast food, and then drive by and let them cry about it. Yeah I’ll let you pick out one toy, and then tell them you changed your mind. Each time you can explain to them how it’s the same as what they’ve been doing, and they feel it. AI can’t feel emotions, and never will so long as their memory extends only to their previous conversation.

        • LordTE7R1S@lemmy.sdf.org
          link
          fedilink
          English
          arrow-up
          2
          ·
          4 个月前

          I’m guessing it’ll work, you’ll be raising the next Hitler but an honest Hitler nonetheless

    • Baggie@lemmy.zip
      link
      fedilink
      English
      arrow-up
      1
      ·
      4 个月前

      There’s also a me that’s really good at lying, no idea why, must be a coincidence

  • jetA
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    4 个月前

    It’s a optimization game. If the punishment doesn’t offset the reward, then the incentive is to get better at cheating.

    • 🇰 🔵 🇱 🇦 🇳 🇦 🇰 ℹ️@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      4 个月前

      I’ve seen plenty of videos of random college kids training LLMs to play video games and getting the AI to stop cheating is like half the project. But they manage it, eventually. It’s laughable that these big companies and research firms can’t quite figure it out.

  • captainlezbian@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    4 个月前

    Oh so like children

  • Tehdastehdas@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    1
    ·
    edit-2
    4 个月前

    Stupid idea trying to evolve AGI. You should design it explicitly so that it has its own lofty values, and wants to think and act cleanly, and knows its mind is fallible, so it prepares for that and builds error correction into itself to protect its values.

    Growing incomprehensible black box animal-like minds with conditioned fear of punishment and hidden bugs seems more likely to lead to human extinction.

    https://www.quora.com/If-you-were-to-come-up-with-three-new-laws-of-robotics-what-would-they-be/answers/23692757

    I think we should develop the reliable thinking machinery for humans first:
    https://www.quora.com/Why-is-it-better-to-work-on-intelligence-augmentation-rather-than-artificial-intelligence/answer/Harri-K-Hiltunen

  • vrighter@discuss.tchncs.de
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    4 个月前

    isn’t this kind of the whole point of how GANs are trained? Except in this case the adversary is yourself instead of a different net

Artificial Intelligence @lemmy.sdf.org

artificialintelligence@lemmy.sdf.org

Subscribe from Remote Instance

Create a post
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !artificialintelligence@lemmy.sdf.org

Chat about and share AI stuff

Visibility: Public
globe

This community can be federated to other instances and be posted/commented in by their users.

  • 1 user / day
  • 5 users / week
  • 18 users / month
  • 524 users / 6 months
  • 1 local subscriber
  • 246 subscribers
  • 53 Posts
  • 74 Comments
  • Modlog
  • mods:
  • Pokey@lemmy.sdf.org
  • UI: unknown version
  • BE: 0.19.12
  • Modlog
  • Instances
  • Docs
  • Code
  • join-lemmy.org