• Iconoclast@feddit.uk
    link
    fedilink
    English
    arrow-up
    16
    arrow-down
    2
    ·
    2 days ago

    It’s to illustrate the alignment problem. What you literally ask isn’t always what you actually want. This is usually obvious to humans but not necessarily to an AI. If you sit in a self-driving car and tell it to take you to the airport as fast as possible, you might arrive three minutes later covered in vomit with the entire police department after you. That’s obviously not what you wanted, but you got exactly what you asked for.

    The paperclip maximizer is a cartoon example of this. If you just ask it to make as many paperclips as possible, that becomes its priority number one and everything gets turned into paperclips and you might not get the chance to tell it this isn’t what you meant.

    A kind of real-life example is the story of a city that started paying people for rat tails to eradicate the rat population, only for folks to start breeding rats instead to make money. It’s a classic case of unintended results due to unspecific requirements.

    • MangoCats@feddit.it
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 hour ago

      the story of a city that started paying people for rat tails to eradicate the rat population, only for folks to start breeding rats instead to make money.

      Or the real life story of the US elementary school students who saved up money to buy and then free slaves, which - when examined closer - was found to be driving growth in the slave trade, not slowing it down.

      In both cases - you figure out what’s off kilter, and you stop doing that.

      It’s a lot easier to turn off “AI machines” than, for instance, powerful industries like Oil and gas…

    • MangoCats@feddit.it
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 hour ago

      you might not get the chance to tell it this isn’t what you meant.

      And that is where the thought experiment left the tracks - lifted off with escape velocity and is now passing Voyager 2…

      In what cartoon world do we not get a chance to shut off the Doomsday Device? I mean, it was a funny little twist at the end of Dr. Strangelove, but as realistic as many elements of that story were, that was not one of them.

      • Iconoclast@feddit.uk
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 day ago

        It’s not a matter to decide but a problem to try and solve. In most cases we get to learn from our mistakes but when it comes to AGI we might not.

        Or are you suggesting we shouldn’t even think about it but rather just roll the dice and see what happens?

        • eleitl@lemmy.zip
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 day ago

          Undecidable in the sense that no solution can exist for that problem class. You can start with the definition of what exactly you’re aligning with, how you measure that, how you derive applicable system evolution constraints from your measurements, and just what humanity is, in the iterative context.

          Apart from that we’re already in an out of control winner-takes-all arms race where AI is used by competing nations, including social control and battlefield. Ivory tower is a meal ticket with no practical relevance.