instead of being killed by murdering robots controlled by an AI, we’re blackmailed by an AI

  • redtea@lemmygrad.ml
    link
    fedilink
    arrow-up
    6
    arrow-down
    1
    ·
    edit-2
    1 year ago

    So they can’t execute code when receiving certain prompts? I know what you mean about not being ‘programmed’ but do they now do more than regurgitate text? What if someone were to ask gpt for something illegal, would it not flag that with a user-profile report? Sounds like a huge flaw, if it can’t do that.

    • albigu@lemmygrad.ml
      link
      fedilink
      arrow-up
      5
      ·
      edit-2
      1 year ago

      I don’t know the internals of Bing, but they have some triggers which themselves seem to be made with NLP. They use it a lot to fetch web info. That means that if the model somehow produces some creative version of a crime that doesn’t get caught, it’ll just send.

      I think this is why Bing sometimes refuses to continue the conversation, or ChatGPT will flag its own text as against their terms sometimes. But yeah, they definitely can encourage how to crime sometimes, I’ve made ChatGPT explicitly tell me how to replicate some crimes like the Armin Meiwes cannibalism one while bored.

      • redtea@lemmygrad.ml
        link
        fedilink
        arrow-up
        4
        arrow-down
        1
        ·
        1 year ago

        The legal cases are going to be fun reading when they come out!

        The AI made me do it