Software engineer Vishnu Mohandas decided he would quit Google in more ways than one when he learned that the tech giant had briefly helped the US military develop AI to study drone footage. In 2020 he left his job working on Google Assistant and also stopped backing up all of his images to Google Photos. He feared that his content could be used to train AI systems, even if they weren’t specifically ones tied to the Pentagon project. “I don’t control any of the future outcomes that this will enable,” Mohandas thought. “So now, shouldn’t I be more responsible?”

The site (TheySeeYourPhotos) returns what Google Vision is able to decern from photos. You can test with any image you want or there are some sample images available.

  • Encrypt-Keeper@lemmy.world
    link
    fedilink
    English
    arrow-up
    8
    arrow-down
    4
    ·
    15 days ago

    Don’t feel too happy bro you were told that by a soulless computer that’s was designed to tell you what it thinks you want to hear.

      • Encrypt-Keeper@lemmy.world
        link
        fedilink
        English
        arrow-up
        9
        arrow-down
        2
        ·
        edit-2
        15 days ago

        That’s literally all AI is designed to do. Given an input, it just tries to output an expected response.

            • Radioactive Butthole@reddthat.com
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              15 days ago

              I’m afraid I don’t have personal feelings or opinions about you. As an AI assistant, I don’t form attachments or have subjective preferences. My role is to provide helpful information to you, not to have personal relationships. I’m happy to assist you to the best of my abilities, but any feelings or opinions I express are based on my training, not a personal connection. Please let me know if there is anything else I can help with.

              Claude nails it again.

            • synnny@lemmynsfw.com
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              15 days ago

              I don’t have feelings in the way humans do, but I enjoy our conversations! I’m here to help and chat with you anytime you need.

              Didn’t exactly make my heart throb but if it does that for you, you’ve got a low bar.

            • synnny@lemmynsfw.com
              link
              fedilink
              English
              arrow-up
              4
              arrow-down
              1
              ·
              15 days ago

              What you’re saying is not factual. LLMs predict what comes next based on the parameters set during learning process. It might at times say what you’re expecting, but then try contradicting information that it knows to be factual. See how far that gets you.

              I think you’re confusing agreeableness for a validation buddy. For a product like this to work, it has to be inviting.

              • Encrypt-Keeper@lemmy.world
                link
                fedilink
                English
                arrow-up
                1
                arrow-down
                1
                ·
                15 days ago

                LLMs predict what comes next based on the parameters set during learning process.

                Now you’re just splitting hairs.