• ImplyingImplications@lemmy.ca
    link
    fedilink
    arrow-up
    2
    ·
    1 year ago

    There are thousands of sci-fi novels where sentient robots are treated terribly by humans and apparently the people at Boston Dynamics have read absolutely zero of them as they spend all day finding new ways to torment their creations.

    • NιƙƙιDιɱҽʂ@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      1 year ago

      Those are just brainless bodies, currently. They don’t have sentience and have no ability to suffer. They’re nothing more than hydraulics, servos, and gyros. I’d be more concerned about mistreatment of advanced AI in disembodied form, something we’re dabbling potentially close to currently.

        • NιƙƙιDιɱҽʂ@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          1 year ago

          I disagree. I care greatly about not mistreating anything with consciousness and worry of where that line is and how we’ll even be able to tell that we’ve crossed it.

          I also recognized that a machinized body without a brain is exactly that - a cluster of unthinking matter. A true artificial intelligence wouldn’t be offended by the mistreatment of inanimate gears and servos any more than I would be. The mistreatment of an intelligent entity, however, is a different story.

      • LillyPip@lemmy.ca
        link
        fedilink
        arrow-up
        0
        ·
        1 year ago

        Food for thought, though: we thought the same thing about all other animals until only a couple of decades ago, and are still struggling over the topic.

        • NιƙƙιDιɱҽʂ@lemmy.world
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          …Just no. Animals are complex organic beings. Of course, we don’t understand them. Machines, though? We built machines from the literal Earth. Their level of complexity is incomparable to that of anything made by nature.

          Now, take a sufficiently advanced neural network that’s essentially a black box that no human can possibly understand entirely and put it inside of that machine? Then you’re absolutely right. We’ll get there soon, I’m sure. For now, however, a physical robotic body is just a machine, no different than a car.

    • LillyPip@lemmy.ca
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      1 year ago

      People think I’m crazy for apologising to my roomba when I trip on it and for saying please and thank you to Alexa and Siri, but I won’t be surprised at all when the robots rise up, considering how our scientists are treating them. I’ll have a track record of being nice, and that has to count for something, right?

  • sleepy@reddthat.com
    link
    fedilink
    arrow-up
    0
    ·
    1 year ago

    Isn’t that a part of the ai marketing though? That whole “this thing could destroy us” stuff?

      • visak@lemmy.world
        link
        fedilink
        arrow-up
        0
        ·
        1 year ago

        The current stuff is smoke and mirrors and not intelligent in any meaningful sense, but that doesn’t mean it isn’t dangerous. It doesn’t have to be robots with guns to screw over people. Just imagine trying to get PharmaGPT to let you refill your meds, or having to deal with BankGPT trying to figure out why it transfered your rent payment twice. And companies are sure as hell thinking about using this stuff to get rid of human decisionmakers.

        • Square Singer@feddit.de
          link
          fedilink
          arrow-up
          0
          ·
          1 year ago

          That is totally true but that’s a different direction than the danger in the marketing as discussed above.

          The media is full of “AI is so amazingly great, we are all going to lose our jobs and it will take over the world.”

          That’s a quite different message than what’s really the case, which is “AI is so shitty, that it will literaly kill people with bad advice when given the chance. And business leaders are so shit that they willingly trust AI, just because it’s cheaper.”

          • Baylahoo@sh.itjust.works
            link
            fedilink
            arrow-up
            0
            ·
            1 year ago

            This is my biggest concern. I’m in a position where (potentially in the near future) I see AI being used as an excuse to do work quicker so we can focus on other things more but still have to review the AI response before agreeing/signing off. Reviewing for accuracy takes just as long as doing it yourself when it’s strongly regulated and it comes down to revisions and document numbers. Much less making a sound argument that actually is up to date with that documentation. So either I trust the AI short cut and open myself up to errors, or redo all the work for them. No gain in time efficiency with shorter timelines. I’d rather make something and have it flag things that I can check so I’m more sure of my own work. What I do shouldn’t be faster, but it can be more error free. It would take a lot of training and updating of training with each iteration of documentation change. I could be the slave of change, with more expectations, with no actual improvement of the tools I have (in fact more risk of issues with the tools being used).

            • psud@aussie.zone
              link
              fedilink
              arrow-up
              0
              ·
              1 year ago

              I’m in agile development, in a reasonably safe-from-AI position (scrum master).

              There has already been a trial of software development by AI, with different generative AIs in each agile role; and it worked.

              Bard claims to be able to write unit tests

              I can imagine many IT jobs becoming less skilled

              • Baylahoo@sh.itjust.works
                link
                fedilink
                arrow-up
                1
                ·
                9 months ago

                Sorry this is months after, but it’s cool to see it worked. I use a software called XXX Agile and it’s not the worst I work with but when ported to my company has some flaws. There’s a long project to switch somewhere else for document control and people who should know much better than me are worried it will fill some gaps but open us up to way more.