• ApfelstrudelWAKASAGI@feddit.de
    link
    fedilink
    arrow-up
    17
    ·
    11 months ago

    It does create the funny paradox where, up to a certain point, a rational utilitarian would choose to kill and a rational mass murderer trying to maximise deaths would choose to double it.

        • interdimensionalmeme@lemmy.ml
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          edit-2
          11 months ago

          Doubling action forever minimizes human deaths.

          Unless someone decide to hit kill. In that case, it’s them doing it. I’m invalidating the argument that pre-empting imaginary future mass murders justifies killing one person today.

          • ApfelstrudelWAKASAGI@feddit.de
            link
            fedilink
            arrow-up
            3
            ·
            11 months ago

            Idk which moral system you operate under, but I’m concerned with minimising human suffering. That implies hitting kill because chances of a mass murderer are too high not to. You also don’t follow traffic laws to a t, but exercise caution because you don’t really care whose fault it ends up being, you want to avoid bad outcomes (in this case the extinction of humankind).

            • interdimensionalmeme@lemmy.ml
              link
              fedilink
              arrow-up
              1
              ·
              11 months ago

              My moral system somehow does not chose to kill people through action against an imagined threat and is therefore objectively superior as is it not susceptible to hostile memetic manipulation (Molloch, Pascal’s wager, Pascal’s mugging, basilisks, social hysteria etc.) and is capable of escaping false choices and other contrived scenarios, breaking premise and the rules of the game as needed to obtain the desired outcome.

              • ApfelstrudelWAKASAGI@feddit.de
                link
                fedilink
                arrow-up
                1
                ·
                11 months ago

                Even if your moral system solves those “problems”, you just “solved” them by substituting the obvious and logical base of utility through personal responsibility. Personal responsibility is no inherent good, unlike utility, if people are unhappy/“feel bad”, it doesn’t matter how personally responsible everyone is being, that world is still a shit place.

                Also, the threat isn’t imagined. I can assure you that there are a lot more than one person on earth who would choose to kill as many people as possible if given the option.