Moral calculus

Pareto optimality, utilitarianism, and murderbots. Staple of science fiction since the first robot, and probably since all the holy books of all the religions. cf Golems, contracts with devils. This has all become much more legible and quantifiable now that the golems are weaponised 3d-printable downloads.

The trolley problem in the age of machine agency, war drones and smart cars. (Also, what is “agency” anyway?) Hell, even if we can design robots to solve ethical dilemmas, do we want to? Do instinctual human ethics have an especially good track record? What are the universals specifically?

<iframe width="560" height="315" src="" frameborder="0" allowfullscreen></iframe>

To file: Journal of practical ethics.

For machines

For humans as cogs in the machine

Try moral philosophy.

No comments yet. Why not leave one?

GitHub-flavored Markdown & a sane subset of HTML is supported.