"In your article on the moral dangers of autonomous, lethally armed robots, Peter Asaro says "most people now feel that it is unacceptable for robots to kill people without human intervention." (18th April, p7). The moral reasoning behind this view is intriguing. How is sending a programmed, armed robot into an area designated as 'enemy occupied' any worse than, say, bombing the area from ten thousand feet? In fact, the level of precision and the amount of human judgement involved in target selection with the robot would be arguably greater."
"There is an even stranger moral angle. Someone who is ordered to go and kill strangers in a war can suffer severe emotional trauma and other mental distress as a result. In the future, there may be societies that decide, on moral grounds, to delegate all killing of the enemy in their wars to fully autonomous robots so as to protect their citizens from such emotional trauma. In that unnerving scenario, the robots wouldn't be seen by those citizens as devils, but heroic guardians."