My previous post was about the remote controlled drones the US military are increasingly using to bomb the shit out of remote tribes and people who don’t agree that their countries are annexed bu the US and it’s coalition of the willing killing. Now while all this remote killing of the natives is good old fashioned fun for all the oldie top guns being recalled from retirement it is less atractive when some of the other robot remote control killers go “Terminator” on it’s owners. At least that is the point this report is trying to point out. So while it is OK to kill everybody who does not agree woith you it is of the utmost importance that the robots you use to do it with have some form of “Do not kill your owner” software written to prevent that from happening.
Robots must learn to obey a warrior code, but increasing intelligence may make keeping the robots from turning on their masters increasingly difficult
Robots gone rogue killing their human masters is rich science fiction fodder, but could it become reality? Some researchers are beginning to ask that question as artificial intelligence advances continue, and the world’s high-tech nations begin to deploy war-robots to the battlefront. Currently, the U.S. armed forces use many robots, but they all ultimately have a human behind the trigger. However, there are many plans to develop and deploy fully independent solutions as the technology improves.
Some mistakenly believe that such robots would only be able to operate within a defined set of behaviors. Describes Patrick Lin, the chief compiler of a new U.S. Navy-funded report, “There is a common misconception that robots will do only what we have programmed them to do. Unfortunately, such a belief is sorely outdated, harking back to a time when . . . programs could be written and understood by a single person.”