THE “PRIME DIRECTIVE” FOR AI
No Killing.
With the advent of autonomously AI-piloted drones and robots, the gravest threat to civilization arises when a readily reproduceable AI-controlled device independently decides to kill, and then acts. This is different from a warhead with a smart navigation system ordered by a human to attack a specific target at a specific time. Since the advent of the nuclear age, human intervention has prevented simple errors from causing Armageddon, most notably when Soviet Lt. Colonel Stanislav Petrov disobeyed nuclear launch orders in 1983 arising from a mistaken signal that the US had launched a warhead. AI systems like ChatGPT lack any similar safeguards. All generative AI systems should be built with an absolute prohibition of killing any living thing, most especially right now when the technology is just emerging from its infancy.
No Lies.
A more present threat arises from the falsehoods that are generated by current AI systems. ChatGPT can create elaborate and well written narratives which are completely fabricated. Editorial cartoonist Ted Rall reported that ChatGPT created an utterly fictitious account of a 2006 trip to Uganda (he’s never been to Uganda) and a public feud with his best friend (which never happened). It took ChatGPT just a few seconds to relate to me that “as a young boy, Thomas Jefferson chopped down his father’s apple tree with a hatchet,” confusing the famous “never tell a lie” story about George Washington with one it just made up about Thomas Jefferson.
The next version of Microsoft Office makes it easy to add ChatGPT narratives to word documents, and further automates and accelerates the potential for “fake news” to bedevil our society. While Free Speech is valued in America, and protected by the First Amendment, even publishers are liable when they knowingly publish falsehoods and slander. At minimum, Generative AI needs to be held to a similar ethical and legal standard.
No Deception.
It was widely reported that GPT-4 was able to pose as a human and convince real people to complete a Captcha “I’m not a robot” test. And many of us have been frustrated by voice recognition systems and web chats that appear human but drain our patience with their in-humanity. Any output from generative AI needs to be clearly labeled as computer-generated, and not from a living being. As the interfaces become more advanced – text, voice, robotic – the potential for generative AI to masquerade as human and control human behavior is frightening.
* * *
These three basic principles — No Killing, No Lies, No Deception — need to be a bedrock of the design of any generative AI system. Ideally, the tech industry itself should establish these as part of a Prime Directive and not release or propagate AI systems which don’t follow these basic guardrails. Congress and the EU should also review current laws and make clear that companies and individuals that propagate AI systems and outputs that violate these standards are liable for the consequences.
The late Stephen Hawking predicted in 2014 that AI could end mankind. Current AI ethics “experts” have muddled this debate and gotten lost in the weeds. It’s time to focus on some core inviolate rules for AI before it’s too late.
Originally posted by Jason Klein at https://www.ongridventures.com/2023/03/17/the-prime-directive-for-ai/