After a great recent weekend at Future Assembly in Melbourne, I was compelled to repeat my talk, “Design Ethics for Artificial Intelligence” (slightly abridged).
Artificial Intelligence has been edging its way into our reality for a while now and it’s a topic that’s been discussed for decades. The fear of humans becoming slaves to AI is an interesting fear; some observers would say we’re already slaves to our devices and gadgets, therefore slaves to intelligence outside of ourselves.
The purpose of my talk was not to paint the expected doomsday view of AI and its possible effects on humanity but to open up the discussions on the complexity of embedding value systems in relation to decision making.
Knowing that to take action we need to assess a situation and make a judgement call. Where do these judgement calls come from? They’re our value systems and they’re complicated.
Using Isaac Asimov’s Three Laws of Robotics, I posed a single scenario (with some variants each time) and asked the audience to make a judgement call based on different value systems:
With a central character, Caitlin, our robotic butler, we posed these scenarios and presented a choice that she had to make based on the above value systems.
A great Q & A followed, exploring the frustrating, flawed, emotional and highly subjective complexity of the human condition aka, our value systems.
When designing Artificial Intelligence, what are we really designing? Further more, what happens when Artificial Intelligence is no longer Artificial and can ponder its very existence?
Download the presentation and discuss it at work and home. I’m not at all worried about AI, if it’s left alone; I’m concerned when humans, who are fundamentally flawed, design decision making into immature intelligence.