Thanks Asimov, but what robots need is Shakespeare
With true Artificial Intelligence in our not too distant future, how to do we ensure we end up with Wall-E and not Skynet?
In the Guardian article, We need robots to have morals. Could Shakespeare and Austen help? written by John Mullan, we learn that scientists are starting to think about how artificial intelligence, or AI, are going to make their decisions.
What code will they follow? How will they learn right from wrong? This is a consideration which is as old as the idea of AI itself. It’s been covered famously by Isaac Asimov and his three robot laws, and also by movies like A.I., Transcendence and Blade Runner.
Scientists at the School of Interactive Computing at the Georgia Institute of Technology have created a system which is intended to extrapolate moral guidance and life lessons from works of literature. The system is named Quixote, which, we can only hope, was named ironically.
Quixote (as in Don Quixote) for those of you who don’t know, is an insane Spanish gentleman who came to believe he was, in fact, a chivalric knight and rode around performing honourable if archaic, deeds, duelling windmills and generally being a nuisance. Fun fact: the book makes a point of saying this delusion/brain fever happens if you stay up at night reading books. Readers be warned
These scientists in Georgia are giving Quixote literature like Shakespeare and Jane Austen and is asked to take moral lessons and judgements from them.
The article ends pretty negatively. As an English literature academic, John Mullan is probably an expert worth listening to. He gives a damning assessment of literature as a moral compass for these robots and suggests that Isaac Asimov’s laws are probably wiser.
But I can’t help but disagree.
Literature can be a moral guide
Why shouldn’t we use literature as a moral guide and encourage AI to use it too? What we want is for these robots to be autonomous, but never to hurt humans. Basically, the message we’re giving our children. We are autonomous, free-willed individuals, who understand that kicking a guy in the ribs because he got the last seat on the tube doesn’t lead to a well-functioning society.
And yet, I don’t ever remember anything telling me not to kick strangers in the ribs. I’m just assumed to know.
The argument against Shakespeare as a moral teacher is obvious. Giving a computer Othello and asking it to create a code of conduct from that is… questionable. As is the morality you could derive from the titular protagonists of Hamlet and King Lear. In fact, most of Shakespeare’s plays give us a shady and out of date message of good through his varied lead characters.
We, on the other hand, are able to read Shakespeare and pick out the good bits, rather than generalising that all actions from the protagonist are acceptable behaviour. But that’s because we’re not starting out in life, tabula rasa style, with Shakespeare alone.
Imagine the world if we were.
Subscribe here for more!
In literature and television aimed at children, you’ll see simple messages dressed up in a story. Moral messages. It’s good to share. And bad to upset people. It’s good to help with chores. Fairy tales like Little Red Riding Hood tell us it’s dangerous to talk to strangers. Especially if they're an anthropomorphic wolf. E. Nesbit’s The Railway Children has an overwhelming message of putting others before yourself. Harry Potter repeats again and again the value of loyalty, friendship, family and teamwork.
These sorts of moral lessons taught early on give you a framework of how the world should work. It’s with this basic understanding we can approach more complicated situations and texts. The fact is, whether you’re a natural voracious reader, telly watcher or even someone who has learned wrong and right by others, literature plays a critical part in our morality.
Doctor Who. Tackling these issues so you don’t have to!
What fictional literature also does is give us bigger problems to apply mortality to, in a way we’ll never encounter in our day to day lives. Decisions which are far more complicated than simply ‘do no harm’.
Your mother can tell you not to hit another child, but she doesn’t tell you what the correct thing to do is if that child is actually a secret alien baby come to destroy the planet putting billions of lives at risk. Hypothetically. We’ll never have to face the thorny question of what we’d do if we could go back in time and meet Hitler as a child.
Are we giving Quixote the wrong reading list?
I don’t think there’s anything wrong in giving the system Quixote Shakespeare to read. Or Austen, Agatha Christe, J.K. Rowling or even Pratchett. Maybe skip G.R.R. Martin. But I don’t think it’s fair to throw Shakespeare at, what is basically, a robotic child. The moral guidance they contain is too obscured in history and context.
Why not start it with Dr Seuss and How the Grinch Stole Christmas? Or Roald Dahl’s The Witches, where children were turned into mice? Maybe Enid Blyton who wrote vaguely uncomfortable stories about dolls with black faces? Or Not Now Bernard by David Mckee which told us if you ignore your child for long enough, he’ll be eaten by a monster.
What I’m saying is, if AIs are ever going to be more like us, they need to be every bit as screwed up as we are. So start with the classics.
If you love the idea of A.I. as much as I do, this is the blog for you.
Because I seem to talk about it constantly! Check out these posts, Westworld, Do Androids Dream of Electric Sheep and Artificial Intelligence.