Banking Fintech

Three rules for banking bots

Three rules for banking bots. Image by Cath Vectorielle, Shutterstock.com
Written by Chris Skinner

Sci-fi author Isaac Asimov wrote the Three Laws of Robotics. In this article, Chris Skinner proposes three rules for banking bots.

Nadia, the avatar based on IBM's Watson technology.Writing more and more about chatbots, robots and AI, I can see a day coming in the not too distant future where we won’t be able to tell the difference between the human and the machine. That’s not scary science fiction, as it’s now almost science fact. However, there’s still a long way to go, as demonstrated by IBM’s Watson Avatar.

Their AI human agent – usually a woman, as all the folks creating chatbots are sexist – looks human-ish, but no avatar or chatbot today feels properly human. In IBM Watson’s case, it’s the mouth movements that give away the fact she’s a machine, even though she’s in HD. 15 years ago, AT&T was working on exactly the same idea, but with scripts for avatars.

AT&T avatarAs you can see, the idea (and sexism) hasn’t changed much in that time. What is changing, however, is the technology behind this idea, and as with all great technological innovation, if something looks worthwhile to develop – biometrics, communicators, health technologies, life sciences, artificial intelligence, robots and so on – then eventually these technologies will develop sufficiently to become mainstream and acceptable.

That will be somewhere over the next 10-25 years to achieve. It feels nearer, but when a group of experts on artificial general intelligence were asked this question in 2012, their view was that it wouldn’t be achieved until 2040. The question here being: When will we achieve the singularity? From Wikipedia:

The technological singularity (also, simply, the singularity) is the hypothesis that the invention of artificial superintelligence will abruptly trigger runaway technological growth, resulting in unfathomable changes to human civilization. According to this hypothesis, an upgradable intelligent agent (such as a computer running software-based artificial general intelligence) would enter a ‘runaway reaction’ of self-improvement cycles, with each new and more intelligent generation appearing more and more rapidly, causing an intelligence explosion and resulting in a powerful superintelligence that would, qualitatively, far surpass all human intelligence. John von Neumann first uses the term “singularity” (c. 1950s), in the context of technological progress causing accelerating change: “The accelerating progress of technology and changes in the mode of human life, give the appearance of approaching some essential singularity in the history of the race beyond which human affairs, as we know them, can not continue”. Subsequent authors have echoed this viewpoint. I. J. Good‘s “intelligence explosion”, predicted that a future superintelligence would trigger a singularity. Science fiction author Vernor Vinge said in his 1993 essay The Coming Technological Singularity that this would signal the end of the human era, as the new superintelligence would continue to upgrade itself and would advance technologically at an incomprehensible rate.

The Three Laws of Robotics

As I thought about these things, it made me think along the lines of other science fiction authors and visionaries, such as Isaac Asimov. An amazing thinker about the future, he wrote a fantastic series about robots, and came up with the Three Laws of Robotics:

  • A robot may not injure a human being or, through inaction, allow a human being to come to harm.
  • A robot must obey orders given it by human beings, except where such orders would conflict with the First Law.
  • A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.

This makes so much sense, and has been used in many other places, notably in the film Robocop, who has three prime directives:

  1. Serve the public trust.
  2. Protect the innocent.
  3. Uphold the law.

This means Robocop can kill humans, but only if it conforms to the above. It made me wonder what the three rules for banking robots should be? We need to have some, otherwise they may run away with our money. After all, robots won’t adhere to national laws and rules, only the way they program and reprogram themselves. So, here’s my speculative take on three rules for banking robots:

  1. Do not rip customers off.
  2. Ensure you are secure.
  3. Freeze any activity that breaks the law.

I think they speak for themselves, but I’m happy to take comments or questions. Finally, if you missed it, this was my favourite recent article about robots.

READ NEXT: Chatbots – saviour or disintermediator?

– This article is reproduced with kind permission. Some minor changes have been made to reflect BankNXT style considerations. Read more here. Image by Cath Vectorielle, Shutterstock.com

About the author

Chris Skinner

Chris Skinner is an independent commentator on the financial markets through the Finanser, and chair of the European networking forum the Financial Services Club, which he founded in 2004. He is an author of numerous books covering everything from European regulations in banking through to the credit crisis, to the future of banking.

Leave a Comment