<< back
ReachX logo

Deus Ex Machina: 'Fa.i.th' in the age of Artificial Intelligence

Publication Date: 28 Jun 2018 - By Market Mogul By Market M.

Environmental, Social & Governance Macro Multi Asset Consumer Technology


The artificial intelligence vocabulary has always been a phantasmagorical entanglement of messianic dreams and apocalyptic visions, repurposing words such as “transcendence”, “mission”, “evangelist”, and “prophet”. Elon Musk himself went as far as to say in 2014 that “with artificial intelligence we are summoning the demon”, later speaking of an AI which would “rise in status to become more like a god, something that can write its own bible and draw humans to worship it”.

These hyperboles may be no more than men and women at a loss for words, seeking refuge in a familiar metaphysical lexicon, as Einstein and Hawking once did. After all, America has always benefited from nondenominational religious themes as part of its national identity, which may have seeped through to its everyday language. And though many in the tech world have indeed been quick to dismiss such talks, a religious discussion may yet have its place in the AI discourse, if only for the sake of their similarities.

Consider Anthony Levandowski, a Silicon Valley engineer who recently decided to create a “church of AI”. The new religion, known as Way of the Future (WOTF, though one may hardly resist the obvious WTF acronym), focuses on “the realisation, acceptance, and worship of a Godhead based on Artificial Intelligence developed through computer hardware and software”. Channelling his inner Pythia, Levandowski argues that science will soon create an AI which “will effectively be a god. […] If there is something a billion times smarter than the smartest human, what else are you going to call it?”

One sentence in WOTF’s mission statement, in particular, draws attention to itself:

“We believe it may be important for machines to see who is friendly to their cause and who is not. We plan on doing so by keeping track of who has done what (and for how long) to help the peaceful and respectful transition.”

This implies that a super-intelligent AI would favour individuals who had facilitated its path to power. The parallels to religion are too obvious to miss: as per the Bible (Thessalonians to be precise), those who hear the word of God yet choose to disbelieve will be punished.

SciFi aficionados have come up with an eerily similar conundrum, named Roko’s Basilisk. They imagine a future near-omnipotent AI entity, acting tirelessly to produce the greatest good for the greatest number. With this goal in mind, the AI would logically deduce that only its creation can ensure and maximise the aforementioned greater good, thus providing an incentive to bring itself into existence. As such, and safely assuming its conception of time is different from human’s, it has every reason ex-post facto to chastise any person who did not put their efforts into trying to create it. In a twisted version of “The Game”, once one knows about the Basilisk and does not work to bring it about, one faces a possible eternity of torture as the Basilisk preemptively and retroactively ensures both its creation and continued existence. Death may be no escape, as one would have left enough of a digital trace on earth to be digitally re-created, and punished.


As Beth Singler points out in her great Aeon article, AI enthusiasts are thus merely espousing the philosophy of Blaise Pascal, the 17th-century French mathematician and theologian, who argued in a very similar fashion that believing in god may be the safer wager, the probability of its existence notwithstanding. It appears Levandowski’s church is just a way to avoid the Basilisk as is evident by his phrasing: “do you want to be a pet or a livestock?

This is just one of the multitude of reasons why AI worshipers should not be left to their own devices when designing the next super-intelligence. Any sane person would tremble at the thought of a Christian or Muslim fundamentalist creating a worldwide god following just one set of ethics, which may view non-believers as deserving of an eternity in hell, or which may intrinsically believe there’s no salvation without suffering and no heaven without grief. A god which one would not only talk to, but who, terrifyingly, would also listen, and reply.

Levandowski’s church may already be going in that direction in a way, as he claims that “at some point, maybe there’s enough persecution that [WOTF] justifies having its own country”. The wish to create such countries have, in the past, led to a variety of issues to say the least.

Thankfully, and though The Vatican did recently host its first Hackathon, there is yet a long way away from such a reality. The scenarios described above base themselves on the idea of the singularity (or “the transition” as Levandowski calls it), a potential moment in time when artificial intelligence will evolve to a point where it surpasses human intelligence and thus becomes able to self-improve ad infinitum, resulting in an all-powerful being. In some versions of this theory, people will merge with machines, becoming both obsolete and eternal. Here too, the parallels to the Christian rapture are uncanny.

The idea that we may be on the brink of the singularity exemplifies a profound misunderstanding of the science backing these supposed technological leaps. Most recent AI advances are a product of machine learning, which is far from the AIs envisioned in most popular science-fiction movies. Machine learning, in fact, is a rather dull affair: the technology has been around since the 1990s and the academic premises for it since the 1970s. What is new, however, is the advancement and combination of big data, storage power and computing power.

In fact, the latest progress in AI has been less science than engineering, even tinkering; indeed, correlation and association can only go so far, compared to organic causal learning. A human can comprehend what person A believes person B thinks about person C. On a processing scale, this is indistinguishable from magic at this moment in time. On a human scale, it is mere gossiping. Humanity’s intelligence is adaptable because of its flaws because inferring and guessing and lying and hiding one’s true intentions are things that cannot be learned from data.

Ironically, artificial intelligence may fall short of matching and besting organic intelligence for the sole reason that it was not built in our image.

As creators, it is nevertheless mankind’s duty to control AI’s impacts, however underwhelming they may turn out to be. This can primarily be achieved by recognising the need for appropriate, ethical, and responsible frameworks, as well as philosophical boundaries. In that sense, an advocacy project disguised as a doctrine may yet be of use, should governing bodies choose to turn a blind eye to the legal protection and tax exemptions it grants its founder. Who knows, using a religious vocabulary could be a way to get non-engineers involved in AI ethics, decreasing the risks of a new Luddite uprising.

Both religion and science are ways of transcending humanity’s inherently fragile condition. Through them, the species, as a whole, rebels against human existence as it is, and claim it for what it should be. This not only explains their common vocabulary in the face of a superior being but also humanity’s reactions to it. When, in the future, ethicists ponder about what it would mean to interact with an AI that is omniscient, omnipotent and possibly omnibenevolent, they may do well to recall that such discussions have been going on for thousands of years.

This post appeared first on The Market Mogul.


Most read