• About Us
  • Subscribe
  • Contact us
Operations - BLOG

The End of Human Risk Management?

Gilles Hilary, INSEAD Professor of Accounting and Control and Arnaud Lagarde, Chief Risk Officer, Mandarin Capital |

Automation only works well when humans are highly involved.

Eric (not his real name) was under pressure from his sales department. He was hesitant to close a large financing deal with a Chinese corporate but had little beyond his intuition to back up his position.  The company’s stock price had gained a whopping 600 percent in one year. Nevertheless, Eric followed his intuition and ran a software analysis of the company trading activity. It didn’t take long for a strange pattern to emerge. There was strong activity at the end of most trading days that was pushing the stock up. He had enough to kill the deal.

A few weeks later, the stock crashed nearly 50 percent in one day, triggering an extended trading suspension pending an investigation by the local regulator. Unstructured data analysis combined with human intuition had saved Eric’s firm from a severe financial and reputational loss.

This example, far from being isolated, stresses the opportunities of automated data management. The monetary cost of conserving data has plummeted over the last few decades while the processing technology has dramatically expanded.  By now, machine-driven analysis has become as ubiquitous as Amazon or Google. 

This has paved the way for automated risk management.  Take cyber-security, for example.  It has evolved from passive protection based on anti-viruses and other similar technologies to real time monitoring based on behavioral indicators and then to dynamic cyber-defense. Machines now make decisions that used to be the purview of IT specialists.  As the amount of data to be treated in real time increased, the role of humans has been shrinking.

People power

However, big strategic decisions are still made by people, not machines. Hardly anyone feel passionate about an algorithm outcome.  Moreover, very few current senior managers understand concepts such as machine learning. This creates the perception of a “black box” that is detrimental to the credibility of machine-generated analysis.  Raw, or even processed, data needs to be converted into something that makes sense for humans such as stories or pictures. Data visualisation, for example, has become a booming industry.  This market is expected to exceed $6 billion by 2019 with a yearly growth rate of 10 percent. For instance, over 92,000 articles from major newswires in the U.S. are posted to the web each day. Naturally, it is impossible for a person to comprehend or summarise such dense data feed. Companies such as Amareos provide heat maps with end-of-day generic summaries. They provide a measure of sentiment after analysing the data and then a graphical representation of the output. Humans can then harness their intuition and their imagination to tell stories based on the graphics, a skill still not readily available to machines.

More broadly, the key point is to establish complementary relationship with machines.   For example, inside financial institutions, automated processing creates correlations between risk models that may induce systemic instability. Recently, flash crashes have started to affect financial markets on a regular basis.  Algorithmic models, often executed at high-frequency, are designed to divide large trades into several smaller ones to manage market impact and risk.  However, algorithmic trading speed and interconnectedness can result in the loss of billions of dollars in a matter of minutes when its cumulative effect reaches a tipping point.

This stresses the need for circuit breakers; a role for humans to play.  Acting as such, an officer by the name of Stanislav Petrov is said to have prevented WWIII in 1983.  Petrov was the duty officer for the newly installed soviet nuclear early-warning system when it detected a few nuclear missiles incoming from the United States. Considering a limited nuclear strike to be implausible, Petrov surmised it was a false alarm.  His judgment, later proven correct, prevented a potential soviet counter-strike.

Beat this

Another issue with machines is their performances in fuzzy environments. While IBM’s Deep Blue supercomputer scored a victory over Chess world champion, Garry Kasparov, as early as 1997, supercomputers have yet to achieve similar victory against grandmasters of Go, a traditional Chinese strategy board game involving squares and pins. While Chess is well suited for algorithmic analysis, Go is based on principles that rely more on qualitative judgment and are somehow easier to visualise for humans. Human intuition is still hard to beat for machines.

This became clear during a 2005 Chess tournament opened to human players, computers, or team composed of both.  Chess grandmasters and best-in-class machines (similar to Deep Blue) competed. The winner turned out to be a pair of amateur chess players using relatively weak laptops. Their comparative advantage was not to play Chess better than the grandmasters or to enjoy better raw computer power from their machines, but to deploy computers more efficiently to help them make the right decisions. Kasparov concluded from that tournament that “Weak human + machine + better process was superior to a strong computer alone and, more remarkably, superior to a strong human + machine + inferior process.”  That is good news for human risk managers.  They are probably not disappearing any time soon but their role will be largely redefined into a more strategic partnership with machines.

Gilles Hilary is an INSEAD Professor of Accounting and Control and The Mubadala Chaired Professor in Corporate Governance and Strategy. He is also a contributing faculty member to the INSEAD Corporate Governance Initiative.

Arnaud Lagarde is the Chief Risk Officer of Mandarin Capital. He is currently studying INSEAD’s International Directors Programme.

Add a comment Already a member?
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.