Commentary & Insights

CFM Talks To: Dr. Bruno Dupire

We had the pleasure of welcoming Dr. Bruno Dupire at our offices in Paris for a candid discussion about the world of finance in general, the status of quantitative finance and research in particular, and his views on a variety of developments set to shape the industry. Bruno is currently the Head of Bloomberg Quantitative Research, and a former head of derivative research at Société Générale, BNP Paribas and Nikko FP. Bruno also lectures in the Courant Master of Sciences Program at NYU. He is probably best known for his ground breaking research in derivative pricing, having shown how to derive a local volatility model consistent with a surface of option prices across strikes and maturities – establishing the so-called Dupire’s approach to local volatility for modelling the volatility smile. The Dupire equation has since become a standard tool in the industry and has been used to price trillions of dollars of options over the years. Bruno also developed the Functional Itô Calculus which is a framework for path dependency with many applications, notably, the decomposition of volatility risk across strikes and maturities. He was awarded the Lifetime Achievement Award by Risk magazine in 2008 for his pioneering work in local volatility modelling.

CFM:

You are head of Quantitative Research at Bloomberg. Can you briefly discuss the nature of you and your team’s research? What is it that Bloomberg Quantitative Researchers typically do?

BD:

We serve the needs of a broad community of over 300,000 users with incredibly varied needs. We cover derivatives, machine learning (ML), portfolio construction, pricing of illiquid assets, electronic trading, visualisation methods, election prediction and much more… We collaborate with numerous teams internally, but we also develop our own initiatives: for instance we have originated an ambitious project – a quant platform with innovative tools for visualisation and interaction. It enables the Bloomberg user to create elaborate studies.

We are also deeply involved in thought leadership – we present our work in the industry and academic conferences, publish research papers and we are running the Bloomberg Quant (aka BBQ) seminar: the world’s largest (200 attendees on average) monthly quantitative finance series. It is a recognised intellectual forum for the community.

CFM:

What do you regard as the most discernible shifts in the focus of quantitative research within the industry? What do you see as the next frontier in quantitative research?

BD:

Over the past years we have witnessed a shift from the sell side to the buy side. From eliminating risk premium for the risk neutral pricing of derivatives, to seeking risk premium for investments. From theory to data. From stochastics to statistics. Indeed, over the past five years the quant community has massively embraced ML, AI and the use of alternative data sets.

At Bloomberg we deploy a massive effort in data science: we now employ hundreds of engineers and we are very active on the communication side as well, together with the heads of ML and of Data Science. I have been talking about it in Europe, Russia, Asia, North and South America – everywhere only over the past year.

CFM:

When did you first get involved in AI?

BD:

My own involvement with AI predates my last 30 years in finance. In the early 80s I obtained a Master’s degree in AI in Paris and in the mid-80s I worked with Francisco Varela on self-organising systems when we started experimenting with neural nets. Then I had the idea of applying neural nets to time series forecasting and sold, in 1987, a study on currency exchange rates to Caisse des Dépôts. It was probably a first. After that I mostly underwent the “winter of AI” until the early 2010s when the company and my group became active in it.

CFM:

You mentioned the advent and large-scale uptake of machine learning and big data as key developments that have the potential of upending the industry. In fact, in a recent article in Risk magazine, entitled “Do or die – asset managers take up data science”, the author stresses a belief amongst many managers that gaining expertise and keeping ahead of the curve in data science (and machine learning, amongst others) is indispensable for staying competitive in the industry. Do you share this view?

BD:

Some say it is the future, others say it still needs to prove its relevance in finance. In my view, ML in finance is here to stay, but it will not solve everything. Currently the hype is intense, so intense that many feel compelled to pepper their speeches with phrases such as AI, ML, big data, predictive analytics, deep strategies, which is all often a varnish to hide a void.

Finance has always tried to relate the available current information to future behaviour in order to improve investment or risk management decisions and ML is the approach of choice to mechanically establish these links. ML relies on three pillars: data, algorithms and computing power. There is a galore of alternative data in finance and the algorithms and computing power have improved substantially. It is thus natural that finance has embraced ML. ML/AI can be used for Natural Learning Processing (NLP) (sentiment extraction, queries, comment generation), option pricing, outlier detection, data compression, covariance matrix cleaning, regime linked strategy selection, pricing of illiquid assets, transaction cost analysis, among other examples.

Some aspects of finance are similar to physics, with experiments that can be repeated, other aspects are more grounded on game theory, with a circle of anticipating the anticipations. In the first category we can find option pricing. Computing one option price as a function of its parameters and of the model parameter may be complicated but learning the whole pricing function can be easier once enough examples have been presented.

On the game theory side one can find trading and investment. It is impossible to repeat an experiment exactly under the same conditions. Regularities such as performance of strategies according to the market regime can be observed, but there is no guarantee of their persistence. The physical side can clearly benefit from ML, which is less obvious for the “game theoretical” one.

CFM:

The explosion in ‘alternative data’ and, as a consequence, a proliferation of alternative data providers is a challenging undertaking for most asset managers, not to mention separating the chaff from the wheat. How do you see managers positioning themselves in this era of an ever-growing amount of data and data providers?

BD:

Opportunities rotate quickly and one has to be nimble to identify and exploit them. Many people will be late to the game but the providers of alternative data will definitely be busy for quite some time.

Examples of alternative data are text, satellite images, supply chain, Environmental, Social, and Governance (ESG), vessel routes, credit cards, geolocation data, etc. But data is certainly not enough. When most market participants have access to the same data, what makes the difference is the ideas and the tools.

CFM:

Having access to all this new, and often very enticing data sources, is embraced as a revolutionary boon for data science and financial research. Do you harbour any reservations about the use and application of data?

BD:

There is a belief, or illusion, that everything can emerge from the data itself, just let the data speak. But data also benefits from rules. Machines can learn from examples but they certainly can benefit from explanations and guidance.

Another issue is the use of the data. It is very difficult to read causation from data. If when I observe that when the value of X is A then Y tends to be around B, does it mean that when I force X to be A then Y will be around B? This question cannot be resolved by the data itself. In the absence of randomly controlled trials, it is necessary to have a causal assumption, as brilliantly explained by “The Book of Why” of Judea Pearl. This is how one can reveal the structure of dependencies without intervening.

CFM:

You included ESG data as another area of focus for, and of alternative data providers, with new sources springing up all the time. Which, if any, of the common reservations about ESG data do you share?

BD:

There is a pressure for transparency, a premium for reporting and there is a massive shift towards increased disclosure. Currently the best data is probably from G, governance (independence, entrenchment, shareholder rights, diversity…), with some solid reporting especially from the early 2010s. However, it is difficult to harmonise data across providers and cultures; for instance the notion of board independence differs according to each region. Sustainability data depend partly on the sector and S, social, is especially patchy. A lot of effort is made to backfill and curate the data and push for better future disclosure.

The data, even if not a complete set, is voluminous. There are time series for hundreds of fields for thousands of stocks. Whether you intend to define a “goodness” score or to find value in the data for factor investing, you need good tools to represent and navigate the data. My team is working on novel ways to visualise and navigate the data that makes it easier to reveal associations.

CFM:

ESG, it would seem, has now become fully embedded in the industry framework. INDEFI, a strategic advisor for the asset management industry, entitled their latest survey, “The Rapid Mainstreaming of ESG Investment Practices among European Asset Owners”. What is your take on the ability of asset managers, especially quantitative, and systematic managers to respond to the ever-increasing ESG demands set by investors?

BD:

Whichever composite factor can be extracted by inspecting past data, the future is bound to be different. ESG data will deeply affect investment decisions due to its ethical dimension and regulatory pressure. Numerous sovereign funds and asset managers express explicitly their preference for good ESG investments, sometimes acting as activists to change corporation practices. Many millennials want to invest only in good ESG stocks. This will generate massive asset migrations, which opens a whole array of opportunities.

CFM:

It is not uncommon to hear that certain institutional investors espouse the mantra that ‘simpler is better’, finding it much more palatable to avoid what they deem to be ‘black box, or black box like” strategies. Do you perceive any disenchantment from the industry towards quantitative strategies?

BD:

Data snooping, overfitting, or apophenia (tendency to interpret noise as a pattern) is indeed a huge pitfall. An algorithm that optimises over the past offers no guarantee of good future performance. The more complex, the more prone to overfitting. The way out is a parsimonious model and a solid narrative that can support statements such as “It has worked in the past and here is the reason it is likely to carry on working”.

The market is a machine that destroys signals. If a quant finds a signal that has worked in the past, he is likely to not be the only one and quickly competitors will try to grab this opportunity away from him. Risk premia are not a law of nature. Opportunities vanish quickly and the investor needs to be creative and to have efficient tools.

However, there are some more resilient strategies, for instance based on behavioural principles. Cognitive biases are here to stay: even to be aware of your own cognitive biases does not prevent you from following them. This means that over-reaction, disposition and endowment effects, conjunction fallacy, remorse aversion, anchoring, herding and reaction to sunk costs will not disappear.

CFM:

What do you see as the key risks, or pitfalls for quantitative investment managers?

BD:

As mentioned earlier, a major issue is overfitting. Beyond that, a pervasive problem, whether it be in biology, social sciences or quantitative finance, is a clash of culture between the domain expert and the data scientist, with relevance being the collateral damage. Using NLP to analyse DNA sequences without knowledge of the field will generate biological nonsense. Similarly, using convolution nets to link returns to characteristics is perilous. The portfolio manager and his data scientist have to both make an effort to create a compatible conceptual platform.

CFM:

Do you have any pet peeves about the industry?

BD:

A dimension of quantitative finance that I find sorely missing is what financial engineering was supposed to address: use techniques to bring solutions to different economic agents. Whether it be in derivative product design or in asset allocation, it seems that the needs of the individual agents are somewhat disregarded. As an example, what matters for a retiree is not if midcap pharmaceuticals in Asia will outperform the market but rather that she will not over live her savings. Her risk is her own longevity, not the performance of the value factor. The asset allocator should definitely adapt to the investor and his or her personal benchmark.

CFM:

Any interesting projects you and your team are focussing on at the moment?

BD:

Beyond its daily tasks and its modelling role, my group explores many different directions: perceptual tests, colour theory, hand movements to input parameters, assistive technology. Moreover we have open sourced bqplot, our graphical library. It is important to give back to the community; the best researchers do not like to do just one thing and they want to have a purpose.

DISCLAIMER
The views and opinions expressed in this interview are those of Dr. Dupire and do not necessarily reflect the official policy or position of either CFM or Bloomberg lP or any of their affiliates. The information provided herein is general information only and does not constitute investment or other advice. Any statements regarding market events, future events or other similar statements constitute only subjective views, are based upon expectations or beliefs, involve inherent risks and uncertainties and should therefore not be relied on. Future evidence and actual results could differ materially from those set forth, contemplated by or underlying these statements. In light of these risks and uncertainties, there can be no assurance that these statements are or will prove to be accurate or complete in any way.