Steps Ahead


Points of discussion:

February 4th

February’s seminar was given by Dr Nicolas Gaussel, Metori Capital Management. He discussed Environmental, Social and Governance (ESG) investing at a portfolio level, resulting from joint research with Laurent Le Saint, with a talk titled, “ESG risk rating of alternative strategies”.
Firstly, Dr Gaussel gave some background of ESG investing, its recent growth, how it is measured and giving examples of ratings. He then stressed the need for a methodology to aggregate ESG ratings at a portfolio level, since the common approach of using a formula based on the weighted average of the scores for individual ratings, was meaningless for hedge funds, who may have leveraged or short positions in an asset.

Dr Gaussel then considered the problem of ESG ratings on long/short and leveraged positions, proposing several interpretations of ESG risk and ESG rating. He showed that the risk of a short position is similar to that of a long position. He proposed that the ESG correlation can be inferred from the correlation of returns on a portfolio, which can then be used to define the overall and ESG specific risk. The ESG rating is then calculated as the ratio of the ESG specific risk to the total risk.

Finally, Dr Gaussel gave an example of applying this methodology to a fund which included futures on stock indices, bonds, currencies and short-term interest rates. Dr Gaussel concluded the talk by reiterating that it is necessary to consider how to produce an ESG rating at a portfolio level and the methodology that he proposed using a risk score based on the ESG ratings of its constituents.

Frontiers in Quantitative Finance

A monthly event hosted by the Oxford Mathematical and Computational Finance Group. Providing a forum for academics and practitioners to discuss and debate new ideas in the modelling, management and regulation of financial risks and an opportunity for networking. Students, academics and professionals from the finance and government sector are welcome to join the online seminar.

Attendance is free of charge but requires prior registration.


Thursday 4th February, 18:00-19:15


Dr Nicolas Gaussel (Metori Capital).

Nicolas Gaussel is co-founder and CEO of Metori Capital Management, a Paris based research-driven systematic investment manager. Nicolas has 25 years’ experience in the financial industry, including 15 years at Societe Generale Asset Management. He was Chief Investment Officer at Lyxor from 2012 to 2016. He holds an engineering degree from French Ecole Centrale de Paris, and a PhD in Economics from Paris La Sorbonne.

Event synopsis:

ESG Risk Rating of Alternative Portfolios

Abstract: How to determine the rating of a portfolio, for which the ESG ratings of its underlying constituents are known? For mutual funds, Morningstar proposes a weighted average approach. Unfortunately, this approach cannot be extended to more general portfolios with leveraged or short positions.

We propose here a methodology to answer that question. We introduce the notion of ESG risk and define the ESG rating as the ratio between the portfolio’s ESG risk and its total risk. We show how both ESG and total risk can be computed with a variance-based formula. The empirical correlation matrix has to go through a prior shrinkage procedure to ensure that the ESG risk cannot be larger than the total risk. The methodology is eventually tested on the Lyxor Epsilon Global Trend Fund.

Points of discussion:

January 7th

January’s Frontiers in Quantitative Finance Seminar was given by Professor Darrell Duffie, Stanford University, where he discussed his latest research, “New approaches to dynamic credit-spread benchmarks”, a collaboration with Professor Antje Berndt & Dr Yichao Zhu from Australian National University and also with research student Zachry Wang, Stanford University.

Professor Duffie began his talk by referring to the new risk-free interest rates benchmarks that were introduced to replace LIBOR. The replacement U.S. benchmark, SOFR, resulted in dissatisfaction from a number of banks due to the increase in their funding costs as credit spreads increase. To address this issue, Professor Duffie proposed his idea for a new credit spread benchmark that he called an, “Across-the-Curve Credit Spread Index” (AXI), which he stressed is just one example of an index with more desirable properties than SOFR.

Firstly, Professor Duffie outlined his list of desirable criteria that would need to be met by any new benchmark: highly correlated with the bank’s cost of funding; statistically robust and difficult to manipulate by being based on a large number of transactions; adaptable as bank’s change their approach to funding. He then explained how AXI meets all these requirements, an index of the credit spreads on unsecured debt instruments issued by U.S. bank holding companies and their commercial bank subsidiaries. The index is a weighted average of credit spreads with maturities ranging from overnight to five years, where the weights reflect both issuance volumes in the primary market and transaction volumes in the secondary market.

Professor Duffie suggested that there is an additional efficiency gain if banks using a credit spread index in their lending, however, he pointed out this was preliminary work and further research was needed.

Professor Duffie found that extending this concept to investment grade corporate bonds increased the volume of transactions to nearly five-fold, and that by increasing the number of transactions further improved the robustness of the index since it was less prone to statistical noise and manipulation. He referred to this as the Financial Conditions Spread Index (FXI) which he found to be highly correlated with AXI spreads over the last few years. 

Points of discussion:

December 10th

Modelling the Oil Squeeze: Storage and Trading Opportunities in Oil Derivatives

December’s Frontiers in Quantitative Finance Seminar was given by Dr Ilia Bouchouev, Managing Partner of Pentathlon Investments and former Global Head of Derivatives at Koch Supply & Trading. He discussed the background and impact of negative oil prices in his talk titled “Modelling the Oil Squeeze: Storage and Trading Opportunities in Oil Derivatives”.

Firstly, Dr Bouchouev gave some background, referring to the event in April 2020, where for first time in history, the price of WTI oil future was negative. He pointed out that whilst negative prices are common in commodities markets, the magnitude of the negative price was unusual, closing at minus $40 per barrel.

Dr Bouchouev then went on to dispel a number of misconceptions such as this only affected the futures market, whereas in fact the impact to the spot price meant that oil producers were paying their consumers to buy oil. He then gave an overview of oil markets, explaining that the roll yield is the main contributor to returns and crucially that “inventory hedgers” are the largest participant in the oil futures market.

He stressed that oil price models must take into consideration storage cost and capacity limits. Dr Bouchouev showed that in early 2020 the impact of COVID-19 resulted in falling demand for oil which resulted in increasing inventories. In addition, transporting oil incurs a cost which is reflected in the spread between storage facilities.

To conclude, Dr Bouchouev explained that storage is important and has an impact on futures, reasoning that as inventories increase, inventory hedgers sell futures leading to systematic bias in trading signals. He suggested that for options, a Bachelier normal model gave an improved performance than compared to a Black lognormal model and that the method of linearisation provides a simple analytic solution.

Points of discussion:

November 19th

November’s seminar was given by Dr Richard Martin, Visiting Professor at Imperial College London.  He discussed negative prices and strikes in the Black Model with a talk titled, “Embedded Optionalities in Commodity Markets”.

Firstly, Dr Martin gave some background, referring to the event in April 2020, where for first time in history, the price of WTI crude was negative as a consequence of storage reaching maximum capacity. He discussed the implications of this on option pricing and the solution that was proposed to abandon the lognormality assumptions in the Black Model (which cannot handle negative prices/strikes), in favour of the Bachelier Model.

Dr Martin suggested that the Bachelier Model approach is problematic due to market familiarity with the Black Scholes Model, (which until this time, has worked well across all commodity products), and related properties of the Black model such as building volatility surfaces and the use of Greeks.

Dr Martin proposed that the Black model can be repaired given its ability to obtain closed-form solutions of pricing and its extension to the Merton jump-diffusions specified as exponential Levy processes which are easier to implement. He therefore reasoned that there was little merit in moving to the Bachelier Model. Finally, he suggested that rather than using local volatility, a better approach is to look for embedded optionality in the assets.

Points of discussion:

October 29th

On the 29th October, the Frontiers in Quantitative Finance Seminar Series, hosted by the Mathematical Computational Finance Group, the University of Oxford and sponsored by Mosaic Smart Data, met for its monthly webcast. This month’s speaker was Dr Charles-Albert Lehalle, Head of Data Analytics, Capital Fund Management Paris, who gave a talk on Reinforcement Learning in High Frequency Finance. This talk was focused around his joint work with Dr Othmane Mounjid (University of California, Berkeley).

Dr Lehalle started his talk by mentioning how Reinforcement Learning has been recently used for application on financial markets such as optimal trading and deep hedging. He stressed the importance of finding the optimal learning rate, the exploitation-exploration trade-off and highlighted results from the literature to obtain convergence. He explained his selection methodology for the learning rate and how to enhance the level of convergence.

Dr Lehalle then described his implementation, called a PASS algorithm, equivalent to a line-search for exploration-exploitation problems. He then showed results from the application of the PASS algorithm to several problems taken from the optimal trading literature: optimal placement of a limit order and the optimisation of liquidation of a large number of shares and compared the convergence rates to a benchmark.

Dr Lehalle showed that on both problems the PASS algorithm was more efficient and demonstrated faster convergence at the beginning of the learning phase.

Points of discussion:

September 24th

On the 24th September, the monthly Frontiers in Quantitative Finance Seminar Series, commenced for the forthcoming academic year, hosted by Mathematical Computational Finance Group, the University of Oxford and sponsored by Mosaic Smart Data.

This month’s speaker was Dr Vladimir Piterbarg, Head of Quantitative Research Analytics at NatWest Markets and his talk was about LIBOR reform and the Arc-Sine Law.

First Dr Piterbarg gave a brief history of LIBOR rates and the need to replace them with alternative benchmarks based on Overnight Risk-Free Rates (RFR). The biggest challenge this will pose is the trillions of existing derivatives contracts that reference the LIBOR rates. He then outlined the ISDA Fallback protocol, a mechanism for switching legacy contracts from LIBOR to RFRs, with most requiring a fixed spread, that he termed the LIBOR Adjustment Spread, to be applied to the RFR. This spread is defined as the median of the spread between LIBOR and the compounded Overnight Index Swap (OIS) rates, over the last five years from the LIBOR cessation announcement date, which is expected at the end of 2021.

Dr Piterbarg proposed a model of the future evolution of spreads using a numerically efficient algorithm to approximate the expected median value using the known historical data. The future values of the spread were modelled as a Brownian motion with time-dependent mean where a critical parameter was determined using the Arc-Sine Law.

Results highlighted differences between implied expectations of the spread in the market versus maximum theoretical values determined by the model. 

Future work is still in progress.

Points of discussion:

June 25th

For the second Frontiers in Quantitative Finance webcast in June, Professor Loriana Pelizzon from Goethe University, Frankfurt, gave a presentation on ‘Loss sharing in Central Counterparties: Winners and Losers’. Her research on the subject is a collaboration with Christian Kubitza from University of Bonn and Mila Getmansky Sherman from University of Massachusetts.

Their aim of their work is to determine the relative benefits of central clearing to different participants following default losses in derivatives transactions. Professor Pelizzon started the webcast by giving a brief background of the over-the-counter (OTC) derivatives market and default losses before the role of the Central Counterparty (CCP) was mandated. She described the role of the CCP and outlined two central clearing mechanisms: multilateral netting which offsets gains and losses across multiple counterparties; and loss-sharing in the event of a member defaulting, in which losses are shared among the remaining members.

For their model, Professor Pelizzon and her co-researchers have considered systematic risk and portfolio directionality using a network consisting of dealers only, and another with all users, and then observed how losses are managed or shared by the CCP.

Their research concludes that central clearing favours dealers over other market participants such as asset managers, hedge funds and non-dealing banks. Their reasoning is that the latter are penalised because of the rule of loss-sharing adopted by the CCPs. Furthermore, this imbalance in benefits to all participants is a reason why there is reluctance to choose to centrally clear when it is not mandated.

Points of discussion:

June 11th

For June’s Frontiers in Quantitative Finance webcast, Professor Marco Avellaneda from New York University explored Statistical Clustering, Hierarchical Principal Component Analysis and Portfolio Management. The research presented was carried out in collaboration with Juan Serur from New York University.

Professor Avellaneda discussed different types of factor models comparing those based on explicit versus mathematical factors. He went on to describe Principal Component Analysis (PCA) and the first eigen portfolio, which is the market portfolio. He then showed some results of PCA applied to stock markets and concluded that it is difficult to find a financial explanation for higher order eigen portfolios.

Professor Avellaneda continued by introducing Hierarchical Principal Component Analysis (HPCA) which aims to partition the stock market into clusters, explaining how he and his research partner constructed the correlation matrix, selecting a representative portfolio for each sector, which extends the work of Cont and Kan [2011]. He showed the results when applied to S&P 500 returns. Further analysis on four major global equity markets, in the US, Europe, Emerging Markets and China, showed that HCPA gave clear high order eigen problems.

Comparing PCA against HPCA using the percentage of variance explained, he concluded that PCA rises faster since it is a less greedy algorithm and the HPCA has a lower concentration given a number of components. HPCA is therefore more interpretable.

He ended the talk with a brief explanation of Statistically Generated Clusters and gave some suggestions for future work.

Points of discussion:

May 7th

May’s Frontiers in Quantitative Finance webcast, co-sponsored by Mosaic Smart Data, was led by Dr Brian Healy from Decision Science Ltd and Professor Andrew Papanicolaou from New York University, who gave a joint presentation on Principal Component Analysis (PCA) of Implied Volatility Surfaces.

The presentation was based on new research carried out in collaboration with Professor Marco Avellaneda from New York University and Professor George Papanicolaou from Stanford University.

As an introduction to the subject, Dr Healy began by explaining how they applied PCA to U.S. equities’ returns. He discussed issues such as how many components should be removed for the residuals to be random, how to construct the explanatory factors and the natural structure of the data. He then demonstrated this by applying some results from random matrix theory to the data.

Professor Andrew Papanicolaou followed by discussing the application of PCA using tensors to implied volatility surface data, which is a higher-dimensional problem. When performing matrix PCA on the volatility surface data, they concluded that the number of significant components is 9. As a result of the structure contained in option prices, this is lower than is typical for equity returns, where the number is usually 20.

In a similar approach to a market capitalisation weighted equity factor or index, Healy and Papanicolaou used the open interest on the option and its vega to create a portfolio of implied volatility returns that tracks the volatility surface ‘EigenPortfolio’ – suggesting this can be used in similar ways to the VIX, the S&P 500 volatility index.

Healy and Papanicolaou concluded the talk by explaining how retaining the tensor structure in the ‘EigenPortfolio’ improves how well it tracks the open interest weighted implied volatility returns portfolio.

The online event was hosted by the Oxford Mathematical and Computational Finance Group and was attended by more than 130 participants from around the world. Follow Mosaic Smart Data on LinkedIn to ensure you don’t miss information on the next Group meeting which will be in June.

To read the full presentation of Hedging with Neural Networks, by Professor Johannes Ruf and Weiguan Wang, click here.

Points of discussion:

April 16th

In April’s Frontiers in Quantitative Finance webcast, co-sponsored by Mosaic Smart Data, Professor Johannes Ruf from the London School of Economics (LSE) spoke about the use of neural networks as estimation tools for the hedging of options.

As Professor of Mathematics at the LSE, he has conducted research with one of his PhD students, Weiguan Wang. The aim of Ruf and Wang’s work was to investigate the application of neural networks (NNs) as a tool for non-parametric estimation, discussing their implementation, comparing the results to several benchmarks and analysing the results, giving possible explanation.

Professor Ruf discussed how NNs have been applied to the pricing and hedging of derivatives over the past 30 years, commenting that most research papers have claimed that NNs have superior performance over other benchmarks.

He and Wang have implemented a NN, ‘HedgeNet’, which is trained to minimise the hedging error rather than the pricing error. Professor Ruf described their implementation of the model, the choice of input features and time periods. The data sets used were the end-of-day and tick prices of S&P 500 and Euro Stoxx 50 options.

He compared their results using Black-Scholes and linear regression as benchmarks. The results showed that HedgeNet outperformed the Black-Scholes benchmark significantly but not the linear regression model, which incorporated the leverage effect. 

Professor Ruf argued that the outperformance of NNs previously reported in mathematical literature is most likely due to a lack of data hygiene such as the incorrect treatment of the time series data or using an incorrect split of the test and training data.

The online event was hosted by the Oxford Mathematical and Computational Finance Group and was attended by 167 people from around the world.

To read the full presentation of Hedging with Neural Networks, by Professor Johannes Ruf and Weiguan Wang, click here.

Points of discussion:

February 20th

In February’s Frontiers in Quantitative Finance event co-sponsored by Mosaic Smart Data, Professor Olivier Guéant from the Université Paris 1 spoke about how automation in financial trading is taking the next step in market making.

Professor Guéant discussed research that he and his team have conducted on how to build market making algorithms for options on liquid assets. He discussed problems faced by market makers and the models that price-setting is based upon.

He formulated the problem as one of optimisation which is complicated due to the ‘curse of dimensionality’. His research uses ‘The Greeks’ to obtain dimensionality reduction, reducing the problem down to a system of linear ordinary differential equations which becomes easier to solve.

In giving numerical results he referred to assumptions he and his team have made, as well as ideas for further work in the algo market making space.

The event was hosted by the Oxford Mathematical and Computational Finance Group and held at Citi Stirling Square.