top of page

Publications

with Yannick Lucotte (LÉO) and Sessi Tokpavi (LÉO), 2019

Journal of Economic Dynamics and Control, 100, 86-114, 2019

Abstract: The aim of this paper is to propose a new network measure of systemic risk contributions that combines the pair-wise Granger causality approach with the leave-one-out concept. This measure is based on a conditional Granger causality test and consists of measuring how far the proportion of statistically significant connections in the system breaks down when a given financial institution is excluded. We analyse the performance of our measure of systemic risk by considering a sample of the largest banks worldwide over the 2003-2018 period. We obtain three important results. First, we show that our measure is able to identify a large number of banks classified as global systemically important banks (G-SIBs) by the Financial Stability Board (FSB). Second, we find that our measure is a robust and statistically significant early-warning indicator of downside returns during the last financial crisis. Finally, we investigate the potential determinants of our measure of systemic risk and find similar results to the existing literature. In particular, our empirical results suggest that the size and the business model of banks are significant drivers of systemic risk.

with Elena Dumitrescu (CRED), Christophe Hurlin (LÉO) and Sessi Tokpavi (LÉO), 2022

European Journal of Operational Research, Volume 297, Issue 3, 16 March 2022, Pages 1178-1192

Abstract: In the context of credit scoring, ensemble methods based on decision trees, such as the random forest method, provide better classification performance than standard logistic regression models. However, logistic regression remains the benchmark in the credit risk industry mainly because the lack of interpretability of ensemble methods is incompatible with the requirements of financial regulators. In this paper, we propose a high-performance and interpretable credit scoring method called penalised logistic tree regression (PLTR), which uses information from decision trees to improve the performance of logistic regression. Formally, rules extracted from various short-depth decision trees built with original predictive variables are used as predictors in a penalised logistic regression model. PLTR allows us to capture non-linear effects that can arise in credit scoring data while preserving the intrinsic interpretability of the logistic regression model. Monte Carlo simulations and empirical applications using four real credit default datasets show that PLTR predicts credit risk significantly more accurately than logistic regression and compares competitively to the random forest method.

with Emmanuel Flachaire (AMSE), Sébastien Laurent (AMSE) and Gilles Hacheme, 2024

Oxford Bulletin of Economics and Statistics, Volume 86, Issue 3, June 2024, Pages 519-590 

Abstract: Despite their high predictive performance, random forest and gradient boostingare often considered as black boxes which has raised concerns from practitioners  and regulators. As an alternative, we suggest using partial linear
models that are inherently interpretable. Specifically, we propose to combine parametric and non-parametric functions to accurately capture linearities and non-linearities prevailing between dependent and explanatory variables, and a variable selection procedure to control for overfitting issues. Estimation relies on a two-step procedure building upon the double residual method. We illustrate the predictive performance and interpretability of our approach on a regression problem.

joint with Christophe Hurlin (LEO), Christophe Pérignon (HEC) and Sébastien Saurin (HEC)

forthcoming in Management Science - Package in Python - Towards Data Science blog

As they play an increasingly important role in determining access to credit, credit scoring models are under growing scrutiny from banking supervisors and internal model validators. These authorities need to monitor the model performance and identify its key drivers. To facilitate this, we introduce the XPER methodology to decompose a performance metric (e.g., AUC, R2) into specific contributions associated with the various features of a forecasting model. XPER is theoretically grounded on Shapley values and is both model-agnostic and performance metric-agnostic. Furthermore, it can be implemented either at the model level or at the individual level. Using a novel dataset of car loans, we decompose the AUC of a machine-learning model trained to forecast the default probability of loan applicants. We show that a small number of features can explain a surprisingly large part of the model performance. Notably, the features that contribute the most to the predictive performance of the model may not be the ones that contribute the most to individual forecasts (SHAP). Finally, we show how XPER can be used to deal with heterogeneity issues and improve performance.

Working Papers

joint with Christophe Hurlin (LEO) and Yang Lu (Concordia)

joint with Sébastien Laurent (AMSE), Ulrich Aiounou (AMSE) and Emmanuel Flachaire (AMSE)

joint with Jérémy Leymarie (Clermont School of Business & CleRMa)

  • Latent Factor Model for Tail Risk: an Integrated Approach to Systemic Risk Evaluation

bottom of page