No Arabic abstract
In Artificial Intelligence, interpreting the results of a Machine Learning technique often termed as a black box is a difficult task. A counterfactual explanation of a particular black box attempts to find the smallest change to the input values that modifies the prediction to a particular output, other than the original one. In this work we formulate the problem of finding a counterfactual explanation as an optimization problem. We propose a new sparsity algorithm which solves the optimization problem, while also maximizing the sparsity of the counterfactual explanation. We apply the sparsity algorithm to provide a simple suggestion to publicly traded companies in order to improve their credit ratings. We validate the sparsity algorithm with a synthetically generated dataset and we further apply it to quarterly financial statements from companies in financial, healthcare and IT sectors of the US market. We provide evidence that the counterfactual explanation can capture the nature of the real statement features that changed between the current quarter and the following quarter when ratings improved. The empirical results show that the higher the rating of a company the greater the effort required to further improve credit rating.
Recent literature implements machine learning techniques to assess corporate credit rating based on financial statement reports. In this work, we analyze the performance of four neural network architectures (MLP, CNN, CNN2D, LSTM) in predicting corporate credit rating as issued by Standard and Poors. We analyze companies from the energy, financial and healthcare sectors in US. The goal of the analysis is to improve application of machine learning algorithms to credit assessment. To this end, we focus on three questions. First, we investigate if the algorithms perform better when using a selected subset of features, or if it is better to allow the algorithms to select features themselves. Second, is the temporal aspect inherent in financial data important for the results obtained by a machine learning algorithm? Third, is there a particular neural network architecture that consistently outperforms others with respect to input features, sectors and holdout set? We create several case studies to answer these questions and analyze the results using ANOVA and multiple comparison testing procedure.
We introduce the general arbitrage-free valuation framework for counterparty risk adjustments in presence of bilateral default risk, including default of the investor. We illustrate the symmetry in the valuation and show that the adjustment involves a long position in a put option plus a short position in a call option, both with zero strike and written on the residual net value of the contract at the relevant default times. We allow for correlation between the default times of the investor, counterparty and underlying portfolio risk factors. We use arbitrage-free stochastic dynamical models. We then specialize our analysis to Credit Default Swaps (CDS) as underlying portfolio, generalizing the work of Brigo and Chourdakis (2008) [5] who deal with unilateral and asymmetric counterparty risk. We introduce stochastic intensity models and a trivariate copula function on the default times exponential variables to model default dependence. Similarly to [5], we find that both default correlation and credit spread volatilities have a relevant and structured impact on the adjustment. Differently from [5], the two parties will now agree on the credit valuation adjustment. We study a case involving British Airways, Lehman Brothers and Royal Dutch Shell, illustrating the bilateral adjustments in concrete crisis situations.
The use of CVA to cover credit risk is widely spread, but has its limitations. Namely, dealers face the problem of the illiquidity of instruments used for hedging it, hence forced to warehouse credit risk. As a result, dealers tend to offer a limited OTC derivatives market to highly risky counterparties. Consequently, those highly risky entities rarely have access to hedging services precisely when they need them most. In this paper we propose a method to overcome this limitation. We propose to extend the CVA risk-neutral framework to compute an initial margin (IM) specific to each counterparty, which depends on the credit quality of the entity at stake, transforming the effective credit rating of a given netting set to AAA, regardless of the credit rating of the counterparty. By transforming CVA requirement into IM ones, as proposed in this paper, an institution could rely on the existing mechanisms for posting and calling of IM, hence ensuring the operational viability of this new form of managing warehoused risk. The main difference with the currently standard framework is the creation of a Specific Initial Margin, that depends in the credit rating of the counterparty and the characteristics of the netting set in question. In this paper we propose a methodology for such transformation in a sound manner, and hence this method overcomes some of the limitations of the CVA framework.
The 2008 financial crisis has been attributed to excessive complexity of the financial system due to financial innovation. We employ computational complexity theory to make this notion precise. Specifically, we consider the problem of clearing a financial network after a shock. Prior work has shown that when banks can only enter into simple debt contracts with each other, then this problem can be solved in polynomial time. In contrast, if they can also enter into credit default swaps (CDSs), i.e., financial derivative contracts that depend on the default of another bank, a solution may not even exist. In this work, we show that deciding if a solution exists is NP-complete if CDSs are allowed. This remains true if we relax the problem to $varepsilon$-approximate solutions, for a constant $varepsilon$. We further show that, under sufficient conditions where a solution is guaranteed to exist, the approximate search problem is PPAD-complete for constant $varepsilon$. We then try to isolate the origin of the complexity. It turns out that already determining which banks default is hard. Further, we show that the complexity is not driven by the dependence of counterparties on each other, but rather hinges on the presence of so-called naked CDSs. If naked CDSs are not present, we receive a simple polynomial-time algorithm. Our results are of practical importance for regulators stress tests and regulatory policy.
The paper examines the potential of deep learning to support decisions in financial risk management. We develop a deep learning model for predicting whether individual spread traders secure profits from future trades. This task embodies typical modeling challenges faced in risk and behavior forecasting. Conventional machine learning requires data that is representative of the feature-target relationship and relies on the often costly development, maintenance, and revision of handcrafted features. Consequently, modeling highly variable, heterogeneous patterns such as trader behavior is challenging. Deep learning promises a remedy. Learning hierarchical distributed representations of the data in an automatic manner (e.g. risk taking behavior), it uncovers generative features that determine the target (e.g., traders profitability), avoids manual feature engineering, and is more robust toward change (e.g. dynamic market conditions). The results of employing a deep network for operational risk forecasting confirm the feature learning capability of deep learning, provide guidance on designing a suitable network architecture and demonstrate the superiority of deep learning over machine learning and rule-based benchmarks.