How to Implement AMA for the Calculation of Capital Requirement for
Transcription
How to Implement AMA for the Calculation of Capital Requirement for
How to Implement AMA for the Calculation of Capital Requirement for Operational Risk GABRIELLA GREPE IVAN THOMSON Master of Science Thesis Stockholm, Sweden 2008 2 Master of Science Thesis How to Implement AMA for the Calculation of Capital Requirement for Operational Risk Gabriella Grepe & Ivan Thomson INDEK 2008:102 KTH Industrial Engineering and Management Industrial Management SE-100 44 STOCKHOLM 3 Master of Science Thesis INDEK 2008:102 How to Implement AMA for the Calculation of Capital Requirement for Operational Risk Gabriella Grepe Ivan Thomson Approved Examiner Supervisor 2008-12-11 Thomas Westin Birger Ljung Abstract This report presents a case study of the development of an AMA 1 model for the calculation of capital requirement for operational risk. The case study was based on an assignment commissioned by one of Sweden’s largest banks. We were given full scope regarding the choice of the model and the assumptions, and began the process with a literature study of previously research in the area of operational risk. We ultimately chose a loss distribution approach for the base; the frequencies of the losses were assumed to be Poisson distributed, whereas the severities were assumed to follow a lognormal distribution. Instead of the correlation approach suggested in the loss distribution approach, we used a bottom-up approach where some losses were assumed to be jointly Poisson distributed. We also presented a way of scaling data in order to better fit external data to the internal ones. Please not that the model is only a suggestion since the lack of data made it impossible to test the ultimate results. What came to our attention during the development of the model is the lack of univocal guidelines for how the model has to be designed. Our opinion is that operational risk does not differ that much between banks. Hence, it might be a good idea to have uniform rules for the model development and only leave the assumptions about scaling and correlation to the bank. AMA = Advanced Measurement Approaches. The most advanced model for calculating the capital requirement for operational risk. A more thorough explanation will be presented in 4.1.3. 1 4 Acknowledgements We would especially like to thank three people for their significant contribution throughout the process of writing this thesis. Birger Ljung, our supervisor at INDEK, for his traditional finance view as well as his insightful comments about the structure of the report. Filip Lindskog, our supervisor at Mathematical Statistics, for his suggestions to the mathematical parts of the thesis. And last but not least Carl Larsson, our supervisor at the bank, for his servings as a sounding board and his valuable help with the implementation of the model. Stockholm, December 1, 2008 Gabriella Grepe Ivan Thomson 5 Table of Contents 1 Introduction.................................................................................................................................................... 8 1.1 Background ............................................................................................................................................. 8 1.2 Problem Statement................................................................................................................................. 9 1.3 Commissioner & Assignment............................................................................................................... 9 1.4 Aim of Thesis.......................................................................................................................................... 9 1.5 Delimitations......................................................................................................................................... 10 1.6 Disposition ............................................................................................................................................ 10 1.7 Terms & Definitions............................................................................................................................ 12 2 Methodology................................................................................................................................................. 13 2.1 Research Approach .............................................................................................................................. 13 2.2 Work Procedure ................................................................................................................................... 13 2.3 Choices of Path..................................................................................................................................... 14 3 Theoretical Framework ...............................................................................................................................15 3.1 Risk Types ............................................................................................................................................. 15 3.1.1 Credit Risk ..................................................................................................................................... 15 3.1.2 Market Risk.................................................................................................................................... 15 3.1.3 Operational Risk ........................................................................................................................... 15 3.2 Risk Measures and Methods ............................................................................................................... 16 3.2.1 Value-at-Risk ................................................................................................................................. 16 3.2.2. Expected Shortfall ....................................................................................................................... 17 3.2.3 Extreme Value Theory................................................................................................................. 18 3.3 Distribution Functions ........................................................................................................................ 21 3.3.1 Poisson Distribution .................................................................................................................... 22 3.3.2 Lognormal Distribution............................................................................................................... 23 3.4 Copula .................................................................................................................................................... 24 3.5 Simulation Methods for Computing VaR/ES ................................................................................. 26 3.5.1 Historical Simulation.................................................................................................................... 26 3.5.2 Monte Carlo Simulation............................................................................................................... 26 4 Basel Rules & Previous Research .............................................................................................................. 28 4.1 Three Types of Models........................................................................................................................ 28 4.1.1 Basic Indicator Approach (BIA)................................................................................................. 28 4.1.2 Standard Approach (SA).............................................................................................................. 29 4.1.3 Advanced Measurement Approaches (AMA) .......................................................................... 30 4.2 Three Types of AMA Models ............................................................................................................ 30 4.2.1 Internal Measurement Approach (IMA) ................................................................................... 31 4.2.2 Loss Distribution Approach (LDA) .......................................................................................... 31 4.2.3 Scorecard Approach (SCA) ......................................................................................................... 32 4.3 Previous Research – AMA.................................................................................................................. 32 5 Model Choice & Assumptions................................................................................................................... 34 5.1 What is a Good Model?....................................................................................................................... 34 5.2 Overall Approach Choice ................................................................................................................... 35 5.3 Distribution Assumptions................................................................................................................... 35 5.4 Correlation Assumptions .................................................................................................................... 37 5.5 Assumptions regarding Scaling, Inflation & Truncation Bias ....................................................... 39 5.6 Risk & Simulation Method.................................................................................................................. 39 6 Description of Our Model.......................................................................................................................... 41 6 6.1 Model Overview ................................................................................................................................... 41 6.2 Data Import .......................................................................................................................................... 41 6.3 Data Processing .................................................................................................................................... 43 6.4 Simulation.............................................................................................................................................. 52 7 Model Testing............................................................................................................................................... 53 8 Concluding Chapter..................................................................................................................................... 54 8.1 Insights................................................................................................................................................... 54 8.2 Future Required Updates .................................................................................................................... 54 8.3 Validity & Critique ...............................................................................................................................56 8.4 Further Research .................................................................................................................................. 58 9. References .................................................................................................................................................... 60 9.1 Written Sources .................................................................................................................................... 60 9.2 Interviews .............................................................................................................................................. 63 7 1 Introduction 1.1 Background This master thesis has been written during a time of financial turbulence. The acquisitions and/or bankruptcies of the five largest American investment banks have been an indirect result of the burst of the real estate bubble in 2006. The disturbance has spread with falling stock markets and concerns about the survival of other banks as a result. The media has speculated about the risk exposure of European banks and the public has expressed their worries about their saved capital. During such time, the attention has focused on the risk management of the banks. Can the banks stand up against the turbulence, and do they have capital to cover for the unexpected losses? The cooperation of international central banks entered the history books with the formation of the Bank for International Settlements (BIS) in 1930 (Bank of International Settlements. BIS History Overview, 2008). The institution, which initially was formed within the context of World War I reparation payments, came over time to focus on fostering monetary policy cooperation. The regulatory supervision of internationally active banks commenced as an important issue following the oil- and international debt crisis of the 70’s. As a result, The Basel Capital Accord was published in 1988 with the purpose of imposing a minimum capital standard in the banking industry (Basel Committee of Banking Supervision, 1988). Since then, banks need to reserve a specific amount of equity that reflects the risk they are exposed to. This Capital Accord was 13 years later replaced by Basel II, which contains more risk sensitive rules and requirements on the management and board of directors (Finansförbundet. Finansvärlden, 2005). One of the news with Basel II was the inclusion of operational risk as a separate category in the capital requirement framework (KPMG International, 2003). The risk is thus divided into three different categories; credit risk, market risk and operational risk. Operational risk is a fairly new modelling phenomenon and there are several complications in the process. The difficulty of predicting the events combined with the limit of historical data do not form a solid statistic platform on which to build the model. The internal fraud that caused the collapse of Barings Bank in the 90’s (BBC News, 1999) is one example of this. How is it possible to cover for a loss exceeding £ 800 million when it has not happened before? There is not a straightforward approach for how the model has to be designed. The Basel Committee proposes 8 three different types of methods with increasing degree of sophistication; banks are encouraged to move on to the more sophisticated ones in order to more precisely measure their operational risk. 1.2 Problem Statement How can a model for the calculation of capital requirement for operational risk be designed? The problem of aiming for an advanced method causes issues to the modelling process. There are several approaches suggested in the literature and some appear more common than others. Neither does however enclose all modelling issues and possible solutions to these. 1.3 Commissioner & Assignment The assignment has been commissioned by one of Sweden’s largest banks. The bank is active mainly in the Nordic region with considerable exposure in the Baltic area. The development process has been accomplished with guidance from the department Group Operational Risk Control, liable for the operational risks and the security in the bank. The Basel Committee on Banking Supervision (2006) has presented three different types of operational risk models with an increasing degree of sophistication. Since the bank used one of the less sophisticated methods for the capital requirement calculation for operational risk (the Standard Approach), a more advanced method was desirable. The assignment has thus concerned the development of the most advanced method called Advanced Measurement Approaches (AMA). 1.4 Aim of Thesis The aim of this study has been to make a suggestion of how AMA can be implemented. The development has been illustrated with the help of a case. The aim of the case study has been to map out the possibilities, the mathematical difficulties and highlight how the model may be designed in order to give a good measure for the capital requirement for operational risk. The aim is also to give an instruction of how to implement the model at a level that can be conducted at any bank or financial institution. The usage of this report is thus not limited to “our” bank; it can in theory be applied at any financial institution with internal loss data captured over a substantial time period. The way in which this report may be used is either directly as a 9 programming template or as discussion material for how to manage operational risk in the organization. By looking at the model development as a case study, we believe that our report may contribute to the research regarding how operational risk may be measured. 1.5 Delimitations The AMA model includes more than just the programming product. In order to get the model approved by Finansinspektionen (FI), the Swedish Financial Supervisory Authority, the model has to fulfil not only quantitative but also qualitative requirements. These include (among others) information about how the model is to be implemented, how the capital is distributed across the operational risks and how the time plan is designed (FI, 2007). Since our aim has been to map out the mathematical difficulties in the development of the model, the qualitative requirements have been considered to be out of scope. The focus has instead been limited to the quantitative requirements that have to be fulfilled in order to get the model approved. What needs to be acknowledged is that due to issues of confidentiality, the Matlab 2 code will not be included in this report. 1.6 Disposition This thesis is divided into nine chapters as follows: Introduction. This part of the thesis explains the background and the nature of the problem. The aim of the thesis, the assignor and the assignment are described as well as necessary delimitations to the problem statement. Methodology. The research approach and the work procedure are presented in this chapter. These include descriptions of how the assignment is related to the problem statement as well as a description of how the work has developed in the process of time. Theoretical Framework. The basics of Risk Management, as presented in traditional finance literature, are described in this chapter. The aim is to provide a basic understanding of what kind of 2 Matlab is a high-level programming language. See http://www.mathworks.com/products/pfo/ for more information. 10 risks banks are exposed to and how these can be measured. An overview of the distributions and simulations used in the model are presented as well. Basel Rules & Previous Research. This chapter focuses on the base for the capital requirement calculations – the rules of the Basel Committee. It presents the regulatory requirement for the AMAmodel as well as a description of the alternatives. The different approaches for how to construct the AMA and previous research about the model are presented as well. Model Choice & Assumptions. Considering the many different approaches that are presented, this chapter focuses on the method chosen in our case. Critical assumptions about loss distributions, correlations and scaling factors are described and argued for in this chapter. The question of what characterizes a good model is also discussed. Model Description. This part describes the features of our developed model. It explains how the data is imported and processed. It also provides a detailed description of how the correlation and scaling is handled and how the simulation is conducted. Model Testing. This chapter focuses on the tests the model has undergone. Concluding Chapter. The thesis is finished by a chapter that concludes the insights and the updates that has to be made. It also discusses the validity of the model and presents recommendations for further research in the area of operational risk. References. This part contains a list of references as well as a list of the interviewees. The figures that have been used in the report have either been collected from literature (and the sources are thus in the reference list) or made by us. In the case of the latter, the source is solely left out in the figure text. 11 1.7 Terms & Definitions This section provides a list of the abbreviations and some of the technical terms used in this report. AMA (Advanced Measurement Approaches). AMA is a summation name for the approaches that fall into the category of being the most sophisticated kind of model for the calculation of the capital requirement of operational risk. BIA (Basic Indicator Approach). The simplest approach for the calculation of the capital requirement of operational risk. Copula. A dependence measure that defines a correlation structure between two correlated variables. EVT (Extreme Value Theory). A theory of how to estimate the tail (in this case; the high severity losses) of a distribution. IMA (Internal Measurement Approach). An approach for how to build the AMA; an alternative approach to the LDA and SCA approaches. LDA (Loss Distribution Approach). An approach for how to build the AMA; an alternative to the IMA and SCA approaches. Monte Carlo simulation. A simulation technique where one assumes a distribution for how the losses are distributed and then simulates new losses according to it. SA (Standard Approach). An approach calculating the capital requirement of operational risk. It is less sophisticated than AMA, but more sophisticated than BIA. SCA (Scorecard approach). An approach for building the AMA; an alternative approach to the LDA and IMA approaches. Truncation bias. The bias that appears due to the fact that the internal dataset is truncated (i.e. that only losses that exceed a specific threshold are reported) VaR (Value-at-Risk). A risk measure answering the question of how much it is possible to loose within a specified amount of days with a certain certainty. 12 2 Methodology This chapter aims to provide an overview of how the work was conducted; what kind of research approaches that was used and how the work procedure was designed. The important choices of path, including the required mathematical and regulation knowledge, are also explained here. 2.1 Research Approach The research approach of this thesis has been conducted with the aid of a case study. Since the assignment came prior to the idea of research, we had to go backwards and find a way in which this assignment could be scientifically contributing. The case itself is interesting since it touches upon one of the new areas of modelling within risk analysis. There are numerous reports regarding different parts of the AMA-model, but we considered there to be a research gap in the overall view; the model issues, how they may be handled and how the model itself may be criticized. It must however be emphasized that this is only a suggestion of how such a model may be designed. We have tested the programming, but because of the lack of data, we have not been able to test its numerical results. Thus, we have not been able to test the sensitivity of the assumptions made, and this report should be read with this concern in mind. 2.2 Work Procedure The process was initiated with a literature review covering working papers from the Basel Committee as well as academic reports of the area. Most of the material was collected from the Internet, with Social Science Research Network (SSRN) as the main source. Material concerning the theoretical framework was read in order to map out applicable risk measurement techniques and simulation methods that could be used in the model. The next step in the process was to make a choice about the necessary assumptions. These regarded the distributions of the losses (how often they occur, and how much that was lost when they occurred), the correlation (is there any reason to believe that the loss in a may influence a loss in b?) and the potential influence of business line income on operational risk. In order to thoroughly revise these assumptions, a number of semi structured interviews were made both with mathematicians at KTH and with employees within the bank. The interviews with mathematicians at KTH regarded mathematical and modelling issues that 13 were approached in the development process. The interviews within the bank regarded different areas dependent on who was interviewed. The discussions with the risk analysts regarded the mathematical aspects and the discussions with employees earlier employed by FI, regarded FI’s work process and point of view. The interviews with experienced employees within the bank concerned the potential correlation of losses that was important to the development of the model. The programming of the model was conducted continuously throughout the development process. 2.3 Choices of Path In writing this thesis, several important standpoints had to be taken. One regards our theoretical base. The choice of several theoretical sources is made because of an ambition to make the report as pedagogical as possible. We have chosen the, in our view, best description of the phenomenon from different books in order to ease the understanding of the measures and methods used. What needs to be emphasized is that the different authors do not contradict each other regarding the theoretical framework. Another standpoint regards to whom this report is directed to. Our aim is to write this report so that someone with at least some knowledge of probability theory and traditional risk analysis may find it possible to understand. The reader does not have to have any awareness of the Basel Committee and its work; the necessary parts will be thoroughly described in chapter 4. The last, but probably the most important, choice of path is the one regarding our underlying assumptions about risk analysis. We have chosen to accede to the traditional finance theory and its interpretation of how risk ought to be measured. We have thus chosen not to criticize the traditional view of analysing risk, where the history is seen as representative for the future. The discussion of alternatives is considered to be out of scope. 14 3 Theoretical Framework In order to simplify for the reader, this part presents an overview of the risk types, risk measures and mathematical phenomena encountered and involved in the modelling process. The discussion of the methods is saved for chapter 5; the aim of this part is only to provide a background for the methods mentioned further on. It is thus recommended to skip this chapter when reading the report for the first time, and instead use it as a book of reference when reading chapters 5-8. 3.1 Risk Types Hult and Lindskog (2007) present a general definition of risk for an organization as “any event or action that may adversely affect an organization to achieve its obligations and execute its strategies”. When it comes to capital requirement, the risk is divided into three categories; market, credit and operational risk. 3.1.1 Credit Risk The Bank of International Settlement (BIS) defines, according to Gallati (2003, p. 130), credit risk as “the risk that a counterparty will not settle an obligation for full value, either when due or at any time thereafter. In exchange-for-value systems, the risk is generally defined to include replacement risk and principal risk”. 3.1.2 Market Risk The Bank for International Settlement (BIS) defines market risk as “the risk of losses in on- and offbalance-sheet positions arising from movements in market prices”, according to (Gallati, 2003, p. 34). He describes the main factors that contribute to market risk as equity, interest rate, foreign exchange and commodity risk. The aggregation of all risk factors is the total market risk. 3.1.3 Operational Risk There is no established way of how to define operational risk. On the one side of the spectrum is the broad definition of operational risk as “anything in the bank that is not market risk or credit risk” (Hull, 2007). On the other side of the spectrum is the narrow definition of operational risk as 15 any risk arising from operations. The definition that is widely used in banking terms is that of the Basel Committee (Basel Committee on Banking Supervision, 2006, p. 144): “The risk of loss resulting from inadequate or failed internal processes, people and systems or from external events. This definition includes legal risk, but excludes strategic and reputational risk.” Examples of operational risks include among others internal and external fraud, IT-breakdowns and natural disasters. The exposure to operational risk is less predictable than the other risks that the bank faces. While some of the operational risks are measurable, some may thoroughly escape detection (Jobst, 2007). The operational risks do mainly occur as rare high severity events or frequent low severity events. The rare high severity events are for example uncontrolled gambling and natural disasters, whereas frequent low severity events may be exemplified by stolen credit card payments. When it comes to the expected loss in the bank, it is easier to quantify the expected loss for frequent low severity events than it is for rare high severity events (Jobst, 2007). The incorporation of the latter is thus the most complicated, and probably also the most important, part of the risk model. 3.2 Risk Measures and Methods The relevant measures and methods to use regarding the calculation of capital requirement are introduced in this section. These are presented in order to give the reader an understanding of how they are used in the model developed. 3.2.1 Value-at-Risk The Value-at-Risk (VaR) measure is by Hult and Lindskog (2007, p. 14) defined as: “Given a loss L and a confidence level α ∈ (0,1), VaRα(L) is given by the smallest number l such that the probability that the loss L exceeds l is no larger than 1- α, i.e.:” VaRα ( L ) = inf {l ∈ R : P ( L > l ) ≤ 1 − α } = inf {l ∈ R : 1 − FL (l ) ≤ 1 − α } = inf {l ∈ R : FL (l ) ≥ α }. 16 The VaR measure thus presents the value V in the following statement: “We are X percent certain that we will not loose more than V dollars in the next N days. “ (Hull, 2007). X is here the confidence level, whereas N is the time horizon of the statement. One example will illustrate this statement. Imagine a portfolio of assets, whose total risk we want to measure with VaR. Then VaR is the loss over the next N days corresponding to the (100-X)th percentile of the distribution of changes in the value of the portfolio (Hull, 2007). With 98 % confidence level, the VaR-value is the second percentile of the distribution. By simulating from this distribution 100 times, there should on average be two values exceeding this value. Even though Value-at-Risk gives an easily understandable answer, there is a drawback. The VaRmeasure only provides a value for the maximum loss with a certain probability; it does not say anything about the loss if that value is exceeded. Hull (2007) demonstrates that this may have a practical drawback in terms of a trader’s risk limits. If the trader is being told that the one-day 99 % VaR has to be kept at a maximum of 10 million, the trader can construct a portfolio where there is a 99 % chance of losing less than 10 million, but a 1 % chance of losing much more. 3.2.2. Expected Shortfall The previously mentioned weakness may be eased by the use of Expected Shortfall (ES). This measure is presented by Hult and Lindskog (2007) as: “For a loss L with continuous loss distribution function FL the expected shortfall at confidence level α ∈ (0,1), is given by” ESα ( L ) = E ( L L ≥ VaRα ( L )) Instead of asking how bad things can get, ES measures what the expected loss is if things do get bad (Hull, 2007). ES is thus the average amount that is lost over the N day-period, if the loss has exceeded the (100-X)th percentile of the distribution (i.e. the VaR-value). The following figure illustrates the difference between Value-at-Risk and Expected Shortfall. 17 Figure 1. Illustration of Value-at-Risk and Expected Shortfall. See Hull (2007, p. 198) for a similar version. 3.2.3 Extreme Value Theory Extreme Value Theory (EVT) is the science of estimating the tails of a distribution (Hull, 2007). The tail does here refer to the ultimate end of the distribution that, in terms of a loss distribution, consists of very high losses. If the tail is heavy, it means that there is a considerable amount of high losses at the end of the tail. Hult and Lindskog (2007, p. 28) point out that even though there is no definition of what is meant by a heavy tail, it is “common to consider the right tail F ( x ) = 1 − F ( x ) , x large, of the distribution function F heavy if: ” lim x →∞ F (x ) = ∞ , for every λ > 0 e − λx Hult and Lindskog (2007) further present the quantile-quantile plot (qq-plot) as a method to achieve an indication of the heaviness of the tail. It is a useful tool for assessing the distribution of identically and independently distributed variables. By using a reference distribution F, it is possible to plot the sample against the reference distribution and thereby get an understanding of how our sample behaves in reference to F. If Xk,n is our sample, the qq-plot consists of the points (Hult & Lindskog, 2007, p. 29): ⎧⎛ ⎫ ←⎛ n − k +1⎞⎞ ⎟ ⎟⎟ : K = 1,..., n⎬ ⎨⎜⎜ X k ,n , F ⎜ ⎝ n +1 ⎠⎠ ⎩⎝ ⎭ 18 If the plot achieved is linear, it means that the sample has a distribution similar to the reference distribution (Hult & Lindskog, 2007). If the sample has heavier tails than the reference distribution, the plot will curve up at the left and/or down at the right. If the sample has lighter tails than the reference distribution, the plot will curve down at the left and/or up at the right (Hult & Lindskog, 2007). The figure below is a qq-plot that compares the sample of historical losses to a simulation of lognormal ones. Since the plot achieved is linear, it is fair to assume the losses to be approximately lognormally distributed. Figure 2. QQ-plot. Hull (2007) mentions that EVT can be used to improve VaR estimates, since it smoothes and extrapolates the tails of the empirical distribution. This means that EVT may be useful where extreme historical data is relatively scarce. EVT may also be used to estimate VaR when the VaR confidence level is very high, and thus provide a closer estimate (Hull, 2007). One major result within EVT, concerning the right tail of the probability distribution, was proved by Gnedenko in 1943 (Hull, 2007). He found that the right tail of the probability distribution behaves according to the generalized Pareto distribution as the threshold u is increased. Hult and Lindskog (2007, p. 38) present the generalized Pareto distribution (GPD) function as: ⎛ γ x⎞ ⎟ Gγ , β ( x ) = 1 − ⎜⎜1 + β ⎟⎠ ⎝ 19 − 1 λ for x ≥ 0 The estimator of the tail (i.e. of the part of the distribution that exceeds u), is presented by Hult and Lindskog (2007, p. 39) as: N Fˆ (u + x ) = u n ⎛ x⎞ ⎜1 + γˆ ⎟ ⎜ βˆ ⎟⎠ ⎝ − 1 γˆ (1) The estimator of the quantile is presented by Hult and Lindskog (2007, p. 39) as: 3 −γˆ ⎞ ⎞ βˆ ⎛ ⎛ n qˆ p (F ) = u + ⎜ ⎜⎜ (1 − p )⎟⎟ − 1⎟ ⎟ γˆ ⎜ ⎝ Nu ⎠ ⎝ (2) ⎠ In order to use the Peaks Over Threshold (POT) method for the estimation of tail probabilities and quantiles, one first have to choose the high threshold u and count the number Nu that exceed that value. See figure below. Figure 3. Illustration of the POT-method. Given this sample of exceedances (the parts of the losses that exceed u), the next step is to estimate the parameters γ and β. These are found by using a maximum likelihood estimation of the parameters, assuming the losses exceeding u are generalized Pareto distributed. 4 The values of u, Nu,, γ and β may then be inserted into (1) and (2). Simulation from (1) now gives a smoother tail than the original sample. 3 4 For proof, see Hult & Lindskog (2007. p. 39). For a more detailed description, see Hult & Lindskog (2007, p. 41). 20 By using the POT-method, the VaR and ES 5 may be calculated directly as well. Hult and Lindskog (2007, p. 44) present the following formulas, based on the above estimated parameters: −γˆ ⎞ ⎞ βˆ ⎛⎜ ⎛ n ˆ ⎜ ⎟ ( ) ( ) VaRp, POT X = u + ⎜ 1 − p ⎟ −1⎟ ⎜ ⎟ γˆ ⎝ Nu ⎠ ⎝ ⎠ βˆ + γˆ (qˆ p − u ) Eˆ S p , POT ( X ) = qˆ p + 1 − γˆ 3.3 Distribution Functions This report concerns two types of distributions; the discrete distribution (for example binomial, geometric and Poisson distributions) and the continuous distribution (for example exponential, gamma and normal distributions). Why these two are of current interest is explained in more detail in chapter 5.3. The definition of a distribution function FX, of the random variable X, is defined as (Gut, 1995, p. 7): FX ( x ) = P ( X = x ),−∞ < x < ∞ For discrete distributions we introduce the probability function pX, defined as: p X ( x ) =P ( X = x ), for all x For the discrete probability function, the following is true (Gut, 1995, p. 7): FX ( x ) = ∑ p X ( y ) , − ∞ < x < ∞ y≤ x For continuous distributions we introduce the density function fX, for which (Gut, 1995, p. 7): 5 For proof, see Hult & Lindskog (2007, p. 41). 21 FX ( x ) = x ∫ f ( y ) dy , − ∞ < x < ∞ X −∞ What is described above is how one random variable, X, may be distributed. There may also be the case that several variables are distributed together. If we introduce another random variable, Y, the description of how the pair (X,Y) is distributed is given by the joint distribution function, FX,Y (Gut, 1995, p. 9): p X ,Y (x, y ) = P ( X = x, Y = y ) , − ∞ < x, y < ∞ There exists a joint probability function in the discrete case, given by (Gut, 1995, p. 9): FX ,Y ( x, y ) = P ( X ≤ x , Y ≤ y ) , − ∞ < x, y < ∞ There exists a joint density function in the continuous case, given by (Gut, 1995, p. 9) f X ,Y ( x, y ) = ∂ 2 FX ,Y (x, y ) ∂x∂y , − ∞ < x, y < ∞ 3.3.1 Poisson Distribution The Poisson distribution is commonly referred to as Po(m), where m > 0. It belongs to the discrete distribution functions and its probability function is defined as (Gut, 1995, p. 259): p(k ) = e −m mk , k = 0,1,2,... k! Blom (2005) describes the Poisson distribution as a distribution that appears in the study of randomly occurrences in time or room. Example of this is for example the occurrence of a traffic accident or that someone is calling SOS. These may occur at any given time point and are independent of each other (Blom, 2005). k in the formula above is thus the number of events that appear in a given interval of time length t; causing k to be Po(m)-distributed, where m is the expected value of occurrences in that interval. 22 By simulating 1000 times from a Po(5)-distribution, the following plot is achieved. It should be read as the number of occurrences per outcome. I.e. the outcome 0 occurs approximately 48 times, the outcome 1 approximately 92 times etc. Figure 4. Illustration of the Poisson distribution with µ=5. If the variables themselves are Poisson-distributed, they occur in time according to a Poissonprocess. Gut (1995) describes the Poisson process as a discrete stochastic process in continuous time that is commonly used to describe phenomena as the above (the SOS-call and the traffic accident). The Poisson process is by Gut (1995, p. 196) defined as “a stochastic process, {X(t), t ≥ 0}, with independent, stationary, Poisson-distributed increments. Also, X(0) = 0.” X(t) in the definition is the number of occurrences in the time interval between 0 and t. 3.3.2 Lognormal Distribution The lognormal distribution is commonly referred to as LN(μ,σ2), where -∞ < μ < ∞, σ > 0. It belongs to the continuous distribution functions and its density function is defined as (Gut, 1995, p. 261): f (x ) = 1 σ x 2π e 1 − (log x − μ ) 2 / σ 2 2 23 ,x>0 In the lognormal distribution, the log of the random variable is normally distributed. This is easier to grasp by comparing the above formula for the lognormal distribution with the below one for the more common normal distribution. f (x ) = 1 − ( x − μ )2 / σ 1 e 2 , −∞ < x < ∞ σ 2π By simulating 1000 times from a LN(12,2)-distribution and sorting the severity in increasing order, the following plot is achieved: Figure 5. Illustration of the lognormal distribution with µ=12 and σ=2. 3.4 Copula Copula is a well-known concept within Risk Management theory because of its usefulness for building multivariate models with non-standard dependence structures (Hult & Lindskog, 2007). The copula is a function that defines a correlation structure between two correlated variables, each with its own marginal distribution. If the marginal distribution for example is a normal distribution, it is convenient to assume that the variables together are bivariate normally distributed (Hull, 2007). The problem however is that there exists no natural way of defining a correlation structure between the two marginal distributions; the bivariate normally distribution is not the only way the two 24 variables could be jointly distributed. Hull (2007) explains that this is where the use of copulas comes in. This section presents two different ways of explaining what a copula is; the first one is a mathematical explanation by Hult and Lindskog (2007), the second one a descriptive explanation by Hull (2007). Hult and Lindskog (2007, p. 63) present two complementing definitions of the copula (where the second one includes Sklar’s theorem): “A d-dimensional copula is a distribution function on [0,1]d with standard uniform marginal distributions. […] This means that a copula is the distribution function P(U1 ≤ u1, …, Ud ≤ ud) of a random vector (U1, …, Ud) with the property that for all k it holds that P(Uk ≤ u) = u for u ∈ [0,1].” “Let F be a joint distribution function with marginal distribution functions F1, …, Fd. Then there exists a copula C such that for all x1, …, xd ℜ ∈ [- ∞ , ∞ ], F ( x1 ,..., xd ) = C (F1 ( x1 ),..., Fd ( xd )) (10.1) If F1, …, Fd are continuous, then C is unique. Conversely, if C is a copula and F1, …, Fd are distribution functions, then F defined by (10.1) is a joint distribution function with marginal distribution functions F1, …, Fd. […] Let F be a joint distribution function with continuous marginal distribution functions F1, …, Fd. Then the copula C in (10.1) is called the copula of F. If X is a random vector with distribution function F, then we also call C the copula of X. “ Hull (2007) explains the copula phenomenon by taking two variables with triangular probability distributions as an example, he calls them V1 and V2. In order to find out the correlation structure between these two variables, the author uses the Gaussian copula. In order to do so, the variables are each mapped into two new variables (call them U1 and U2) that are standard normally distributed. This is done by looking at V1’s value in each percentile of the distribution and calculating what that is according to the standard normal distribution. By assuming that the variables U1 and U2 are jointly bivariate normal, it implies a joint distribution and correlation structure between V1 and V2 (Hull, 2007). What is done here is what the copula does in practice. Instead of 25 determining the correlation directly out of the original variables (V1 and V2 in this case), these are mapped into other variables (U1 and U2) with well-behaved distributions for which a correlation structure is easily defined. The key property of the copula model here is that the marginal distributions are preserved at the same time as the correlation structure between them is defined (Hull, 2007). 3.5 Simulation Methods for Computing VaR/ES 3.5.1 Historical Simulation In this simulation approach, it is supposed that the historical risk factor changes are representative for the future. I.e. it is supposed that the same changes will occur over the next period (Hult & Lindskog, 2007). With the aid of the risk factor changes we achieve the loss vector l. The losses are then sorted in increasing size, in order to compute the empirical VaR and ES using the formulas below (Hult & Lindskog, 2007, p. 25). n represents the number of losses in the loss vector. ( ) VˆaRα (L ) = qˆ FLn = l[n (1−α )]+1,n [n (1−α )]+1 Eˆ Sα (L ) = ∑l i =1 i ,n [n(1 − α )] + 1 Hult and Lindskog (2007) however mention the disadvantages of the worst case never being worse than what has already happened in the past. “We need a very large sample of relevant historical data to get reliable estimates of Value-at-Risk and expected shortfall” (Hult & Lindskog, 2007, p. 25). 3.5.2 Monte Carlo Simulation The use of the Monte Carlo simulation method requires an assumption of the distribution of the previous mentioned risk factor changes. The real loss distribution is thus approximated by the empirical distribution FLN. Instead of using historical observations, the risk factor changes are in this method simulated according to the distribution chosen. The empirical VaR and ES are then computed the same way as shown above. 26 The authors consider the simulation method to be flexible, since it is possible to choose any model from which it is possible to simulate. It is however computationally intensive as a large amount of simulations have to be done in order to give a statistically correct value of the confidence level chosen (Hult & Lindskog, 2007). 27 4 Basel Rules & Previous Research This chapter aims to give an overview of the Basel Committee’s regulation regarding the calculation of the capital requirement. It provides a description of the three types of models that may be used in this process, where AMA is one. Three different approach of how to build the AMA is also presented in order to give an overview of the alternatives available. The chapter also aims to give an overview of the previous research regarding AMA and its implementation. 4.1 Three Types of Models The Basel Committee on Banking Supervision (2006) has identified three types of models for the calculation of the capital requirement for operational risk. These are, in order of increased sophistication; the Basic Indicator Approach (BIA), the Standard Approach (SA) and the Advanced Measurement Approaches (AMA). 4.1.1 Basic Indicator Approach (BIA) BIA is the simplest approach of the ones available. The Basel Committee on Banking Supervision (2006, p. 144) explains it as follows: “Internationally active banks and banks with significant operational risk exposures (for example, specialized processing banks) are expected to use an approach that is more sophisticated than the Basic Indicator Approach and that is appropriate for the risk profile of the institution”. The bank must thus hold capital for operational risk equal to the average over the previous three years of a fixed percentage (15 %) of positive annual gross income 6 . It is expressed as follows (Basel Committee on Banking Supervision, 2006, p. 144): K BIA = α N N ∑ GI i 1 Where KBIA= the capital requirement under the Basic Indicator Approach, α = 15 % (which is set by the committee), N = number of the three previous years for which the gross income is positive, GIi = annual gross income (if positive). Gross Income = Net Interest Income + Net Non-Interest Income (Basel Committee on Banking Supervision, 2001, p. 7). 6 28 4.1.2 Standard Approach (SA) “In the Standardised Approach, banks’ activities are divided into eight business lines: corporate finance, trading & sales, retail banking, commercial banking, payment & settlement, agency services, asset management, and retail brokerage” (Basel Committee on Banking Supervision, 2006, p. 147). Just like in the Basic Indicator Approach (BIA), you look at an average over the three last years. But instead of using only one fixed percentage for the entire bank, there are different percentages for all of the eight business lines. Similarly to the BIA, if the aggregate gross income for all business lines any given year is negative, it counts as zero. In detail, this is done as follows: ⎛ ⎡ 8 ⎤⎞ ⎜ max , 0 GI β ⎢∑ i , j j ⎥ ⎟⎟ ∑ ⎜ i =1 ⎝ ⎣ j =1 ⎦⎠ = 3 3 K SA Where, KSA = the capital requirement under the Standardized Approach, GIi,j= annual gross income in a given year for each of the eight business lines, βj= a fixed percentage, set by the committee, for each of the eight business lines. The β-values are detailed below: Business Lines Beta Factors Corporate Finance (β1) 18 % Trading & Sales (β2) 18 % Retail Banking (β3) 12 % Commercial Banking (β4) 15 % Payment & Settlement (β5) 18 % Agency Services (β6) 15 % Asset Management (β7) 12 % Retail Brokerage (β8) 12 % Figure 6. List of the β-values in the SA (Basel Committee on Banking Supervision, 2006, p. 147). 29 4.1.3 Advanced Measurement Approaches (AMA) The focus of this study is the development of an Advanced Measurement Approach (AMA). This approach implies that the bank should by itself model the operational risk; there is no direct formula available as above. The measurement system must, according to the Basel Committee on Banking Supervision (2006, p. 144), “reasonably estimate unexpected losses based on the combined use of internal and relevant external loss data, scenario analysis and bank-specific business environment and internal control factors”. Figure 7. The requirements for the AMA-model. The internal data contains the losses that the bank has experienced and reported into its own system. N.B. that all of the reported losses exceed a specific amount, the reporting threshold. That means that all losses that are not as severe as to reach that threshold are not reported into the system. One can thus say that the internal dataset is “truncated” below a specific value. The external data are for example losses from several banks that have been pooled in order for member banks to collect supplementary information. ORX is one of these pools where members are expected to report their full loss data history on a quarterly basis (ORX Association, 2007). The threshold for these data is € 20000. Scenario analysis implies the use of expert opinion in form of risk management experts and experienced managers in order to evaluate exposure to high-severity events. The analysis should also cover how deviations of correlation assumptions may impact the outcome (Basel Committee of Banking Supervision, 2006). Business environment and internal control factors are drivers of risk that should be translatable to quantitative measures. Changes of risk control in the bank should thus be captured by the model. 4.2 Three Types of AMA Models Since there is no specific regulations of how to develop the AMA model, except that it has to fulfil the above requirements, there have been different approaches at different banks. The Basel committee (Basel Committee on Banking Supervision, 2001) has acknowledged three broad types of 30 approaches that are currently under development; the internal measurement approach (IMA), the loss distribution approach (LDA) and the scorecard approach (SCA). 4.2.1 Internal Measurement Approach (IMA) In this type of approach the bank’s losses is generally divided into business lines and event types. For each of these combinations, the expected loss is calculated by the combination of frequency and severity estimates. The capital charge for each of the combinations Ki,j is (Basel Committee of Banking Supervision, 2001): K i , j = γ i , j * EI i , j * PDi , j * LGDi , j = γ i , j * ELi , j Where γi,j= a parameter whose value depends on the combination of business line / event type, EIi,j = the Exposure indicator. The higher EIi,j the larger exposure in that particular business line, PDi,j= Probability of default, LGDi,j= the loss given default. IMA thus assumes the relationship between expected and unexpected loss to be fixed, regardless of the level of expected losses and how the frequency and severity are combined (Basel Committee of Banking Supervision, 2001). 4.2.2 Loss Distribution Approach (LDA) The loss distribution approach has over the recent years emerged as the most common of the three approaches (Jobst, 2007). As in the case of IMA, the bank’s losses are divided into business lines and event types. The approach is founded on certain assumptions about the distributions for the severity and frequency of events (Basel Committee on Banking Supervision, 2001). These are commonly simulated using the Monte Carlo simulation technique (ITWG, 2003). The aggregated loss distribution is the combination of these two distributions. By using the Value-At-Risk of the loss distribution one finds the total risk exposure (EL and UL) for 1 year at 99.9 % statistical confidence level (Jobst, 2007). In other words, this amount should cover all expected and unexpected losses the bank faces during a year with 99.9 % probability. I.e. if the model is correct, the bank’s equity should not be sufficient one year out of 1000. What distinguishes this approach is that it assesses unexpected losses directly. It does not, as in the case of IMA, make an assumption about the relationship 31 between EL and UL (Basel Committee of Banking Supervision, 2001). Haas and Kaiser (Cruz, 2004) present an overview of the Loss Distribution Approach, illustrated as follows: Figure 8. Overview of the Loss Distribution Approach. (Cruz, 2004, p. 18) 4.2.3 Scorecard Approach (SCA) This approach is based on the idea that banks determine an initial level of operational risk capital, which they modify over time as the risk profile changes. The scorecards that are used to reflect improvements of risk control may be based on either actual measures of risks (for example LDA or IMA) or indicators as proxies for particular risk types (Basel Committee on Banking Supervision, 2001). Even though the model has to incorporate both internal and external data in order to be qualified as an AMA, this approach is quite different from LDA and IMA regarding the use of historical information. According to the Basel Committee on Banking Supervision (2001), once the initial amount is set the modification can be made solely on qualitative information. 4.3 Previous Research – AMA The research regarding AMA and its implementation regards everything from qualitative requirements to detailed mathematical derivations. Since we have limited our thesis to the quantitative requirements, we have only searched for reports concerning these. A selection of the empirical research that was found is presented below. Frachot and Roncalli (2002) propose a way of how to mix internal and external data in the AMA; both the frequencies and the severities. 32 Baud, Frachot and Roncalli (2002) present a way to avoid the overestimation of the capital charge in the mix of truncated internal and external data. With the argument that the truncation gives falsely high parameter values, they propose a way to adjust for the unreasonably high capital charge that is calculated. Shevchenko (2004) mentions important aspects in the quantification of the operational risk. These are among others the dependence between risks, the scaling of the internal data and the modelling of the tail. Embrechts, Furrer and Kaufmann (2003) provide a thoroughly explanation of how extreme value theory can be applied in the model. They also propose 200 exceedances above u in the POT-method to be ideal for reliable estimation of the tail. Frachot, Georges and Roncalli (2001) explore the LDA approach. They calculate the capital charge, compare the approach with the IMA and relate it to the economic capital allocation in the bank. Frachot, Roncalli and Salomon (2004) examine the correlation problem in the modelling of operational risk. They present a simplified formula for the total capital charge where the correlation parameter is incorporated. The correlation parameter is, according to their calculations, not higher than 5-10 %. Jobst (2007, p. 32) focuses on the extreme value theory and found that a bank with “distinct lowfrequency, high-severity loss profile” only loses approximately 1.75 % of gross income over five years. This can be compared with the SA where the capital charge approximates 15 % of gross income. Thus, Jobst (2007) argues that business volume may not be a good measure of the drivers behind operational risk. 33 5 Model Choice & Assumptions The literature concerning operational risk, and the AMA in particular, presents a wide range of model approaches as shown. The Basel Committee’s list of approaches (IMA, SCA and LDA) is an attempt to categorize the different models suggested. Since the model choice as well as the assumptions made is critical for the model design, this section concerns these issues. 5.1 What is a Good Model? Since our case study is the development of a model, one may ask what a good model is in this respect. And if there is a difference in the answer depending on who is asked. The reason for turning from simpler to more advanced methods from Finansinspektionen’s point of view is to achieve a more precise calculation of the operational risk that the bank is exposed to. For the bank to have an incentive to run the development from simple to advanced model (more than to get a model that more precisely estimates the risk) the more simple methods have overestimated the risk taken. By developing an AMA-model, the bank has a chance to lower the capital requirement, and thus the equity that has to be held. The question of how to design a good model is thus two folded. Both the bank and Finansinspektionen wants a model that estimates the risk as precisely as possible. From the bank’s perspective, the capital requirement calculated however also has to be lower than with the Standard approach in order to be implemented in the bank. We have tried to rid ourselves off the fact that we work for the bank and have aimed to make the model as precise and “academically correct” as possible. Since we lack the data necessary, it is not possible to figure out if the capital requirement actually turns out to be lower than with the model currently used. A good model according to us is further a model that gives good incentives to improve the risk exposure of the bank. One may argue that the main role for the model itself is not to measure, but to have a driving power on how the risk is managed within the bank. Key employees should further on be able to comment and influence some parameters of the model so as to make it as bank-specific as possible. It should also be flexible to business changes over time; if the risk or size increases, it should be possible to adjust some parameter so that it affects the result. 34 5.2 Overall Approach Choice Our overall approach choice, based on the Basel Committee’s list of approaches, is the Loss Distribution Approach (LDA). The main reason for the choice is the fact that it appears to be more of a “best practice” among academics than the others. Frachot, Roncalli and Salomon (2004, p. 1) emphasize that “we firmly believe that LDA models should be at the root of the AMA method”. Shevchenko (2004, p. 1) mentions that ”Emerging best practices share a common view that AMA based on the Loss Distribution Approach (LDA) is the soundest method”. There are however drawbacks to the “basic” LDA approach presented in chapter 4.2.2. Nyström and Skoglund point out, in the collection by Cruz (2004), that the operational risk event in the basic version is completely exogenous, leaving the risk manager with no control over the risks and the capital charge. It also implies perfect positive dependence between the cells (BL/ET), something that is both unrealistic and gives too high of a capital requirement. We have chosen this model, but have made a few exceptions due to these issues. In our model, the total losses from the 56 cells are aggregated each year; the Value-at-Risk is thus measured on the aggregated number, not on the 56 cells separately. That implies that we do not consider them to be perfectly correlated as in the “basic” LDA case. All cells, except for some as explained in section 5.3, are instead considered independent. We have also added a scaling factor that causes risk managers influence to be reflected in the model. 5.3 Distribution Assumptions The loss frequency is assumed to be Poisson distributed. This is in line with “best practice” in the area; Rao and Dev states in the collection by Davis (2006, p. 278) for example that “Most commonly, a frequency distribution is posited as a Poisson distribution”, Kühn and Neu (2008, p. 3) say that “Common choices for the loss frequency distribution function are the Poisson or negative binomial distribution” and Mignola and Ugoccioni (2006, p. 34) mention that “The frequency distribution is assumed to be a Poisson distribution …”. There is no consensus about how the severity is distributed; but in order to reflect high severities, the tail has to be long and heavy tailed. The most common ones to use are according to Rao and Dev (as presented in the collection by Davis 2006) the lognormal distribution and the Weibull 35 distribution. The authors however mention that the lack of data at the right tail (the very high losses) implies the use of extreme value theory. They refer to a study conducted by De Koker (2005) which implies a solid basis for using generalized Pareto distribution (the POT-method in EVT) in the tail and lognormal or Weibull in the body. Moscadelli (2004) also reports that the extreme value theory, in its POT-representation, “explains the behaviour of the operational risk data in the tail area well. “ The problem in our model is however that the use of the generalized Pareto distribution (i.e. the POT-method) implies that one has to make a choice about the value of u. Embrechts, Furrer and Kaufmann (2003) emphasize that this choice is crucial, but not easily made. The u has to be significantly high, and there has to be a significant number of losses exceeding this value. Embrechts, Furrrer and Kaufmann (2003) refer to a simulation study, worked out by McNeil and Saladin, which considered 200 exceedances over u to be necessary; both if one chooses u to be 70 % and 90 % of the largest value. Since the bank’s internal data do not fulfil these requirements and since we lack external data, we are not able to fulfil that requirement in the choice of u. Thus, we can not use the POT-method for the tail distribution and are left with only one distribution for the whole body and tail. In the instruction for future updates, a remark about the POT-method is conducted. In order to make a choice about the lognormal or Weibull distribution for the severity distribution, two qq-plots were made: Figure 9. qq-plot of the internal data and the lognormal distribution. 36 Figure 10. qq-plot of the internal data and the Weibull distribution. Since a straight line indicates that our sample is distributed as the reference distribution (as explained in section 3.2.3) we decided, by looking at the qq-plots, that the lognormal distribution was the most appropriate choice for the severity of the data. 5.4 Correlation Assumptions The correlation among the cells is one major issue that is illustrated in several research reports. When referring to the correlation; there may be both frequency correlations and severity correlations. We have in our model tried to make a bottom-up approach, indicating that the cells have to have obvious dependence in order to be considered correlated. We have thus considered some cells to be correlated out of the assumption that they have occurred simultaneously by a reason. No correlation has been applied afterwards. The assumptions are proposed by key employees within the bank. In order for event types to be correlated but not reported as one loss, there has to be an “indirect” correlation. Thus, there has to be something that triggers the loss in the other event type, but nothing that is all too apparent so that they will be reported as one major loss instead of two separate ones. The Event types that are considered correlated are Business Disruptions & System Failures & Internal Fraud, Execution, Deliver & 37 Process Management & Internal Fraud and Business Disruption & System Failures & External Fraud. Examples of these are IT-crashes that cause internal and external fraud, and control failures that lead to internal fraud. The correlations are illustrated in the figure below. Each colour represents one type of correlation and there are only correlations within each business line. The number of correlation pairs thus totals 24. Figure 11. Overview of the correlation pairs assumed in “our” bank. The frequencies of the “x-marked” cells are assumed to follow a joint Poisson distribution. That means that the pairs are Poisson-distributed “together”, recall the definition p X ,Y (x, y ) = P ( X = x, Y = y ) , − ∞ < x, y < ∞ How the correlations are handled in the model is presented in greater detail in chapter 6. The potential correlation between severities may be modelled by the use of copulas. Recall that the copula is a tool that defines a correlation structure between two correlated variables, each with their own marginal distributions. Marginal distributions refer to the own frequency distribution that each cell has (even if they both have marginal Poisson distributions, they probably have different parameter values). By using a copula, the marginal distributions are preserved at the same time as the correlation structure between the cells is defined. Here, we however made the assumption that the severities of the cells are not likely to be correlated. I.e. a large loss in Business Disruptions & System Failures does not imply a large loss in Internal Fraud, only that the frequencies of the losses are 38 dependent. There is however a possibility of implementing a severity correlation in the model, if it can be shown that there is a substantial positive dependence between them. 5.5 Assumptions regarding Scaling, Inflation & Truncation Bias When it comes to the mixing of internal and external loss data, there is a potential problem relating to size. “Our” bank is relatively small compared to the other banks in the ORX database (ORX Association, 2007). We assume that larger banks have larger operational losses and that the income thus reflects the amount of risk the bank is exposed to. This is in line with the less sophisticated models proposed by the Basel Committee, which calculate the capital requirement solely on the amount of gross income in the bank. Since we want to build the model as flexible as possible, we want to scale each combination of Business Line/Event Type. That is done by adding a cap for how large the severities can be in each cell. Unreasonable great losses above a certain threshold are scaled down in order to better reflect the risk that “our” bank is exposed to in that particular Business Line/Event Type. We have decided to adjust for the inflation effect by multiplying the severities with (1 + average inflation)k. k denotes the number of years since the loss occurred, where the average inflation is calculated over the last five years. Different banks have different levels of truncated internal loss data. That means that only losses that exceed the value of h is reported, causing the dataset to be truncated at the value of h. Baud, Frachot and Roncalli (2002) report that the truncation affects the estimated parameter values used in the simulation. That in turn causes an overestimation of the capital charge calculated by Value-at-Risk. Since “our” bank’s truncation level h is relatively low, we assume there to be a marginal impact on the parameters. 5.6 Risk & Simulation Method We have chosen the Value-at-Risk measure in the model, despite Expected Shortfall’s advantages described in chapter 3.2.2. The reason behind this rationale is three folded; Finansinspektionen does not require the use of ES in the model, it is best practice to use Value-at-Risk (especially in the loss distribution approach chosen) and the security level of 99.9 % is expected to be sufficiently high. 39 The simulation method chosen is the Monte Carlo simulation. It is superior to the historical simulation, since the latter is limited to the losses that have occurred in the bank. By using a Monte Carlo simulation, it is possible to simulate larger losses than what has happened historically. Major simulated losses that exceed our yet largest loss are thus possible according to the mean and variance in the historically data, with the only exception that they have yet not occurred. What nonetheless needs to be emphasized is that the distribution assumption causes a limit (even though not as severe as in the historical simulation case) to the losses that can be simulated; the simulated values are completely dependent on the distribution that has been chosen. 40 6 Description of Our Model This section describes how our model has been designed. The description is divided into four parts; an overall view of how the model is designed, description of the import of data, the processing of data and how the simulation is conducted. N.B. that the Matlab code has not been published due to confidential reasons. 6.1 Model Overview The calculations in this model are tested only on internal historical loss data. The model is however constructed so that it is possible to import external loss data and mix them with the internal ones as soon as they become available. The data is first imported into Matlab where it is sorted and structured to facilitate further calculations. Irrelevant data is removed at this stage. After that, we fit the data into different distributions according to our assumptions. Ultimately we simulate losses from these distributions and identify VaR0.999 (the Value-at-Risk at confidence level 99.9 %). Each step is thoroughly explained below. 6.2 Data Import The historical losses are delivered in excel-sheets. In order to use the data in Matlab, it has to be “recoded”. This is done as follows: 1. Every Business Line is given a number. 1 2 3 4 5 6 7 8 Business Lines Corporate Finance Trading & Sales Retail Banking Commercial Banking Payment & Settlement Agency Services Asset Management Retail Brokerage Figure 12. Business Line overview. 41 2. Every Event Type is given a number. 1 2 3 4 5 6 7 Event Types Internal Fraud External Fraud Employment Practices & Workplace Safety Clients, Products & Business Practices Damage to Physical Assets Business Disruption & System Failures Execution, Delivery & Process Management Figure 13. Event Type overview. This gives us a matrix of 56 different kinds of losses looking as follows: Figure 14. Matrix overview. 3. Dates are converted into numbers, where 1900-01-01 is the first day and the all the following dates are counted up by one (i.e. 2008-09-11 is converted into 39702). This is done, in the Swedish version of Excel, by using the command ‘=DATUMVÄRDE("2008-09-11")’. 4. Every Region is given a number. 1 2 3 Area Sweden Baltic States International Figure 15. Area overview. 42 5. Events that have neither BL nor ET are removed since they cannot contribute to the calculations. There must not be any extra spaces in the data, i.e. a loss of ten thousand SEK must for example be listed as “10000” and not as “10 000,00”. 6. Once all the text has been coded into digits, the data is imported into Matlab. 6.3 Data Processing Once imported, the data needs to be processed. This is done as follows: 1. Direct and Indirect (Other) losses are added together. Recoveries are not taken into consideration, since it is the direct loss that is of interest, not the long-term effect. 2. Data older than date x 7 is removed from the data set. That, since the bank’s data collection procedures were not fully reliable until year x. 3. When the data is delivered it is sorted in the order of reporting date. However, the reporting date is not very important, what matters the most is the time of occurrence. The data is thus resorted in order of when it actually happened, starting with the most recent event. It is then divided into separate years y1, …, yn. Each year gets two matrices, one for frequency and one for severity. Figure 16. Overview of how the data is sorted. 4. The average frequency per year and the average severity per occurrence are calculated. This is done for each Business Line and Event Type. The yearly losses are calculated by starting with the latest event (not necessarily today) and counting a year from that date. See figure below. 7 The date x is confidential. 43 Figure 17. Frequency calculation. 5. Average frequency and severity for each Business Line and Event Type is calculated for the different regions as well. As for the frequency calculation, we assume that the data available represents the same number of years for all regions. That, even though it in reality may not always be the case. See figure below: Figure 18. Frequency calculation across different areas. 6. Frequencies are adjusted depending on the relative size of that specific Business Line compared to the average size of a bank involved in the external data. Size is calculated as the average turnover per bank per year over the last three years. 7. In order to compensate for losses that are unreasonably large for a bank of “our” bank’s size, external severities are capped. Each cell has its own cap based on size (measured in gross income per annum) and expert opinions. The compensated loss for losses exceeding the cap equals the cap size plus x percent of the residual loss 8 . Losses that are large (more than 50 % of the cap size) but not larger than the cap are reduced by a percentage based on the size of the loss compared to the cap. This is done in two steps; at 50 % and at 75 % of the cap size. 8 In our case, we chose 1 %. 44 Figure 19. Illustration of the scaling. To know exactly what reduction to use we would recommend comparing the average of the top 25 % of the internal losses µint,1 with the average of the top 25 % of the external losses µext,1 (see figure 20). Then scale the losses by multiplying the external losses that are more than 75 % of the cap with: 1+ x1 = μ int,1 μ ext ,1 2 Similar calculations can be made with the top 25 % through 50 % of the external and internal losses. This means that if the internal losses in the top quadrant are, on average, 80 % of the external losses in the top quadrant, then we will use 90 % of the external losses that are larger than 75 % of the cap size in our calculations. Please note that the cap sizes will be set relatively high, so this adjustment of losses will only be used if the losses are substantially large and thus unlikely to occur. To smoothen the curve above (i.e. to make sure that losses that are just above a certain threshold is not counted as lower than a loss just below that same threshold) we adjust the model by letting the adaptations on the lower levels be used also for the bigger losses. I.e. a loss that exceeds the cap size C will be adjusted as follows: ~ L = C ⋅ 0.5 + (C ⋅ 0.75 − C ⋅ 0.5) ⋅ x 2 + (C − C ⋅ 0.75) ⋅ x1 + (L − C ) ⋅ x0 45 Where ~ L = Loss used in calculatio ns C = Cap size L = Actual external loss x0 = Constant above cap size x1 = Constant for top 25 % x 2 = Constant for top 25 % - 50 % Figure 20. Illustration of the scaling. 8. All severities are multiplied by (1 + average inflation)k where k is the number of years since the loss occurred. This is done to compensate for inflation (which is calculated as the average inflation over the last five years). 9. The severity data is fitted into a lognormal distribution using maximum likelihood estimation at a 95 % confidence level. 10. The cells that are assumed to be independent, their frequencies are assumed to have a Poisson distribution. 46 11. The pair cells that are identified as dependent are assumed to have a compound Poissondistribution. We take this into consideration by calculating a new combined frequency for the correlated cells. The frequency answers the question of how many times each year a loss occurs in any or both of these two cells. This is not done by just adding the two original frequencies together, since it may happen that events in these particular cells occur at the same time without any dependence being the cause. Just adding them would mean that they occur more often than what is rational to assume. During the simulations, for every occurrence in this new frequency, we simulate the combination of the two that has actually occurred (the first, the second or both of them). Figure 21. Illustration of the assumed correlations. The severities are not assumed to be correlated. This means that once the number of occurrences in each cell is calculated, the severities in each cell are simulated independent of the other cells. Hence, we assume that an occurrence in one cell can lead to an occurrence in another, but having a large loss in the first does not necessarily imply a large loss in the second as well. The calculation of the correlated frequencies is conducted as follows: Event a occurs with frequency λa (assumed given by historical data). Event b occurs with frequency λb (assumed given by historical data). Event c occurs with frequency λc (assumed given by historical data). Event d occurs with frequency λd (assumed given by historical data). 47 a and/or b and/or c and/or d occurs with frequency λ (unknown). However, the combinations (a,c), (a,d) and (b,d) can not occur in this specific part of the model since they are assumed not to be correlated. If a and/or b and/or c and/or d has occurred (λ), how do we know what actually occurred? Assume: P(I a = 1, I b = 0, I c = 0, I d = 0 ) = p1,0, 0,0 P(I a = 0, I b = 1, I c = 0, I d = 0 ) = p0,1, 0,0 P(I a = 0, I b = 0, I c = 1, I d = 0 ) = p0, 0,1,0 P(I a = 0, I b = 0, I c = 0, I d = 1) = p0, 0,0,1 P(I a = 1, I b = 1, I c = 0, I d = 0 ) = p1,1,0, 0 P(I a = 0, I b = 1, I c = 1, I d = 0 ) = p0,1,1, 0 P(I a = 0, I b = 0, I c = 1, I d = 1) = p0,0,1,1 where pa,b,c,d is the probability that event a, b, c or d occurs simultaneously or one or the other. We know that one (and only one) of these will occur. Hence: p1, 0,0, 0 + p0,1, 0,0 + p0,0,1, 0 + p0, 0,0,1 + p1,1,0, 0 + p0,1,1,0 + p0, 0,1,1 = 1 (1) In order to be true to our historical data, we must choose λ so that the following holds: λa = λ ⋅ p1, 0,0, 0 + λ ⋅ p1,1, 0,0 ⇒ p1, 0,0, 0 = λa − λ ⋅ p1,1, 0,0 λ (2) λb = λ ⋅ p0,1,0, 0 + λ ⋅ p1,1, 0,0 + λ ⋅ p0,1,1,0 ⇒ p0,1, 0,0 = λb − λ ⋅ p1,1, 0,0 − λ ⋅ p0,1,1,0 λ (3) λc = λ ⋅ p0,0,1,0 + λ ⋅ p0,1,1, 0 + λ ⋅ p0, 0,1,1 ⇒ p0,0,1,0 = λc − λ ⋅ p0,1,1,0 − λ ⋅ p0, 0,1,1 λ (4) λ d = λ ⋅ p 0, 0,0,1 + λ ⋅ p 0,0,1,1 ⇒ p 0,0 ,0,1 = λ d − λ ⋅ p 0,0 ,1,1 λ (2), (3), (4) and (5) are inserted into (1) which gives: 48 (5) λa − λ ⋅ p1,1,0, 0 λb − λ ⋅ p1,1,0, 0 − λ ⋅ p0,1,1, 0 λc − λ ⋅ p0,1,1,0 − λ ⋅ p0,0,1,1 λd − λ ⋅ p0,0,1,1 + + = + λ λ λ λ = 1 − ( p1,1, 0,0 + p0,1,1, 0 + p0, 0,1,1 ) λ is taken out, which gives: λ= λa + λb + λc + λd 1 + p1,1,0, 0 + p0,1,1, 0 + p0,0,1,1 So, to be able to calculate λ we first need to estimate p1,1,0,0, p0,1,1,0 and p0,0,1,1 which is done based on historical data and expert opinions. Since also p1,0,0,0 through p0,0,0,1 will be calculated based on this, one must be very careful when estimating p1,1,0,0 , p0,1,1,0 and p0,0,1,1, because, for example, setting p1,1,0,0 too high could result in p1,0,0,0 being negative (which, of course, is not allowed). Explicitly, the following must hold for p1,1,0,0 and λb: p1,1, 0,0 + p0,1,1, 0 < λb λ When simulating the correlated losses we first draw from the joint Poisson distribution. Then, for each occurrence, we draw a number between 0 and 1. a, b, c, d or a combination thereof occurs depending on what the drawn number is (see figure 22 below). Figure 22. Illustration of the probabilities in the joint distribution. The value of p0,1,1,0 may be hard to estimate for the key employees in the bank, since they have to answer the question; “Dependent on the fact that either A, B, C or D has occurred, what is the possibility that only B and C have occurred?”. It would have been much easier for the employee to estimate p1,1, i.e. to answer the question “Dependent on the fact that either B or C has 49 occurred, what is the probability that both B and C have occurred?” In order to be able to ask this question instead of the first one, we have to find what p0,1,1,0 is dependent of the value of p1,1. The derivation of this formula is expressed below. Figure 23. Illustration of four events that are not mutually exclusive. Assume the four events as given in figure 23, we thus want to find: ⎧ P( X ∩ Y ) ⎫ P((B ∩ C ) ∩ ( A ∪ B ∪ C ∪ D )) p0,1,1, 0 = P(B ∩ C | A ∪ B ∪ C ∪ D ) = ⎨ P( X | Y ) = ⎬= P(Y ) ⎭ P( A ∪ B ∪ C ∪ D ) ⎩ In this case P ( A ∪ B ∪ C ∪ D ) = P ( A) + P (B ) + P (C ) + P (D ) − P ( A ∩ B ) − P (B ∩ C ) − P (C ∩ D ) and P ((B ∩ C ) ∩ ( A ∪ B ∪ C ∪ D )) = P (B ∩ C ) since if (B ∩ C ) is true then (A ∪ B ∪ C ∪ D) must also be true. Hence, p0,1,1, 0 = P (B ∩ C ) P ( A) + P (B ) + P (C ) + P (D ) − P ( A ∩ B ) − P (B ∩ C ) − P (C ∩ D ) (1) We need to find P ( A ∩ B ) , P (B ∩ C ) and P (C ∩ D ) . The question we want answered above gives us p1,1 where, for example p1b,1,c = P(B ∩ C | B ∪ C ) . We can also express this as: P (B ∩ C | B ∪ C ) = P ((B ∩ C ) ∩ (B ∪ C )) , but P ((B ∩ C ) ∩ (B ∪ C )) = P (B ∩ C ) since if P (B ∪ C ) 50 (B ∩ C ) is true then (B ∪ C ) must also be true. Hence, P (B ∩ C | B ∪ C ) = P ((B ∩ C ) ∩ (B ∪ C )) P (B ∩ C ) = ⇒ P (B ∪ C ) P (B ∪ C ) ⇒ P(B ∩ C ) = P(B ∩ C | B ∪ C ) ⋅ 1442443 = p1b,,1c P (B ∪ C ) 1424 3 ⇒ =( P ( B )+ P (C )− P ( B∩C )) ⇒ P(B ∩ C ) = p1b,1,c ⋅ (P(B ) + P(C ) − P(B ∩ C )) ⇒ ⇒ P(B ∩ C ) + p1b,1,c ⋅ P(B ∩ C ) = p1b,1,c ⋅ (P(B ) + P(C )) ⇒ p1b,1,c ⋅ (P (B ) + P(C )) p1b,1,c ⋅ (P (B ) + P(C )) P(B ) + P (C ) = = ⇒ P (B ∩ C ) = 1 1 + p1b,1,c ⎛ 1 ⎞ b ,c 1 + b ,c p1,1 ⎜⎜ b ,c + 1⎟⎟ p1,1 ⎝ p1,1 ⎠ (2) Note that (2) also holds for P ( A ∩ B ) and P (C ∩ D ) . (2) is inserted into (1) and we get p 0,1,1, 0 P(B ) + P(C ) 1 1 + b ,c p1,1 = P( A) + P(B ) P(B ) + P(C ) P(C ) + P(D ) P( A) + P(B ) + P(C ) + P(D ) − − − 1 1 1 1 + a ,b 1 + b ,c 1 + c,d p1,1 p1,1 p1,1 Since P(A) through P(D) are unknown (actually, we need p0,1,1,0 to get them), we have to use something else. P(A) through P(D) must be based on our historical data, λa through λd, hence we can use these to get estimations of P(X) which we call P’(X). If we assume that P ′( X ) = λx λ a + λb + λc + λ d where X = A, B, C, D. Then we can calculate our p0,1,1,0 and then calculate or “real” P(A) through P(D). Hence 51 p 0,1,1, 0 P ′(B ) + P ′(C ) 1 1 + b ,c p1,1 = P ′( A) + P ′(B ) P ′(B ) + P ′(C ) P ′(C ) + P ′(D ) P ′( A) + P ′(B ) + P ′(C ) + P ′(D ) − − − 1 1 1 1 + a ,b 1 + b ,c 1 + c,d p1,1 p1,1 p1,1 This formula provides the value of p0,1,1,0 dependent on the values of p1,1 for each of the alternatives. Thus, we are able to ask the question “Dependent on the fact that either B or C has occurred, what is the probability that both B and C have occurred?” and insert the answer in the formula above. 6.4 Simulation We are now ready to begin our Monte Carlo simulation. The simulation is conducted as follows: 1. First we simulate the number of occurrences in each cell (BT and ET) each year. By this we mean how many losses that occur in each cell per year. For the uncorrelated cells, this is done by assuming a Poisson distribution using our frequencies as λ, and drawing random numbers from this distribution. For the correlated cells, we use the combined λ calculated above to draw the number of occurrences for the combination of the different cells. Then, for each occurrence, we simulate what actually happened using the different pa,b,c,d we have. This gives us the occurrences for each cell. 2. For each cell we check the number of simulated losses and simulate the size of those losses using a lognormal distribution. The losses are then saved. 3. For each year, the losses in each cell are added up, giving us the total loss for that specific year. 4. This is repeated a great number of times (105), and simulated losses are saved in a vector l. 5. The vector is sorted in order of increasing size and we identify VaR0.999. Recall from section 3.5.2 that this is calculated by using VˆaRα = l[n (1−α )]+1 , n 52 7 Model Testing Usually when it comes to building a model, the testing part of the process is of considerable importance. One not only wants to test the results; if they are reasonable and how sensitive they are to changes in the parameters. One also wants to test the programming itself; does it give any results, are the results logical and is the model flexible enough to handle changes? One major issue that we were facing in our modelling process was the lack of data. The internal data are scarce and we did not have access to any external data. This made it impossible to test and verify the results. It was for example not possible to test what kind of influence different weighing factors, scaling assumptions, distribution choices and correlation coefficients would have had on the result. The calculation of the capital requirement itself did not provide any reliable results, since the estimation of the severity and frequency parameters can not be considered to be statistically valid. It was further on also impossible to stress test the result, since the lack of data did not give a reliable result. Our model was thus made as a “shell” for how to calculate the capital requirement once the necessary data is available. As for the ultimate result, the model does give an output of a relatively reasonable amount at this point. The programming could however be tested and that was done thoroughly during the process of time. When something was added to the code, there was continuously testing of how the model worked. The same can be said about the logics of the code, so as to be sure that it worked as it was supposed to. We also continuously tried to make the code more concise, easier to understand and more flexible. Hard coding was avoided as much as possible. 53 8 Concluding Chapter What kind of insights has this modelling process provided us? To what level is our model valid and how may it be criticized? These questions are answered in this last part of the report together with suggestions of further research in the area of operational risk. 8.1 Insights Since the bank is given full scope, except for the four requirements previously mentioned in chapter 4.1.3, in the development of an AMA, there is no univocal way of how to develop the model. During the process it was apparent that it may not always be positive to leave all assumptions free of choice. It may sound like a fair approach to leave it up to the bank by assuming that the bank gets more knowledge of its risk exposure by developing the model. One could however argue that it would be helpful to have more strict guidelines from Finansinspektionen, and that the model itself would be more fair if there was one approach for all banks to use. We believe that operational risk may not differ that much between banks so as to argue that totally different approaches would be appropriate. It could be a better idea to just leave the scaling and correlation assumptions to the banks, in order to make the data more reliable to the bank’s specific exposure. During the modelling process it came to our attention how difficult it is to model operational risk. The majority of the interviewees questioned if it is even possible to model operational risk appropriately and if it was not better to keep it in Pillar 2 of the Basel II rules 9 . Compared to credit risk and market risk, operational risk is far more unpredictable; “Hur modellerar man Leeson?” [How can Leeson be modelled?] was asked by one of the interviewees. The insight that operational risk is a new and relatively unexplored area of risk modelling was thus central in the development process. 8.2 Future Required Updates The lack of data caused, as previously mentioned, problems to the test of the model. The lack of data was an issue for the choice of method as well; since we lacked reliable data, we could not use Basel II is divided into 3 Pillars; the calculation of the capital requirement belongs to Pillar 1, the aggregated risk assessment belongs to Pillar 2 and the rules for publishing information belong to Pillar 3. See Finansinspektionen (2002) for a more detailed explanation. 9 54 the POT-method for the distribution of the tail. The use of extreme value theory (i.e. the POTmethod) is recommended in literature, but was something we had to forsake at this point. When data becomes available, it is thus recommended to implement a generalized Pareto distribution for the tail of the severity distribution. That is, to use extreme value theory’s POT-method in order to get more high severity data. If it becomes apparent that external data comes from a substantial number of banks from other geographical areas than the ones that “our” bank operates in, it is possible to implement a weighing factor. It may for example be a good idea to weigh non-European banks’ losses as less likely than European banks’ losses. But it depends on what kind of factors is assumed to drive operational risk; if the geographical area has a substantial impact on the frequency and severity of events. Even though this model is only a “mathematical shell”, there are a few quantitative factors that have to be updated in the future. These include expert opinions for the correlation assumptions as well as expert opinions for the caps of the cells’ exposures. It is fair to assume that these are factors that may change in the future and thus need to be updated by key employees within the bank. It would further on be a good idea to make a qq-plot of the severities of the external data once they become available. Our qq-plot, with the following assumption of choosing the lognormal distribution for the data, is based only on the bank’s internal data. To be sure that the external data behaves similarly, a qq-plot of the external data as the sample and the lognormal distribution as the reference distribution is thus recommended. The same thing can be said about the frequency of the data. As it is now, we have not tested the suitability of the Poisson distribution for the frequency of the losses. One way of testing if the Poisson distribution is appropriate is to do a goodness of fit-test 10 , this could however not be conducted at this point since the data points are too few. As presented by the Basel Committee on Banking Supervision (2006) it is necessary to have scenario analysis in the model. We have fulfilled a part of the requirement by using expert opinion in the scaling part of the model (recall that the scenario analysis according to the Basel Committee on Banking Supervision (2006) incorporates the use of expert opinion in order to evaluate exposure to high-severity events, see chapter 4.1.3). The Basel Committee on Banking Supervision (2006, p. 154) 10 See http://www.zoology.ubc.ca/~whitlock/bio300/LectureNotes/GoF/GoF.html for an explanation of the test. 55 further mentions that it is necessary to have scenario analysis that “should be used to assess the impact of deviations from the correlation assumptions”. This is something that has to be added to the model after the correlation parameters have been chosen and the external data is available. It would also be a good idea to stress test the model. It would for example be useful to see how sensitive our parameters are to the addition of one extra large historical loss. It would be a problem if the capital requirement calculated would change dramatically as a result. Another test that could be conducted in the stress test is what effect a rise in the cap of one business line has on the capital requirement. If say Corporate Finance’s income triples, the cap for that business line would rise in our model. It would be interesting to see how sensitive our model is for the level of these caps. The stress testing of the model is however nothing that can be done without access to external data. 8.3 Validity & Critique Validity, as presented by Colorado State University, concerns “the degree to which a study accurately reflects or assesses the specific concept that the researcher is attempting to measure”. Regarding the validity of the model, there are a number of things to comment on. First, one may discuss how valid an AMA model is on the whole. The model shall provide a number that the bank will not exceed more than one year out of 1000 (i.e. with a 99.9 % confidence level). As a base for that calculation, the required observation period is only three years (Finansinspektionen, 2007). To make a good estimation, we would need about a million years of data (another problem would be that the data has to be identically distributed as well; so if we actually would have a bank that was a million years old, it would be highly questionable if the loss data from the early years could reflect the situation of today). Because of the lack of data, the number that our model provides will only at best be a “qualified guess”. Even though the use of external data provides more losses, one must keep in mind that the observation period for these is as short as for “our” bank’s internal losses. Second, one may criticize how the model itself is built. The fact that we do not use extreme value theory for the tail is one distinct weakness, since potential high losses are the most harmful losses for the bank. The scaling assumptions that are made are also arguable. Here, we rely on the assumption that the larger the bank is, the larger the potential losses. A bank of “our” bank’s size is assumed not able to meet with the high losses that other large banks have had. What really is the driver of operational risk is however far from clear. In our model we rely on key employees within 56 the bank to decide a cap for the exposure in each cell; for example “it is not possible to loose more than x kr of Internal Fraud in Corporate Finance”. Here, subjectivity is a large part in the decision of the maximum loss. The same can be said about the correlation assumption between the cells, which of the cells that are supposed to be correlated and to which extent they are correlated, is highly arguable. Third, there are details in the models that may be “theoretically wrong”. The yearly losses are calculated by starting with the latest event (not necessarily today) and counting backwards one year from that date. If there has gone a long time since the last event has occurred, we will be overestimating the risk. However, the risk/chance that nothing happens for a long time is very small. See figure below for an illustration of the problem: Figure 24. Illustration of the weakness in the calculation of the frequency. A similar problem appears when we calculate the frequency and severity for different regions. We assume that the data available represents the same number of years for all regions. This may not be the case, and it becomes a problem if the red area in the figure below becomes large. Figure 25. Illustration of the weakness in the calculation of the frequency across different areas. The inflation is calculated as the average inflation over the last five years. A more theoretically correct way to adjust for the inflation would be to implement that the losses would be multiplied with the actual inflation for the specific year. That would give a more correct value if, say the inflation rises dramatically five years in a row. 57 Fourth, the intentions behind the model development may be criticized. One may question the appropriateness of developing this model for the bank where we work. During the process we had to find the balance between developing a reasonable model and developing a beneficential one. On the one hand, if we had tried to make the model as appropriate and reliable as possible, there would have been a risk of creating a model that no one would have wanted to use (since there are less sophisticated models giving a lower requirement). On the other hand, if we would have developed the model as generous as possible, we would have put the bank in a situation where it would have risked having too low equity for the risks it was exposed to. From a mathematical standpoint it all comes down to what kind of assumptions that are made, since they have a direct effect on how ethically correct the model becomes in the end. 8.4 Further Research As more banks develop an AMA-model, there will be more interest in research concerning this area. We have encountered some research gaps that would have been good to look into in order to ease the modelling process. One possible research area concerns the reporting structure in the bank. In order to build and rely on a model like the AMA, it is required that the data set available is correct. That means that all data above a certain threshold should be reported, and reported as to the true loss size. One may however assume that employees within the bank are more reluctant to report some losses than others, internal fraud being one example. Some research concerning the employees’ view of reporting such losses may thus be valuable as to know how reliable the internal data really is. Another possible research area concerns the bank’s geographic location and what affect that has on its operational losses. That becomes an issue when one wants to mix external data with internal data; are the external ones as relevant as our internal data? It is for example possible to assume that banks in areas where tornados and other nature disasters are relatively frequent are more exposed to losses caused by Damage of Physical Assets than a bank in the Nordic region. If it could be shown that banks in specific parts of the world are exposed to greater operational risks, it is possible to scale down their part of the external data and thus give a more reliable set of data to work with. 58 Another possible research area concerns the impact the age of the bank has on the operational losses. Are new banks/bank divisions more exposed to operational risk than older banks? This question becomes an issue as one mixes the internal and external data. Different weights related to age could be applied to the data if it could be shown to have an impact on the losses. When we in our model considered correlation among severities to be irrelevant, it was more a logical assumption than a result from a research analysis conducted. By studying large losses and their possible dependence to other events, it may be concluded that such correlations really exist. It may also be arguable that there exists dependence among losses over time; for example that risk control makes it less probable for a large loss to be followed by another large loss. Research regarding such correlations could thus be helpful in the modelling process and propose the use of copulas in our model. An interesting research approach, regarding the AMA model in overall, is to study the degree of driving power that this kind of model has on the organization. It would be interesting to see to which extent such a model goes beyond its “main role” as a pure measuring model, and actually influences how the risk is managed in the bank. The last possible future research that we want to discuss concerns the driving factors of operational risk. In BIA and SA, it is assumed that the size of the firm, in terms of income, is the main driver. The higher the gross income, the higher the capital requirement ought to be. There is however one problem; if the bank’s income declines as a result from the economic situation, the capital requirement will get lower as a result. The question is if the operational losses in reality really do decline as well. Even if it is not a main assumption in the AMA-model, we have used the fact that large banks are more exposed to high severity losses when we scale down large losses for “our” bank. By conducting research with that as a base, one of the major problems within operational risk modelling could be eased. 59 9. References 9.1 Written Sources Bank of International Settlements. BIS History Overview. [Online] Available from: http://www.bis.org/about/history.htm [Accessed September 24, 2008] Basel Committee on Banking Supervision. (2001). Working Paper on the Regulatory Treatment of Operational Risk. [Online] Available from: http://www.bis.org/publ/bcbs_wp8.pdf [Accessed September 24, 2008] Basel Committee on Banking Supervision. (2006). Basel II: International Convergence of Capital Measurement and Capital Standards: A Revised Framework – Comprehensive Version. [Online] Available from: http://www.bis.org/publ/bcbs128.pdf. [Accessed September 23, 2008] Baud, N., Frachot, A. & Roncalli, T. (2002). How to Avoid Over-estimating Capital Charge for Operational Risk? Crédit Lyonnais, France. [Online] Available from: http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1032591 [Accessed September 29, 2008] BBC News. (1999). How Leeson broke the bank. [Online] Available from: http://news.bbc.co.uk/2/hi/business/375259.stm [Accessed September 25, 2008] Blom, G. (2005). Sannolikhetsteori och statistikteori med tillämpningar. Studentlitteratur, Lund. Cruz, M. (ed). (2004). Operational Risk Modelling and Analysis: Theory and Practice. Risk Books, London. Colorado State University. Writing Guides Reliability & Validity. [Online] Available from: http://writing.colostate.edu/guides/research/relval/pop2b.cfm [Accessed December 1, 2008] Davis, E. (ed). (2006). The Advanced Measurement Approach to Operational Risk. Riskbooks, London. 60 Embrechts, P., Furrer, H. & Kaufmann, R. (2003). Quantifying Regulatory Capital for Operational Risk. RiskLab, Switzerland. [Online] Available from: http://www.bis.org/bcbs/cp3/embfurkau.pdf [Accessed September 29, 2008] Finansförbundet, Finansvärlden. (2005). Vad är Basel II? [Online] Available from: http://www.finansforbundet.se/Resource.phx/pubman/templates/4.htx?id=478 [Accessed September 24, 2008] Finansinspektionen. (2002). Riskmätning och kapitalkrav II. [Online] Available from: www.fi.se/Templates/Page____3746.aspx - 21k [Accessed December 2, 2008] Finansinspektionen. (2007). Ansökan om internmätningsmetod, operativ risk. [Online] Available from: http://www.fi.se/upload/30_Regler/50_Kapitaltackning/070131/Ansokan_internmatningsmetode r_operativa_risker070126.pdf [Accessed September 25, 2008] Frachot, A., Georges, P. & Roncalli, T. (2001). Loss Distribution Approach for operational risk. Crédit Lyonnais, France. Frachot, A. & Roncalli, T. (2002). Mixing internal and external data for managing operational risk. Crédit Lyonnais, France. [Online] Available from: http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1032525 [Accessed November 31, 2008] Frachot, A., Roncalli, T. & Salomon, E. (2004). The Correlation Problem in Operational Risk. Crédit Agricole SA, France. [Online] Available from: http://gro.creditlyonnais.fr/content/wp/ldacorrelations.pdf [Accessed September 19, 2008] Gallati, R. (2003). Risk Management and Capital Adequacy. The McGraw-Hill Companies. Gut, A. (1995). An Intermediate Course in Probability. Springer Science+Business Media, Inc., New York. Hull, J.C. (2007). Risk Management and Financial Institutions. Pearson International Edition. Pearson Education, Inc. New Jersey. 61 Hult, H. & Lindskog, F. (2007). Mathematical Modeling and Statistical Methods for Risk Management: Lecture Notes. Royal Institute of Technology, Stockholm. [Online] Available from: http://www.math.kth.se/matstat/gru/5b1580/RMlecturenotes07B.pdf [Accessed October 31, 2008] Industry Technical Working Group (ITWG). (2003). The LDA-based Advanced Measurement Approach for Operational Risk – Current and In Progress Practice. RMG Conference, May 29. [Online] Available from: http://fednewyork.org/newsevents/events/banking/2003/con0529e.ppt [Accessed September 26, 2008] Jobst, A. A. (2007). Operational Risk – The Sting is Still in the Tail But the Poison Depends on the Dose. [Online] Available from: http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1087157 [Accessed September 26, 2008] KPMG International. (2003). Basel II: A Closer Look: Managing Operational Risk. [Online] Available from: http://www.kpmg.ca/en/industries/fs/banking/documents/Basel%20II_%20A%20closer%20look _%20managing%20operational%20risk.pdf [Accessed September 24, 2008] Kühn, R. & Neu, P. (2008). Functional Correlation Approach to Operational Risk in Banking Organizations. Institut für Theoretische Physik, Universität Heidelberg & Dresdner Bank AG, Germany. [Online] Available from: http://www.sciencedirect.com/science?_ob=ArticleURL&_udi=B6TVG-47G3FV22&_user=10&_rdoc=1&_fmt=&_orig=search&_sort=d&view=c&_version=1&_urlVersion=0&_ userid=10&md5=e97b061eacbe8065db66bf2bdbbb8cf9 [Accessed September 19, 2008] Mignola, G. & Ugoccioni, R. (2006). Sources of uncertainty in modelling operational risk losses. Journal of Operational Risk (33-50). Volume 1 / Number 2. Italy. [Online] Available from: http://www.journalofoperationalrisk.com/data/jop/pdf/jop_v1n2a3.pdf [Accessed September 29, 2008] 62 Moscadelli, M. (2004). The modelling of operational risk: experience with the analysis of the data collected by the Basel Committee. Banca D’Italia, Italy. [Online] Available from: http://papers.ssrn.com/sol3/papers.cfm?abstract_id=557214 [Accessed September 19, 2008] ORX Association. (2007). ORX Reporting Standards: An ORX Members’ Guide to Operational Risk Event/Loss Reporting. [Online] Available from: http://www.orx.org/lib/dam/1000/ORRS-Feb07.pdf [Accessed September 25, 2008] Shevchenko, P.V. (2004). Valuation and Modelling Operational Risk: Advanced Measurement Approach. [Online] Available from: http://www.cmis.csiro.au/Pavel.Shevchenko/docs/ValuationandModelingOperationalRisk.pdf [Accessed November 5, 2008] 9.2 Interviews Andersson, H. PhD & Risk Manager at the commissioner’s department Asset management, Stockholm. October 23, 2008. Degen, W. Head of compliance & operational risk at the commissioner’s department Swedish banking, Stockholm, October 19, 2008. Elsnitz, C. Risk Analyst at the commissioner’s department Group risk control, Stockholm. October 15, 2008. Keisu, T. Head of compliance, operational risk & security at the commissioner’s IT department Stockholm, October 20, 2008. Lang, H. Associate Professor at KTH Mathematic’s division of mathematical statistics, Stockholm. September 30, 2008. Lindskog, F. Associate Professor at KTH Mathematic’s division of mathematical statistics, Stockholm. September 30, October 16 & October 25, 2008. 63 Gustavsson, A. Risk Analyst at the commissioner’s Group risk control, Stockholm. September 20 & November 27, 2008. Lundberg, K. Risk Analyst at the commissioner’s Group risk control, Stockholm. October 15, 2008. Masourati, G. Risk Analyst at the commissioner’s Group risk control, Stockholm. September 20 & November 27, 2008. 64