Estimation quality for ordinal insolvency prognoses

from Wikipedia, the free encyclopedia

Estimates of the quality of ordinal insolvency prognoses measure the quality of ordinal insolvency prognoses . Ordinal insolvency prognoses are prognoses about the relative default probabilities of the rated companies like "company Y is more likely to fail than company X but less likely than company Z".

Ordinal insolvency forecasts could theoretically be differentiated as desired, but ordinal rating systems have prevailed in practice, which communicate their results on a discrete, 7- or 17-point scale in a notation adopted by Standard & Poor’s (S&P).

Importance of ordinal appraisal measures

Even if ordinal insolvency prognoses are more general than categorical insolvency prognoses , only comparative statements about the relative default risk of companies are insufficient for most applications. Rather, quantitative default forecasts are also required, for example to be able to specify whether a risk premium of 1.5% p. a. for a bullet , unsecured and senior credit with three years maturity is appropriate for a company whose rating classification stating that it "currently has the capacity to meet its financial obligations. However, adverse business, financial or macroeconomic conditions would likely affect his ability and willingness to meet his financial obligations ”.

Average historical one-year default rates according to modified rating classes according to Standard and Poor's (1981–2005) and Moody's (1983–2005), source: S&P (2006, p. 18), Moody's (2006, p. 35)

Nevertheless, it is worth examining the quality of estimates for ordinal insolvency prognoses:

  • In terms of their intention, they correspond to what the rating agencies promise to deliver.
  • They are now a dominant method for assessing the quality of rating results.
  • There are numerous possibilities of graphical representation for the ordinal estimation quality of rating systems, which also facilitates the communicability of the key figures derived from these representations.
  • A high degree of selectivity , as measured by quality measures for ordinal insolvency prognoses, is also important for the quality of cardinal insolvency prognoses - more important than correct calibration . It is easier to generate "correct" (calibrated) than selective failure probabilities.
  • In the empirical comparison of different rating procedures on the basis of identical samples, the order of the quality of the individual procedures according to ordinal estimated quality measures often also corresponds to the order of quality based on cardinal estimated quality measures. Differences in quality of different methods seem to be less in a different ability to issue calibrated failure prognoses, which is only relevant for cardinal quality measures, but rather in a different ability to deliver selective prognoses, which is relevant for both ordinal and cardinal estimated quality measures.
  • For those aspects of cardinal insolvency prognoses that cannot already be measured with the instruments developed for assessing ordinal insolvency prognoses, especially for the aspect of calibrating insolvency prognoses, no meaningful test procedures are currently available - if the default probabilities of the various companies are not are uncorrelated . However, the actual relevance of this theoretical objection is still controversial. There are empirical indications that the corresponding segment-specific correlation parameters assumed within the framework of Basel II are set too high by a factor of 15 to 120 (on average around 50).
Average historical cumulative one to ten year default rates according to modified rating classes according to Standard and Poor's (1981–2005) and Moody's (1983–2005) Source: S&P (2006, p. 18), Moody's (2006, p. 35)

Even if the rating agencies do not want to issue cardinal default forecasts for any specified period ex ante with the ordinal rating grades they have assigned, they demonstrate the quality of their rating assessments and the like. a. with their ability to identify companies with significantly different, monotonously increasing failure frequencies (see the two following figures for the average rating class-specific one- and multi-year failure rates according to S&P and Moody's).

The two figures show that the rating agencies were able, with their ex-ante rating grades , to separate ex-post groups of companies with very different realized default rates.

This is a necessary prerequisite in order to be able to make selective failure forecasts, but not a sufficient one. For example, the following extreme cases would be compatible with the representations in the figures above:

  • Extreme case I: The rating system assigns almost all companies to a medium rating level (see the first figure), for example BB or Ba2, and only very few companies to other rating levels.
  • Extreme case II: The rating assigns almost all companies to the extreme rating levels, i. H. either AAA or CCC / C, one and only a few in the middle rating levels.

In case I, the rating would be almost worthless, as it practically did not allow any differentiation between the various debtors. In case II, on the other hand, the information benefit would be considerable: the rating would always make "extreme forecasts" - that is, predict either an extremely low or an extremely high probability of default - and the forecasts would mostly be correct - an AAA rating would almost never and within a year CCC / C rating would be followed by default in at least 25% –30% of all cases.

In order to determine the ordinal quality of a rating system, it is therefore not only necessary to know the rating class-specific default rates, but also the distribution of the companies in the individual rating classes.

Graphical appraisal

The ability of a rating system to separate “good” and “bad” debtors with great reliability can be visualized using ROC or CAP curves (cumulative accuracy profile) and quantified using various key figures that can be converted into one another. Further names for ROC or CAP curves are reconnaissance profile, power curve, Lorenz curve, Gini curve, lift curve, dubbed curve or ordinal dominance graph.

ROC and CAP curves

The ROC curve of a procedure results from the set of all combinations of hit rates (100% - error I type ) and " false alarm rates " ( error II. Type ), which an insolvency forecast procedure based on different cut-off scores, ie . H. when converting cardinal or ordinal insolvency forecasts into categorical insolvency forecasts . If the cut-off value was “too sharp”, all companies would be predicted to be insolvent (100% hit rate (= 0% error I type); 100% false alarm rate (= 100% error II type)). Cutoff value, none of the companies would be forecast to be likely to be insolvent (0% hit rate; 0% false alarm rate). Even between these extreme cases there is a balance between hit and false alarm rates. The quality of ordinal insolvency prognoses is reflected in the nature of this compensation. A perfect forecasting procedure would not have to “wrongly” exclude a single debtor in order to record 100% of all defaults (vertical course of the ROC curve from (0%; 0%) to (0%; 100%)), a subsequent tightening of the cut-off value would only lead to an increase in the false alarm rate (horizontal course of the ROC curve from (0%; 100%) to (100%; 100%)). The ROC curve of a rating procedure, on the other hand, whose evaluations were purely random, would run along the main diagonal - every percentage point in the improvement of the hit rate would have to be paid for with one percentage point in the deterioration of the false alarm rate.

Real bankruptcy forecasting ROC curves are concave in shape. The concavity of an ROC curve implies that the realized default rates decrease with better creditworthiness, i.e. This means that the underlying rating system is "semi-calibrated".

CAP curves result from the application of a construction principle that is only slightly modified compared to ROC curves. The X-axis does not show the "false alarm rates" (type II errors), but the proportion of companies that have to be excluded from the forecasting process - regardless of whether they are actual failures or non-failures to achieve a certain "hit rate" (100% error of the first type). The CAP curve of a perfect rating runs steeply to the top right, starting from the point (0%; 0%) - but not vertically, as at least PD% of all companies must be excluded in order to record all failures (see the graphic above, right) .

non-overlapping and intersecting ROC curves

If the ROC / CAP curve of a method A1 is regularly to the left above the ROC / CAP curve of another method A2 when assessing the same companies, method A1 delivers better forecasts than method A2 for every conceivable cut-off value (see the graphic above) . If the relative advantage of a method only depends on its relative estimation quality, then method A1 would always be preferable to method A2 - regardless of the individual specifics of the decision-maker, for example with regard to its costs for errors I and II. Art. However, the ROC curves of two intersect Methods B1 and B2 (see the figure above, right), neither of the two methods is objectively superior to the other. In the example chosen, method B1 is better able than B2 to differentiate within the “good companies”, while method B2 has a higher degree of selectivity in the area of ​​“bad companies”. If the decision maker is not forced to use either only one or only the other method, he can possibly combine methods B1 and B2 to form a third method B3 in such a way that this is superior to both B1 and B2. The same applies to methods A1 and A2: although A2 is strictly dominated by A1, a combination of A1 and A2 can possibly generate a method A3 that is superior not only to A2 but also to A1. In empirical studies, for example, the forecast quality of bank ratings based exclusively on key figures could be improved by including “soft factors” , although the quality of the forecast of the key figure-based ratings was better than the quality of the forecast of the soft factor ratings.

In addition to graphical representations, key figures are also required for practical use in measuring the estimation quality of insolvency forecasts in order to present the information contained in CAP or ROC curves as compactly as possible. The minimum requirement for such an index is that this index must be defined for all ROC curves and that, if the ROC curve of a method A1 is on the left above the ROC curve A2, the quality measure associated with A1 must have a higher value as the A2 associated quality measure.

Quantitative appraisal

The quality measure usually used in connection with ROC curves is the content of the area under the ROC curve:

Formula 1:

The area under the ROC curve corresponds to the probability that for two randomly selected individuals, one drawn from the defaulting companies and one from the non-defaulting companies, the non-defaulting company has a better score than the defaulting company. With a perfect rating system, AUC is always 100%; with a purely random score (naive rating system), the expected value for AUC is 50%. To adjust the value range of the quality measure to the interval [-100%; +100%] instead of [0%; 100%] and to show "naive forecasts" with an expected value of 0% instead of 50%, the Accuracy Ratio is determined on the basis of the following:

Formula 2:

The parameter AR can also be represented as a special case of other common ordinal parameters. In contrast to the procedure for ROC curves, it is not usual to show the AUC dimensions calculated on the basis of CAP curves. Only the calculation of the CAP accuracy ratio, which is slightly modified compared to the ROC curve, with:

Formula 3:

Formula 3b:

where A´ stands for the area between the CAP curve and the diagonal and B´ for the area between the diagonal and the CAP curve, which a perfect rating procedure could achieve at most, see the graphic on the right in the figure above. It can be shown that and are identical.

For a rating system with g discrete classes, the integral ∫CAP (x) dx can be broken down into g triangular and rectangular sub-areas with known side lengths (see the following figure). The figure shows the area for a rectangular and a triangular area.

Formula 4:

Legend
g ... number of discrete rating classes,
a j ... share of companies in rating class j,
PD j ... realized failure frequency in rating class j,
PD ... average probability of failure
vertical decomposition of the area under the CAP curve, source: Bemmann (2005, p. 22)

By rearranging and inserting it in formula 3b, the result is:

Formula 5: with

Formula 5b:

where kumPD i stands for the proportion of defaults in rating classes 1..i in relation to the total of all defaults. The quality measures presented are defined equally for continuous scores as for scores with a finite number of possible characteristics (rating classes). The combination of continuous scores into discrete rating classes is associated with only minor loss of information, so that the quality of ordinal insolvency forecasts can be determined solely on the basis of the relative frequencies of the individual rating classes and the information on the realized default rates per rating class. This data is, for example, by the rating agencies S & P and Moody's published, but also for other rating systems, such as the credit rating of Creditreform available.

swell

  1. This article is based on Bemmann (2005).
  2. 17-step scale: 1 = "AAA", 2 = "AA +", 3 = "AA", 4 = "AA-", 5 = "A +", 6 = "A", 7 = "A-", 8 = "BBB +", 9 = "BBB", 10 = "BBB-", 11 = "BB +", 12 = "BB", 13 = "BB-", 14 = "B +", 15 = "B", 16 = "B-", 17 = "CCC / C", 7-point scale: "AAA", "AA", "A", "BBB", "BB", "B", "C".
  3. The Basler Committee (2000c, p. 23f) examines the rating symbols of 30 rating procedures from different international rating agencies . 22 agencies use letter combinations to express their rating assessment (approx. 75%), 6 agencies communicate their ratings in the form of grades (20%), only two agencies give their assessment in the form of default probabilities. Of the 22 “letter ratings”, 14 correspond exactly to the detailed S&P notation (17-step scale), 2 to the shortened S&P notation (7-step scale). In the case of banks, on the other hand, numerical rating class designations predominate (approx. 85%) instead of letter combinations (approx. 15%), see English, Nelson (1998, p. 4).
  4. Standard & Poor's has been using the rating class representation, modified by plus and minus signs, since 1974. Moody's, on the other hand, has only been around since 1982, see Cantor, Packer (1994, p. 2)
  5. See the definition for rating level B according to S&P (2003b, p. 7, own translation)
  6. See also Frerichs, Wahrenburg (2003, p. 13): “Under what circumstances is such a measure area Under the ROC-curve useful? The ranking of borrowers is sufficient for credit risk management if banks are not able to charge different credit risk premiums for different customers in the market. In this case, banks maximize their risk-adjusted returns by not granting credit to customers with negative expected returns which is equivalent to defining a minimum credit score. Yet, this line of thought does not lead us to the AUC as a measure of system quality, but to the concept of minimized expected error costs. The AUC measures the quality of the complete ranking and not only of one threshold. Only if the threshold is difficult to define in practice, the AUC may be a sensible measure. "- or in short and concise:" There are no bad loans, only bad prices. ", See Falkenstein, Boral, Kocagil (2000, p. 5).
  7. Cantor, Mann (2003, p. 6): "Moody's primary objective is for its ratings to provide an accurate relative (ie, ordinal) ranking of credit risk at each point in time, without reference to an explicit time horizon." And Cantor, Mann (2003, p. 1): "Moody's does not target specific default rates for the individual rating categories.", But also: "Moody's also tracks investment-grade default rates and the average rating of defaulting issuers prior to their defaults . These metrics measure Moody's success at meeting a secondary cardinal or absolute rating system objective, namely that ratings are useful to investors who employ simple rating 'cutoffs' in their investment eligibility guidelines. ", See ibid.
  8. Basel Committee (2000c, p. 2) "Most firms report that they rate risk on a relative - rather than absolute - scale, and most indicate that they rate 'across the business cycle", suggesting that ratings should not be significantly in principle affected by purely cyclical influences. ”Of the 15 rating agencies that provided information on this, 13 stated that they measure the“ relative risk ”of companies with their rating, only 2 agencies measure the“ absolute risk ”of companies according to their own information ( KMV Corporation and Upplysningscentralen AB), see ibid. (P. 23f).
  9. See for example McQuown (1993, p. 5ff), Keenan, Sobehart (1999, p. 5ff), Stein (2002, p. 5ff.), Fahrmeir, Henking, Hüls (2002, p. 22f), Engelmann, Hayden, Taschen (2003, p. 3ff.), Deutsche Bundesbank (2003a, p. 71ff.), OeNB (2004a, p. 113ff.). The rating agencies Standard and Poor's (2006, p. 48) and Moody's (2006, p. 11ff.) Also measure the quality of their rating systems using the methods and key figures presented in this section.
  10. Blochwitz, Liebig, Nyberg (2000, p. 3): “It is usually much easier to recalibrate a more powerful model than to add statistical power to a calibrated model. For this reason, tests of power are more important in evaluating credit models than tests of calibration. This does not imply that calibration is not important, only that it is easier to carry out. ", Analogous to Stein (2002, p. 9)
  11. The calibration of a rating system is described in Sobehart et al. (2000, p. 23f.) Or Stein (2002, p. 8ff.).
  12. See, for example, Krämer, Güttler (2003) for comparing the estimated quality of default forecasts by S&P and Moody's or Sobehart, Keenan, Stein (2000, p. 14) for comparing the estimated quality of six different methods using ordinal and cardinal estimation quality indicators.
  13. see Basel Committee (2005a, p. 31f.): “These [cardinal] measures appear to be of limited use only for validation purposes as no generally applicable statistical tests for comparisons are available. [...] The [Basle Committee's Validation] Group has found that the Accuracy Ratio (AR) and the ROC measure appear to be more meaningful than the other above-mentioned indices because of their statistical properties. For both summary statistics, it is possible to calculate confidence intervals in a simple way. [...] However, due to the lack of statistical test procedures applicable to the Brier score, the usefulness of this metric for validation purposes is limited. "And ibid., P. 34:" At present no really powerful tests of adequate calibration are currently available. Due to the correlation effects that have to be respected there even seems to be no way to develop such tests. Existing tests are rather conservative [...] or will only detect the most obvious cases of miscalibration [...] "
  14. Even with moderate default correlations (at least compared to the benchmark values ​​assumed by the supervisory authority), the realized default rates - even for portfolios of any size - can deviate considerably from the expected default rates, see Huschens, Höse (2003, p. 152f.) Or Blochwitz et al. (2004, p. 10).
  15. See Scheule (2003, p. 149). The investigations are based on a dataset from the Deutsche Bundesbank, which covers the period 1987-2000 and contains over 200,000 annual financial statements from over 50,000 (West German) companies, 1,500 of which became insolvent, see ibid. (P. 113ff.). On the basis of of Scheule (2003, p.156) portfolio loss loss distribution determined (a concrete model portfolio with an expected failure rate of 1.34%) is largely symmetrical and differs significantly from the typical left share-right skewed distribution type that can occur with strong default correlations results . The 99.9% quantile of both distributions, which is used to determine the capital requirements according to the Basel II rules , differs in the example chosen by a factor of 5 to 6!
  16. S&P (2005, p. 28): "Many practitioners utilize statistics from this default study and CreditPro® to estimate probability of default and probability of rating transition. It is important to note that Standard & Poor's ratings do not imply a specific probability of default; However, Standard & Poor's historical default rates are frequently used to estimate these characteristics. "
  17. The failure studies by S&P and Moody's are based on issuer ratings (also issuer ratings, corporate credit ratings, implied senior-most ratings, default ratings, natural ratings, estimated senior ratings), which are intended to represent a measure of the companies' expected probability of default, but without specify an explicit forecast horizon, see S&P (2003b, p. 3ff., 61ff.), Cantor, Mann (2003, p. 6f.), Moody's (2004b, p. 8) and Moody's (2005, p. 39). The ratings of specific liabilities of enterprises, so-called issue ratings (issue rating), consider not only the probability of default and the expected loss severity and can therefore be interpreted as a measure of the expected default costs. Depending on the rank of the assessed liability within the capital structure of the company, the security granted and other influencing factors, the issue rating results from surcharges or discounts (usually 1 to 2 points) from the issue rating. The issuer rating of a company generally corresponds to the issue rating of first-class, unsecured liabilities, see ibid.
  18. see Cantor, Mann (2003, p. 14)
  19. ROC curves have been used in experimental psychology since the early 1950s, see Swets (1988, p. 1287).
  20. see Blochwitz, Liebig, Nyberg (2000, p. 33): power curve, lift-curve, dubbed-curve, receiver-operator curve; Falkenstein, Boral, Kocagil (2000, p. 25): Gini curve, Lorenz curve, ordinal dominance graph; Schwaiger (2002, p. 27): Enlightenment profile
  21. See Bemmann (2005, Appendix I) for numerous examples of ROC curves for real insolvency forecasting methods.
  22. see Krämer (2003, p. 403)
  23. Another criterion for the relative advantageousness of the proceedings could be the costs of the proceedings themselves. While banks often use automated (and tend to have few selective) procedures for the classification of smaller loans, more labor-intensive evaluation processes are also worthwhile for larger loans, see Treacy, Carey (2000, p. 905) and Basler Committee (2000b, p. 18f. )
  24. see Blochwitz, Liebig, Nyberg (2000, p. 7f.)
  25. According to Krämer (2003, p. 402), when comparing rating systems, the “normal case” consists of intersecting ROC and CAP curves: “In this respect, the concept of failure dominance does not help in many applications. [...] The concept of failure dominance [recommends] above all for sorting out substandard systems. ”Predictions with weaker separation (failure-dominated) can be derived from more distinctive forecasts, see Krämer (2003, p. 397f.).
  26. See the study by Lehmann (2003, p. 21) in Section 3.4. See Grunert, Norden, Weber (2005), which however only show aggregated key figures for the different models (key figure rating, soft factor rating, combined rating) and no CAP or ROC curves. In the latter study, the predictability of the soft factor rating was even greater than that of the key figure rating (see ibid., p. 519).
  27. The reversal of the minimum requirement (“If A1> A2, then the ROC curve of A1 lies to the left above A2”) does not apply. It is only excluded that the A1 associated CAP / ROC curve lies to the right below the A2 associated CAP / ROC curve. It is also conceivable that the CAP / ROC curves of A1 and A2 intersect if A1> A2. Cantor, Mann (2003, p. 12): "Although the accuracy ratio is a good summary measure, not every increase in the accuracy ratio implies an unambiguous improvement in accuracy."
  28. see Deutsche Bundesbank (2003a, p. 71ff.)
  29. For further, easily interpretable indicators see Lee (1999).
  30. see Lee (1999, p. 455)
  31. The lowest possible value, 0%, would be achieved by a rating system whose forecasts are always wrong. By negating the forecasts, this rating could be converted into a perfect rating.
  32. See Hamerle, Rauhmeier, Rösch (2003, p. 21f.), Which show that the accuracy ratio is a special case of the more general key figure Somer's D, which is also used for variables to be explained with more than two possible characteristics (failure vs. non-failure) is defined. See Somer (1962, pp. 804f.) For an illustration of the relation of Somer's D to other ordinal measures of quality such as Kendall's Tau or Goodman and Kruskal's Gamma.
  33. see Engelmann, Hayden, Tasche (2003, p. 23)
  34. see Bemmann (2005, p. 22f.)
  35. see Bemmann (2005, Section 2.3.4)
  36. see Lawrenz, Schwaiger (2002)

literature