via email
Citigroup
Comment 1
Document 2: Draft supervisory guidance on Operational
Risk Advanced Measurement Approaches for Regulatory Capital
The comments to this paper are indexed
to the numbering of the standards in the document.
S 1. The institution’s operational
risk framework must include an independent firm wide operational risk
management function, line of business management oversight, and
independent testing and verification functions.
Clarification of the role of the
independent testing and verification functions is required. Our
Operational risk framework is reviewed by our independent Audit and Risk
Review (ARR) organization. However, testing of controls within each
business, as prescribed by our Risk and Control Self-Assessment
standards, is performed by individuals within the business. We consider
this to be appropriate because the businesses are ultimately responsible
for managing and controlling their operational risks.
S 2. The board of directors must
oversee the development of the firm-wide operational risk framework, as
well as major changes to the framework. Management roles and
accountability must be clearly established.
S 3. The board of directors and
management must ensure that appropriate resources are allocated to
support the operational risk framework.
The Board of Directors does have an
important role in reviewing Citigroup’s Operational Risk, however, roles
such as resource allocation are more appropriately executed by senior
management, rather than the Board.
S 4. The institution must have an
independent operational risk management function that is responsible for
overseeing the operational risk framework at the firm level to ensure
the development and consistent application of operational risk policies,
processes, and procedures throughout the institution.
S 5. The firm-wide operational risk
management function must ensure appropriate reporting of operational
risk exposures and loss data to the board of directors and senior
management.
The wording of these standards has been
improved substantially and now represents an appropriate division of
responsibilities – in particular, the term “framework” describes the
role as we have implemented it. With this change, we now feel that this
is one area in which we are already well positioned.
S 6. Line of business management is
responsible for the day-to-day management of operational risk within
each business unit.
S 7. Line of business management must
ensure that internal controls and practices within their line of
business are consistent with firm-wide policies and procedures to
support the management and measurement of the institution’s operational
risk.
Again, we support the division of
responsibilities as being a suitable basis on which to organise the
operational risk management function.
S 8. The institution must have
policies and procedures that clearly describe the major elements of the
operational risk management framework, including identifying, measuring,
monitoring, and controlling operational risk.
We have no fundamental disagreement with
any of the aspects of the operational risk management framework that is
listed under this standard. We would only comment that if the external
data comes from a consortium comprising a fairly small number of banks,
it is quite probable that this external loss data may well not include
any large potential events. Useful coverage of large events is more
likely if the external data comes from a database of large public loss
events. The two types of external data have rather different uses.
S 9. Operational risk management
reports must address both firm wide and line of business results. These
reports must summarize operational risk exposure, loss experience,
relevant business environment and internal control assessments, and must
be produced no less often than quarterly.
S 10. Operational risk reports must
also be provided periodically to senior management and the board of
directors, summarizing relevant firm-wide operational risk information.
This is work in progress. We are
confident that the result will be that we meet the required standards
for reporting.
S 11. An institution’s internal
control structure must meet or exceed minimum regulatory standards
established by the Agencies.
S 12. The institution must demonstrate
that it has appropriate internal loss event data, relevant external loss
event data, assessments of business environment and internal controls
factors, and results from scenario analysis to support its operational
risk management and measurement framework.
S 13. The institution must include the
regulatory definition of operational risk as the baseline for capturing
the elements of the AMA framework and determining its operational risk
exposure.
S 14. The institution must have clear
standards for the collection and modification of the elements of the
operational risk AMA framework.
The four elements of the AMA framework
will play a significant role in both the management and measurement of
operational. We object to the requirement that any risk measurement
system must include the use of all four elements - internal data,
relevant external data, scenario analysis and factors reflecting the
business environment and internal control systems. Certainly, each of
these elements is well worth considering as part of the management
framework, but a requirement to include all of them in the quantitative
measurement may be excessively burdensome. Consider a business that has
an internal data set that is sufficient for modeling the risk using an
allowable AMA methodology. Such a business should be permitted to
proceed without using external data. Similarly, scenario analysis might
be an appropriate way to evaluate the results of an AMA model for some
business lines, but should not be a required element in every AMA
calculation.
To reiterate, only some of these elements
may be appropriate for the measurement of the operational risk of a
given business unit, though all the elements should be considered in the
management of that operational risk.
The significant use of overrides for
internal loss data should not be required, other than to correct input
errors. However, if external data is used, then there may be many events
in the external database that are simply not relevant. Since only
relevant external events are required, this could lead to a significant
workload to decide and document exactly which events are relevant and
which are not.
S 15. The institution must have at
least five years of internal operational risk loss data captured across
all material business lines, events, product types, and geographic
locations.
Initially, less than five years worth
of data will be available at the time that the accord is scheduled to
become effective. The flexibility described in footnote 12 is essential.
S 16. The institution must be able to
map internal operational risk losses to the seven loss-event type
categories.
S 17. The institution must have a
policy that identifies when an operational risk loss becomes a loss
event and must be added to the loss event database. The policy must
provide for consistent treatment across the institution.
S 18. The institution must establish
appropriate operational risk data thresholds.
S 19. Losses that have any
characteristics of credit risk, including fraud-related credit losses,
must be treated as credit risk for regulatory capital purposes. The
institution must have a clear policy that allows for the consistent
treatment of loss event classifications (e.g., credit, market, or
operational risk) across the organization.
We opposed the specification in CP3 of a
loss data collection threshold because we believed that the threshold
should be established by line of business at a level that would be
appropriate for the quantification methodology being use there. Thus we
particularly welcome the flexibility that the Agencies have incorporated
in that we will have the ability to use different data thresholds in
different businesses. However, we are concerned that this flexibility
will not benefit our card business, for example, which can be typified
as having a large number of small losses, all similar in nature, but
which in total do represent a significant proportion of the total
operational risk losses. Although the number of losses and the size of
the losses are already captured with precision, we do not feel that
there is a need to capture the detailed information on each individual
loss event.
The implication is that the
quantification of operational risk will require modeling of individual
events, whereas in fact other models may be more suitable for certain
businesses, such as the credit card business. We request clarification
that the allowable models will not be limited to those that can be
considered to model individual events.
We do not see that the cost of capturing
comprehensive data on “near misses” in a central database will be
warranted, although it is certainly important that the business line
management to be aware of significant occurrences of this type.
We do not see adequate benefit, relative
to the costs, to justify capturing, in our operational loss database,
information data that is already being captured and capitalized as
credit or market risk. The cost of the effort to collect this data would
be a burden, yet the data would not be used to calculate economic
capital or regulatory capital requirements. The implementation of such a
process would require resources but not produce a clear benefit where
these events are already well managed, e.g., as credit risk. The
definition of the regulatory boundary between operational risk and
credit risk is a welcome clarification.
S 20. The institution must have
policies and procedures that provide for the use of external loss data
in the operational risk framework.
S 21. Management must systematically
review external data to ensure an understanding of industry experience.
We particularly welcome the fact that
external data no longer has to be used as an explicit input into our
loss data set. In some instances, we expect to use external data only as
a benchmark or perhaps as a form of scenario analysis.
S 22. The institution must have a
system to identify and assess business environment and internal control
factors.
S 23. Management must periodically
compare the results of their business environment and internal control
factor assessments against actual operational risk loss experience.
S 24. Management must have policies
and procedures that identify how scenario analysis will be incorporated
into the operational risk framework.
Again, we do not believe that there is
always a necessity to incorporate scenario analysis into the measurement
of operational risk regulatory capital. In some instances, scenario
analysis is more appropriately used in the management of operational
risk, for example to investigate whether the response to certain
scenarios would be appropriate. We understand that, by using the term
“framework” in this standard, such a use would be acceptable to ensure
compliance with this standard.
S 25. The institution must have a
comprehensive operational risk analytical framework that provides an
estimate of the institution’s operational risk exposure, which is the
aggregate operational loss that it faces over a one-year period at a
soundness standard consistent with a 99.9 per cent confidence level.
S 26. Management must document the
rationale for all assumptions underpinning its chosen analytical
framework, including the choice of inputs, distributional assumptions,
and the weighting across qualitative and quantitative elements.
Management must also document and justify any subsequent changes to
these assumptions.
S 27. The institution’s operational
risk analytical framework must use a combination of internal operational
loss event data, relevant external operational loss event data, business
environment and internal control factor assessments, and scenario
analysis. The institution must combine these elements in a manner that
most effectively enables it to quantify its operational risk exposure.
The institution can choose the analytical framework that is most
appropriate to its business model.
S 28. The institution’s capital
requirement for operational risk will be the sum of expected and
unexpected losses unless the institution can demonstrate, consistent
with supervisory standards, the expected loss offset.
It should be recognized that direct
calculation of specific risk results at a 99.9% confidence level will
not be possible for most business lines, given the available data. Any
such calculation will be subject to significant errors. We request
clarification that the regulatory standards will reflect the practical
necessity to generate results at lower confidence levels which can then
be scaled to a higher target confidence level using an estimated scaling
variable.
We very much doubt that the comparison of
the exposure estimate with actual loss experience will enable us to
prove that that the outputs are reasonable. The model is intended to
produce a figure that could occur once every thousand years.
Statistically speaking, it is unlikely that a few years or even a few
decades will be sufficient time to make such a validation, so judgment
will need to be employed in the process for approval of the AMA model.
The inclusion of Expected Losses in the
capital requirements will result in punitive capital requirements in
higher Expected Loss businesses such as credit cards and some consumer
lending, without taking into account the fact that such businesses have
fairly stable losses and therefore are less volatile. The same
fundamental issues apply to a broader set of businesses in the context
of Operational Risk where Expected Losses are routinely built into
pricing. The document states that an institution will not be permitted
to recognize EL offsets on budgeted loss contingencies that fall below
the established data thresholds, and that this is relevant as many
institutions currently budget for low severity, high frequency events
that are more likely to fall below most institutions’ thresholds.
Indeed, this is exactly the case for some of our consumer businesses,
where individual losses are small and below the threshold, yet gross
losses are high and fairly stable and covered by future margin income.
We strongly oppose this guidance. We regard it as critically important
that such expected losses be recognized, and that we are not required to
cover such losses twice, once through reserves or pricing, and once
through capital and that we are not required to capture details
individually about these small losses.
S 29. Management must document how its
chosen analytical framework accounts for dependence (e.g., correlations)
among operational losses across and within business lines. The
institution must demonstrate that its explicit and embedded dependence
assumptions are appropriate, and where dependence assumptions are
uncertain, the institution must use conservative estimates.
Diversification does reduce overall risk
levels and Citigroup believes that the AMA must include the opportunity
to capture the risk-reducing benefits of diversification and
efficiencies of scale. Although correlation of operational risks is
certainly less than perfect, empirical data to demonstrate this
mathematically will always remain scarce. Therefore, we welcome the new
language in this standard and trust that we can demonstrate
appropriateness without having to demonstrate validity.
However, this does raise the difficult
issue of diversification. If we have a number of legal entities, each of
which has to have sufficient capital to cover losses at the 99.9 %
confidence level, then the total corporation will be carrying capital
sufficient to cover losses at an excessively high confidence level. We
see that this could be a sufficiently large problem to impede the use of
the AMA altogether. Subsidiary legal vehicles might not warrant the
complexity of an AMA, and there might be no point in having an AMA at
the group level if the capital requirement at that level is simply the
sum of the capital requirements at the lowest level. A solution that
addresses the issue of diversification is required.
S 30. Institutions may reduce their
operational risk exposure results by no more than 20% to reflect the
impact of risk mitigants. Institutions must demonstrate that mitigation
products are sufficiently capital-like to warrant inclusion in the
adjustment to the operational risk exposure.
In principle, we object to floors and
caps and welcome their elimination over time, including the 20% limit on
insurance-related capital benefits. The recognition of risk mitigation
is welcome, but should be expanded beyond insurance in due course, as we
believe is implied in this ANPR. However, we favor an initial increase
in the amount of the cap above 20%, followed by its eventual
elimination.
It is not sound from an economic
perspective to deny both the benefits of using a captive insurance
company and the consolidation of their capital. If the risk has to be
passed through the captive insurer, then the capital of that insurer
should be recognized. The approach should be changed so that the capital
in the captive is recognized as available to cover firm risks. The
current draft denies most of the benefits of using a captive insurer,
while on the other hand it restricts the recognition of the capital held
in that insurer.
S 31. Institutions using the AMA
approach for regulatory capital purposes must use advanced data
management practices to produce credible and reliable operational risk
estimates.
S 32. The institution must test and
verify the accuracy and appropriateness of the operational risk
framework and results.
S 33. Testing and verification must be
done independently of the firm-wide operational risk management function
and the institution’s lines of business.
This again raises the question of exactly
what is meant by independence, which was discussed earlier.
Comment 2
Document 3: Draft
supervisory guidance on Internal Ratings –Based Systems for Corporate
Credit
In general, we find the Draft Guidance to
be highly prescriptive for the corporate credit rating systems of
Advanced Banks. These prescriptions could lead to “less-than-best
practice” rating systems, multiple ratings systems, onerous processes
and in some cases, may introduce systemic risk into the banking system.
At times, the Draft Guidance appeared to be written with extreme focus
on each section but with minimal appreciation of how all of the sections
would work together. In addition, some of the key points appear to be
drawn from evidence based on bond defaults, which can vary significantly
from outcomes in the loan segment. There are other indications that the
guidelines are meant to apply mainly to banks that operate only within
North America and/or Europe, where rating agency data is more relevant,
where external benchmarks are available and where single business cycles
can be applied. These conditions do not apply to a global bank such as
Citigroup, which operates in over 100 countries.
• Best Practice vs. Conservatism:
Although one of the stated requirements is that the “(r)atings used for
regulatory capital must be the same ratings used to guide day-to-day
credit risk management activities”, the Guidance simultaneously states
"Parameter estimates must incorporate a degree of conservatism that is
appropriate for the overall robustness of the quantification process"
and “the bank must adjust estimates conservatively in the presence of
uncertainty or potential error”. We could not find any delineation of
how a bank is to square the standard of adhering to internal credit risk
management with the proscriptive rules on “conservatism”. Clearly, any
type of modeling of credit risk involves a degree of uncertainty, given
the relative rarity of default. Adjusting all the parameters
conservatively, as well as following the prescriptions listed below
will result in overly conservative ratings, rather than best estimates
of the risk, affecting our ability to compete in the marketplace (where
we compete against many different intermediaries, many of whom do not
fall under these regulations):
o The prohibition against the use of
joint default probabilities despite recognition of the favorable
risk-mitigation effect.
o The prohibition against implied
support or verbal assurances, even in the presence of supporting
empirical evidence.
o The prohibition against LGDs of zero.
Our empirical studies indicate that LGDs of zero are relatively
frequent and, in some cases we actually have found negative LGDs. For
instance, trade loans guaranteed by the Exim Bank, where the guarantee
covers any interest drag during the 6-month filing period.
o The required reliance on stressed PDs.
As such, the risk measures move away from the most probable estimates
of individual obligor defaults toward the worst case scenarios, no
longer producing a good measure of expected loss of an obligor or of
the economic risk for a global portfolio. A measure of economic
capital for corporate credit risk that was based on stressed PDs for
all obligors in all the industries and countries around the world we
operate in would materially exaggerate our risks. With regard to the
PDs, the ANPR asserts that ratings must “take into account possible
adverse events that might increase an obligor’s likelihood of
default.” There is little guidance as to what is appropriate within
the “possible adverse events” schema.
o Required reliance on stressed LGDs,
in addition to stressed PDs: the ANPR states that loss severity
ratings must “reflect losses expected during periods with a relatively
high number of defaults”. Although research based on bond default and
recovery rates have shown a positive correlation between the total
number of bond defaults in the economy within a year and the average
LGD, such a relationship has not been established for loans – at least
based on our own internal work (more on the reliance on bond data
further on). Historically, there has been a material difference
between how our bank has typically managed corporate loans after
default and how defaulted bonds are treated in the market.
• Reliance on Agency Processes and
Vendor Models
o The guidance indicates a regulatory
preference for agency practices or vendors over that of banks. The
multiple requirements to map, validate and define rating practices
using external ratings as benchmarks is troubling for several reasons:
• Lack of clear ratings definitions
and transparent processes at the agencies or vendors. It is unclear
what validation standards are to be applied to the agencies and
vendors that are consistent with requirements on the internal
ratings processes of banks. In our own research, we have found
agency ratings to be inconsistent across industries, for instance,
in terms of implied default rates. The published studies from the
agencies lack that level of granularity. Similarly, the output from
some of our validated internal models varies considerably from some
vendor models.
• Rating agencies have focused on the
bond markets, not on loans. For instance, the studies cited
regarding the correlation on defaults and losses are generally based
on an analysis of bond defaults and losses.
• The focus and experience of rating
agencies are largely limited to North America and Europe. The
empirical data on ratings and recovery are heavily weighted toward
these two markets. Rating agencies have limited experience and data
in many markets we operate in.
o The guidance implies a reliance on
the agencies and other “third parties” for validating ratings
processes without providing the standards that to which the third
parties will be held. For instance, it is unclear how the supervisors
would view a rating process where the conceptual practices are sound
and the validation against defaults, for instance, proves quite
compelling but the comparison to external ratings produces divergent
outcomes.
o The guidance places considerable
importance on benchmarking, often to external agencies, however
ratings vary for many reasons and, except against actual default/loss
events, it is near impossible to determine what an individual rating
should be. Indeed, one supervisory standard speaks about a bank
adopting and defending a ratings philosophy, but the Guidance gives
overly broad definitions of two different philosophies
("through-the-cycle" and "point-in-time"). Later, though, the Guidance
states "The ratings agencies are commonly believed to use
through-the-cycle rating approaches." As such, requiring a convergence
to agency ratings or any other external benchmark may introduce a
higher degree of systemic risk.
o The guidance also states “banks will
eventually be expected to use variables that are widely recognized as
the most reliable predictors of default risk in mapping exercises”.
Who is the arbiter of “most reliable predictors” and how is “most
reliable” determined? This would seem to go very much against the
premise that credit risk analysis is evolving and the availability of
data will allow enhancements to the current state of credit risk
analysis. For instance, the guidance cites that borrower size is
predictive but less so than leverage and cash flow, without citing the
source of that statement. This citation seems extremely broad and
sweeping. We have developed a large number of rating models for
different industries and global regions based on and validated by the
empirical data from that industry/region. In some of our internal
modeling efforts, size proved to be the most significant determinant
of credit quality. And yet, in other internal modeling efforts, we
found the key determinants of credit quality and their relative
importance to vary from industry to industry, market to market. We
think it is inappropriate and naïve for the ANPR to assume that the
relative importance of the particular input variables needed to
estimate an obligor’s PD is universal for all corporations, in all
industries and countries of the world. Finally, this emphasis on use
of widely recognized variables may introduce systematic risk into the
banking system. To wit: if all banks use widely-recognized variables
in their models, then banks may all move in tandem in and out of
markets, heightening volatility and potentially damaging whole sectors
of the economy.
• Practical Issues for Global Banks
There are many processes included in the
guidance, which are quite burdensome, without appearing to add
significant value.
• Re-estimating or validating the
model/process risk parameters on an annual basis: Except for the
largest markets, such as the United States, there will rarely be
sufficient new defaults and resolved defaults (for LGD/EAD) to justify
the re-estimation and reprogramming, testing, training and
distribution of models and processes used to assign ratings. Even
within the United States, there are generally few defaults within a
particular industry/geography segment of a portfolio on an annual
basis, much less a quarterly basis, except in the SME. On average,
defaults take more than a year to resolve, so even during periods of
higher-than-average defaults, the additional information would not
seem to justify the required changes. Constant turnover in rating
methodologies and processes could jeopardize the quality of the
ratings. In addition, given that the Guidance states a minimum of five
years of reference data, which must include periods of economic stress
to estimate PDs, re-estimating or validating year after year during an
economic expansion may dilute the stressed periods' data and weaken a
model's ability to accurately assess risks when a downturn occurs.
• Re-rating the portfolio for each
change to the rating process: In addition, the process becomes
even more onerous with the requirement to re-rate the entire portfolio
every time a rating process changes. It is possible to re-rate a
portfolio to the degree that all inputs into a rating are
quantifiable. However, a system that relies both on models and
expertise is difficult to replicate. The risk manager may not apply
the same adjustment to a rating once the model changes.
• Gauging impact of changes in
actual economic circumstances on PDs and LGDs: This guidance may
seem relatively straightforward for a bank operating in a single or
relatively few markets. However, in building models or establishing
LGDs that cover multiple countries, economic cycles often diverge or
are not easily identifiable.
• Calibrate models to fit customer
base: We request more information on this requirement. This
requirement could become very burdensome for a large global bank where
the portfolios change rapidly. Our goal is to build models, for
instance, that are appropriate for rating across the credit spectrum,
even if all of the current customer base are rated investment grade.
Models built on one industry or geography can be tested for ratings
accuracy on other customers, assuming that the concepts are sound for
the other business. If for instance, a rating model is built on
corporates in Eastern Europe and due to an acquisition in Poland,
companies from that country become more prevalent in the portfolio, is
the existing model invalid? If the acquired bank did not have 5 years
of data on the customers, should the model be recalibrated?
• Comparability of reference data to
current credit portfolio: As above, more clarification is
requested, as well as some examples of how to establish that
comparability.
• Potential Erosion of Comparative
Advantage: We need to understand the exact nature of what would be
disclosed in "Summaries of (trends developed from obligor and facility
risk rating data)…included…public disclosures."
• Appropriate Control and Oversight
Mechanisms. In contrast to market risk, the guidance requires an
additional level of internal review of ratings and all ratings
processes, models, and data aside from Audit and regulatory oversight.
For banks where the responsibilities are already distributed across
independent risk functions, this additional level of review is
superfluous, expensive and bureaucratic.
• While we believe in independent
oversight, the prescripts found in the Control and Oversight sections
is inconsistent, impractical, and contradicts current best practices
in the industry, and as outlined by the Agency’s own on-site
supervisory staff. In fact, the guidance is internally inconsistent
and arbitrary—for example section 213 describes ‘flexibility’ while
table 4.1 mandates the creation of a ‘ratings system review’ area. We
ask the Agencies to reconsider their approach in its entirety.
• We are concerned that the Agencies,
in spite of their good intentions, are shockingly naïve in their
underlying premise regarding how a bank typically operates
effectively. In sections 215-216 and 220, two very different
institutions are described (one with independent model-based ratings
development groups vs. one without), but in BOTH circumstances a
separate ratings review function is necessary. In the latter case, we
agree. In the former case, we completely disagree; the prescribed role
is typically performed by the ‘loan review group’ that is often found
in Internal Audit. In this situation, there are lending officers that
have only modest discretion to adjust ratings, there is an independent
ratings-setting group that reports to independent risk management, and
there is Internal Audit/Loan Review.
• For banking institutions that
primarily use models to assign ratings, only two independent
organizational units are necessary to create the ‘checks and
balances’—an Independent Ratings Group and Internal Audit’s Loan
Review Group. A ‘Ratings Review Group’ as described in Table 4.2 in
the text is completely superfluous. Specifically, all responsibilities
are already accounted for in an organization such as ours:
Responsibility—per Table
4.2
|
Group Responsible
|
Design of ratings systems |
Independent Ratings Group |
Compliance with policies |
Audit/Loan Review |
Check risk rating grades |
Audit/Loan Review |
Consistency across industries |
Indep Ratings Group and Audit/Loan
Review |
Model development |
Audit/Loan Review |
Model use |
Audit/Loan Review |
Overrides and policy exceptions |
Audit/Loan Review |
Quantification process |
Independent Ratings Group |
Back testing |
Audit/Loan Review |
Actual and predicted ratings trans. |
[meaning is unclear] |
Benchmarking |
Independent Ratings Group |
Adequacy of data maintenance |
Audit/Loan Review |
Identify errors and
flaws |
[meaning is
unclear] |
Recommend corrective
actions |
Audit/Loan Review
|
• The agencies fail to realize that when a Ratings Group reports to
independent risk management, they have no incentive to sacrifice ratings
accuracy for sales and marketing purposes. In fact, just the opposite is
true: the ethos that develops in such a group is one of economic logic,
empirical facts, thoughtful modeling, which is the best ‘check and
balance’ against the influence of sales and marketing. And then
Audit/Loan Review creates even more independence.
• The Ratings system oversight suggested
by the guidance is impractical and obfuscates the role of management.
The risk-rating system can best be understood by practitioners, and not
by directors who will be unable to understand the data and detail
inherent in this task.
|