Skip to main content
U.S. flag
An official website of the United States government
Dot gov
The .gov means it’s official. 
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.
Https
The site is secure. 
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.
Federal Register Publications

FDIC Federal Register Citations



Home > Regulation & Examinations > Laws & Regulations > FDIC Federal Register Citations




FDIC Federal Register Citations

[Federal Register: August 4, 2003 (Volume 68, Number 149)]

[Notices]

[Page 45949-45988]

From the Federal Register Online via GPO Access [wais.access.gpo.gov]

[DOCID:fr04au03-137]

-----------------------------------------------------------------------

DEPARTMENT OF THE TREASURY

Office of the Comptroller of the Currency

[Docket No. 03-15]

FEDERAL RESERVE SYSTEM

[Docket No. OP-1153]

FEDERAL DEPOSIT INSURANCE CORPORATION

DEPARTMENT OF THE TREASURY

Office of Thrift Supervision

[No. 2003-28]

Internal Ratings-Based Systems for Corporate Credit and

Operational Risk Advanced Measurement Approaches for Regulatory Capital

AGENCIES: Office of the Comptroller of the Currency (OCC), Treasury;

Board of Governors of the Federal Reserve System (Board); Federal

Deposit Insurance Corporation (FDIC); and Office of Thrift Supervision

(OTS), Treasury.

ACTION: Draft supervisory guidance with request for comment.

-----------------------------------------------------------------------

SUMMARY: The OCC, Board, FDIC, and OTS (the Agencies) are publishing

for industry comment two documents that set forth draft supervisory

guidance for implementing proposed revisions to the risk-based capital

standards in the United States. These proposed revisions, which would

implement the New Basel Capital Accord in the United States, are

published as an advance notice of proposed rulemaking (ANPR) elsewhere

in today's Federal Register. Under the advanced approaches for credit

and operational risk described in the ANPR, banking organizations would

use internal estimates of certain risk components as key inputs in the

determination of their regulatory capital requirements. The Agencies

believe that supervisory guidance is necessary to balance the

flexibility inherent in the advanced approaches with high standards

that promote safety and soundness and encourage comparability across

institutions.

The first document sets forth Draft Supervisory Guidance on

Internal Ratings-Based Systems for Corporate Credit (corporate IRB

guidance). This document describes supervisory expectations for

institutions that intend to adopt the advanced internal ratings-based

approach (A-IRB) for credit risk as set forth in today's ANPR. The

corporate IRB guidance is intended to provide supervisors and

institutions with a clear description of the essential components and

characteristics of an acceptable A-IRB framework. The guidance focuses

specifically on corporate credit portfolios; further guidance is

expected at a later date on other credit portfolios (including, for

example, retail and commercial real estate portfolios).

The second document sets forth Draft Supervisory Guidance on

Operational Risk Advanced Measurement Approaches for Operational Risk

(AMA guidance). This document outlines supervisory expectations for

institutions that intend to adopt an advanced measurement approach

(AMA) for operational risk as set forth in today's ANPR.

The Agencies are seeking comments on the supervisory standards set

forth in both documents. In addition to seeking comment on specific

aspects of the supervisory guidance set forth in the documents, the

Agencies are seeking comment on the extent to which the supervisory

guidance strikes the appropriate balance between flexibility and

specificity. Likewise, the Agencies are seeking comment on whether an

appropriate balance has been struck between the regulatory requirements

set forth in the ANPR and the supervisory standards set forth in these

documents.

DATES: Comments must be received no later than November 3, 2003.

ADDRESSES: Comments should be directed to:

OCC: Please direct your comments to: Office of the Comptroller of

the Currency, 250 E Street, SW., Public Information Room, Mailstop 1-5,

Washington, DC 20219, Attention: Docket No. 03-15; fax number (202)

874-4448; or Internet address: regs.comments@occ.treas.gov. Due to

delays in paper mail delivery in the Washington area, we encourage the

submission of comments by fax or e-mail whenever possible. Comments may

be inspected and photocopied at the OCC's Public Information Room, 250

E Street, SW., Washington, DC. You may make an appointment to inspect

comments by calling (202) 874-5043.

Board: Comments should refer to Docket No. OP-1153 and may be

mailed to Ms. Jennifer J. Johnson, Secretary, Board of Governors of the

Federal Reserve System, 20th Street and Constitution Avenue, NW.,

Washington, DC, 20551. However, because paper mail in the Washington

area and at the Board of Governors is subject to delay, please consider

submitting your comments by e-mail to regs.comments@federalreserve.gov,

or faxing them to the Office of the Secretary at 202/452-3819 or 202/

452-3102. Members of the public may inspect comments in Room MP-500 of

the Martin Building between 9 a.m. and 5 p.m. on weekdays pursuant to

Sec. 261.12, except as provided in Sec. 261.14, of the Board's Rules

Regarding Availability of Information, 12 CFR 261.12 and 261.14.

FDIC: Written comments should be addressed to Robert E. Feldman,

Executive Secretary, Attention: Comments, Federal Deposit Insurance

Corporation, 550 17th Street, NW., Washington, DC, 20429. Commenters

are encouraged to submit comments by facsimile transmission to (202)

898-3838 or by electronic mail to Comments@FDIC.gov. Comments also may

be hand-delivered to the guard station at the rear of the 550 17th

Street Building (located on F Street) on business days between 8:30

a.m. and 5 p.m. Comments may be inspected and photocopied at the FDIC's

Public Information Center, Room 100, 801 17th Street, NW., Washington,

DC between 9 a.m. and 4:30 p.m. on business days.

OTS: Send comments to Regulation Comments, Chief Counsel's Office,

Office of Thrift Supervision, 1700 G Street, NW., Washington, DC 20552,

Attention: No. 2003-28. Delivery: Hand deliver comments to the Guard's

desk, east lobby entrance, 1700 G Street, NW., from 9 a.m. to 4 p.m. on

business days, Attention: Regulation Comments, Chief Counsel's Office,

Attention: No. 2003-28. Facsimiles: Send facsimile transmissions to FAX

Number (202) 906-6518, Attention: No 2003-28. e-mail: Send e-mails to

regs.comments@ots.treas.gov, Attention: No. 2003-28, and include your

name and telephone number. Due to temporary disruptions in mail service

in the Washington, DC area, commenters are encouraged to send comments

by fax or e-mail, if possible.

FOR FURTHER INFORMATION CONTACT:

OCC: Corporate IRB guidance: Jim Vesely, National Bank Examiner,

Large Bank Supervision (202/874-5170 or james.vesely@occ.treas.gov);

AMA guidance: Tanya Smith, Senior International Advisor, International

Banking & Finance (202/874-4735 or tanya.smith@occ.treas.gov).

Board: Corporate IRB guidance: David Palmer, Supervisory Financial

Analyst, Division of Banking Supervision and Regulation (202/452-2904

or david.e.palmer@frb.gov); AMA guidance: T. Kirk Odegard, Supervisory

Financial Analyst, Division of Banking Supervision and Regulation (202/

530-6225 or thomas.k.odegard@frb.gov). For users of Telecommunications

Device for

[[Page 45950]]

the Deaf (``TDD'') only, contact 202/263-4869.

FDIC: Corporate IRB guidance and AMA guidance: Pete D. Hirsch,

Basel Project Manager, Division of Supervision and Consumer Protection

(202/898-6751 or phirsch@fdic.gov).

OTS: Corporate IRB guidance and AMA guidance: Michael D. Solomon,

Senior Program Manager for Capital Policy (202/906-5654); David W.

Riley, Project Manager (202/906-6669), Supervision Policy; Teresa A.

Scott, Counsel (Banking and Finance) (202/906-6478); or Eric

Hirschhorn, Principal Financial Economist (202/906-7350), Regulations

and Legislation Division, Office of the Chief Counsel, Office of Thrift

Supervision, 1700 G Street, NW., Washington, DC 20552.

Document 1: Draft Supervisory Guidance on Internal Ratings-Based

Systems for Corporate Credit

Table of Contents

I. Introduction

A. Purpose

B. Overview of Supervisory Expectations

1. Ratings Assignment

2. Quantification

3. Data Maintenance

4. Control and Oversight Mechanisms

C. Scope of Guidance

D. Timing

II. Ratings for IRB Systems

A. Overview

B. Credit Ratings

1. Rating Assignment Techniques

a. Expert Judgment

b. Models

c. Constrained Judgment

C. IRB Ratings System Architecture

1. Two-Dimensional Rating System

a. Definition of Default

b. Obligor Ratings

c. Loss Severity Ratings

2. Other Considerations of IRB Rating System Architecture

a. Timeliness of Ratings

b. Multiple Ratings Systems

c. Recognition of the Risk Mitigation Benefits of Guarantees

3. Validation Process

a. Ratings System Developmental Evidence

b. Ratings System Ongoing Validation

c. Back Testing

III. Quantification of IRB Systems

A. Introduction

1. Stages of the Quantification Process

2. General Principles for Sound IRB Quantification

B. Probability of Default (PD)

1. Data

2. Estimation

3. Mapping

4. Application

C. Loss Given Default (LGD)

1. Data

2. Estimation

3. Mapping

4. Application

D. Exposure at Default (EAD)

1. Data

2. Estimation

3. Mapping

4. Application

E. Maturity (M)

F. Validation

Appendix to Part III: Illustrations of the Quantification Process

IV. Data Maintenance

A. Overview

B. Data Maintenance Framework

1. Life Cycle Tracking

2. Rating Assignment Data

3. Example Data Elements

C. Data Element Functions

1. Validation and Refinement

2. Developing Parameter Estimates

3. Applying Rating System Improvements Historically

4. Calculating Capital Ratios and Reporting to the Public

5. Supporting Risk Management

D. Managing data quality and integrity

1. Documentation and Definitions

2. Electronic Storage

3. Data Gaps

V. Control and Oversight Mechanisms

A. Overview

B. Independence in the Rating Approval Process

C. Transparency

D. Accountability

1. Responsibility for Assigning Ratings

2. Responsibility for Rating System Performance

E. Use of Ratings

F. Rating System Review (RSR)

G. Internal Audit

1. External Audit

H. Corporate Oversight

I. Introduction

A. Purpose

This document describes supervisory expectations for banking

organizations (institutions) adopting the advanced internal ratings-

based approach (IRB) for the determination of minimum regulatory risk-

based capital requirements. The focus of this guidance is corporate

credit portfolios. Retail, commercial real estate, securitizations, and

other portfolios will be the focus of later guidance. This draft

guidance should be considered with the advance notice of proposed

rulemaking (ANPR) on revisions to the risk-based capital standard

published elsewhere in today's Federal Register.

The primary objective of IRB is to enhance the sensitivity of

regulatory capital requirements to credit risk. To accomplish that

objective, IRB harnesses a bank's own risk rating and quantification

capabilities. In general, the IRB approach reflects and extends recent

developments in risk management and banking supervision. However, the

degree to which any individual bank will need to modify its own credit

risk management practices to deliver accurate and consistent IRB risk

parameters will vary from institution to institution.

This guidance is intended to provide supervisors and institutions

with a clear description of the essential components and

characteristics of an acceptable IRB framework. Toward that end, this

document sets forth IRB system supervisory standards that are

highlighted in bold and designated by the prefix ``S.'' Whenever

possible, these supervisory standards are principle-based to enable

institutions to implement the framework flexibly. However, when

prudential concerns or the need for standardization override the desire

for flexibility, the supervisory standards are more detailed.

Ultimately, institutions must have credit risk management practices

that are consistent with the substance and spirit of the standards in

this guidance.

The IRB conceptual framework outlined in this document is intended

neither to dictate the precise manner by which institutions should seek

to meet supervisory expectations, nor to provide technical guidance on

how to develop such a framework. As institutions develop their IRB

systems in anticipation of adopting them for regulatory capital

purposes, supervisors will be evaluating, on an individual bank basis,

the extent to which institutions meet the standards outlined in this

document. In evaluating institutions, supervisors will rely on this

supervisory guidance as well as examination procedures, which will be

developed separately. This document assumes that readers are familiar

with the proposed IRB approach to calculating minimum regulatory

capital articulated in the ANPR.

B. Overview of Supervisory Expectations

Rigorous credit risk measurement is a necessary element of advanced

risk management. Qualifying institutions will use their internal rating

systems to associate a probability of default (PD) with each obligor

grade, as well as a loss given default (LGD) with each credit facility.

In addition, institutions will estimate exposure at default (EAD) and

will calculate the effective remaining maturity (M) of credit

facilities.

Qualifying institutions will be expected to have an IRB system

consisting of four interdependent components:

[sbull] A system that assigns ratings and validates their accuracy

(Chapter 1),

[sbull] A quantification process that translates risk ratings into

IRB parameters (Chapter 2),

[sbull] A data maintenance system that supports the IRB system

(Chapter 3), and,

[[Page 45951]]

[sbull] Oversight and control mechanisms that ensure the system is

functioning as intended and producing accurate ratings (Chapter 4).

Together these rating, quantification, data, and oversight

mechanisms present a framework for defining and improving the

evaluation of credit risk.

It is expected that rating systems will operate dynamically. As

ratings are assigned, quantified and used, estimates will be compared

with actual results and data will be maintained and updated to support

oversight and validation efforts and to better inform future estimates.

The rating system review and internal audit functions will serve as

control mechanisms that ensure that the process of ratings assignment

and quantification function according to policy and design and that

noncompliance and weaknesses are identified, communicated to senior

management and the board, and addressed. Rating systems with

appropriate data and oversight feedback mechanisms foster a learning

environment that promotes integrity in the rating system and continuing

refinement.

IRB systems need the support and oversight of the board and senior

management to ensure that the various components fit together

seamlessly and that incentives to make the system rigorous extend

across line, risk management, and other control groups. Without strong

board and senior management support and involvement, rating systems are

unlikely to provide accurate and consistent risk estimates during both

good and bad times.

The new regulatory minimum capital requirement is predicated on an

institution's internal systems being sufficiently advanced to allow a

full and accurate assessment of its risk exposures. Under the new

framework, an institution could experience a considerable capital

shortfall in the most difficult of times if its risk estimates are

materially understated. Consequently, the IRB framework demands a

greater level of validation work and controls than supervisors have

required in the past. When properly implemented, the new framework

holds the potential for better aligning minimum capital requirements

with the risk taken, pushing capital requirements higher for

institutions that specialize in riskier types of lending, and lower for

those that specialize in safer risk exposures.

Supervisors will evaluate compliance with the supervisory standards

for each of the four components of an IRB system. However, evaluating

compliance with each of the standards individually will not be

sufficient to determine an institution's overall compliance. Rather,

supervisors and institutions must also evaluate how well the various

components of an institution's IRB system complement and reinforce one

another to achieve the overall objective of accurate measures of risk.

In performing their evaluation, supervisors will need to exercise

considerable supervisory judgment, both in evaluating the individual

components and the overall IRB framework. A summary of the key

supervisory expectations for each of the IRB components follows.

Ratings Assignment

The first component of an IRB system involves the assignment and

validation of ratings (see Chapter 1). Ratings must be accurately and

consistently applied to all corporate credit exposures and be subject

to initial and ongoing validation. Institutions will have latitude in

designing and operating IRB rating systems subject to five broad

standards:

Two-dimensional risk-rating system--IRB institutions must be able

to make meaningful and consistent differentiations among credit

exposures along two dimensions--obligor default risk and loss severity

in the event of a default.

Rank order risks--IRB institutions must rank obligors by their

likelihood of default, and facilities by the loss severity expected in

default.

Calibration--IRB obligor ratings must be calibrated to values of

the probability of default (PD) parameter and loss severity ratings

must be calibrated to values of the loss given default (LGD) parameter.

Accuracy--Actual long-run actual default frequencies for obligor

rating grades must closely approximate the PDs assigned to those grades

and realized loss rates on loss severity grades must closely

approximate the LGDs assigned to those grades.

Validation process--IRB institutions must have ongoing validation

processes for rating systems that include the evaluation of

developmental evidence, process verification, benchmarking, and the

comparison of predicted parameter values to actual outcomes (back-

testing).

Quantification

The second component of an IRB system is a quantification process

(see Chapter 2). Since obligor and facility ratings may be assigned

separately from the quantification of the associated PD and LGD

parameters, quantification is addressed as a separate process. The

quantification process must produce values not only for PD and LGD but

also for EAD and for the effective remaining maturity (M). The

quantification of those four parameters is expected to be the result of

a disciplined process. The key considerations for effective

quantification are as follows:

Process--IRB institutions must have a fully specified process

covering all aspects of quantification (reference data, estimation,

mapping, and application).

Documentation--The quantification process, including the role and

scope of expert judgment, must be fully documented and updated

periodically.

Updating--Parameter estimates and related documentation must be

updated regularly.

Review--A bank must subject all aspects of the quantification

process, including design and implementation, to an appropriate degree

of independent review and validation.

Constraints on Judgment--Judgmental adjustments may be an

appropriate part of the quantification process, but must not be biased

toward lower risk estimates.

Conservatism--Parameter estimates must incorporate a degree of

conservatism that is appropriate for the overall robustness of the

quantification process.

Data Maintenance

The third component of an IRB system is an advanced data management

system that produces credible and reliable risk estimates (see Chapter

3). The broad standard governing an IRB data maintenance system is that

it supports the requirements for the other IRB system components, as

well as the institution's broader risk management and reporting needs.

Institutions will have latitude in managing their data, subject to the

following key data maintenance standards:

Life Cycle Tracking--Institutions must collect, maintain, and

analyze essential data for obligors and facilities throughout the life

and disposition of the credit exposure.

Rating Assignment Data--Institutions must capture all significant

quantitative and qualitative factors used to assign the obligor and

loss severity rating.

Support of IRB System--Data collected by institutions must be of

sufficient depth, scope, and reliability to:

[sbull] Validate IRB system processes,

[sbull] Validate parameters,

[sbull] Refine the IRB system,

[sbull] Develop internal parameter estimates,

[sbull] Apply improvements historically,

[sbull] Calculate capital ratios,

[sbull] Produce internal and public reports, and

[[Page 45952]]

[sbull] Support risk management.

Control and Oversight Mechanisms

The fourth component of an IRB system is comprised of control and

oversight mechanisms that ensure that the various components of the IRB

system are functioning as intended (see Chapter 4). Given the various

uses of internal risk ratings, including their direct link to

regulatory capital requirements, there is enormous, sometimes

conflicting, pressure on banks' internal rating systems. Control

structures are subject to the following broad standards:

Interdependent System of Controls--IRB institutions must implement

a system of interdependent controls that include the following

elements:

[sbull] Independence,

[sbull] Transparency,

[sbull] Accountability,

[sbull] Use of ratings,

[sbull] Rating system review,

[sbull] Internal audit, and

[sbull] Board and senior management oversight.

Checks and Balances--Institutions must combine the various control

mechanisms in a way that provides checks and balances for ensuring IRB

system integrity.

The system of oversight and controls required for an effective IRB

system may operate in various ways within individual institutions. This

guidance does not prescribe any particular organizational structure for

IRB oversight and control mechanisms. Banks have broad latitude to

implement structures that are most effective for their individual

circumstances, as long as those structures support and enhance the

institution's ability to satisfy the supervisory standards expressed in

this document.

C. Scope of Guidance

This draft guidance reflects work performed by supervisors to

evaluate and compare current practices at institutions with the

concepts and requirements for an IRB framework. For instances in which

a range of practice was observable, examples are provided on how

certain practices may or may not qualify. However, in many other

instances, practices were at such an early stage of development that it

was not feasible to describe specific examples. In those cases,

requirements tend to be principle-based and without examples. Given

that institutions are still in the early stages of developing

qualifying IRB systems, it is expected that this guidance will evolve

over time to more explicitly take into account new and improving

practices.

D. Timing

S. An IRB system must be operating fully at least one year prior to

the institution's intended start date for the advanced approach.

As noted in the ANPR, the significant challenge of implementing a

fully complying IRB system requires that institutions and supervisors

have sufficient time to observe whether the IRB system is delivering

risk-based capital figures with a high level of integrity. The ability

to observe the institution's ratings architecture, validation, data

maintenance and control functions in a fully operating environment

prior to implementation will help identify how well the IRB system

design functions in practice. This will be particularly important given

that in the first year of implementation institutions will not only be

subject to the new minimum capital requirements, but will also be

disclosing risk-based capital ratios for the public to rely upon in the

assessment of the institution's financial health.

II. Ratings for IRB Systems

A. Overview

This chapter describes the design and operation of risk-rating

systems that will be acceptable in an internal ratings-based (IRB)

framework. Banks will have latitude in designing and operating IRB

rating systems, subject to five broad standards:

Two-dimensional risk-rating system--IRB institutions must be able

to make meaningful and consistent differentiations among credit

exposures along two dimensions--obligor default risk and loss severity

in the event of a default.

Rank order risks--IRB institutions must rank obligors by their

likelihood of default, and facilities by the loss severity expected in

default.

Calibration--IRB obligor ratings must be calibrated to values of

the probability of default (PD) parameter and loss severity ratings

must be calibrated to values of the loss given default (LGD) parameter.

Accuracy--Actual long-run actual default frequencies for obligor

rating grades must closely approximate the PDs assigned to those grades

and actual loss rates on loss severity grades must closely approximate

the LGDs assigned to those grades.

Validation process--IRB institutions must have ongoing validation

processes for rating systems that include the evaluation of

developmental evidence, process verification, benchmarking, and the

comparison of predicted parameter values to actual outcomes (back-

testing).

B. Credit Ratings

In general, a credit rating is a summary indicator of the relative

risk on a credit exposure. Credit ratings can take many forms. The most

widely known credit ratings are the public agency ratings, which are

expressed as letters; bank internal ratings tend to be expressed as

whole numbers--for example, 1 through 10. Some rating model outputs are

expressed in terms of probability of default or expected default

frequency, in which case they may be more than relative measures of

risk. Regardless of the form, meaningful credit ratings share two

characteristics:

[sbull] They group credits to discriminate among possible outcomes.

[sbull] They rank the perceived levels of credit risk.

Banks have used credit ratings of various types for a variety of

purposes. Some ratings are intended to rank obligors by risk of default

and some are intended to rank facilities\1\ by expected loss, which

incorporates risk of default and loss severity. Bank rating systems

that are geared solely to expected loss will need to be amended to meet

the two-dimensional requirements of the IRB approach.

Rating Assignment Techniques

Banks use different techniques, such as expert judgment and models,

to assign credit risk ratings. For banks using the IRB approach, how

ratings are assigned is important because different techniques will

require different validation processes and control mechanisms to ensure

the integrity of the rating system. To assist the discussion of rating

architecture requirements, described below are some of the current

rating assignment techniques. Any of these techniques--expert judgment,

models, constrained judgment, or a combination thereof--could be

acceptable within an IRB system, provided the bank meets the standards

outlined in this document.

---------------------------------------------------------------------------

\1\ Facilities--loans, lines, or other separate extensions of

credit to an obligor.

---------------------------------------------------------------------------

Expert Judgment

Historically, banks have used expert judgment to assign ratings to

commercial credits. With this technique, an individual weighs relevant

information and reaches a conclusion about the appropriate risk rating.

Presumably, the rater makes informed judgments based on knowledge

gained through experience and training.

[[Page 45953]]

The key feature of expert-judgment systems is flexibility. The

prevalence of judgmental rating systems reflects the view that the

determinants of default are too complicated to be captured by a single

quantitative model. The quality of management is often cited as an

example of a risk determinant that is difficult to assess through a

quantitative model. In order to foster internal consistency, banks

employing expert judgment rating systems typically provide narrative

guidelines that set out ratings criteria. However, the expert must

decide how narrative guidelines apply to a given set of circumstances.

The flexibility possible in the assignment of judgmental ratings

has implications for the types of ratings review that are feasible. As

part of the ratings validation process, banks will attempt to confirm

that raters follow bank policy. However, two individuals exercising

judgment can use the same information to support different ratings.

Thus, the review of an expert judgment rating system will require an

expert who can identify the impact of policy and the impact of judgment

on a rating.

Models

In recent years, models have been developed for use in rating

commercial credits. In a model-based approach, inputs are numeric and

provide quantitative and qualitative information about an obligor. The

inputs are combined using mathematical equations to produce a number

that is translated into a categorical rating. An important feature of

models is that the rating is perfectly replicable by another party,

given the same inputs.

The models used in credit rating can be distinguished by the

techniques used to develop them. Some models may rely on statistical

techniques while others rely on expert-judgment techniques.

Statistical models. Statistically developed models are the result

of statistical optimization, in which well-defined mathematical

criteria are used to choose the model that has the closest fit to the

observed data. Numerous techniques can be used to build statistical

models; regression is one widely recognized example. Regardless of the

specific statistical technique, a knowledgeable independent reviewer

will have to exercise judgment in evaluating the reasonableness of a

model's development, including its underlying logic, the techniques

used to handle the data, and the statistical model building techniques.

Expert-derived models.\2\ Several banks have built rating models by

asking their experts to decide what weights to assign to critical

variables in the models. Drawing on their experience, the experts first

identify the observable variables that affect the likelihood of

default. They then reach agreement on the weights to be assigned to

each of the variables. Unlike statistical optimization, the experts are

not necessarily using clear, consistent criteria to select the weights

attached to the variables. Indeed, expert-judgment model building is

often a practical choice when there is not enough data to support a

statistical model building. Despite its dependence on expert judgment,

this method can be called model-based as long as the result--the

equation, most likely with linear weights--is used as the basis to rate

the credits. Once the equation is set, the model shares the feature of

replicability with statistically derived models. Generally, independent

credit experts use judgment to evaluate the reasonableness of the

development of these models.

---------------------------------------------------------------------------

\2\ Some banks have developed credit rating models that they

refer to as ``scorecards,'' but they have used expert judgment to

derive the weights. While they are models, they are not scoring

models in the now conventional use of the term. In its conventional

use, the term scoring model is reserved for a rating model derived

using statistical techniques.

---------------------------------------------------------------------------

Constrained Judgment

The alternatives just described present the extremes, but in

practice, many banks use rating systems that combine models with

judgment. Two approaches are common.

Judgmental systems with quantitative guidelines or model results as

inputs. Historically, the most common approach to rating has involved

individuals exercising judgment about risks, subject to policy

guidelines containing quantitative criteria such as minimum values for

particular financial ratios. Banks develop quantitative criteria to

guide individuals in assigning ratings, but often believe that those

criteria do not adequately reflect the information needed to assign a

rating.

One version of this constrained judgment approach features a model

output as one among several criteria that an individual may consider in

assigning ratings. The individual assigning the rating is responsible

for prioritizing the criteria, reconciling conflicts between criteria,

and if warranted, overriding some criteria. Even if individuals

incorporate model results as one of the factors in their ratings, they

will exercise judgment in deciding what weight to attach to the model

result. The appeal of this approach is that the model combines many

pieces of information into a single output, which simplifies analysis,

while the rater retains flexibility regarding the use of the model

output.

Model-based ratings with judgmental overrides. When banks use

rating models, individuals are generally permitted to override the

results under certain conditions and within tolerance levels for

frequency. Credit-rating systems in which individuals can override

models raise many of the same issues presented separately by pure

judgment and model-based systems. If overrides are rare, the system can

be evaluated largely as if it is a model-based system. If, however,

overrides are prevalent, the system will be evaluated more like a

judgmental system.

Since constrained judgment systems combine features of both expert

judgment and model-based systems, their evaluation will require the

skills required to evaluate both of these other systems.

C. IRB Ratings System Architecture

Two-Dimensional Rating System

S. IRB risk rating systems must have two rating dimensions--obligor

and loss severity ratings.

S. IRB obligor and loss severity ratings must be calibrated to

values of the probability of default (PD) and the loss given default

(LGD), respectively.

Regardless of the type of rating system(s) used by an institution,

the IRB approach imposes some specific requirements. The first

requirement is that an IRB rating system must be two-dimensional. Banks

will assign obligor ratings, which will be associated with a PD. They

will also either assign a loss severity rating, which will be

associated with LGD values, or directly assign LGD values to each

facility. The process of assigning the obligor and loss severity

ratings--hereafter referred to as the rating system--is discussed

below, and the process of calibrating obligor and loss severity ratings

to PD and LGD parameters is discussed in Chapter 2.

S. Banks must record obligor defaults in accordance with the IRB

definition of default.

Definition of Default

The consistent identification of defaults is fundamental to any IRB

rating system. For IRB purposes, a default is considered to have

occurred with regard to a particular obligor when either or both of the

two following events have taken place:

[sbull] The obligor is past due more than 90 days on any material

credit

[[Page 45954]]

obligation to the banking group. Overdrafts will be considered as being

past due once the customer has breached an advised limit or been

advised of a limit smaller than current outstandings.

[sbull] The bank considers that the obligor is unlikely to pay its

credit obligations to the banking group in full, without recourse by

the bank to actions such as liquidating collateral (if held).

Any obligor (or its underlying credit facilities) that meets one or

more of the following conditions is considered unlikely to pay and

therefore in default:

[sbull] The bank puts the credit obligation on non-accrual status.

[sbull] The bank makes a charge-off or account-specific provision

resulting from a significant perceived decline in credit quality

subsequent to the bank taking on the exposure.

[sbull] The bank sells the credit obligation at a material credit-

related economic loss.

[sbull] The bank consents to a distressed restructuring of the

credit obligation where this is likely to result in a diminished

financial obligation caused by the material forgiveness, or

postponement, of principal, interest or (where relevant) fees.

[sbull] The bank has filed for the obligor's bankruptcy or a

similar order in respect of the obligor's credit obligation to the

banking group.

[sbull] The obligor has sought or has been placed in bankruptcy or

similar protection where this would avoid or delay repayment of the

credit obligation to the banking group.

While most conditions of default currently are identified by bank

reporting systems, institutions will need to augment data capture

systems to collect those default circumstances that may not have been

traditionally identified. These include facilities that are current and

still accruing but where the obligor declared or was placed in

bankruptcy. They must also capture so called ``silent defaults''--

defaults when the loss on a facility was avoided by liquidating

collateral.

Loan sales on which a bank experiences a material loss due to

credit deterioration are considered a default. Material credit related

losses are defined as XX. (The agencies seek comment on how to define

``material'' loss in the case of loans sold at a discount). Banks

should ensure that they have adequate systems to identify such

transactions and to maintain adequate records so that reviewers can

assess the adequacy of the institution's decision-making process in

this area.

Obligor Ratings

S. Banks must assign discrete obligor grades.

While banks may use models to estimate probabilities of default for

individual obligors, the IRB approach requires banks to group the

obligors into discrete grades. Each obligor grade, in turn, must be

associated with a single PD.

S. The obligor-rating system must result in a ranking of obligors

by likelihood of default.

The proper operation of the obligor-rating system will feature a

ranking of obligors by likelihood of default. For example, if a bank

uses a rating system based on a 10-point scale, with 1 representing

obligors of highest financial strength and 10 representing defaulted

obligors, grades 2 through 9 should represent groups of ever-increasing

risk. In a rating system in which risk increases with the grade, an

obligor with a grade 4 is riskier than an obligor with a grade 2, but

need not be twice as risky.

S. Separate exposures to the same obligor must be assigned to the

same obligor rating grade.

As noted above, the IRB framework requires that the obligor rating

be distinct from the loss severity rating, which is assigned to the

facility. Collateral and other facility characteristics should not

influence the obligor rating. For example, in a 1-to-10 rating system,

where risk increases with the number grade, a defaulted borrower with a

fully cash-secured transaction should be rated a 10--defaulted--

regardless of the remote expectation of loss. Likewise, a borrower

whose financial condition warrants the highest investment grade rating

should be rated a 1 even if the bank's transactions are subordinate to

other creditors and unsecured. Since the rating is assigned to the

obligor and not the facility, separate exposures to the same obligor

must be assigned to the same obligor rating grade.

At the bottom of any IRB system rating scale is a default grade.

Once an obligor is considered to be in default for IRB purposes, that

obligor must be assigned a default grade until such time as its

financial condition and performance improve sufficiently to clearly

meet the bank's internal rating definition for one of its non-default

grades. Once an obligor is in default on any material credit obligation

to the subject bank, all of its facilities at that institution are

considered to be in default.

S. In assigning an obligor to a rating category, the bank must

assess the risk of obligor default over a period of at least one year.

S. Obligor ratings must reflect the impact of financial distress.

In assigning an obligor to a rating category, the bank must assess

the risk of obligor default over a period of at least one year. This

use of a one-year assessment horizon does not mean that a bank should

limit its consideration to outcomes for that obligor that are most

likely over that year; the rating must take into account possible

adverse events that might increase an obligor's likelihood of default.

Rating Philosophy--Decisions Underlying Ratings Architecture

S. Banks must adopt a ratings philosophy. Policy guidelines should

describe the ratings philosophy, particularly how quickly ratings are

expected to migrate in response to economic cycles.

S. A bank's capital management policy must be consistent with its

ratings philosophy in order to avoid capital shortfalls in times of

systematic economic stress.

In the IRB framework, banks assign obligors to groups that are

expected to share common default frequencies. That general description,

however, still leaves open different possible implementations,

depending on how the bank defines the set of possible adverse events

that the obligor might face. A bank must decide whether obligors are

grouped by expected common default frequency over the next year (a so-

called point-in-time rating system) or by an expected common default

frequency over a wider range of possible stress outcomes (a so-called

through-the-cycle rating system). Choosing between a point-in-time

system and a through-the-cycle system yields a rating philosophy.

In point in time rating systems, obligors are assigned to groups

that are expected to share a common default frequency in a particular

year. Point-in-time ratings change from year to year as borrowers'

circumstances change, including changes due to the economic

possibilities faced by the borrowers. Since the economic circumstances

of many borrowers reflect the common impact of the general economic

environment, the transitions in point-in-time ratings will reflect that

systematic influence. A Merton-style probability of default prediction

model is commonly believed to be an example of a point-in-time approach

to rating (although that may depend on the specific implementation of

the model).

Through-the-cycle rating systems do not ask the question, what is

the probability of default over the next year.

[[Page 45955]]

Instead, they assign obligors to groups that would be expected to share

a common default frequency if the borrowers in them were to experience

distress, regardless of whether that distress is in the next year.

Thus, as the descriptive title suggests, this rating philosophy

abstracts from the near-term economic possibilities and considers a

richer assessment of the possibilities. Like point-in-time ratings,

through the cycle ratings will change from year to year due to changes

in borrower circumstance. However, since this rating philosophy

abstracts from the immediate economic circumstance and considers the

implications of hypothetical stress circumstances, year to year

transitions in ratings will be less influenced by changes in the actual

economic environment. The ratings agencies are commonly believed to use

through-the-cycle rating approaches.

Current practice in many banks in the U.S. is to rate obligors

using an approach that combines aspects of both point-in-time and

through the cycle approaches. The explanation provided by banks that

combine those approaches is that they want rating transitions to

reflect the directional impact of changes in the economic environment,

but that they do not want all of the volatility in ratings associated

with a point-in-time approach.

Regardless of which ratings philosophy a bank chooses, an IRB bank

must articulate clearly its approach and the implications of that

choice. As part of the choice of rating philosophy, the bank must

decide whether the same ratings philosophy will be employed for all of

the bank's portfolios. And management must articulate the implications

that the bank's ratings philosophy has on the bank's capital planning

process. If a bank chooses a ratings philosophy that is likely to

result in ratings transitions that reflect the impact of the economic

cycle, its capital management policy must be designed to avoid capital

shortfalls in times of systematic economic stress.

Obligor-Rating Granularity

S. An institution must have at least seven obligor grades that

contain only non-defaulted borrowers and at least one grade to which

only defaulted borrowers are assigned.

The number of grades used in a rating system should be sufficient

to reasonably ensure that management can meaningfully differentiate

risk in the portfolio, without being so large that it limits the

practical use of the rating system. To determine the appropriate number

of grades beyond the minimum seven non-default grades, each institution

must perform its own internal analysis.

S. An institution must justify the number of obligor grades used in

its rating system and the distribution of obligors across those grades.

The mere existence of an exposure concentration in a grade (or

grades) does not, by itself, reflect weakness in a rating system. For

example, banks may focus on a particular type of lending, such as

asset-based lending, in which the borrowers may have similar default

risk. Banks with such focused lending activities may use close to the

minimum number of obligor grades, while banks with a broad range of

lending activities should have more grades. However, banks with a high

concentration of obligors in a particular grade are expected to perform

a thorough analysis that supports such a concentration.

A significant concentration within an obligor grade may be

suspected if the financial strength of the borrowers within that grade

varies considerably. If obligors seem unduly concentrated, then

management should ask themselves the following questions:

[sbull] Are the criteria for each grade clear? Those rating

criteria may be too vague to allow raters to make clear distinctions.

Ambiguity may be an issue throughout the rating scale or it may be

limited to the most commonly used ratings.

[sbull] How diverse are the obligors? That is how many market

segments (for example, large commercial, middle market, private

banking, small business, geography, etc.) are significantly represented

in the bank's borrower population? If a bank's commercial loan

portfolio is not concentrated in one market segment, its risk rating

distribution is not likely to be concentrated.

[sbull] How broad are the bank's internal rating categories

compared to those of other lenders? The bank may be able to learn

enough from publicly available information to adjust its rating

criteria.

Some banks use ``modifiers'' to provide more risk differentiation

to a given rating system. A risk rating modified with a plus, minus or

other indicator does not constitute a separate grade unless the bank

has developed a distinct rating definition and criteria for the

modified grade. In the absence of such distinctions, grades such as 5,

5+, and 5- are viewed as a single grade for regulatory capital purposes

regardless of the existence of the modifiers.

Loss Severity Ratings

S. Banks must rank facilities by the expected severity of the loss

upon default.

The second dimension of an IRB system is the loss severity rating,

which is calibrated to LGD. A facility's LGD estimate is the loss the

bank is likely to incur in the event that the obligor defaults, and is

expressed as a percentage of exposure at the time of default. LGD

estimates can be assigned either through the use of a loss severity

rating system or they can be directly assigned to each facility.

LGD analysis is still in very early stages of development relative

to default risk modeling. Academic research in this area is relatively

sparse, data are not abundant, and industry practice is still widely

varying and evolving. Given the lack of data and the lack of research

into LGD modeling, some banks are likely, as a first step, to segment

their portfolios by a handful of available characteristics and

determine the appropriate LGDs for those segments. Over time, banks'

LGD methodologies are expected to evolve. Long-standing banking

experience and existing research on LGD, while preliminary, suggests

that collateral values, seniority, industry, etc. are predictive of

loss severity.

S. Banks must have empirical support for LGD rating systems

regardless of whether they use an LGD grading system or directly assign

LGD estimates.

Whether a bank chooses to assign LGD values directly or,

alternatively, to rate facilities and then quantify the LGD for the

rating grades, the key requirement is that it will need to identify

facility characteristics that influence LGD. Each of the loss severity

rating categories must be associated with an empirically supported LGD

estimate. In much the same way an obligor-rating system ranks exposures

by the probability of default, a facility rating system must rank

facilities by the likely loss severity.

Regardless of the method used to assign LGDs (loss severity grades

or direct LGD estimation), data used to support the methodology must be

gathered systematically. For many banks, the quality and quantity of

data available to support the LGD estimation process will have an

influence on the method they choose.

Stress Condition LGDs

S. Loss severity ratings must reflect losses expected during

periods with a relatively high number of defaults.

Like obligor ratings, which group obligors by expected default

frequency, loss severity ratings assign facilities to groups that are

expected to experience a common loss severity. However, the different

treatment accorded to PD and LGD in the model used to calculate IRB

capital requirements mandates an

[[Page 45956]]

asymmetric treatment of obligor and loss severity ratings. Obligor

ratings assign obligors to groups that are expected to experience

common default frequencies across a number of years, some of which are

years of general economic stress and some of which are not. In

contrast, loss severity ratings (or estimates) must pertain to losses

expected during periods with a high number of defaults--particular

years that can be called stress conditions. For cases in which loss

severities do not have a material degree of cyclical variability, use

of a long-run default weighted average is appropriate, although stress

condition LGD generally exceeds these averages.

Loss Severity Rating/LGD Granularity

S. Banks must have a sufficiently fine loss severity grading system

or prediction model to avoid grouping facilities with widely varying

LGDs together.

While there is no stated minimum number of loss severity grades,

the systems that provide LGD estimates must be flexible enough to

adequately segment facilities with significantly varying LGDs. Banks

should have a sufficiently fine LGD grading system or LGD prediction

model to avoid grouping facilities with widely varying LGDs together.

For example, a bank using a loss severity rating-scale approach that

has credit products with a variety of collateral packages or financing

structures would be expected to have more LGD grades than those

institutions with fewer options in their credit products.

Other Considerations of IRB Rating System Architecture

Timeliness of Ratings

S. All risk ratings must be updated whenever new relevant

information is received, but must be updated at least annually.

A bank must have a policy that requires a dynamic ratings approach

ensuring that obligor and loss severity ratings reflect current

information. That policy must also specify minimum financial reporting

and collateral valuation requirements. For example, at the time of

servicing events, banks typically receive updated financial information

on obligors. For cases in which loss severity grades or estimates are

dependent on collateral values or other factors that change

periodically, that policy must take into account the need to update

these factors.

Banks' policies may include an alternative rating update timetable

for exposures below a de minimus amount that is justified by the lack

of materiality of the potential impact on capital. For example, some

banks use triggering events to prompt an update of their ratings on de

minimus exposures rather than adhering to a specific timetable.

Multiple Ratings Systems

Some banks may develop one risk-rating system that can be used

across the entire commercial loan portfolio. However, a bank can choose

to deploy any number of rating systems as long as all exposures are

assigned PD and LGD values. A different rating system could be used for

each business line and each rating system could use a different rating

scale. A bank could also use a different rating system for each

business line with each system using a common rating scale. Rating

models could be used for some portfolios and expert judgment systems

for others. An institution's complexity and sophistication, as well as

the size and range of products offered, will affect the types and

numbers of rating systems employed.

While using a number of rating systems is feasible, such a practice

might make it more difficult to meet supervisory standards. Each rating

system must conform to the standards in this guidance and must be

validated for accuracy and consistency. The requirement that each

rating systems be calibrated to parameter values imposes the ultimate

constraint, which is that ratings be applied consistently.

Recognition of the Risk Mitigation Benefits of Guarantees

S. Banks reflecting the risk-mitigating effect of guarantees must

do so by either adjusting PDs or LGDs, but not both.

S. To recognize the risk-mitigating effects of guarantees,

institutions must ensure that the written guarantee is evidenced by an

unconditional and legally enforceable commitment to pay that remains in

force until the debt is satisfied in full.

Adjustments for guarantees must be made in accordance with specific

criteria contained in the bank's credit policy. The criteria should be

plausible and intuitive, and should address the guarantor's ability and

willingness to meet its obligations. Banks are expected to gather

evidence that confirms the risk-mitigating effect of guarantees.

Other forms of written third-party support (for example, comfort

letters or letters of awareness) that are not legally binding should

not be used to adjust PD or LGD unless a bank can demonstrate through

analysis of internal data the risk-mitigating effect of such support.

Banks may not adjust PDs or LGDs to reflect implied support or verbal

assurances.

Regardless of the method used to recognize the risk-mitigating

effects of guarantees, a bank must adopt an approach that is applied

consistently over time and across the portfolio. Moreover, the onus is

on the bank to demonstrate that its approach is supported by logic and

empirical results. While guarantees may provide grounds for adjusting

PD or LGD, they cannot result in a lower risk weight than that assigned

to a similar direct obligation of the guarantor.\3\

---------------------------------------------------------------------------

\3\ The probability that an obligor and a guarantor (who

supports the obligor's debt) will both default on a debt is lower

than the probability that either the obligor or the guarantor will

default. This favorable risk-mitigation effect is known as the

reduced likelihood of ``double default.'' In determining their

rating criteria and procedures, banks are not permitted to consider

possible favorable effects of imperfect expected correlation between

default events for the borrower and guarantor for purposes of

regulatory capital requirements. Thus, the adjusted risk weight

cannot reflect the risk mitigation of double default. The ANPR

solicits public comment on the double-default issues.

---------------------------------------------------------------------------

Validation Process

S. IRB rating system architecture must be designed to ensure rating

system accuracy.

As part of their IRB rating system architecture, banks must

implement a process to ensure the accuracy of their rating systems.

Rating system accuracy is defined as the combination of the following

outcomes:

[sbull] The actual long-run average default frequency for each

rating grade is not significantly greater than the PD assigned to that

grade.

[sbull] The actual stress-condition loss rates experienced on

defaulted facilities are not significantly greater than the LGD

estimates assigned to those facilities.

Some differences across individual grades between observed outcomes

and the estimated parameter inputs to the IRB equations can be

expected. But if systematic differences suggest a bias toward lowering

regulatory capital requirements, the integrity of the rating system (of

either the PD or LGD dimensions or of both) becomes suspect. Validation

is the set of activities designed to give the greatest possible

assurances of ratings system accuracy.

S. Banks must have ongoing validation processes that include the

review of developmental evidence, ongoing monitoring, and the

comparison of predicted parameter values to actual outcomes (back-

testing).

Validation is an integral part of the rating system architecture.

Banks must have processes designed to give

[[Page 45957]]

reasonable assurances of their rating systems' accuracy. The ongoing

process to confirm and ensure rating system accuracy consists of:

[sbull] The evaluation of developmental evidence,

[sbull] Ongoing monitoring of system implementation and

reasonableness (verification and benchmarking), and

[sbull] Back-testing (comparing actual to predicted outcomes).

IRB institutions are expected to employ all of the components of

this process. However, the data to perform comprehensive back-testing

will not be available in the early stages of implementing an IRB rating

system. Therefore, banks will have to rely more heavily on

developmental evidence, quality control tests, and benchmarking to

assure themselves and other interested parties that their rating

systems are likely to be accurate. Since the time delay before rating

systems can be back-tested is likely to be an important issue--because

of the rarity of defaults in most years and the bunching of defaults in

a few years--the other parts of the validation process will assume

greater importance. If rating processes are developed in a learning

environment in which banks attempt to change and improve ratings, back

testing may be delayed even further. Validation in its early stages

will depend on bank management's exercising informed judgment about the

likelihood of the rating system working--not simply on empirical tests.

Ratings System Developmental Evidence

The first source of support for the validity of a bank's rating

system is developmental evidence. Evaluating developmental evidence

involves making a reasonable assessment of the quality of the rating

system by analyzing its design and construction. Developmental evidence

is intended to answer the question, Could the rating system be expected

to work reasonably if it is implemented as designed? That evidence will

have to be revisited whenever the bank makes a change to its rating

system. If a bank adopts a rating system and does not make changes,

this step will not have to be revisited. However, since rating systems

are likely to change over time as the bank learns about the

effectiveness of the system and incorporates the results of those

analyses, the evaluation of developmental evidence is likely to be an

ongoing part of the process. The particular steps taken in evaluating

developmental evidence will depend on the type of rating system.

Generally, the evaluation of developmental evidence will include a

body of expert opinion. For example, developmental evidence in support

of a statistical rating model must include information on the logic

that supports the model and an analysis of the statistical model-

building techniques. In contrast, developmental evidence in support of

a constrained-judgment system that features guidance values of

financial ratios might include a description of the logic and evidence

relating the values of the ratios to past default and loss outcomes.

Regardless of the type of rating system, the developmental evidence

will be more persuasive when it includes empirical evidence on how well

the ratings might have worked in the past. This evidence should be

available for a statistical model since such models are chosen to

maximize the fit to outcomes in the development sample. In addition,

statistical models should be supported by evidence that they work well

outside the development sample. Use of ``holdout'' sample evidence is a

good model-building practice to ensure that the model is not merely a

statistical quirk of the particular data set used to build the model.

Empirical developmental evidence of rating effectiveness will be

more difficult to produce for a judgmental rating system. Such evidence

would require asking raters how they would have rated past credits for

which they did not know the outcomes. Those retrospective ratings could

then be compared to the outcomes to determine whether the ratings were

correct on average. Conducting such tests, however, will be difficult

because historical data sets may not include all of the information

that an individual would have actually used in making a judgment about

a rating.

The sufficiency of the developmental evidence will itself be a

matter of informed expert opinion. Even if the rating system is model-

based, an evaluation of developmental evidence will entail judging the

merits of the model-building technique. Although no bright line tests

are feasible because expert judgment is essential to the evaluation of

rating system development, experts will be able to draw conclusions

about whether a well-implemented system would be likely to perform

satisfactorily.

Ratings System Ongoing Validation

The second source of analytical support for the validity of a bank

rating system is the ongoing analysis intended to confirm that the

rating system is being implemented and continues to perform as

intended. Such analysis involves process verification and benchmarking.

Process Verification

Verification activities address the question, Are the ratings being

assigned as intended? Specific verification activities will depend on

the rating approach. If a model is used for rating, verification

analysis begins by confirming that the computer code used to deploy the

model is correct. The computer code can be verified in a number of

established ways. For example, a qualified expert can duplicate the

code or check the code line by line. Process verification for a model

will also include confirmation that the correct data are being used in

the model.

For expert-judgment and constrained-judgment systems, verification

requires other individual reviewers to evaluate whether the rater

followed rating policy. The primary requirements for verification of

ratings assigned by individuals are:

[sbull] A transparent rating process,

[sbull] A database with information used by the rater, and

[sbull] Documentation of how the decisions were made.

The specific steps will depend on how much the process incorporates

specific guidelines and how much the exercise of judgment is allowed.

As the dependence on specific guidelines increases, other individuals

can more easily confirm that guidelines were followed by reference to

sufficient documentation. As the dependence on judgment rises, the

ratings review function will have to be staffed increasingly by experts

with appropriate skills and knowledge about the rating policies of the

bank.

Ratings process verification also includes override monitoring. If

individuals have the ability to override either models or policies in a

constrained-judgment system, the bank should have both a policy stating

the tolerance for overrides and a monitoring system for identifying the

occurrence of overrides. A reporting system capturing data on reasons

for overrides will facilitate learning about whether overrides improve

accuracy.

Benchmarking

S. Banks must benchmark their internal ratings against internal,

market and other third-party ratings.

Benchmarking is the set of activities that uses alternative tools

to draw inferences about the correctness of ratings before outcomes are

actually

[[Page 45958]]

known. The most important type of benchmarking of a rating system is to

ask whether another rater or rating method attaches the same rating to

a particular obligor or facility. Regardless of the rating approach,

the benchmark can be either a judgmental or a model-based rating.

Examples of such benchmarking include:

[sbull] Ratings reviewers who completely re-rate a sample of

credits rated by individuals in a judgmental system.

[sbull] An internally developed model is used to rate credits rated

earlier in a judgmental system.

[sbull] Individuals rate a sample of credits rated by a model.

[sbull] Internal ratings are compared against results from external

agencies or external models.

Because it will take considerable time before outcomes will be

available, using alternative ratings as benchmarks will be a very

important validation device. Such benchmarking must be applied to all

rating approaches, and the benchmark can be either a model or judgment.

At a minimum, banks must establish a process in which a representative

sample of its internal ratings is compared to third-party ratings

(e.g., independent internal raters, external rating agencies, models,

or other market data sources) of the same credits.

Benchmarking also includes activities designed to draw broader

inferences about whether the rating system--as opposed to individual

ratings--is working as expected. The bank can look for consistency in

ranking or consistency in the values of rating characteristics for

similarly rated credits. Examples of such benchmarking activities

include:

[sbull] Analyzing the characteristics of obligors that have

received common ratings.

[sbull] Monitoring changes in the distribution of ratings over

time.

[sbull] Calculating a transition matrix calculated from changes in

ratings in a bank's portfolio and comparing it to historical transition

matrices from internal bank data or publicly available ratings.

While benchmarking activities allow for inferences about the

correctness of the ratings system, they are the not same thing as back-

testing. The benchmark itself is a prediction and may be in error. If

benchmarking evidence suggests a pattern of rating differences, it

should lead the bank to investigate the source of the differences.

Thus, the benchmarking process illustrates the possibility of feedback

from ongoing validation to model development, underscoring the

characterization of validation as a process.

Back Testing

S. Banks must develop statistical tests to back-test their IRB

rating systems.

S. Banks must establish internal tolerance limits for differences

between expected and actual outcomes.

S. Banks must have a policy that requires remedial actions be taken

when policy tolerances are exceeded.

The third component of a validation process is back-testing, which

is the comparison of predictions with actual outcomes. Back-testing of

IRB systems is the empirical test of the accuracy of the parameter

values, PD and LGD, associated with obligor and loss severity ratings,

respectively. For IRB rating systems, back-testing addresses the

combined effectiveness of the assignment of obligor and loss severity

ratings and the calibration of the parameters PD and LGD attached to

those ratings.

At this time, there is no generally agreed-upon statistical test of

the accuracy of IRB systems. Banks must develop statistical tests to

back-test their IRB rating systems. In addition, banks must have a

policy that specifies internal tolerance limits for comparing back-

testing results. Importantly, that policy must outline the actions that

would be taken whenever policy limits are exceeded.

As a combined test of ratings effectiveness, back-testing is a

conceptual bridge between the ratings system architecture discussed in

this chapter and the quantification of parameters, discussed in Chapter

2. The final section of Chapter 2 discusses back-testing as one type of

quantitative test required to validate the quantification of parameter

values.

III. Quantification of IRB Systems

Ratings quantification is the process of assigning numerical values

to the four key components for internal ratings-based assessments of

credit-risk capital: probability of default (PD), the expected loss

given default (LGD), the expected exposure at default (EAD), and

maturity (M). Section I establishes an organizing framework for

considering IRB quantification and develops general principles that

apply to the entire process. Sections II through IV cover specific

principles or supervisory standards that apply to PD, LGD, and EAD

respectively. The maturity component, which is much less dependent on

statistical estimates and the use of data, receives somewhat different

treatment in section V. Validation of the quantification process is

covered in section VI.

A. Introduction

Stages of the Quantification Process

With the exception of maturity, the risk components are

unobservable and must be estimated. The estimation must be consistent

with sound practice and supervisory standards. In addition, a bank must

have processes to ensure that these estimates remain valid.

Calculation of risk components for IRB involves two sets of data:

the bank's actual portfolio data, consisting of current credit

exposures assigned to internal grades, and a ``reference data set,''

consisting of a set of defaulted credits (in the case of LGD and EAD

estimation) or both defaulted and non-defaulted credits (in the case of

PD estimation). The bank estimates a relationship between the reference

data set and probability of default, loss severity, or exposure; then

this estimated relationship is applied to the actual portfolio data for

which capital is being assessed.

Quantification proceeds through four logical stages: obtaining

reference data; estimating the reference data's relationship to the

parameters; mapping the correspondence between the reference data and

the portfolio's data; and applying the relationship between reference

data and parameters to the portfolio's data. (Readers may find it

helpful to refer to the appendix to this chapter, which illustrates how

this four-stage framework can be applied to ratings quantification

approaches in practice.) An evaluation of any bank's IRB quantification

process focuses on understanding how the bank implements each stage for

each of the key parameters, and on assessing the adequacy of the bank's

approach.

Data--First, the bank constructs a reference data set, or source of

data, from which parameters can be estimated.

Reference data sets include internal data, external data, and

pooled internal/external data. Important considerations include the

comparability of the reference data to the current credit portfolio,

whether the sample period ``appropriately'' includes periods of stress,

and the definition of default used in the reference data. The reference

data must be described using a set of observed characteristics;

consequently, the data set must contain variables that can be used for

this characterization. Relevant characteristics might include external

debt ratings, financial measures, geographic regions, or any other

factors that are believed to be

[[Page 45959]]

related in some way to PD, LGD, or EAD. More than one reference data

set may be used.

Estimation--Second, the bank applies statistical techniques to the

reference data to determine a relationship between characteristics of

the reference data and the parameters (PD, LGD, or EAD).

The result of this step is a model that ties descriptive

characteristics of the obligor or facility in the reference data set to

PD, LGD, or EAD estimates. In this context, the term `models' is used

in the most general sense; a model may be simple, such as the

calculation of averages, or more complicated, such as an approach based

on advanced regression techniques. This step may include adjustments

for differences between the IRB definition of default and the default

definition in the reference data set, or adjustments for data

limitations. More than one estimation technique may be used to generate

estimates of the risk components, especially if there are multiple sets

of reference data or multiple sample periods.

Mapping--Third, the bank creates a link between its portfolio data

and the reference data based on common characteristics.

Variables or characteristics that are available for the current

portfolio must be mapped to the variables used in the default, loss-

severity, or exposure model. (In some cases, the bank constructs the

link for a representative exposure in each internal grade, and the

mapping is then applied to all credits within a grade.) An important

element of mapping is making adjustments for differences between

reference data sets and the bank's portfolio. The bank must create a

mapping for each reference data set and for each combination of

variables used in any estimation model.

Application--Fourth, the bank applies the relationship estimated

for the reference data to the actual portfolio data.

The ultimate aim of quantification is to attribute a PD, LGD, or

EAD to each exposure within the portfolio, or to each internal grade if

the mapping was done at the grade level. This step may include

adjustments to default frequencies or loss rates to ``smooth'' the

final parameter estimates. If the estimates are applied to individual

transactions, the bank must in some way aggregate the estimates at the

grade level. In addition, if multiple data sets or estimation methods

are used, the bank must adopt a means of combining the various

estimates.

A number of examples are given in this chapter to aid exposition

and interpretation. None of the examples is sufficiently detailed to

incorporate all the considerations discussed in this chapter. Moreover,

technical progress in the area of quantification is rapid. Thus, banks

should not interpret an example that is consistent with the standard

being discussed, and that resembles the bank's current practice, as

creation of a ``safe harbor'' or as an indication that the bank's

practice will be approved as-is. Banks should consider this guidance in

its entirety when determining whether systems and practices are

adequate.

General Principles for Sound IRB Quantification

Several core principles apply to all elements of the overall

ratings quantification process; those general principles are discussed

in this introductory section. Each of these principles is, in effect, a

supervisory standard for IRB systems. Other supervisory standards,

specific to particular elements or parameters, are discussed in the

relevant sections.

Supervisory evaluation of IRB quantification requires consideration

of all of these principles and standards, both general and specific.

Particular practical approaches to ratings quantification may be highly

consistent with some standards, and less so with others. In any

particular case, an ultimate assessment relies on the judgment of

supervisors to weigh the strengths and weaknesses of a bank's chosen

approach, using these supervisory standards as a guide.

S. IRB institutions must have a fully specified process covering

all aspects of quantification (reference data, estimation, mapping, and

application). The quantification process, including the role and scope

of expert judgment, must be fully documented and updated periodically.

A fully specified quantification process must describe how all four

stages (data, estimation, mapping, and application) are implemented for

each parameter. Documentation promotes consistency and allows third

parties to review and replicate the entire process. Examples of third

parties that might use the documentation include rating-system

reviewers, auditors, and bank supervisors. Periodic updates to the

process must be conducted to ensure that new data, analytical

techniques, and evolving industry practice are incorporated into the

quantification process.

S. Parameter estimates and related documentation must be updated

regularly.

The parameter estimates must be updated at least annually, and the

process for doing so must be documented in bank policy. The update

should also evaluate the judgmental adjustments embedded in the

estimates; new data or techniques may suggest a need to modify those

adjustments. Particular attention should be given to new business lines

or portfolios in which the mix of obligors is believed to have changed

substantially. A material merger, acquisition, divestiture, or exit

clearly raises questions about the continued applicability of the

process and should trigger an intensive review and updating.

The updating process is particularly relevant for the reference

data stage because new data become available all the time. New data

must be incorporated, into the PD, LGD, and EAD estimates, using a

well-defined process.

S. A bank must subject all aspects of the quantification process,

including design and implementation, to an appropriate degree of

independent review and validation.

An independent review is an assessment conducted by persons not

accountable for the work being reviewed. The reviewers may be either

internal or external parties. The review serves as a check that the

quantification process is sound and works as intended; it should be

broad-based, and must include all of the elements of the quantification

process that lead to the ultimate estimates of PD, LGD, and EAD. The

review must cover the full scope of validation: evaluation of the

integrity of data inputs, analysis of the internal logic and

consistency of the process, comparison with relevant benchmarks, and

appropriate back-testing based on actual outcomes.

S. Judgmental adjustments may be an appropriate part of the

quantification process, but must not be biased toward lower estimates

of risk.

Judgment will inevitably play a role in the quantification process

and may materially affect the estimates. Judgmental adjustments to

estimates are often necessary because of some limitations on available

reference data or because of inherent differences between the reference

data and the bank's portfolio data. The bank must ensure that

adjustments are not biased toward optimistically low parameter

estimates for PD, LGD, and EAD. Individual assumptions are less

important than broad patterns; consistent signs of judgmental decisions

that lower parameter estimates materially may be evidence of bias.

[[Page 45960]]

The reasoning and empirical support for any adjustments, as well as

the mechanics of the calculation, must be documented. The bank should

conduct sensitivity analysis to demonstrate that the adjustment

procedure is not biased toward reducing capital requirements. The

analysis must consider the impact of any judgmental adjustments on

estimates and risk weights, and must be fully documented.

S. Parameter estimates must incorporate a degree of conservatism

that is appropriate for the overall robustness of the quantification

process.

In estimating values of PD, LGD, and EAD should be as precise and

accurate as possible. However, estimates of PD, LGD and EAD are

statistics, and thus inherently subject to uncertainty and potential

error. It is often possible to be reasonably confident that a risk

component or other parameter lies within a particular range, but

greater precision is difficult to achieve. Aspects of the ratings

quantification process that are apt to introduce uncertainty and

potential error include the following:

The estimation of coefficients of particular variables in a

regression-based statistical default or severity model.

[sbull] The calculation of average default or loss rates for

particular categories of credits in external default databases.

[sbull] The mapping between portfolio obligors or facilities and

reference data when the set of common characteristics does not align

exactly.

A general principle of the IRB approach is that a bank must adjust

estimates conservatively in the presence of uncertainty or potential

error. In many cases this corresponds to assigning a final parameter

estimate that increases required capital relative to the best estimate

produced through sound-practice estimation techniques. The extent of

this conservative adjustment should be related to factors such as the

relevance of the reference data, the quality of the mapping, the

precision of the statistical estimates, and the amount of judgment used

throughout the process. Margins of conservatism need not be added at

each step; indeed, that could produce an excessively conservative

result. The overall margin of conservatism should adequately account

for all uncertainties and weaknesses; this is the general

interpretation of requirements to incorporate appropriate degrees of

conservatism. Improvements in the quantification process (use of better

data, estimation techniques, and so on) may reduce the appropriate

degree of conservatism over time.

Estimates of PD, LGD, EAD, or other parameters or coefficients

should be presented with an accompanying sense of the statistical

precision of the estimates; this facilitates an assessment of the

appropriate degree of conservatism.

B. Probability of Default (PD)

Data

To estimate PD accurately, a bank must have a comprehensive

reference data set with observations that are comparable to the bank's

current portfolio of obligors. Clearly, the data set used for

estimation should be similar to the portfolio to which such estimates

will be applied. The same comparability standard applies to both

internal and external data sets.

To ensure ongoing applicability of the reference data, a bank must

assess the characteristics of its current obligors relative to the

characteristics of obligors in the reference data. Such variables might

include qualitative and quantitative obligor information, internal and

external rating, rating dates, and line of business or geography. To

this end, a bank must maintain documentation that fully describes all

explanatory variables in the data set, including any changes to those

variables over time. A well-defined and documented process must be in

place to ensure that the reference data are updated as frequently as is

practical, as fresh data become available or portfolio changes make

necessary.

S. The sample for the reference data must be at least five years,

and must include periods of economic stress during which default rates

were relatively high.

To foster more robust estimation, banks should use longer time

series when more than five years of data are available. However, the

benefits of using a longer time series (longer than five years) may

have to be weighed against a possible loss of data comparability. The

older the reference data, the less similar they are likely to be to the

bank's current portfolio; striking the correct balance is a matter of

judgment. Reference obligors must not differ from the current portfolio

obligors systematically in ways that seem likely to be related to

obligor default risk. Otherwise, the derived PD estimates may not be

applicable to the current portfolio.

Note that this principle does not simply restate the requirement

for five years of data: periods of stress during which default rates

are relatively high must be included in the data sample. Exclusion of

such periods biases PD estimates downward and unjustifiably lowers

regulatory capital requirements.

Example. A bank's reference data set covers the years 1987

through 2001. Each year includes identical data elements, and each

year is similarly populated. For its grade PD estimates, the bank

relies upon data from a sub-sample covering 1992 through 2001. The

bank provides no justification for dropping the years from 1987

through 1991. The bank contends that it is not necessary to include

those data, as the reference sample they use for estimation

satisfies the five-year requirement. This practice is not consistent

with the standard because the bank has not supported its decision to

ignore available data. The fact that the excluded years include a

recession would raise particular concerns.

S. The definition of default within the reference data must be

reasonably consistent with the IRB definition of default.

Regardless of the source of the reference data, a bank must apply

the same default definition throughout the quantification processes.

This fosters consistent estimation across parameters and reduces the

potential for undesired bias. In addition, consistent application of

the same definition across banks will permit true horizontal analysis

by supervisors and engaged market participants.

This standard applies to both internal and external reference data.

For internal data, a bank's default definition is expected to be

consistent with the IRB definition going forward. Banks will be

expected to make appropriate adjustments to their data systems such

that all defaults as defined for IRB are captured by the time a bank

fully implements its IRB system. For any historical or external data

that do not fully comply with the IRB definition of default, a bank

must make conservative adjustments to reflect such discrepancies.

Larger discrepancies require larger adjustments for conservatism.

Example. To identify defaults in its historical data, a bank

applies a consistent definition of ``placed on nonaccrual.'' This

definition is used in the bank's quantification exercises to

estimate PD, LGD, and EAD. The bank recognizes that use of the

nonaccrual definition fails to capture certain defaults as

identified in the IRB rules. Specifically, the bank indicates that

the following kinds of defaulted facilities would not have been

placed on nonaccrual: (1) Credit obligations that were sold at a

material credit-related economic loss, and (2) distressed

restructurings. To be consistent with the standard, the bank must

make a well-supported adjustment to its grade PD estimates to

reflect the difference in the default definitions.

Estimation

Estimation of PD is the process by which characteristics of the

reference

[[Page 45961]]

data are related to default frequencies.\4\ The relevant

characteristics that help to determine the likelihood of default are

referred to as ``drivers of default''. Drivers might include variables

such as financial ratios, management expertise, industry, and

geography.

---------------------------------------------------------------------------

\4\ The New Basel Capital Accord produced by the Basel Committee

on Banking Supervision discusses three techniques for PD estimation.

IRB banks are not constrained to select from among these three

techniques; they have broad flexibility to implement appropriate

approaches to quantification. The three Basel techniques are best

regarded not as a complete taxonomy of the possible approaches to PD

estimation, but rather as illustrations of a few of the many

possible approaches.

---------------------------------------------------------------------------

S. Estimates of default rates must be empirically based and must

represent a long-run average.

Estimates must capture average default experience over a reasonable

mix of high-default and low-default years of the economic cycle. The

average is labeled ``long-run'' because a long observation period would

span both peaks and valleys of the economic cycle. The emphasis should

not be on time-span; the long-run average concept captures the breadth,

not the length, of experience.

If the reference data are characterized by internal or external

rating grades, one estimation approach is to calculate the mean of one-

year realized default rates for each grade, giving equal weight to each

year's realized default rate. PD estimates generally should be

calculated in this manner.

Another approach is to pool obligors in a given grade over a number

of years and then calculate the mean default rate. In this case, each

year's default rate is weighted by the number of obligors. This

approach may underestimate default rates. For example, if lending

declines in recessions so that obligors are fewer in those years than

in others, weighting by number of obligors would dilute the effect of

the recession year on the overall mean. The obligor-weighted

calculation, or another approach, will be allowed only if the bank can

demonstrate that this approach provides a better estimate of the long-

run average PD. At a minimum, this would involve comparing the results

of both methods.

Statistical default prediction models may also play a role in PD

estimation. For example, the characteristics of the reference data

might include financial ratios or a distance-to-default measure, as

defined by a specific implementation of a Merton-style structural

model.

For a model-based approach to meet the requirement that ultimate

grade PD estimates be long-run averages, the reference data used in the

default model must meet the long-run requirement. For example, a model

can be used to relate financial ratios to likelihood of default based

on the outcome for the firms--default or non-default. Such a model must

be calibrated to capture the default experience over a reasonable mix

of good and bad years of the economic cycle. The same requirement would

hold for a structural model; distance to default must be calibrated to

default frequency using long-run experience. This applies to both

internal and vendor models, and a bank must verify that this

requirement is met.

Example 1. A bank uses external data from a rating agency to

estimate PD. The PD estimate for each agency grade is calculated as

the mean of yearly realized default rates over a time period (1980

through 2001) that includes several recessions and high-default

years. The bank provides support that this time period adequately

represents long-run experience. This illustrates an estimation

method that is consistent with the standard.

Example 2a. Like the institution in example 1, a bank maps

internal ratings to agency grades. The estimates for the agency

grades are set indirectly, using the default probabilities from a

default prediction model. The bank does so because although it links

internal and agency grades, the bank views the default model's

results as more predictive than the historical agency default

experience. For each agency grade, the bank calculates a PD estimate

as the mean of the model-based default probabilities for the agency-

rated obligors. In order to meet the long-run requirement, the bank

calculates the estimates over the seven years from 1995 through

2001. The bank demonstrates that this time period includes a

reasonable mix of high-default and low-default experience. This

estimation method is consistent with the standard.

Example 2b. In a variant of example 2a, a bank uses the mean

default frequency per agency rating grade for a single year, such as

2001. Empirical evidence shows that the mean default frequency for

agency grades varies substantially from year to year. A single year

thus does not reflect the full range of experience, because a long-

run average should be relatively stable year to year. Such

instability makes this estimation method unacceptable.

Example 2c. Another bank calculates the agency grade PD

estimates as the median default probability of companies in that

grade. The bank does so without demonstrating that the median is a

better statistical estimator than the mean. This estimation method

is not consistent with the standard. A median gives less weight to

obligors with high estimated default probabilities than a simple

mean does. The difference between mean and median can be material

because distributions of credits within grades often are

substantially skewed toward higher default probabilities: the

riskier obligors within a grade tend to have individual default

probabilities that are substantially worse than the median, while

the least risky have default probabilities only somewhat better than

the median.

S. Judgmental adjustments may play an appropriate role in PD

estimation, but must not be biased toward lower estimates.

The following examples illustrate how supervisors will evaluate

adjustments:

Example 1. A bank uses the last five years of internal default

history to estimate grade PDs. However, they recognize that the

internal experience does not include any high-default years. In

order to remedy this and still take advantage of its experience, the

bank uses external agency data to adjust the estimates upward. Using

the agency data, the bank calculates the ratio between the long-run

average and the mean default rate per grade over the last five

years. The bank assumes that the relationship observed in the agency

data applies to its portfolio, and adjusts the estimates for the

internal data accordingly. This practice is consistent with the

standard.

Example 2. A bank uses internal default experience to estimate

grade PDs. However, the bank has historically failed to recognize

defaults when the loss on the default obligation was avoided by

seizing collateral. The bank makes no adjustment for such missing

defaults. The realized default rate using the more inclusive

definition would be higher than that observed by the bank (and loss

severity rates would be correspondingly lower). This practice would

not be consistent with the standard, unless the bank demonstrates

that the necessary adjustment is immaterial.

Mapping

Mapping is the process of establishing a correspondence between the

bank's current obligors and the reference obligor data used in the

default model. Hence, mapping involves identifying how default-related

characteristics of the current portfolio correspond to the

characteristics of reference obligors. Such characteristics might

include financial and nonfinancial variables, and assigned ratings or

grades.

Mapping can be thought of as taking each obligor in the bank's

portfolio and characterizing it as if it were part of the reference

data. There are two broad approaches to the mapping process:

Obligor mapping: Each portfolio obligor is mapped to the reference

data based on its individual characteristics. For example, if a bank

applies a default model, a default probability will be generated for

each obligor. That individual default probability is then used to

assign each obligor to a particular internal grade, based on the bank's

established criteria. To obtain a final estimate of the grade PD in the

subsequent application stage, the bank averages the default

probabilities of individual obligors within each grade.

Grade mapping: Characteristics of the obligors within an internal

grade are

[[Page 45962]]

averaged or otherwise summarized to construct a ``typical'' or

representative obligor for each grade. Then, the bank maps that

representative obligor to the reference data. For example, if the bank

uses a default model, the default probability associated with that

typical obligor will serve as the grade PD in the application stage.

Alternatively, the bank may map the typical obligor to a particular

external rating grade based on quantitative and qualitative

characteristics, and assign the long-run default rate for that rating

to the internal grade in the application stage.

Either grade mapping or obligor mapping can be part of the

quantification process; either method can produce a single PD estimate

for each grade in the application stage. However, in the absence of

other compelling considerations, banks should use obligor mapping for

two reasons:

[sbull] First, default probabilities are nonlinear under many

estimation approaches. As a result, the default probability of the

typical obligor--the result of a grade mapping approach--is often lower

than the mean of the individual obligor default probabilities from the

obligor mapping approach. For example, consider a bank that maps to the

S&P scale and uses historical S&P bond default rates. For ease of

illustration, suppose that one internal grade contains only three

obligors that individually map to BB, BB-, and B+. The historical

default rates for these three grades are 1.07, 1.76, and 3.24 percent,

respectively (based on 1981-2001 data). Using obligor mapping, those

rates would be assigned directly to the three obligors, yielding a mean

PD of 2.02 percent for the grade. Using grade mapping, the grade PD

would be only 1.76, because the grade's typical obligor is rated BB-.

[sbull] Second, a hypothetical obligor with a grade's average

characteristics may not represent well the risks presented by the

grade's typical obligor. For example, a bank might observe that

obligors with high leverage and low earnings variability have about the

same default risk as obligors with low leverage and high earnings

variability. These two types of obligors might both end up in the same

grade, for example, Grade 6. If so, the typical obligor in Grade 6

would have moderate leverage and moderate earnings variability--a

combination that might fail to reflect any of the individual obligors

in Grade 6, and that could easily result in a PD for the grade that is

too low.

A bank electing to use grade mapping instead of obligor mapping

should be especially careful in choosing a ``typical'' obligor for each

grade. Doing so typically requires that the bank examine the actual

distribution of obligors within each grade, as well as the

characteristics of those obligors. Banks should be aware that different

measures of central tendency (such as mean, median, or mode) will give

different results, and that these different results may have a material

effect on a grade's PD; they must be able to justify their choice of a

measure. Banks must have a clear and consistent policy toward the

calculation.

S. The mapping must be based on a robust comparison of available

data elements that are common to the portfolio and the reference data.

Sound mapping practice uses all common elements that are available

in the data as the basis for mapping. If a bank chooses to ignore

certain common variables or to weight some variables more heavily than

others, those choices must be supported. Mapping should also take into

account differences in rating philosophy (for example, point-in-time or

through-the-cycle) between any ratings embedded in the reference data

set and the bank's own rating regime.

A mapping should be plausible, and should be consistent with the

rating philosophy established by the bank as part of its obligor rating

policy. For a bank that uses grade mapping, levels and ranges of key

variables within each internal grade should be close to values of

similar variables for corresponding obligors within the reference data.

The standard allows for use of a limited set of common variables

that are predictive of default risk, in part to permit flexibility in

early years when data may be far from ideal. Nevertheless, banks will

eventually be expected to use variables that are widely recognized as

the most reliable predictors of default risk in mapping exercises. In

the meantime, banks relying on data elements that are weak predictors

must compensate by making their estimates more conservative. For

example, leverage and cash flow are widely recognized to be reliable

predictors of corporate default risk. Borrower size is also predictive,

but less so. A mapping based solely on size is by nature less reliable

than one based on leverage, cash flow, and size.

Example 1. In estimating PD, a bank relies on observed default

rates on bonds in various agency grades for PD quantification. To

map its internal grades to the agency grades, the bank identifies

variables that together explain much of the rating variation in the

bond sample. The bank then conducts a statistical analysis of those

same variables within its portfolio of obligors, using a

multivariate distance calculation to assign each portfolio obligor

to the external rating whose characteristics it matches most closely

(for example, assigning obligors to ratings so that the sum of

squared differences between the external grade averages and the

obligor's characteristics is minimized). This practice is broadly

consistent with the standard.

Example 2. A bank uses grade mapping to link portfolio obligors

to the reference data set described by agency ratings. The bank

looks at publicly rated portfolio obligors within an internal grade

to determine the most common external rating, does the same for all

grades, and creates a correspondence between internal and external

ratings. The strength of the correspondence is a function of the

number of externally rated obligors within each grade, the

distribution of those external ratings within each grade and the

similarity of externally rated obligors in the grade to those not

externally rated. This practice is broadly consistent with this

standard, but would require a comparison of rating philosophies and

may require adjustments and the addition of margins of conservatism.

S. A mapping process must be established for each reference data

set and for each estimation model.

Banks should never assume that a mapping is self-evident. Even a

rating system that has been explicitly designed to replicate external

agency ratings may or may not be effective in producing a replica;

formal mapping is still necessary. Indeed, in such a system the kind of

analysis involved in mapping may help identify inconsistencies in the

rating process itself.

A mapping process is needed even where the reference obligors come

from internal historical experience. Banks must not assume that

internal data do not require mapping, because changes in bank strategy

or external economic forces may alter the composition of internal

grades or the nature of the obligors in those grades over time.

Mappings must be reaffirmed regardless of whether rating criteria or

other aspects of the ratings system have undergone explicit changes

during the period covered by the reference data set.

Banks often use multiple reference data sets, and then combine the

resulting estimates to get a grade PD. A bank that does that must

conduct a rigorous mapping process for each data set.

Supervisors expect all meaningful characteristics of obligors to be

factored directly into the rating process; this should include

characteristics like the obligor's industry or physical location. But

in some circumstances, certain effects related to industry, geography,

or other factors are not reflected in rating assignments or default

estimates. In such cases, it may be appropriate for banks to capture

the impact of the

[[Page 45963]]

omissions by using different mappings for different business lines or

types of obligors. Supervisors expect this practice to be transitional;

banks will eventually be required to incorporate the omitted effects

into the rating system and the estimation process as they are uncovered

and documented, rather than adjusting the mapping.

Example 1. The bank maps its internal grades carefully to one

rating agency, and then assumes a correspondence to another agency's

scale despite known differences in the rating methods of the two

agencies. The bank then applies a mean of the grade default rates

from these two public debt-rating agencies to its internal grades.

This practice is not consistent with the standard, because the bank

should map to each agency's scale separately.

Example 2. A bank uses internal historical data as its reference

data. The bank computes a mean default rate for each grade as the

grade PD for capital purposes, and asserts that mapping is

unnecessary because ``its strong credit culture ensures that a 4 is

always a 4.'' This practice is not consistent with the standard,

because no mapping has been done; there is no assurance that a

representative obligor in a grade today is comparable to an obligor

in that same grade in the past.

S. The mapping must be updated and independently validated

regularly.

The appropriate mapping between a bank's portfolio and the

reference data may change over time. For example, relationships between

internal grades and external agency grades may change during the

economic cycle because of differences in rating philosophy. Similarly,

distance-to-default measures for obligors in a given grade may not be

constant over time. These likely changes make it imperative that the

bank update all mappings regularly.

Sound validation practices may include tests for internal

consistency such as ``reverse mapping.'' Using this technique, a bank

evaluates obligors from the reference data set as if they were subject

to the bank's rating system (that is, part of the bank's current

portfolio). The bank's mapping is then applied to these reverse-mapped

obligors to see whether the mapped characterization of the reference

obligor is consistent with that of the initial evaluation.\5\ Another

valuable technique is to apply different mapping methods and compare

the results. For example, mappings based on financial ratio comparisons

can be rechecked using mappings based on available external ratings.

---------------------------------------------------------------------------

\5\ For example, suppose a bank asserts that its Grade 3

corresponds to an S&P rating of A. Applying reverse mapping, the

bank would take a sample of A-rated obligors from the reference

data, run them through the bank's rating process (perhaps a

simplified version), and check to see that those obligors usually

receive a grade of 3 on the bank's internal scale.

Example. A bank mapped its internal grades to the rating scale

of one public debt-rating agency in 1992. Since then, the bank has

completed a major acquisition of another large bank and

significantly changed its business mix in other ways. The bank

continues to use the same mapping, without reassessing its validity.

This practice is not consistent with the standard.

Application

In the application stage, the bank applies the PD estimation method

to the current portfolio of obligors using the mapping process. It

obtains final PD estimates for each rating grade, which will be used to

calculate minimum regulatory capital. To arrive at those estimates, a

bank may adjust the raw results derived from the estimation stage. For

example, it might aggregate individual obligor default probabilities to

the rating grade level, or smooth results because a rating grade's PD

estimate was higher than a lower quality grade. The bank must explain

and support all adjustments when documenting its quantification

process.

Example. A bank uses external data to estimate long-run average

PDs for each grade. The resulting PD estimate for Grade 2 is

slightly higher than the estimate for Grade 3, even though Grade 2

is supposedly of higher credit quality. The bank uses statistics to

demonstrate that this anomaly occurred because defaults are rare in

the highest quality rating grades. The bank judgmentally adjusts the

PD estimates for grades 2 and 3 to preserve the expected

relationship between obligor grade and PD, but requires that total

risk-weighted assets across both grades using the adjusted PD

estimates be no less than total risk-weighted assets based on the

unadjusted estimates, using a typical distribution of obligors

across the two grades. Such an adjustment during the application

stage is consistent with this guidance.

S. IRB institutions that aggregate the default probabilities of

individual portfolio obligors when calculating PD estimates for

internal grades must have a clear policy governing the aggregation

process.

As noted above, mapping may be grade-based or obligor-based. Grade-

based mappings naturally provide a single PD per grade, because the

estimated default model is applied to the representative obligor for

each grade. In contrast, obligor-based mappings must aggregate in some

manner the individual PD estimates to the grade level. The expectation

is that the grade PD estimate will be calculated as the mean. The bank

will be allowed to calculate this estimate differently only if it can

demonstrate that the alternative method provides a better estimate of

the long-run average PD. To obtain this evidence, the bank must at

least compare the results of both methods.

S. IRB institutions that combine estimates from multiple sets of

reference data must have a clear policy governing the combination

process, and must examine the sensitivity of the results to alternative

combinations.

Because a bank should make use of as much information as possible

when mapping, it will usually use multiple data sets. The manner in

which the data or the estimates from those multiple data sets are

combined is extremely important. A bank must document its justification

for the particular combination methods selected. Those methods must be

subject to appropriate approval and oversight.

The data may come from the same basic data source but from

different time periods or from different data sources altogether. For

example, banks often combine internal data with external data, use

external data from different sample periods, or combine results from

corporate-bond default databases with results from equity-based models

of obligor default. Different combinations will produce different PD

estimates. The bank should investigate alternative combinations and

document the impact on the estimates. When ultimate results are highly

sensitive to how estimates from different data sources are combined,

the bank must choose among the alternatives conservatively.

C. Loss Given Default (LGD)

The LGD estimation process is similar to the PD estimation process.

The bank identifies a reference data set of defaulted credits and

relevant descriptive characteristics. Once the bank obtains these data

sets (with the facility characteristics), it must select a technique to

estimate the economic loss per dollar of exposure at default, for a

defaulted exposure with a given array of characteristics. The bank's

portfolio must then be mapped, so that the model can be applied to

generate an estimate of LGD for each portfolio transaction or severity

grade.

Data

Unlike reference data sets used for PD estimation, data sets for

severity estimation contain only exposures to defaulting obligors. At

least two broad categories of data are necessary to produce LGD

estimates.

First, data must be available to calculate the actual economic loss

experienced for each defaulted facility. Such data may include the

market value of the facility at default, which can be

[[Page 45964]]

used to proxy a recovery rate. Alternatively, economic loss may be

calculated using the exposure at the time of default, loss of

principal, interest, and fees, the present value of subsequent

recoveries and related expenses (or the costs as calculated using an

approved allocation method), and the appropriate discount rate.

Second, factors must be available to group the defaulted facilities

in meaningful ways. Characteristics that are likely to be important in

predicting loss rates include whether or not the facility is secured

and the type and coverage of collateral if the facility is secured,

seniority of the claim, general economic conditions, and obligor's

industry. Although these factors have been found to be significant in

existing academic and industry studies, a bank's quantification of LGD

certainly need not be limited to these variables. For example, a bank

might expand its loss severity research by examining many other

potential drivers of severity (characteristics of an obligor that might

help the bank predict the severity of a loss), including obligor size,

line of business, geographic location, facility type, obligor ratings

(internal or external), historical internal severity grade, or tenor of

the relationship.

A bank must ensure that the reference data remains applicable to

its current portfolio of facilities. It must implement established

processes to ensure that reference data sets are updated when new data

become available. All data sources, variables, and the overall

processes concerning data collection and maintenance must be fully

documented, and that documentation should be readily available for

review.

S. The sample period for the reference data must be at least seven

years, and must include periods of economic stress during which

defaults were relatively high.

Seven years is the minimum sample period for the LGD reference

data. A longer sample period is desirable, because more default

observations will be available for analysis and may serve to refine

severity estimates. In any case, a bank must select a sample period

that includes episodes of economic stress, which are defined as periods

with a relatively high number of defaults. Inclusion of stress periods

increases the size and potentially the breadth of the reference data

set. According to some empirical studies, the average loss rate is

higher during periods of stress.

Example. A bank intends to rely primarily on internal data when

quantifying all parameter estimates, including LGD. Its internal

data cover the period 1994 through 2000. The bank will continue to

extend its data set as time progresses. Its current policy mandates

that credits be resolved within two years of default, and the data

set contains the most recent data available. Although the current

data set satisfies the seven-year requirement, the bank is aware

that it does not include stress periods. In comparing its loss

estimates with rates published in external studies for similarly

stratified data, the bank observes that its estimates are

systematically lower. To be consistent with the standard, the bank

must take steps to include stress periods in its estimates.

S. The definition of default within the reference data must be

reasonably consistent with the IRB definition of default.

This standard parallels a similar standard in the section on PD.

The following examples illustrate how it applies in the case of LGD.

Example 1. For LGD estimation, a bank includes in its default

data base only defaulted facilities that actually experience a loss,

and excludes credits for which no loss was recorded because

liquidated collateral covered the loss (effectively applying a

``loss given loss'' concept). This practice is not consistent with

the standard because the bank's default definition for LGD is

narrower than the IRB definition.

Example 2. A bank relies on external data sources to estimate

LGD because it lacks sufficient internal data. One source uses

``bankruptcy filing'' to indicate default while another uses

``missed principal or interest payment,'' and the two sources result

in significantly different loss estimates for the severity grades

defined by the bank. The bank's practice is not consistent with the

standard, and the bank should determine whether the definitions used

in the reference data sets differ substantially from the IRB

definition. If so, and the differences are difficult to quantify,

the bank should seek other sources of reference data. For more minor

differences, the bank may be able to make appropriate adjustments

during the estimation stage.

Estimation

Estimation of LGD is the process by which characteristics of the

reference data are related to loss severity. The relevant

characteristics that help explain how severe losses tend to be upon

default might include variables such as seniority, collateral, facility

type, or business line.

S. The estimates of loss severity must be empirically based and

must reflect the concept of ``economic loss.''

Loss severity is defined as economic loss, which is different from

accounting measures of loss. Economic loss captures the value of

recoveries and direct and indirect costs discounted to the time of

default, and it should be measured for each defaulted facility. The

scope of the cash flows included in recoveries and costs is meant to be

broad. Workout costs that can be clearly attributed to certain

facilities or types of facilities must be reflected in the bank's LGD

assignments for those exposures. When such allocation is not practical,

the bank may assign those costs using factors based on broad averages.

A bank must establish a discount rate that reflects the time value

of money and the opportunity cost of funds to apply to recoveries and

costs. The discount rate must be no less than the contract interest

rate on new originations of a type similar to the transaction in

question, for the lowest-quality grade in which a bank originates such

transactions.\6\ Where possible, the rate should reflect the fixed rate

on newly originated exposures with term corresponding to the average

resolution period of defaulting assets.

---------------------------------------------------------------------------

\6\ The appropriate discount rate for IRB purposes may differ

from the contract rate required under FAS 114 for accounting

purposes.

---------------------------------------------------------------------------

Ideally, severity should be measured once all recoveries and costs

have been realized. However, a bank may not resolve a defaulted

obligation for many years following default. For practical purposes,

banks may choose to close the period of observation before this final

resolution occurs--that is, at a point in time when most costs have

been incurred and when recoveries are substantially complete. Banks

that do so should estimate the additional costs and recoveries that

would likely occur beyond this period and include them in the LGD

estimates. A bank must document its choice of the period of

observation, and how it estimated additional costs and recoveries

beyond this period.

LGD for each type of exposure must be the loss per default

(expressed as a percentage of exposure at default) expected during

periods when default rates are relatively high. This expected loss rate

is referred to as ``stress-condition LGD.'' For cases in which loss

severities do not have a material degree of cyclical variability, use

of the long-run default-weighted average is appropriate, although

stress-condition LGD generally exceeds this average.

The drivers of severity can be linked to loss estimates in a number

of ways. One approach is to segment the reference defaults into groups

that do not overlap. For example, defaults could be grouped by business

line, predominant collateral type, and loan-to-value coverage. The LGD

estimate for each category is the mean loss calculated over the

category's defaulted facilities. Loss must be calculated as the

default-weighted average (where individual defaults receive equal

weight) rather than the average of

[[Page 45965]]

annual loss rates, and must be based on results from periods during

which default rates were relatively numerous if loss rates are

materially cyclical.

Banks can also draw estimates of LGD from a statistical model. For

example, they can build a regression model of severity using data on

loss severity and some quantitative measures of the loss drivers. Any

model must meet the requirements for model validation discussed in

Chapter 1. Other methods for computing LGD could also be appropriate.

Example 1. A bank has internal data on defaulted facilities,

including information on business line, facility type, seniority,

and predominant collateral type (if the facility is secured). The

data allow for a reasonable calculation of economic loss. The data

span eight years and include three years that can be termed high-

default years. After analyzing the economic cycle using internal and

external data, the bank concludes that the data show no evidence of

material cyclical variability in loss severities, and that the

default data span enough experience to allow estimation of a long-

run average. On the basis of preliminary analysis, the bank

determines that the drivers of loss severity for large corporate

facilities are similar to those for middle-market loans, and that

the two groups can be estimated as a pool. Again on the basis of

preliminary analysis, the bank segments this pool by seniority and

by six collateral groupings, including unsecured. These groupings

contain enough defaults to allow reasonably precise estimates. The

loss severity estimates are then calculated by averaging loss rates

within each segment. This practice is consistent with the standard.

Example 2. A bank uses internal data in which information on

security and seniority is lacking. The bank groups corporate and

middle-market defaulted facilities into a single pool and calculates

the LGD estimate as the mean loss rate. No adjustments for the lack

of data are made in the estimation or application steps. This

practice is unacceptable because there is ample external evidence

that security and seniority matter in these segments. A bank with

such limited internal default data must incorporate external or

pooled data into the estimation.

Example 3. A bank determines that a business unit--for example,

a unit dedicated to a particular type of asset-based lending--forms

a homogeneous pool for the purposes of estimating loss severity.

That is, although the facilities in this pool may differ in some

respects, the bank determines that they share a similar loss

experience in default. The bank must provide reasonable support for

this pooling through analysis of lending practices and available

internal and external data. In this example, the mean of a single

segment is consistent with the standard.

S. Judgmental adjustments may play an appropriate role in LGD

estimation, but must not be biased toward lower estimates.

It is difficult to make general statements about good and bad

practices in this area, because adjustments can take many different

forms. The following examples illustrate how supervisors would be

likely to evaluate particular adjustments observed in practice.

Example 1. A bank divides observed defaults into segments

according to collateral type. One of the segments has too few

observations to produce a reliable estimate. Relying on external

data and judgment, the bank determines that the segment's estimated

severity of loss falls somewhere between the estimates for two other

categories. This segment's severity is set judgmentally to be the

mean of the estimates for the other segments. This practice is

consistent with the standard.

Example 2. A bank does not know when recoveries (and related

costs) occurred in a portfolio segment; therefore, it cannot

properly discount the segment's cash flows. However, the bank has

sufficient internal data to calculate economic loss for defaulted

facilities in another portfolio segment. The bank can support the

assumption that the timing of cash flows for the two segments is

comparable. Using the available data and informed judgment, the bank

estimates that the measured loss without discounting should be

grossed up to account for the time value of money and the

opportunity cost of funds. This practice is consistent with the

standard.

Example 3. A bank segments internal defaults in a business unit

by some factors, including collateral. Although the available

internal and external evidence indicates a higher LGD, the bank

judgmentally assigns a loss estimate of 2 percent for facilities

secured by cash collateral. The basis for this adjustment is that

the lower estimate is justified by the expectation that the bank

would do a better job of following policies for monitoring cash

collateral in the future. Such an adjustment is generally not

appropriate because it is based on projections of future performance

rather than realized experience. This practice is not consistent

with the standard.

Mapping

LGD mapping follows the same general principles that PD mapping

does. A mapping must be plausible and must be based on a comparison of

severity-related data elements common to both the reference data and

the current portfolio. The mapping approach is expected to be unbiased,

such that the exercise of judgment does not consistently lower LGD

estimates. The default definitions in the reference data and the

current portfolio of obligors should be comparable. The mapping process

must be updated regularly, well-documented, and independently reviewed.

S. A bank must conduct a robust comparison of available common

elements in the reference data and the portfolio.

Mapping involves matching facility-specific data elements available

in the current portfolio to the factors in the reference data set used

to estimate expected loss severity rates. Examples of factors that

influence loss rates include collateral type and coverage, seniority,

industry, and location.

At least three kinds of mapping challenges may arise. First, even

if similarly named variables are available in the reference data and

portfolio data, they may not be directly comparable. For example, the

definition of particular collateral types, or the meaning of

``secured,'' may vary from one application to another. Hence, a bank

must ensure that linked variables are truly similar. Although

adjustments to enhance comparability can be appropriate, they must be

rigorously developed and documented. Second, levels of aggregation may

vary. For example, the reference data may only broadly identify

collateral types, such as financial and nonfinancial. The bank's

information systems for its portfolio might supply more detail, with a

wide variety of collateral type identifiers. To apply the estimates

derived from the reference data, the internal data must be regrouped to

match the coarser level of aggregation in the reference data. Third,

reference data often do not include workout costs and will often use

different discounting. Judgmental adjustments for such problems must be

well-documented and, as much as possible, empirically based.

S. A mapping process must be established for each reference data

set and for each estimation model.

Mapping is never self-evident. Even when reference data are drawn

from internal default experience, a bank must still link the

characteristics of the reference data with those of the current

portfolio.

Different data sets and different approaches to severity estimation

may be entirely appropriate, especially for different business segments

or product lines. Each mapping process must be specified and

documented.

Application

At the application stage, banks apply the LGD estimation framework

to their current portfolio of credit exposures. Doing so might require

them to aggregate individual LGD estimates into broader averages (for

example, into discrete severity grades) or to combine estimates in

various ways.

The inherent variability of recovery, due in part to unanticipated

circumstances, demonstrates that no facility type is wholly risk-free,

regardless of structure, collateral type, or collateral coverage. The

existence of

[[Page 45966]]

recovery risk dictates that application of a zero percent LGD is not

acceptable.

S. IRB institutions that aggregate LGD estimates for severity

grades from individual exposures within those grades must have a clear

policy governing the aggregation process.

Banks with discrete severity grades compute a single estimate of

LGD for a representative exposure within each of those grades. If a

bank with a discrete scale of severity grades maps those grades to the

reference data using grade mapping, there will be a single estimate of

LGD for each grade, and the bank does not need to aggregate further.

However, if the bank maps at the individual transaction level, the bank

may then choose to aggregate those individual LGD estimates to the

grade level and use the grade LGD in capital calculations. Because

different methods of aggregation are possible, a bank must have a clear

policy regarding how aggregation should be accomplished; in general,

simple averaging is preferred. (This standard is irrelevant for banks

that choose to assign LGD estimates directly to individual exposures

rather than grades, because aggregation is not required in that case.)

S. An IRB institution must have a policy describing how it combines

multiple sets of reference data.

Multiple data sets may produce superior estimates of loss severity,

if the results are appropriately combined. Combining such sets

differently usually produces different estimates of LGD. As a matter of

internal policy, a bank should investigate alternative combinations,

and document the impact on the estimates. If the results are highly

sensitive to the manner in which different data sources are combined,

the bank must choose conservatively among the alternatives.

D. Exposure at Default (EAD)

Compared with PD and LGD quantification, EAD quantification is less

advanced. As such, it is addressed in somewhat less detail in this

guidance than are PD and LGD quantification. Banks should continue to

innovate in the area EAD estimation, refining and improving practices

in EAD measurement and prediction. Additional supervisory guidance will

be provided as more data become available and estimation techniques

evolve.

A bank must provide an estimate of expected EAD for each facility

in its portfolio. EAD is defined as the bank's expected gross dollar

exposure of the facility upon the obligor's default. For fixed

exposures like term loans, EAD is equal to the current amount

outstanding. For variable exposures such as loan commitments or lines

of credit, exposure is equal to current outstandings plus an estimate

of additional drawings up to the time of default. This additional

drawdown, identified as loan equivalent exposure (LEQ) in many

institutions, is typically expressed as a percentage of the current

total committed but undrawn amount. EAD can thus be represented as:

EAD = current outstanding + LEQ x (total committed-current outstanding)

As it is the LEQ that must be estimated, LEQ is the focus of this

guidance.

Even though EAD estimation is less sophisticated than PD and LGD

estimation, a bank still develops EAD estimates by working through the

four stages that produce the other types of quantification: The bank

must use a reference data set; it must apply an estimation technique to

produce an expected total dollar exposure at default for a facility

with a given array of characteristics; it must map its current

portfolio to the reference data; and, by applying the estimation model,

it must generate an EAD estimate for each portfolio facility or

facility-type, as the case may be.

Data

Like reference data sets used for LGD estimation, LEQ data sets

contain only exposures to defaulting obligors. In many cases, the same

reference data may be used for both LGD and LEQ. In addition to

relevant descriptive characteristics (referred to as ``drivers'') that

can be used in estimation, the reference data must include historical

information on the exposure (both drawn and undrawn amounts) as of some

date prior to default, as well as the drawn exposure at the date of

default.

As discussed below under ``Estimation,'' LEQ estimates may be

developed using either a cohort method or a fixed-horizon method. The

bank's reference data set must be structured so that it is consistent

with the estimation method the bank applies. Thus, the data must

include information on the total commitment, the undrawn amount, and

the exposure drivers for each defaulted facility, either at fixed

calendar dates for the cohort method or at a fixed interval prior to

the default date for the fixed-horizon method.

The reference data must contain variables that enable the bank to

group the exposures to defaulted obligors in meaningful ways. Obligor

and facility risk ratings are commonly believed to be significant

characteristics for predicting additional drawdown. Since less

empirical research has been done on EAD estimation, little is known

about other potential drivers of EAD. Among the many possibilities,

banks may consider time from origination, time to expiration or

renewal, economic conditions, risk rating changes, or certain types of

covenants. Some potential drivers may be linked to a bank's credit risk

management skills, while others may be exogenous. Industry practice is

likely to improve as banks extend their research to identify other

meaningful drivers of EAD.

A bank must ensure continued applicability of the reference data to

its current portfolio of facilities. The reference data must include

the types of variable exposures found in a bank's current portfolio.

The definitions of default and exposure in the reference data should be

consistent with the IRB definition of default, and consistent with the

definitions used for PD and LGD quantification. Established processes

must be in place to ensure that reference data sets are updated when

new data are available. All data sources, variables, and the overall

processes governing data collection and maintenance must be fully

documented, and that documentation should be readily available for

review.

Seven years of data are required for EAD (or LEQ) estimation. The

sample should include periods during which default rates were

relatively high, and ideally cover a complete economic cycle.

Estimation

To derive LEQ estimates, characteristics of the reference data are

related to additional drawings preceding a default event. The

estimation process must be capable of producing a plausible estimate of

LEQ to support the EAD calculation for each facility. Two broad types

of estimation methods are used in practice, the cohort method and the

fixed-horizon method.

Under the cohort method, a bank groups defaults into discrete

calendar periods (such as a year or a quarter). The bank then estimates

the relationship between the drivers as of the start of that calendar

period, and EAD or LEQ for each exposure to a defaulter. For each

exposure category (that is, for each combination of exposure drivers

identified by the bank), the LEQ estimate is calculated as the mean

additional drawing for facilities in that category. To combine results

for multiple periods into a single long-run average, the period-by-

period means should be weighted by the proportion of defaults occurring

in each period.

Under the fixed-horizon method, for each exposure to a defaulted

obligor the

[[Page 45967]]

bank compares additional drawdowns to the total commitment but undrawn

amount that existed at the start of a fixed interval prior to the date

of the default (the horizon). For example, the bank might base its

estimates on a reference data set that supplies the actual exposure at

default along with the drawn and undrawn amounts (as well as relevant

drivers) at a date a fixed number of months prior to the date of each

default, regardless of the actual calendar date on which the default

occurred. Estimates of LEQ are computed from the average drawdowns that

occur over the fixed-horizon interval, for whatever combinations of the

driving variables the bank has determined are relevant for explaining

and predicting exposure at default.

Evidence may indicate that LEQ estimates are positively correlated

with economic downturns; that is, it may be that LEQs increase during

high-default periods. If so, the higher drawdowns that occur during

high-default periods are denoted ``stress-condition LEQs,'' analogous

to the ``stress-condition LGDs'' discussed earlier in this chapter. For

any exposure type whose LEQ estimates exhibit material cyclicality, a

bank must use the stress-condition LEQ for purposes of calculating EAD.

In general, all available data should be used; particular

observations or time periods should not be excluded from the data

sample. Any adjustments a bank makes to the estimation results should

be justified and fully documented. The analysis should be refreshed

periodically as new data become available, and a bank should have a

process in place to ensure that advances in analytical techniques and

industry practice are considered as they emerge and are incorporated as

appropriate. LEQ estimates should be updated at least annually.

Detailed documentation, ongoing validation, and adequate oversight are

fundamental controls that support a sound estimation process.

Mapping

If the same variables that drive exposure in the reference data are

also available for facilities in the portfolio, mapping may be

relatively easy. However, the bank must still review the definitions to

ensure that variables that seem to be the same actually are. If the

relevant variables are not available in a bank's current portfolio

information system, the bank will encounter the same mapping

complexities that it does when mapping for PD and LGD in similar

circumstances. A bank should have well-documented policies that govern

the mapping. Any exceptions to mapping policy should be reviewed,

justified and fully documented. Mapping may be done for each exposure

or for broad categories of exposure; the latter would be analogous to

the ``grade mapping'' discussed earlier in this chapter.

Application

In the application stage, the estimated relationship between

drivers and LEQ is applied to the bank's actual portfolio. To ensure

that estimated EAD is at least as large as the currently drawn amount

for all exposures, LEQs must not be negative. Multiple reference data

sets may be used for LEQ estimation and combined at the application

stage; those combinations should be rigorously developed, approved, and

documented. Any smoothing or use of expert judgment to adjust the

results should be well-justified and clearly documented. This includes

any adjustment for definitions of default that do not meet the

supervisory standards. The less robust the process, the more

conservative the result should be.

Some facility types may be treated as exceptions, and assigned an

LEQ that does not vary with characteristics such as line of business or

risk rating. Such exceptional treatment should be clearly justified,

and the justification should be fully documented.

EAD may be particularly sensitive to changes in the way banks

manage individual credits. For example, a change in policy regarding

covenants may have a significant impact on LEQ. When such changes take

place, the bank should consider them when making its estimates--and it

should do so from a conservative point of view. Policy changes likely

to significantly increase LEQ should prompt immediate increases in LEQ

estimates. If a bank's policy changes seem likely to reduce LEQ,

estimates should be reduced only after the bank accumulates a

significant amount of actual experience under the new policy to support

the reductions.

E. Maturity (M)

A bank must assign a value of effective remaining maturity (M) to

each credit exposure in its portfolio. In general, M is the weighted-

average number of years to receipt of the cash flows the bank expects

under the contractual terms of the exposure, where the weights are

equal to the fraction of the total undiscounted cash flow to be

received at each date. Mathematically, M is given by:

[GRAPHIC] [TIFF OMITTED] TN04AU03.008

where wt is the fraction of the total cash flow received at

time t, that is:

[GRAPHIC] [TIFF OMITTED] TN04AU03.009

Ct is the undiscounted cash flow received at time t, with t

measured in years from the date of the calculation of M.

Effective maturity, sometimes referred to as ``average life,'' need

not be a whole number, and often is not. For example, if 33 percent of

the cash flow is expected at the end of one year (t=1) and the other 67

percent two years from today (t=2), then M is calculated as:

M = (1x0.33) + (2x0.67) = 1.67

for an effective maturity of 1.67 years. This value of M would be used

in the IRB capital calculation.

The relevant cash flows are the future payments the bank expects to

receive from the obligor, regardless of form; they may include payments

of interest or fees, principal repayments, or other types of payments

depending on the structure of the transaction. For exposures whose cash

flow schedule is virtually predetermined unless the obligors defaults

(fixed-rate loans, for example), the calculation of the weighted-

average remaining maturity is straightforward, using the scheduled

timing and amounts of the individual undiscounted cash flows. These

cash flows should be the contractually expected payments; the bank

should not take into account the possibility of delayed or reduced cash

flows due to potential future default.

Cash flows associated with other types of credit exposures may be

somewhat less certain. In such cases, the bank must establish a method

of projecting expected cash flows. In general, the method used for any

exposure should be the same as the one used by the bank for purposes of

valuation or risk management. The method must be well-documented and

subject to independent review and approval. A bank must demonstrate

that the method used is standard industry practice, that it is widely

used within the bank for purposes other than regulatory capital

calculations, or both.

To be conservative, a bank may set M equal to the maximum number of

years the obligor could take to fully discharge the contractual

obligation (provided that the maximum is not longer than five years, as

noted below). In many cases, this maximum will correspond to the stated

or nominal maturity of the instrument. Banks must make this

conservative choice (maximum nominal maturity) if the timing and

amounts of

[[Page 45968]]

the cash flows on the exposure cannot be projected with a reasonable

degree of confidence.

Certain over-the-counter derivatives contracts and repurchase

transactions may be subject to master netting agreements. In such

cases, the bank may compute a single value of M for the transactions as

a group by weighting each individual transaction's effective maturity

by that transaction's share of the total notional value subject to the

netting agreement, and summing the result across all of the

transactions.

For IRB capital calculations, the value of M for any exposure is

subject to certain upper and lower limits, regardless of the actual

effective maturity of the exposure. In all cases, the value of M should

be no greater than 5 years. If an exposure clearly has an effective

maturity that exceeds this upper limit, the bank may simply use a value

of M=5 rather than calculating the actual effective maturity.

For most exposures, the value of M must be no less than one year.

For certain short-term exposures (repo-style transactions, money market

transactions, trade finance-related transactions, and exposures arising

from payment and settlement processes) that are not part of a bank's

ongoing financing of a borrower and that have an original maturity of

less than three months, M may be set as low as one day. For over-the-

counter derivative and repurchase-style transactions subject to a

master netting agreement, weighted average maturity must be set at no

less than five days.

F. Validation

Values of PD, LGD, and EAD are estimates with implications for

credit risk and the future performance of a bank's credit portfolio

under IRB; in essence, they are forecasts. ``Validation'' of these

estimates describes the full range of activities used to assess their

quality as forecasts of default rates, loss severity rates, and

exposures at default. Chapter 1 discusses validation of IRB systems in

general; this section focuses specifically on ratings quantification,

which includes the assignment of PD to obligor grades and the

assignment of LGD, EAD, and M to exposures.

S. A validation process must cover all aspects of IRB

quantification.

Banks must have a process for validating IRB quantification; their

policies must state who is accountable for validation, and describe the

actions that will proceed from the different possible results.

Validation should focus on the three estimated IRB parameters (PD, LGD,

and EAD). Although the established validation process should result in

an overall assessment of IRB quantification for each parameter, it also

must cover each of the four stages of the quantification process as

described in preceding sections of this chapter (data, estimation,

mapping, and application). The validation process must be fully

documented, and must be approved by appropriate levels of the bank's

senior management. The process must be updated periodically to

incorporate new developments in validation practices and to ensure that

validation methods remain appropriate; documentation must be updated

whenever validation methods change.

Banks should use a variety of validation approaches or tools; no

single validation tool can completely and conclusively assess IRB

quantification. Three broad types of tools that are useful in this

regard are evaluation of the conceptual soundness of the approach to

quantification (evaluation of logic), comparison to other sources of

data or estimates (benchmarking), and comparisons of actual outcomes to

predictions (back-testing). Each of these types of tools has a role to

play in validation, although the role varies across the four stages of

quantification.

Evaluation of logic is essential in validating all stages of the

quantification process. The quantification process requires banks to

adopt methods, choose variables, and make adjustments; each of these

actions requires an exercise of judgment. Validation should ensure that

these judgments are plausible and informed.

A bank should also validate estimates by comparing them with

relevant external sources, a process broadly described as benchmarking.

``External'' in this context refers to anything other than the specific

reference data, estimation approach, or mapping under consideration.

Reference data can be compared with other data sources; choices of

variables can be compared with similar choices made by others;

estimation results can be compared with the results of alternative

estimation methods using the same reference data. Other data sources

may show that default and severity rates across the economy or the

banking system are high or low relative to other periods, or may reveal

unusual effects in parts of the quality spectrum.

Effective validation must compare actual results with predictions.

Such comparisons, often referred to as ``back-testing,'' are valuable

comprehensive tests of the rating system and its quantification.

However, they are only one element of the broader validation regime,

and should not be a bank's only method of validation. Because they test

the results of the rating system as a whole, they are unlikely to

identify specific reasons for any divergence between expectations and

realizations. Rather they will indicate only that further investigation

is necessary.

By applying back-testing to the reference data set as it is updated

with new data, a bank can improve the estimation process. To further

improve the process, a bank must regularly compare realized default

rates, loss severities, and exposure-at-default experience from its

portfolio with the PD, LGD, and EAD estimates on which capital

calculations are based. Realizations should be compared with expected

ranges based on the estimates. These expected ranges should take into

account the bank's rating philosophy (the relative weight given to

current and stress conditions in assigning ratings). Depending on that

philosophy, year-by-year realized default rates and loss severities may

be expected to differ significantly from the long-run average. If a

bank adjusts final estimates to be conservative, it should likely do

its back-testing on the unadjusted estimates.

A bank's quantitative testing methods and other validation

techniques should be robust to economic cycles. A sound validation

process should take business cycles into account, and any adjustments

for stages of the cycle should be clearly specified in advance and

fully documented as part of the validation policy. The fact that a year

has been ``unusual'' should not be taken as a reason to abandon the

bank's standard validation practices.

S. A bank must comprehensively validate parameter estimates at

least annually, must document the results, and must report these

results to senior management.

A full and comprehensive annual validation is a minimum for

effective risk management under IRB. More frequent validation may be

appropriate for certain parts of the IRB system and in certain

circumstances; for example, during high-default periods, banks should

compute realized default and loss severity rates more frequently,

perhaps quarterly. They must document the results of validation, and

must report them to appropriate levels of senior risk management.

S. The validation policy must outline appropriate remedial

responses to the results of parameter validation.

The goal of validation should be to continually improve the rating

process and its quantification. To this end, the bank should establish

thresholds or accuracy tolerances for validation results. Results that

breach thresholds

[[Page 45969]]

should bring an appropriate response; that response should depend on

the results and should not necessarily be to adjust the parameter

estimates. When realized default, severity, or exposures rates diverge

from expected ranges, those divergences may point to issues in the

estimation or mapping elements of quantification. They may also

indicate potential problems in other parts of the ratings assignment

process. The bank's validation policy must describe (at least in broad

terms) the types of responses that should be considered when relevant

action thresholds are crossed.

Appendix to Part III: Illustrations of the Quantification Process

This appendix provides examples to show how the logical

framework described in this guidance, with its four stages (data,

estimation, mapping, and application), applies when analyzing

typical current bank practices. The framework is broadly

applicable--for PD or LGD or EAD; using internal, external, or

pooled reference data; for simple or complex estimation methods--

although the issues and concerns that arise at each stage depend on

a bank's approach. These examples are intended only to illustrate

the logic of the four-stage IRB quantification framework, and should

not be taken to endorse the particular techniques presented in the

examples. In fact, certain aspects of the examples are not

consistent with the standards outlined in this guidance.

Example 1: PD Estimation From Bond Data

[sbull] A bank establishes a correspondence between its internal

grades and external rating agency grades; the bank has determined

that its Grade 4 is equivalent to \3/4\ BB and \1/4\ B on the

Standard and Poor's scale.

[sbull] The bank regularly obtains published estimates of mean

default frequencies for publicly rated BB and B obligors in North

America from 1970 through 2002.

[sbull] The BB and B historical default frequencies are weighted

75/25, and the result is a preliminary PD for the bank's internal

Grade 4 credits.

[sbull] However, the bank then increases the PD by 10 percent to

account for the fact that the S&P definition of default is more

lenient than the IRB definition.

[sbull] The bank makes a further adjustment to ensure that the

resulting grade PD is greater than the PD attributed to Grade 3 and

less than the PD attributed to Grade 5.

[sbull] The result is the final PD estimate for Grade 4.

Process Analysis for Example 1

Data--The reference data set consists of issuers of publicly

rated debt in North America over the period 1970 through 2002. The

data description is very basic: each issuer in the reference data is

described only by its rating (such as AAA, AA, A, BBB, and so on).

Estimation--The bank could have estimated default rates itself

using a database purchased from Standard and Poor's, but since these

estimates would just be the mean default rates per year for each

grade, the bank could just as well (and in this example does) use

the published historical default rates from S&P; in essence, the

estimation step has been outsourced to S&P. The 10 percent

adjustment of PD is part of the estimation process in this case

because the adjustment was made prior to the application of the

agency default rates to the internal portfolio data.

Mapping--The bank's mapping is an example of a grade mapping;

internal Grade 4 is linked to the 75/25 mix of BB and B. Based on

the limited information presented in the example, this step should

be explored further. Specifically, how did the bank determine the

75/25 mix?

Application--Although the application step is relatively

straightforward in this case, the bank does make the adjustment of

the Grade 4 PD estimate to give it the desired relationship to the

adjacent grades. This adjustment is part of the application stage

because it is made after the adjusted agency default rates are

applied to the internal grades.

Example 2: PD Estimation Using a Merton-Type Equity-Based Model

[sbull] A bank obtains a 20-year database of North American

firms with publicly traded equity, some of which defaulted during

the 20-year period.

[sbull] The bank uses the Merton approach to modeling equity in

these firms as a contingent claim, constructing an estimate of each

firm's distance-to-default at the start of each year in the

database. The bank then ranks the firm-years within the database by

distance-to-default, divides the ordered observations into 20 equal

groups or buckets, and computes a mean historical one-year default

frequency for each bucket. That default frequency is taken as an

estimate of the applicable PD for any obligor within the range of

distance-to-default values represented by each of the 20 buckets.

[sbull] The bank next looks at all obligors with publicly traded

shares within each of its internal grades, applies the same Merton-

type model to compute distance-to-default at quarter-end, sorts

these observations into the 20 buckets from the previous step, and

assigns the corresponding PD estimate.

[sbull] For each internal grade, the bank computes the mean of

the individual obligor default probabilities and uses that average

as the grade PD.

Process Analysis for Example 2

Data--The reference data set consists of the North American

firms with publicly traded equity in the acquired database. The

reference data are described in this case by a single variable,

specifically an identifier of the specific distance-to-default range

from the Merton model (one of the 20 possible in this case) into

which a firm falls in any year.

Estimation--The estimation step is simple: the average default

rate is calculated for each distance-to-default bucket. Since the

data cover 20 years and a wide range of economic conditions, the

resulting estimates satisfy the long-run average requirement.

Mapping--The bank maps selected portfolio obligors to the

reference data set using the distance-to-default generated by the

Merton model. However, not all obligors can be mapped, since not all

have traded equity. This introduces an element of uncertainty into

the mapping that requires additional analysis by the bank: were the

mapped obligors representative of other obligors in the same grade?

The bank would need to demonstrate comparability between the

publicly traded portfolio obligors and those not publicly traded. It

may be appropriate for the bank to make conservative adjustments to

its ultimate PD estimates to compensate for the uncertainty in the

mapping. The bank also would need further analysis to demonstrate

that the implied distance-to-default for each internal grade

represented long-run expectations for obligors assigned to that

grade; this could involve computing the Merton model for portfolio

obligors over several years of relevant history that span a wide

range of credit conditions.

Application--The final step is aggregation of individual

obligors to the grade level through calculation of the mean for each

grade, and application of this grade PD to all obligors in the

grade. The bank might also choose to modify PD assignments further

at this stage, combining PD estimates derived from other sources,

applying adjustments for cyclicality, introducing an appropriate

degree of conservatism, or making other adjustments.

Example 3: LGD Estimation From Internal Default Data

[sbull] For each loan in its portfolio, a bank records

collateral coverage as a percentage, as well as which of four types

of collateral applies.

[sbull] A bank has retained data on all defaulted loans since

1995. For each defaulted loan in the database, the bank has a record

of the collateral type within the same four broad categories.

However, collateral coverage is only recorded at three levels (low,

moderate, or high, depending on the ratio of collateral to exposure

at default).

[sbull] The bank also records the timing and discounted value of

recoveries net of workout costs for each defaulted loan in the

database. Cash flows are tracked from the date of default to a

``resolution date,'' defined as the point at which the remaining

balance is less than 5 percent of the exposure at the time of

default. A recovery percentage is computed, equal to the value of

recoveries discounted to the date of default, divided by the

exposure at default.

[sbull] For each cell (each of the 12 combinations of collateral

type and coverage), the bank computes a simple mean LGD percentage

as the mean of one minus the recovery percentage. One of the

categories has a mean LGD of less than zero (recoveries have

exceeded exposure on average), so the bank sets the LGD at zero to

be conservative.

[sbull] The bank assigns an estimate of expected LGD to each

loan in the current portfolio by using collateral information to

slot it into one of the 12 cells. The bank then applies the mean

historical LGD for that cell and adjusts the result upward by 10

percent to compensate for the fact that the loss data come from a

period believed to be unusually good economic performance.

[[Page 45970]]

Process Analysis for Example 3

Data--The reference data is the collection of historical

defaults with the loss amounts from the bank's historical portfolio.

The reference data are described by the two categorical variables

(levels of collateral coverage and types of collateral). It would be

important to determine whether the defaults over the past few years

are comparable to defaults from the current portfolio. One would

also want to ask why the bank ignores potentially valuable

information by converting the continuous data on collateral coverage

into a trimodal categorical variable.

Estimation--Conceptually, the bank is using a ``loss severity

model'' in which 12 binary variables, one for each loan coverage/

type combination, explain the percentage loss. The coefficients on

the variables are just the mean loss figures from the reference

data.

Mapping--Mapping in this case is fairly straightforward, since

all of the relevant characteristics of the reference data are also

in the loan system for the current portfolio. However, the bank

should determine whether the variables are being recorded in the

same way (for example, the same definitions of collateral types),

otherwise some adjustment might be needed.

Application--The bank is able to apply the loss model by simply

plugging in the relevant values for the current portfolio (or what

amounts to the same thing, looking up the cell mean). The bank's

assignment of zero LGD for one of the cells merits special

attention; while the bank represented this assignment as

conservative, the adjustment does not satisfy the supervisory

requirement that LGD must exceed zero. A larger upward adjustment is

necessary. Finally, the upward adjustment of the LGD numbers to

account for the benign environment in which the reference data were

generated presents one additional wrinkle. The bank must provide a

well-documented, empirically based analysis of why a 10 percent

upward adjustment is sufficient.

IV. Data Maintenance

A. Overview

Institutions using the IRB approach for regulatory capital purposes

will need advanced data management practices to produce credible and

reliable risk estimates. The guiding principle governing an IRB data

maintenance system is that it must support the requirements for the

quantification, validation, control and oversight mechanisms described

in this guidance, as well as the institution's broader risk management

and reporting needs. The precise data elements to be collected will be

dictated by the features and methodology of the IRB system employed by

the institution. The necessary data elements will therefore vary by

institution and even among business lines within an institution.

Institutions will have latitude in managing their data, subject to

the following key data maintenance standards:

Life Cycle Tracking--institutions must collect, maintain, and

analyze essential data for obligors and facilities throughout the life

and disposition of the credit exposure.

Rating Assignment Data--institutions must capture all significant

quantitative and qualitative factors used to assign the obligor and

loss severity ratings.

Support of IRB System--data collected by institutions must be of

sufficient depth, scope, and reliability to:

[sbull] Validate IRB system processes,

[sbull] Validate parameters,

[sbull] Refine the IRB system,

[sbull] Develop internal parameter estimates,

[sbull] Apply improvements historically,

[sbull] Calculate capital ratios,

[sbull] Produce internal and public reports, and

[sbull] Support risk management.

This chapter covers the requirements for maintaining internal data.

Reference data sets used for estimating IRB parameters are discussed in

Chapter 2.

B. Data Maintenance Framework

Life Cycle Tracking

S. Institutions must collect, maintain, and analyze essential data

for obligors and facilities throughout the life and disposition of the

credit exposure.

Using a life cycle or ``cradle to grave'' concept for each obligor

and facility supports front-end validation, back-testing, system

refinements and risk parameter estimates. A depiction of life-cycle

tracking follows:

[GRAPHIC] [TIFF OMITTED] TN04AU03.001

Data elements must be recorded at origination and whenever the

rating is reviewed, regardless of whether the rating is actually

changed. Data elements associated with current and past ratings must be

retained and include the following:

[sbull] Key borrower and facility characteristics,

[sbull] Ratings for obligor and loss severity grades,

[sbull] Key factors used to assign the ratings,

[sbull] Person or model responsible for assigning the rating,

[sbull] Date rating assigned, and

[sbull] Overrides to the rating and authorizing individual.

At disposition, data elements must include:

[sbull] Nature of disposition: renewal, repayment, loan sale,

default, restructuring,

[sbull] For defaults: exposure, actual recoveries, source of

recoveries, costs of workouts and timing,

[sbull] Guarantor support,

[sbull] Sale price for loans sold, and

[sbull] Other key elements that the bank deems necessary.

[[Page 45971]]

Rating Assignment Data

S. Institutions must capture all significant quantitative and

qualitative factors used to assign the obligor and loss severity

rating.

Assigning a rating to an obligor requires the systematic collection

of various borrower characteristics as these factors are critical to

validating the rating system. Obligors are rated using various methods,

as discussed in Chapter 1. Each of these methods presents different

challenges for input collection. For example, in judgmental rating

systems, the factors used in the ratings decision have not

traditionally been explicitly recorded. For purposes of an IRB

approach, institutions that use expert and constrained judgment must

record these factors and deliver them to the data warehouse.

For loss severity estimates, institutions must record the basic

structural characteristics of facilities and the factors used in

developing the facility rating or LGD estimate. These often include the

seniority of the credit, the amount and type of collateral, the most

recent collateral valuation date and its fair value.

Institutions must also track any overrides of the obligor or loss

severity rating. Tracking overrides separately allows risk managers to

identify whether the outcome of such overrides suggests either problems

with rating criteria, or an improper level of discretion in adjusting

the ratings.

Example Data Elements

For illustrative purposes, the following section provides examples

of the kinds of data elements institutions will collect under an IRB

data maintenance framework.

General descriptive obligor and facility data

The data below could be contained within a loan record or derived

from various sources within the data warehouse. Guarantor data

requirements are the same as for the obligor.

Obligor/Guarantor Data

[sbull] General data: name, address, industry

[sbull] ID number (unique for all related parent/sub relationships)

[sbull] Rating, date, and rater

[sbull] PD percentage corresponding to rating

General Facility Characteristics

[sbull] Facility amounts: committed, outstanding

[sbull] Facility type: Term, revolver, bullet, amortizing, etc.

[sbull] Purpose: acquisition, expansion, liquidity, inventory,

working capital

[sbull] Covenants

[sbull] Facility ID number

[sbull] Origination and maturity dates

[sbull] Last renewal date

[sbull] Obligor ID link

[sbull] Rating, date and rater

[sbull] LGD dollar amount or percentage

[sbull] EAD dollar amount or percentage

Rating Assignment Data

The data below provide an example of the categories and types of

data that institutions must retain in order to continually validate and

improve rating systems. These data items should tie directly to the

documented criteria that the institution employs in assigning ratings,

both qualitative and quantitative. For example, rating criteria often

include ranges of leverage or cash flow for a particular obligor

rating. In addition, qualitative factors, such as management

effectiveness can be recorded in numeric form. For example, a 1 may

equate to exceptionally strong management, and a 5 to very weak. The

rating data elements collected should be complete enough so that others

can review the relevant factors driving the rating decisions.

Quantitative Factors in Obligor Ratings

[sbull] Asset and sale size

[sbull] Key ratios used within rating criteria:

--profitability,

--cash flow,

--leverage,

--liquidity, and

--other relevant factors.

Qualitative Factors in Obligor Ratings

[sbull] Quality of earnings and cash flow

[sbull] Management effectiveness, reliability

[sbull] Strategic direction, industry outlook, position

[sbull] Country factors and political risk

[sbull] Other relevant factors

External Factors in Obligor Ratings

[sbull] Public debt rating and trend

[sbull] External credit model score and trend

Rating Notations

[sbull] Flag for overrides or exceptions

[sbull] Authorized individual for changing rating

Key Facility Factors in LGD Ratings

[sbull] Seniority

[sbull] Collateral type: (cash, marketable securities, AR, stock,

RE, etc.)

[sbull] Collateral value and valuation date

[sbull] Advance rates, LTV

[sbull] Industry

[sbull] Geography

Rating Notations

[sbull] Flag for overrides or exceptions

[sbull] Authorized individual for changing rating

Final Disposition Data

Only recently have institutions begun to collect more complete data

about a loan's disposition. Many institutions maintain subsidiary

systems for their problem credits with details recorded, at times

manually, on systems that were not linked with the institution's

central loan or risk management systems. The unlinked data are a

significant hindrance in developing reliable PD, LGD, and EAD

estimates.

In advanced systems, the ``grave'' portion of obligor and exposure

tracking is an essential component for producing and validating risk

estimates and is an important feedback mechanism for adjusting and

improving risk estimates over time. Essential data elements are

outlined below.

Obligor/Guarantor

[sbull] Default date

[sbull] Circumstances of default (for example, nonaccrual,

bankruptcy chapters 7-11, nonpayment)

Facility

[sbull] Outstandings at default

[sbull] Amounts undrawn and outstanding plus time series prior to

and through default

Disposition

[sbull] Amounts recovered and dates (including source: cash,

collateral, guarantor, etc.)

[sbull] Collection cost and dates

[sbull] Discount factors to determine economic cost of collection

[sbull] Final disposition (for example, restructuring or sale)

[sbull] Sales price, if applicable

[sbull] Accounting items (charge-offs to date, purchased discounts)

C. Data Element Functions

S. Data elements must be of sufficient depth, scope, and

reliability to:

[sbull] Validate IRB system processes,

[sbull] Validate parameters,

[sbull] Refine the IRB system,

[sbull] Develop internal parameter estimates,

[sbull] Apply improvements historically,

[sbull] Calculate capital ratios,

[sbull] Produce internal and public reports, and

[sbull] Support risk management.

Validation and Refinement

The data elements collected by institutions must be capable of

meeting

[[Page 45972]]

the validation requirements described in Chapters 1 and 2. These

requirements include validating the institution's IRB system processes,

including the ``front end'' aspects such as assigning ratings so that

any issues can be identified early. The data must support efforts to

identify whether raters and models are following rating criteria and

policies and whether ratings are consistent across portfolios. In

addition, data must support the validation of parameters, particularly

the comparison of realized outcomes with estimates. Thorough data on

default and disposition characteristics are of paramount importance for

parameter back-testing.

A rich source of data for validation efforts provides insights on

the performance of the IRB system, and contributes to a learning

environment in which refinements can be made to the system. These

potential refinements include enhancements to rating assignment

controls, processes, criteria or model coefficients, rating system

architecture and parameter estimates.

Developing Parameter Estimates

As detailed in Chapter 2, institutions will be developing their PD,

LGD, and EAD parameter estimates using reference data sets comprised of

internal, pooled, and external data. Institutions are expected to work

toward eventually using as much of their own experience as possible in

their reference data sets.

Applying Rating System Improvements Historically

For loss severity estimates, institutions must record the basic

structural characteristics of facilities and the factors used in

developing the facility rating or LGD estimate. These often include the

seniority of the credit, the amount and type of collateral, the most

recent collateral valuation date and its fair value.

To maintain a consistent series of information for credit risk

monitoring and validation purposes, institutions need to be able to

apply historically improvements they make to their rating systems. In

the example below, a bank experiences unexpected and rapid migrations

and defaults in its grade 4 category during 2006. Analysis of the

actual financial condition of borrowers that defaulted compared with

those that did not suggests the debt-to-EBITDA range for its expert

judgment criteria of 3.0 to 5.5 is too broad. Research indicates that

grade 4 should be redefined to include only borrowers with debt-to-

EBITDA ratios of 3.0-4.5 and grade 5 as 4.5-6.5. In 2007, the change is

initiated, but prior years' numbers are not recast (see Exhibit A).

Consequently, a break in the series prevents the bank from evaluating

credit quality changes over several years and from identifying whether

applying the new rating criteria historically provides reasonable

results.

[GRAPHIC] [TIFF OMITTED] TN04AU03.007

Recognizing the need to provide senior managers and board members

with a consistent risk trend, the new criteria are applied historically

to obligors in grades 4 and 5 as reflected in Exhibit B. The original

ratings assigned to the grades are maintained along with notations

describing what the grade would be under the new rating criteria. If

the precise weight an expert has given one of the redefined criteria is

unknown, institutions are expected to make estimates on a best efforts

basis. After the retroactive reallocation process, the bank observes

that the mix of obligors in grade 5 declined somewhat over the past

several years while the mix in grade 4 increased slightly. This

contrasts with the trend identified before the retroactive

reallocation. The result is that the multiyear transition statistics

for grades 4 and 5 provide risk managers a clearer picture of risk.

[[Page 45973]]

[GRAPHIC] [TIFF OMITTED] TN04AU03.002

This example is based on applying ratings historically using data

already collected by the bank. However, for some rating system

refinements, institutions may identify in the future drivers of default

or loss that might not have been collected for borrowers or facilities

in the past. That is why institutions are encouraged to collect data

that they believe may serve as a stronger predictor of default in the

future. For example, certain elements of a borrower's cash flow might

currently be suspected to overstate actual operational health for a

particular industry. In the future, should an institution decide to

deduct this item from cash flow with a resulting downgrade of many

obligor ratings, the institution that collected these data could apply

this rating change for prior years. This would provide the benefit of

providing a consistent picture of risk over time and also present

opportunities to validate the new criteria using historical data.

Recognizing that institutions will not be able to anticipate fully the

data they might find useful in the future, institutions are expected to

reallocate grades on a best efforts basis when practical.

Calculating Capital Ratios and Reporting to the Public

Data retained by the bank will be essential for regulatory risk-

based capital calculations and public reporting under the Pillar 3

disclosures. These uses underscore the need for a well-defined data

maintenance framework and strong controls over data integrity. Control

processes and data elements themselves should also be subject to

periodic verification and testing by internal and external auditors.

Supervisors will rely on these processes and also perform testing as

circumstances warrant.

Supporting Risk Management

The information that can be gleaned from more extensive data

collection will support a broad range of risk management activities.

Risk management functions will rely on accurate and timely data to

track credit quality, make informed portfolio risk mitigation

decisions, and perform portfolio stress tests. Trends developed from

obligor and facility risk rating data will be used to support internal

capital allocation models, pricing models, ALLL calculations, and

performance management measures, among others. Summaries of these are

included in reports to institutions' boards of directors, regulators,

and in public disclosures.

D. Managing Data Quality and Integrity

Because data are collected at so many different stages involving a

variety of groups and individuals, there are numerous challenges to

ensuring the quality of the data. For example:

[sbull] Data will be retained over long timeframes,

[sbull] Qualitative risk-rating variables will have subjective

elements and will be open to interpretation, and

[sbull] Exposures will be acquired through mergers and purchases,

but without an adequate and easily retrievable institutional rating

history.

Documentation and Definitions

S. Institutions must document the process for delivering, retaining

and updating inputs to the data warehouse and ensuring data integrity.

Given the many challenges presented by data for an IRB system, the

management of data must be formalized. Fully documenting how the

institution's flow of data is managed provides a means for evaluating

whether the data maintenance framework is functioning as intended.

Moreover, institutions must be able to communicate to individuals

developing or delivering various data the precise definition of the

items intended to be collected. Consequently, a ``data dictionary'' is

necessary to ensure consistent inputs from individuals and data vendors

and to allow third parties (such as the rating system review function,

auditors, or bank supervisors) to evaluate data quality and integrity.

S. Institutions must develop comprehensive definitions for the data

elements used within each credit group or business line (a ``data

dictionary'').

Electronic Storage

S. Institutions must store data in electronic format to allow

timely retrieval for analysis, validation of risk rating systems, and

required disclosures.

To meet the significant data management challenges presented by the

validation and control features of an IRB system, institutions will

need to store their data electronically. Institutions will have a

variety of storage techniques and potentially a variety of systems to

create their data

[[Page 45974]]

warehouses. IRB data requirements can be achieved by melding together

existing accounting, servicing, processing, workout and risk management

systems, provided the linkages among these systems are well documented

and include sufficient edit and integrity checks to ensure the data can

be used reliably.

Institutions without electronic databases would need to resort to

manual reviews of paper files for ongoing back-testing and ad hoc

``forensic'' data mining and would be unable to perform that work in

the timely and comprehensive manner required of IRB systems. Forensic

mining of paper files to build an initial data warehouse from the

institution's credit history is encouraged. In some instances, paper

research may be necessary to identify data elements or factors not

originally considered significant in estimating the risk of a

particular class of obligor or facility.

Data Gaps

Rating histories are often lost or are irretrievable for loans

acquired through mergers, acquisitions, or portfolio purchases.

Institutions are encouraged wherever practical to collect any missing

historical rating assignment driver data and to re-grade the acquired

obligors and facilities for prior periods. In cases where retrieving

historical data is not practical, institutions may attempt to create a

rating history through a careful mapping of the legacy system and the

new rating structure. Mapped ratings should be reviewed thoroughly for

accuracy. The level of effort placed on filling data gaps should be

commensurate with the size of the new exposures to be newly

incorporated into the institution's IRB system.

V. Control and Oversight Mechanisms

A. Overview

Banks' internal rating systems are the foundation for credit-risk

management practices and play an important role in pricing, reserving,

portfolio management, performance measurement, economic capital

modeling, and long-term capital planning. Banks adopting the IRB

approach will also use their credit-risk ratings to determine

regulatory capital levels. The pivotal and varied uses of such risk

ratings put enormous, sometimes conflicting, pressure on banks'

internal rating systems. The consequences of inaccurate ratings and

their associated estimates are significant, particularly as they affect

minimum regulatory capital requirements.

As risk ratings and their related parameters become better

integrated in institutions' decision making, conflicting incentives

arise that, if not well managed, can lead to overly optimistic or

biased ratings. For example, sales and marketing staff (relationship

managers or RMs) are typically compensated according to the volume of

business they generate. That may predispose the RMs to assign more

favorable ratings in order to achieve rate-of-return and sales

objectives. More favorable ratings may create the appearance of higher

risk-adjusted returns and business line profitability. Banks need to be

aware of the full range of incentive conflicts that arise, and must

develop effective controls to keep these incentive conflicts in check.

Banks will have latitude in designing and implementing their

control structures subject to the following principle:

IRB institutions must implement a system of controls that includes

the following elements: independence, transparency, accountability, use

of ratings, rating system review, internal audit, and board and senior

management oversight. While banks will have flexibility in how these

elements are combined, they must incorporate sufficient checks and

balances to ensure that the credit risk management system is

functioning properly.

Banks additionally will want to embody the following more generic

principles in their control system: separation of duties, balancing

incentives, and layers of review. Table 4.1 lists the key components of

an IRB control and oversight system. How these control mechanisms can

best be combined to reinforce one another is a key challenge for banks

implementing IRB systems:

Table 4.1 Control and Oversight Mechanisms

[GRAPHIC] [TIFF OMITTED] TN04AU03.003

[[Page 45975]]

As the following examples indicate, how a bank conducts its

business will influence how it designs its control structure. A bank

using an expert-judgment system will likely establish a different set

of controls than a bank using mainly models. Recognizing that its

expert-judgment system is less than fully transparent, a bank could

offset this vulnerability by opting for complete independence in the

rating approval process and an enhanced rating system review.

Other considerations would influence the choice of controls when

banks use models to assign ratings. While the ratings produced by

models are transparent, a model's performance depends on how well the

model was developed, the model's logic, and the quality of the data

used to implement the model. Banks that use models to assign ratings

must implement a system of controls that addresses model development,

testing and implementation, data integrity and overrides. These

activities would be covered by a comprehensive and independent rating

system review and by ongoing spot checks on the accuracy of model

inputs. Other control mechanisms such as accountability and audit would

also be required.

B. Independence in the Rating Approval Process

An independent rating process is one in which the parties

responsible for approving ratings and transactions are separate from

sales and marketing and in which the persons approving ratings are

principally compensated on risk-rating accuracy. As relative

independence increases, the likelihood of accurate ratings assignments

grows markedly.

S. Ratings must be subject to independent approval or review.

One way institutions can better achieve objective and accurate risk

ratings is by ensuring that its rating approval process is independent.

Institutions that firmly separate sales/marketing from credit are

better able to manage the conflict between the goal of high sales

volume and the need for good credit quality. An institution whose

rating process is less independent must compensate by strengthening

other control and oversight mechanisms. A significant factor in the

evaluation of the rating system will be the assessment of whether such

compensating controls are sufficient to offset a less-than-independent

ratings process. While the overriding objective is to achieve

independence in the rating approval process, in some instances, the

relative materiality of a portfolio and cost/benefit trade-offs may

support a less rigorous control process.

The degree of independence achieved in the rating process depends

on how an institution is organized and how it conducts its lending

activities.

Rating Approval Processes

Responsibility for recommending and approving ratings varies by

institution and, quite often, by portfolio.\7\ At some institutions,

ratings are assigned and approved by relationship managers (RMs); at

others, deal teams assign ratings that are later approved by credit

officers. Still other institutions have independent credit officers

assign and approve ratings. The culture of an institution and its

business mix generally determine whether the business line or credit

function is ultimately responsible for ratings.

---------------------------------------------------------------------------

\7\ Rating processes vary by institution but generally involve

an ``assignor'' and an ``approver.'' For instance, at many

organizations the rating assignor is the person who ``owns'' the

relationship (such as a ``relationship manager'') and the rating

approver is an individual with credit authority (a ``credit risk

manager''). In some cases, the rating assignor and approver are the

same. Banks that separate the rating assignment and approval

processes do so in order to minimize potential conflicts of interest

and the potential for rating errors.

---------------------------------------------------------------------------

The subsections that follow describe various rating assignment and

approval structures used by banking organizations and the challenges

that emerge in ensuring objective and consistent ratings. Any of the

following structures can work as long as ratings are subject to an

independent approval or review process, and are not unduly influenced

by the line of business:

Relationship Managers. As noted earlier, relationship managers are

primarily responsible for marketing the bank's products and services,

and their compensation is tied to the volume of business they generate.

When RMs also have responsibility for assigning and approving ratings,

there is an inherent conflict of interest. Credit quality and the

ability to produce timely and accurate risk ratings are generally not

major factors in an RM's compensation, even when he or she has

responsibility for assigning and approving ratings. In addition, RMs

also may become too close to the borrower to maintain their objectivity

and remain unbiased. When banks delegate rating responsibility to RMs,

they must offset the lack of independence with rigorous controls to

prevent bias from affecting the rating process. Such controls must

operate in practice, not just on paper, and would include, at a

minimum, a comprehensive, independent post-closing review of ratings by

a rating system review function.

Deal Team. Some major banks employ a ``deal-team'' structure for

credit origination and rating assignment. Using this approach, all

members of the team--credit officers, investment bankers, underwriters,

and others--contribute to analyzing creditworthiness, underwriting the

deal, and assigning ratings.

On the one hand, deal teams increase the access of credit officers

to information on obligors and transactions early in the underwriting

process, enabling them to make more informed credit decisions and to

influence facility structure to address obligors' weaknesses. On the

other hand, participation in the deal team could compromise the credit

officer's objectivity. While credit officers typically report to an

independent credit-risk-management function, they also have allegiance

to the deal team that reports to executives within the sales and

marketing line of business. In addition, credit officers may defer to

the members of the team whose compensation is based on the revenue and

sales volume they generate for the bank. Banks that maintain deal teams

must ensure that the credit officer's independence is safeguarded

through independent reporting lines and well-defined performance

measures (e.g., adherence to policy, rating accuracy and timeliness).

Credit Officers. Some banks give sole responsibility for assigning

and approving ratings to credit officers who report to an independent

credit function. In addition to assigning and approving and assigning

initial ratings, credit officers regularly monitor the condition of

obligors and refresh ratings as necessary. The potential downside of

this structure is that these credit officers may have limited access to

borrower information. Those credit officers that have a separate

reporting line and whose compensation is principally based on their

risk-rating accuracy are typically more independent than RMs or deal

teams.

Models. At some institutions, models assign ratings directly; at

other institutions, models and judgment are combined to rate credits.

Models introduce a high degree of independence to the rating process,

but they too require human oversight and controls. Banks that use

models must incorporate an independent judgmental review of the rating

assignments to ensure that all relevant information is considered and

to identify potential rating errors. Judgmental reviews are also needed

when model outputs are

[[Page 45976]]

overridden. In addition, controls are needed to ensure accuracy of data

inputs. When a bank uses a model to assign risk ratings, an individual

obligor's rating is ``transparent.'' However, the model itself is not

``transparent'' without a great deal of effort to document how the

model functions.

C. Transparency

Transparency is the ability of a third party, such as rating system

reviewers, auditors or bank supervisors, to observe how the rating

system operates and to understand the pertinent characteristics of

individual ratings.

S. IRB institutions must have a transparent rating system.

Transparency in a rating system is achieved through documentation

that covers the following:

[sbull] The rating system's design, purpose, performance horizon,

and performance standards;

[sbull] The rating assignment process, including procedures for

adjustments and overrides;

[sbull] Rating definitions and criteria, scorecard criteria, and

model specifications;

[sbull] Parameter estimates and the process for their estimation;

[sbull] Definition of the data elements to be warehoused to support

controls, oversight, validation, and parameter estimation; and

[sbull] Specific responsibilities of, and performance standards

for, individuals and units involved in the rating system and its

oversight.

Transparency allows third parties (such as rating system review,

auditors, or supervisors) to evaluate whether the rating system is

performing as intended. Without transparency, it is difficult to hold

people accountable for ratings errors and to validate the performance

of the system.

S. Rating criteria must be clear and specific and must include

qualitative and quantitative factors.

To produce transparent individual ratings, a bank's policies must

contain clear, detailed ratings definitions. Banks should specify

criteria for each factor that raters must consider, which may require

unique rating definitions for certain industries. Banks should consider

criteria for factors such as liquidity, sales and profitability, debt

service and fixed charge coverage, minimum equity support, position

within the industry, strength of management. A rating system with vague

criteria or one merely defined by PDs or LGDs is not transparent. For

example, the following rating definitions are not transparent because

they require the rater to do too much interpreting:

Borrower exhibits satisfactory quality and demonstrates acceptable

principal and interest repayment capacity in the near term.

Lower tier company in a cyclical industry. Unbalanced position with

tight liquidity and high leverage. Declining or erratic profitability

and marginal debt service capacity. Management is untested.

D. Accountability

``Accountability'' is holding people responsible for their actions

and establishing adverse consequences for inaccurate ratings.

S. Policies must identify the parties responsible for rating

accuracy and rating system performance.

For accountability to be effective, it should be both observable

and ingrained in the culture. Persons who assign and approve rate

credits, derive parameter estimates, or oversee rating systems must be

held accountable for complying with rating system policies and ensuring

that aspects of the rating system within their control are as unbiased

and accurate as possible. These persons must have the tools and

resources necessary to carry out their responsibilities, and their

performance should be evaluated against clear and specific objectives

documented in policy.

Responsibility for Assigning Ratings

S. Individuals must be held accountable for complying with rating

system policies and for assigning accurate ratings, and their

performance and compensation must be linked to well-defined measurable

performance standards.

Responsibilities of raters should be clear, and performance should

be measured against specific objectives. Performance evaluation and

incentive compensation should be tied to performance goals. Examples of

performance measures include:

[sbull] Number and frequency of rating errors,

[sbull] Significance of errors (for example, multiple downgrades),

and

[sbull] Proper and consistent application of criteria, including

override criteria.

Responsibility for Rating System Performance

Just as individuals will be held accountable for the accuracy of

ratings, an individual must be held responsible for the overall

performance of the rating system. This individual must ensure that the

rating system and all of its component parts--rating assignments,

parameter estimation, data collection, control and oversight

mechanisms--are functioning as intended. While these components often

are housed within separate units of the organization, an individual

must be responsible for ensuring that the parts work together

effectively and efficiently.

E. Use of Ratings

S. Ratings used for regulatory capital must be the same ratings

used to guide day-to-day credit risk management activities.

The different uses and applications of the risk-rating system's

outputs should promote greater accuracy and consistency of credit-risk

evaluations across an organization. Ratings and the associated default,

loss, and EAD estimates need to be incorporated within the credit-risk

management, internal capital allocation, and corporate governance

functions of IRB banks.

S. Banks that use parameter estimates for risk management that are

different from those used for regulatory capital must provide a well-

documented rationale for the differences.

PD and LGD parameters used for regulatory capital purposes may not

be appropriate for other uses purposes. For example, PD estimates used

to estimate reserve needs could reflect current economic conditions

that are different from the longer term view appropriate to

calculations of regulatory capital. When banks employ different

estimates, those parameters must be defensible and supported by the

following:

[sbull] Qualitative and quantitative analysis of the logic and

rationale for the difference(s); and

[sbull] Senior management approval of the difference(s).

F. Rating System Review (RSR)

S. Banks must have a comprehensive, coordinated, independent review

process to ensure that ratings are accurate and that the rating system

is performing as intended.

Rating system review (RSR) ensures that the rating system as a

whole is functioning as intended. A broad range of responsibilities

come under RSR's purview, as outlined in Table 4.2:

Table 4.2.--Responsibilities of Rating System Review

------------------------------------------------------------------------

-------------------------------------------------------------------------

Scope of Review:

Design of the rating system.

Compliance with policies and procedures, including application of

criteria.

Check of all risk-rating grades for accuracy.

Consistency across industries/portfolios/geographies.

[[Page 45977]]

Model development.

Model use, including inputs and outputs.

Overrides and policy exceptions.

Quantification process.

Back-testing (perform or review).

Actual and predicted ratings transitions.

Benchmarking against third-party data sources (perform or review).

Adequacy of data maintenance.

Analysis and Reporting:

Identify errors and flaws.

Recommend corrective action.

------------------------------------------------------------------------

For each of these responsibilities, RSR is largely checking and

confirming the work of others and ensuring that the rating system's

components work well together. RSR's testing and review should identify

current and potential weaknesses and should lead to recommendations and

corrective action such as

[sbull] Adjusting policies and procedures,

[sbull] Requiring additional training of staff,

[sbull] Investing in infrastructure improvements,

[sbull] Adjusting rating criteria, and

[sbull] Adjusting parameter estimates.

S. Rating system review must report significant findings to senior

management and the board quarterly.

RSR's role is to identify issues and areas of concern and report

findings to the area that is accountable. When issues are systematic,

RSR should bring them to the attention of senior management and the

board.

The activities of this function could be distributed across

multiple areas or housed within one unit. Organizations will choose a

structure that fits within their management and oversight framework.

These units must always have high standing within the organization and

should be staffed by individuals possessing the requisite stature,

skills, and experience.

Like internal audit, RSR must be independent from all in-house

designers and developers (that is, system and model designers) and

raters (that is, ratings and parameter assigners) in the risk-rating

process. RSR's independence eliminates potential conflicts of interest

and gives the group credibility when it reports findings and

conclusions to the board and senior management.

G. Internal Audit

S. An independent internal audit function must determine whether

rating system controls function as intended.

S. Internal audit must evaluate annually whether the bank is in

compliance with the risk-based capital regulation and supervisory

guidance.

Internal audit determines whether the bank's system of controls

over internal ratings and the related parameters is robust. In its

evaluation of controls, internal audit must consider any trade-offs

made between the various mechanisms and confirm their continued

appropriateness and relevance. As part of its review of control

mechanisms, audit will evaluate the depth, scope, and quality of RSR's

work and will conduct limited testing to ensure that their conclusions

are well founded. The amount of testing will depend on whether audit is

the primary or secondary reviewer of that work.

Internal audit will report to the board and management on whether

the bank is in compliance with the IRB standards. This report will

allow the board and management to disclose that its rating processes

and the controls surrounding these processes are in compliance with the

IRB standards. This will be critical for public disclosure and ongoing

work of supervisors.

External Audit

As part of the process of certifying financial statements, external

auditors will confirm that the institution's capital position is fairly

presented. To verify that actual capital exceeds regulatory minimums

and to confirm compliance with the IRB rules, the external auditors

must ascertain that the IRB system is rating credit risk appropriately

and linking these ratings to appropriate estimates. Auditors must

evaluate the bank's internal control functions and its compliance with

the risk-based capital regulation and supervisory guidance.

H. Corporate Oversight

S. The full board or a committee of the board must approve key

elements of the IRB system.

Consistent with sound practice, bank management must ensure that a

corporate culture exists in which institutional needs are readily

identified and appropriate resources are brought to bear to rectify

shortcomings. In the IRB context, senior management and the board of

directors must ensure the objectivity and accuracy of the bank's

credit-risk management systems and approach.

Either the full board or a committee of the board should approve

key elements of the risk-rating system. Information provided to the

board should be sufficiently detailed to allow directors to confirm the

continuing appropriateness of the institution's rating approach and to

verify the adequacy of the controls supporting the rating system.

S. Senior management must ensure that all components of the IRB

system, including controls, are functioning as intended and comply with

the risk-based capital regulation and supervisory guidance.

Senior management's oversight should be even more active than that

of the board of directors. Senior management should articulate what it

expects of the technical and operational units of the risk-rating

system, as well as what it expects of the units that manage the

system's controls. To oversee the risk-rating system, senior management

must have an extensive understanding of credit policies, underwriting

standards, lending practices, and collection and recovery practices,

and must be able to understand how these factors affect default and

loss estimates. Senior management should not only oversee the controls

process (its traditional role) but also should periodically meet with

raters and validators to discuss the rating system's performance, areas

needing improvement, and the status of efforts to improve previously

identified deficiencies.

The depth and frequency of information provided to the board and

senior management must be commensurate with their oversight

responsibilities and the condition of the institution. These reports

should include the following information:

[sbull] Risk profile by grade,

[sbull] Risk rating migration across grades with emphasis on

unexpected results,

[sbull] Changes in parameter estimates by grade,

[sbull] Comparison of realized PD, LGD, and EAD rates against

expectations,

[sbull] Reports measuring changes in regulatory and economic

capital,

[sbull] Results of capital stress testing, and

[sbull] Reports generated by rating system review, audit, and other

control units.

Although all of an institution's controls must function smoothly,

independently, and in concert with the others, the direction and

oversight provided by the board and senior management are perhaps most

important to ensure that the IRB system is functioning properly.

Document 2: Draft Supervisory Guidance on Operational Risk Advanced

Measurement Approaches for Regulatory Capital

Table of Contents

I. Purpose

II. Background

III. Definitions

IV. Banking Activities and Operational Risk

V. Corporate Governance

A. Board and Management Oversight

[[Page 45978]]

B. Independent Firm-wide Risk Management Function

C. Line of Business Management

VI. Operational Risk Management Elements

A. Operational Risk Policies and Procedures

B. Identification and Measurement of Operational Risk

C. Monitoring and Reporting

D. Internal Control Environment

VII. Elements of an AMA Framework

A. Internal Operational Risk Loss Event Data

B. External Data

C. Business Environment and Internal Control Factor Assessments

D. Scenario Analysis

VIII. Risk Quantification

A. Analytical Framework

B. Accounting for Dependence

IX. Risk Mitigation

X. Data Maintenance

XI. Testing and Verification

Appendix A: Supervisory Standards for the AMA

I. Purpose

The purpose of this guidance is to set forth the expectations of

the U.S. banking agencies for banking institutions that use Advanced

Measurement Approaches (AMA) for calculating the operational risk

capital charge under the new capital regulation. Institutions using the

AMA will have considerable flexibility to develop operational risk

measurement systems appropriate to the nature of their activities,

business environment, and internal controls. An institution's

operational risk regulatory capital requirement will be calculated as

the amount needed to cover its operational risk at a level of

confidence determined by the supervisors, as discussed below. Use of an

AMA is subject to supervisory approval.

This draft guidance should be considered with the advance notice of

proposed rulemaking (ANPR) on revisions to the risk-based capital

standard published elsewhere in today's Federal Register. As with the

ANPR, the Agencies are seeking industry comment on this draft guidance.

In addition to seeking comment on all specific aspects of this

supervisory guidance, the Agencies are seeking comment on the extent to

which the supervisory guidance strikes the appropriate balance between

flexibility and specificity. Likewise, the Agencies are seeking comment

on whether an appropriate balance has been struck between the

regulatory requirements set forth in the ANPR and the supervisory

standards set forth in this guidance.

II. Background

Effective management of operational risk is integral to the

business of banking and to institutions' roles as financial

intermediaries. Although operational risk is not a new risk,

deregulation and globalization of financial services, together with the

growing sophistication of financial technology, new business activities

and delivery channels, are making institutions' operational risk

profiles (i.e., the level of operational risk across an institution's

activities and risk categories) more complex.

This guidance identifies the supervisory standards (S) that

institutions must meet and maintain to use an AMA for the regulatory

capital charge for operational risk. The purpose of the standards is to

provide the foundation for a sound operational risk framework, while

allowing institutions to identify the most appropriate mechanisms to

meet AMA requirements. Each institution will need to consider its

complexity, range of products and services, organizational structure,

and risk management culture as it develops its AMA. Operational risk

governance processes need to be established on a firm-wide basis to

identify, measure, monitor, and control operational risk in a manner

comparable with the treatment of credit, interest rate, and market

risks.

Institutions will be expected to develop a framework that measures

and quantifies operational risk for regulatory capital purposes. To do

this, institutions will need a systematic process for collecting

operational risk loss data, assessing the risks within the institution,

and adopting an analytical framework that translates the data and risk

assessments into an operational risk exposure (see definition below).

The analytical framework must incorporate a degree of conservatism that

is appropriate for the overall robustness of the quantification

process. Because institutions will be permitted to calculate their

minimum regulatory capital on the basis of internal processes, the

requirements for data capture, risk assessment, and the analytical

framework described below are detailed and specific.

Effective operational risk measurement systems are built on both

quantitative and qualitative risk assessment techniques. While the

output of the regulatory framework for operational risk is a measure of

exposure resulting in a capital number, the integrity of that estimate

depends not only on the soundness of the measurement model, but also on

the robustness of the institution's underlying risk management

processes. In addition, supervisors view the introduction of the AMA as

an important tool to further promote improvements in operational risk

management and controls at large banking institutions.

This document provides both AMA supervisory standards and a

discussion of how those standards should be incorporated into an

operational risk framework. The relevant supervisory standards are

listed at the beginning of each section and a full compilation of the

standards is provided in Appendix A. Not every section has specific

supervisory standards. When spanning more than one section, supervisory

standards are listed only once.

Institutions will be required to meet, and remain in compliance

with, all the supervisory standards to use an AMA framework. However,

evaluating an institution's qualification with each of the individual

supervisory standards will not be sufficient to determine an

institution's overall readiness for AMA. Instead, supervisors and

institutions must also evaluate how well the various components of an

institution's AMA framework complement and reinforce one another to

achieve the overall objectives of an accurate measure and effective

management of operational risk. In performing their evaluation,

supervisors will exercise considerable supervisory judgment, both in

evaluating the individual components and the overall operational risk

framework.

An institution's AMA methodology will be assessed as part of the

ongoing supervision process. This will allow supervisors to incorporate

existing supervisory efforts as much as possible into the AMA

assessments. Some elements of operational risk (e.g., internal controls

and information technology) have long been subject to examination by

supervisors. Where this is the case, supervisors will make every effort

to leverage off these examination activities to assess the

effectiveness of the AMA process. Substantive weaknesses identified in

an examination will be factored into the AMA qualification process.

III. Definitions

There are important definitions that institutions must incorporate

into an AMA framework. They are:

[sbull] Operational risk: The risk of loss resulting from

inadequate or failed internal processes, people and systems, or from

external events. The definition includes legal risk, which is the risk

of loss resulting from failure to comply with laws as well as prudent

ethical standards and contractual obligations. It also includes the

exposure to litigation from all aspects of an institution's

[[Page 45979]]

activities. The definition does not include strategic or reputational

risks.\8\

---------------------------------------------------------------------------

\8\ An institution's definition of risk may encompass other risk

elements as long as the supervisory definition is met.

---------------------------------------------------------------------------

[sbull] Operational risk loss: The financial impact associated with

an operational event that is recorded in the institution's financial

statements consistent with Generally Accepted Accounting Principles

(GAAP). Financial impact includes all out-of-pocket expenses associated

with an operational event but does not include opportunity costs,

foregone revenue, or costs related to investment programs implemented

to prevent subsequent operational risk losses. Operational risk losses

are characterized by seven event factors associated with:

i. Internal fraud: An act of a type intended to defraud,

misappropriate property or circumvent regulations, the law or company

policy, excluding diversity/discrimination events, which involve at

least one internal party.

ii. External fraud: An act of a type intended to defraud,

misappropriate property or circumvent the law, by a third party.

iii. Employment practices and workplace safety: An act inconsistent

with employment, health or safety laws or agreements, from payment of

personal injury claims, or from diversity/discrimination events.

iv. Clients, products, and business practices: An unintentional or

negligent failure to meet a professional obligation to specific clients

(including fiduciary and suitability requirements), or from the nature

or design of a product.

v. Damage to physical assets: The loss or damage to physical assets

from natural disaster or other events.

vi. Business disruption and system failures: Disruption of business

or system failures.

vii. Execution, delivery, and process management: Failed

transaction processing or process management, from relations with trade

counterparties and vendors.

[sbull] Operational risk exposure: An estimate of the potential

operational losses that the banking institution faces at a soundness

standard consistent with a 99.9 per cent confidence level over a one-

year period. The institution will multiply the exposure by 12.5 to

obtain risk-weighted assets for operational risk; this is added to the

risk-weighted assets for credit and market risk to arrive at the

denominator of the regulatory capital ratio.

[sbull] Business environment and internal control factor

assessments: The range of tools that provide a meaningful assessment of

the level and trends in operational risk across the institution. While

the institution may use multiple tools in an AMA framework, they must

all have the same objective of identifying key risks. There are a

number of existing tools, such as audit scores and performance

indicators that may be acceptable under this definition.

IV. Banking Activities and Operational Risk

The above definition of operational risk gives a sense of the

breadth of exposure to operational risk that exists in banking today as

well as the many interdependencies among risk factors that may result

in an operational risk loss. Indeed, operational risk can occur in any

activity, function, or unit of the institution.

The definition of operational risk incorporates the risks stemming

from people, processes, systems and external events. People risk refers

to the risk of management failure, organizational structure or other

human resource failures. These risks may be exacerbated by poor

training, inadequate controls, poor staffing resources, or other

factors. The risk from processes stem from breakdowns in established

processes, failure to follow processes, or inadequate process mapping

within business lines. System risk covers instances of both disruption

and outright system failures in both internal and outsourced

operations. Finally, external events can include natural disasters,

terrorism, and vandalism.

There are a number of areas where operational risks are emerging.

These include:

[sbull] Greater use of automated technology has the potential to

transform risks from manual processing errors to system failure risks,

as greater reliance is placed on globally integrated systems;

[sbull] Proliferation of new and highly complex products;

[sbull] Growth of e-banking transactions and related business

applications expose an institution to potential new risks (e.g.,

internal and external fraud and system security issues);

[sbull] Large-scale acquisitions, mergers, and consolidations test

the viability of new or newly integrated systems;

[sbull] Emergence of institutions acting as large-volume service

providers create the need for continual maintenance of high-grade

internal controls and back-up systems;

[sbull] Development and use of risk mitigation techniques (e.g.,

collateral, insurance, credit derivatives, netting arrangements and

asset securitizations) optimize an institution's exposure to market

risk and credit risk, but potentially create other forms of risk (e.g.,

legal risk); and

[sbull] Greater use of outsourcing arrangements and participation

in clearing and settlement systems mitigate some risks while increasing

others.

The range of banking activities and areas affected by operational

risk must be fully identified and considered in the development of the

institution's risk management and measurement plans. Since operational

risk is not confined to particular business lines \9\, product types,

or organizational units, it should be managed in a consistent and

comprehensive manner across the institution. Consequently, risk

management mechanisms must encompass the full range of risks, as well

as strategies that help to identify, measure, monitor and control those

risks.

---------------------------------------------------------------------------

\9\ Throughout this guidance, terms such as ``business units''

and ``business lines'' are used interchangeably and refer not only

to an institution's revenue-generating businesses, but also to

corporate staff functions such as human resources or information

technology.

---------------------------------------------------------------------------

V. Corporate Governance

Supervisory Standards

S 1. The institution's operational risk framework must include an

independent firm-wide operational risk management function, line of

business management oversight, and independent testing and verification

functions.

The management structure underlying an AMA operational risk

framework may vary between institutions. However, within all AMA

institutions, there are three key components that must be evident--the

firm-wide operational risk management function, lines of business

management, and the testing and verification function. These three

elements are functionally independent \10\ organizational components,

but should work in cooperation to ensure a robust operational risk

framework.

---------------------------------------------------------------------------

\10\ For the purposes of AMA, ``functional independence'' is

defined as the ability to carry out work freely and objectively and

render impartial and unbiased judgments. There should be appropriate

independence between the firm-wide operational risk management

functions, line of business management and staff and the testing/

verification functions. Supervisory assessments of independence

issues will rely upon existing regulatory guidance (e.g. audit,

internal control systems, board of directors/management, etc.)

---------------------------------------------------------------------------

A. Board and Management Oversight

Supervisory Standards

S 2. The board of directors must oversee the development of the

firm-wide operational risk framework, as

[[Page 45980]]

well as major changes to the framework. Management roles and

accountability must be clearly established.

S 3. The board of directors and management must ensure that

appropriate resources are allocated to support the operational risk

framework.

The board is responsible for overseeing the establishment of the

operational risk framework, but may delegate the responsibility for

implementing the framework to management with the authority necessary

to allow for its effective implementation. Other key responsibilities

of the board include:

[sbull] Ensuring appropriate management responsibility,

accountability and reporting;

[sbull] Understanding the major aspects of the institution's

operational risk as a distinct risk category that should be managed;

[sbull] Reviewing periodic high-level reports on the institution's

overall operational risk profile, which identify material risks and

strategic implications for the institution;

[sbull] Overseeing significant changes to the operational risk

framework; and

[sbull] Ensuring compliance with regulatory disclosure

requirements.

Effective board and management oversight forms the cornerstone of

an effective operational risk management process. The board and

management have several broad responsibilities with respect to

operational risk:

[sbull] To establish a framework for assessing operational risk

exposure and identify the institution's tolerance for operational risk;

[sbull] To identify the senior managers who have the authority for

managing operational risk;

[sbull] To monitor the institution's performance and overall

operational risk profile, ensuring that it is maintained at prudent

levels and is supported by adequate capital;

[sbull] To implement sound fundamental risk governance principles

that facilitate the identification, measurement, monitoring, and

control of operational risk;

[sbull] To devote adequate human and technical resources to

operational risk management; and

[sbull] To institute remuneration policies that are consistent with

the institution's appetite for risk and are sufficient to attract

qualified operational risk management and staff.

Management should translate the operational risk management

framework into specific policies, processes and procedures that can be

implemented and verified within the institution's different business

units. Communication of these elements will be essential to the

understanding and consistent treatment of operational risk across the

institution. While each level of management is responsible for

effectively implementing the policies and procedures within its

purview, senior management should clearly assign authority,

responsibilities, and reporting relationships to encourage and maintain

this accountability and ensure that the necessary resources are

available to manage operational risk. Moreover, management should

assess the appropriateness of the operational risk management oversight

process in light of the risks inherent in a business unit's activities.

The testing and verification function is responsible for completing

timely and comprehensive assessments of the effectiveness of

implementation of the institution's operational risk framework at the

line of business and firm-wide levels.

Management collectively is also responsible for ensuring that the

institution has qualified staff and sufficient resources to carry out

the operational risk functions outlined in the operational risk

framework. Additionally, management must communicate operational risk

issues to appropriate staff that may not be directly involved in its

management. Key management responsibilities include ensuring that:

[sbull] Operational risk management activities are conducted by

qualified staff with the necessary experience, technical capabilities

and access to adequate resources;

[sbull] Sufficient resources have been allocated to operational

risk management, in the business lines as well as the independent firm-

wide operational risk management function and verification areas, so as

to sufficiently monitor and enforce compliance with the institution's

operational risk policy and procedures; and

[sbull] Operational risk issues are effectively communicated with

staff responsible for managing credit, market and other risks, as well

as those responsible for purchasing insurance and managing third-party

outsourcing arrangements.

B. Independent Firm-Wide Risk Management Function

Supervisory Standards

S 4. The institution must have an independent operational risk

management function that is responsible for overseeing the operational

risk framework at the firm level to ensure the development and

consistent application of operational risk policies, processes, and

procedures throughout the institution.

S 5. The firm-wide operational risk management function must ensure

appropriate reporting of operational risk exposures and loss data to

the board of directors and senior management.

The institution must have an independent firm-wide operational risk

management function. The roles and responsibilities of the function

will vary between institutions, but must be clearly documented. The

independent firm-wide operational risk function should have

organizational stature commensurate with the institution's operational

risk profile, while remaining independent of the lines of business and

the testing and verification function. At a minimum, the institution's

independent firm-wide operational risk management function should

ensure the development of policies, processes, and procedures that

explicitly manage operational risk as a distinct risk to the

institution's safety and soundness. These policies, processes and

procedures should include principles for how operational risk is to be

identified, measured, monitored, and controlled across the

organization. Additionally, they should provide for the collection of

the data needed to calculate the institution's operational risk

exposure.

Additional responsibilities of the independent firm-wide

operational risk management function include:

[sbull] Assisting in the implementation of the overall firm-wide

operational risk framework;

[sbull] Reviewing the institution's progress towards stated

operational risk objectives, goals and risk tolerances;

[sbull] Periodically reviewing the institution's operational risk

framework to consider the loss experience, effects of external market

changes, other environmental factors, and the potential for new or

changing operational risks associated with new products, activities or

systems. This review process should include an assessment of industry

best practices for the institution's activities, systems and processes;

[sbull] Reviewing and analyzing operational risk data and reports;

and

[sbull] Ensuring appropriate reporting to senior management and the

board.

C. Line of Business Management

Supervisory Standards

S 6. Line of business management is responsible for the day-to-day

management of operational risk within each business unit.

S 7. Line of business management must ensure that internal controls

and

[[Page 45981]]

practices within their line of business are consistent with firm-wide

policies and procedures to support the management and measurement of

the institution's operational risk.

Line of business management is responsible for both managing

operational risk within the business lines and ensuring that policies

and procedures are consistent with and support the firm-wide

operational risk framework. Management should ensure that business-

specific policies, processes, procedures and staff are in place to

manage operational risk for all material products, activities, and

processes. Implementation of the operational risk framework within each

line of business should reflect the scope of that business and its

inherent operational complexity and operational risk profile. Line of

business management must be independent of both the firm-wide

operational risk management and the testing and verification functions.

VI. Operational Risk Management Elements

The operational risk management framework provides the overall

operational risk strategic direction and ensures that an effective

operational risk management and measurement process is adopted

throughout the institution. The framework should provide for the

consistent application of operational risk policies and procedures

throughout the institution and address the roles of both the

independent firm-wide operational risk management function and the

lines of business. The framework should also provide for the consistent

and comprehensive capture of data elements needed to measure and verify

the institution's operational risk exposure, as well as appropriate

operational risk analytical frameworks, reporting systems, and

mitigation strategies. The framework must also include independent

testing and verification to assess the effectiveness of implementation

of the institution's operational risk framework, including compliance

with policies, processes, and procedures.

In practice, an institution's operational risk framework must

reflect the scope and complexity of business lines, as well as the

corporate organizational structure. Each institution's operational risk

profile is unique and requires a tailored risk management approach

appropriate for the scale and materiality of the risks present, and the

size of the institution. There is no single framework that would suit

every institution; different approaches will be needed for different

institutions. In fact, many operational risk management techniques

continue to evolve rapidly to keep pace with new technologies, business

models and applications.

The key elements in the operational risk management process

include:

[sbull] Appropriate policies and procedures;

[sbull] Efforts to identify and measure operational risk;

[sbull] Effective monitoring and reporting;

[sbull] A sound system of internal controls; and

[sbull] Appropriate testing and verification of the operational

risk framework.

A. Operational Risk Policies and Procedures

Supervisory Standards

S 8. The institution must have policies and procedures that clearly

describe the major elements of the operational risk management

framework, including identifying, measuring, monitoring, and

controlling operational risk.

Operational risk management policies, processes, and procedures

should be documented and communicated to appropriate staff. The

policies and procedures should outline all aspects of the institution's

operational risk management framework, including:

[sbull] The roles and responsibilities of the independent firm-wide

operational risk management function and line of business management;

[sbull] A definition for operational risk, including the loss event

types that will be monitored;

[sbull] The capture and use of internal and external operational

risk loss data, including large potential events (including the use of

scenario analysis);

[sbull] The development and incorporation of business environment

and internal control factor assessments into the operational risk

framework;

[sbull] A description of the internally derived analytical

framework that quantifies the operational risk exposure of the

institution;

[sbull] An outline of the reporting framework and the type of data/

information to be included in line of business and firm-wide reporting;

[sbull] A discussion of qualitative factors and risk mitigants and

how they are incorporated into the operational risk framework;

[sbull] A discussion of the testing and verification processes and

procedures;

[sbull] A discussion of other factors that affect the measurement

of operational risk; and

[sbull] Provisions for the review and approval of significant

policy and procedural exceptions.

B. Identification and Measurement of Operational Risk

The result of a comprehensive program to identify and measure

operational risk is an assessment of the institution's operational risk

exposure. Management must establish a process that identifies the

nature and types of operational risk and their causes and resulting

effects on the institution. Proper operational risk identification

supports the reporting and maintenance of capital for operational risk

exposure and events, facilitates the establishment of mechanisms to

mitigate or control the risks, and ensures that management is fully

aware of the sources of emerging operational risk loss events.

C. Monitoring and Reporting

Supervisory Standards

S 9. Operational risk management reports must address both firm-

wide and line of business results. These reports must summarize

operational risk exposure, loss experience, relevant business

environment and internal control assessments, and must be produced no

less often than quarterly.

S 10. Operational risk reports must also be provided periodically

to senior management and the board of directors, summarizing relevant

firm-wide operational risk information.

Ongoing monitoring of operational risk exposures is a key aspect of

an effective operational risk framework. To facilitate monitoring of

operational risk, results from the measurement system should be

summarized in reports that can be used by the firm-wide operational

risk and line of business management functions to understand, manage,

and control operational risk and losses. These reports should serve as

a basis for assessing operational risk and related mitigation

strategies and creating incentives to improve operational risk

management throughout the institution.

Operational risk management reports should summarize:

[sbull] Operational risk loss experience on an institution, line of

business, and event-type basis;

[sbull] Operational risk exposure;

[sbull] Changes in relevant risk and control assessments;

[sbull] Management assessment of early warning factors signaling an

increased risk of future losses;

[sbull] Trend analysis, allowing line of business and independent

firm-wide operational risk management to assess

[[Page 45982]]

and manage operational risk exposures, systemic line of business risk

issues, and other corporate risk issues;

[sbull] Exception reporting; and

[sbull] To the extent developed, operational risk causal factors.

High-level operational risk reports must also be produced

periodically for the board and senior management. These reports must

provide information regarding the operational risk profile of the

institution, including the sources of material risk both from a firm-

wide and line of business perspective, versus established management

expectations.

D. Internal Control Environment

Supervisory Standards

S 11. An institution's internal control structure must meet or

exceed minimum regulatory standards established by the Agencies.

Sound internal controls are essential to an institution's

management of operational risk and are one of the foundations of safe

and sound banking. When properly designed and consistently enforced, a

sound system of internal controls will help management safeguard the

institution's resources, produce reliable financial reports, and comply

with laws and regulations. Sound internal controls will also reduce the

possibility of significant human errors and irregularities in internal

processes and systems, and will assist in their timely detection when

they do occur.

The Agencies are not introducing any new internal control

standards, but rather emphasizing the importance of meeting existing

standards. There is a recognition that internal control systems will

differ among institutions due to the nature and complexity of an

institution's products and services, organizational structure, and risk

management culture. The AMA standards allows for these differences,

while also establishing a baseline standard for the quality of the

internal control structure. Institutions will be expected to at least

meet the minimum interagency standards\11\ relating to internal

controls as a criterion for AMA qualification.

---------------------------------------------------------------------------

\11\ There are a number of interagency standards that cover

topics relevant to the internal control structure. These include,

for example, the Interagency Policy Statement on the Internal Audit

Function and Its Outsourcing (March 2003), the Federal Financial

Institution's Examination Council's (FFIEC's) Business Continuity

Planning Booklet (May 2003), the FFIEC's Information Security

Booklet (January 2003). In addition, each Agency has extensive

guidance on corporate governance, internal controls, and monitoring

and reporting in its respective examination policies and procedures.

---------------------------------------------------------------------------

The extent to which an institution meets or exceeds the minimum

standards will primarily be assessed through current and ongoing

supervisory processes. As noted earlier, the Agencies will leverage off

existing examination processes, to avoid duplication in assessing an

institution's implementation of an AMA framework. Assessing the

internal control environment is clearly an area where the supervisory

authorities already focus considerable attention.

VII. Elements of an AMA Framework

Supervisory Standards

S 12. The institution must demonstrate that it has appropriate

internal loss event data, relevant external loss event data,

assessments of business environment and internal controls factors, and

results from scenario analysis to support its operational risk

management and measurement framework.

S 13. The institution must include the regulatory definition of

operational risk as the baseline for capturing the elements of the AMA

framework and determining its operational risk exposure.

S 14. The institution must have clear standards for the collection

and modification of the elements of the operational risk AMA framework.

Operational risk inputs play a significant role in both the

management and measurement of operational risk. Necessary elements of

an institution's AMA framework include internal loss event data,

relevant external loss event data, results of scenario analysis, and

assessments of the institution's business environment and internal

controls. Operational risk inputs aid the institution in identifying

the level and trend of operational risk, determining the effectiveness

of risk management and control efforts, highlighting opportunities to

better mitigate operational risk, and assessing operational risk on a

forward-looking basis.

To use its AMA framework, an institution must demonstrate that it

has established a consistent and comprehensive process for the capture

of all elements of the AMA framework. The institution must also

demonstrate that it has clear standards for the collection and

modification of all AMA inputs. While the analytical framework will

generally combine these inputs to develop the operational risk

exposure, supervisors must have the capacity to review the individual

inputs as well; specifically, supervisors will need to review the loss

information that is being provided to the analytical framework that

stems from internal loss event data, versus the loss event information

provided by external loss event data capture, scenario analysis, or the

assessments of the business environment and internal control factors.

The capture systems must cover all material business lines,

business activities and corporate functions that could generate

operational risk. The institution must have a defined process that

establishes responsibilities over the systems developed to capture the

AMA elements. In particular, the issue of overriding the data capture

systems must be addressed. Any overrides should be tracked separately

and documented. Tracking overrides separately allows management and

supervisors to identify the nature and rationale, including whether

they stem from simple input errors or, more importantly, from exclusion

because a loss event was not pertinent for the quantitative

measurement. Management should have clear standards for addressing

overrides and should clearly delineate who has authority to override

the data systems and under what circumstances.

As noted earlier, for AMA qualification purposes, an institution's

operational risk framework must, at a minimum, use the definition of

operational risk that is provided in paragraph 10 when capturing the

elements of the AMA framework. Institutions may use an expanded

definition if considered more appropriate for risk management and

measurement efforts. However, for the quantification of operational

risk exposure for regulatory capital purposes, an institution must

demonstrate that the AMA elements are captured so as to meet the

baseline definition.

A. Internal Operational Risk Loss Event Data

Supervisory Standards

S 15. The institution must have at least five years of internal

operational risk loss data \12\ captured across all material business

lines, events, product types, and geographic locations.

---------------------------------------------------------------------------

\12\ With supervisory approval, a shorter initial historical

observation period is acceptable for banks newly authorized to use

an AMA methodology.

---------------------------------------------------------------------------

S 16. The institution must be able to map internal operational risk

losses to the seven loss-event type categories.

S 17. The institution must have a policy that identifies when an

operational risk loss becomes a loss event and must be added to the

loss

[[Page 45983]]

event database. The policy must provide for consistent treatment across

the institution.

S 18. The institution must establish appropriate operational risk

data thresholds.

S 19. Losses that have any characteristics of credit risk,

including fraud-related credit losses, must be treated as credit risk

for regulatory capital purposes. The institution must have a clear

policy that allows for the consistent treatment of loss event

classifications (e.g., credit, market, or operational risk) across the

organization.

The key to internal data integrity is the consistency and

completeness with which loss event data capture processes are

implemented across the institution. Management must ensure that

operational risk loss event information captured is consistent across

the business lines and incorporates any corporate functions that may

also experience operational risk events. Policies and procedures should

be addressed to the appropriate staff to ensure that there is

satisfactory understanding of operational risk and the data capture

requirements under the operational risk framework. Further, the

independent operational risk management function must ensure that the

loss data is captured across all material business lines, products

types, event types, and from all significant geographic locations. The

institution must be able to capture and aggregate internal losses that

cross multiple business lines or event types. If data is not captured

across all business lines or from all geographic locations, the

institution must document and explain the exceptions.

AMA institutions must be able to map operational risk losses into

the seven loss event categories defined in paragraph 10. Institutions

will not be required to produce reports or perform analysis for

internal purposes on the basis of the loss event categories, but will

be expected to use the information about the event-type categories as a

check on the comprehensiveness of the institution's data set.

The institution must have five years of internal loss data,

although a shorter range of historical data may be allowed, subject to

supervisory approval. The extent to which an institution collects

operational risk loss event data will, in part, be dependent upon the

data thresholds that the institution establishes. There are a number of

standards that an institution may use to establish the thresholds. They

may be based on product types, business lines, geographic location, or

other appropriate factors. The Agencies will allow flexibility in this

area, provided the institution can demonstrate that the thresholds are

reasonable, do not exclude important loss events, and capture a

significant proportion of the institution's operational risk losses.

The institution must capture comprehensive data on all loss events

above its established threshold level. Aside from information on the

gross loss amount, the institution should collect information about the

date of the event, any recoveries, and descriptive information about

the drivers or causes of the loss event. The level of detail of any

descriptive information should be commensurate with the size of the

gross loss amount. Examples of the type of information collected

include:

[sbull] Loss amount;

[sbull] Description of loss event;

[sbull] Where the loss is reported and expensed;

[sbull] Loss event type category;

[sbull] Date of the loss;

[sbull] Discovery date of the loss;

[sbull] Event end date;

[sbull] Management actions;

[sbull] Insurance recoveries;

[sbull] Other recoveries; and

[sbull] Adjustments to the loss estimate.

There are a number of additional data elements that may be

captured. It may be appropriate, for example, to capture data on ``near

miss'' events, where no financial loss was incurred. These near misses

will not factor into the regulatory capital calculation, but may be

useful for the operational risk management process.

Institutions will also be permitted and encouraged to capture loss

events in their operational risk databases that are treated as credit

risk for regulatory capital purposes, but have an underlying element of

operational risk failure. These types of events, while not incorporated

into the regulatory capital calculation, may have implications for

operational risk management. It will be essential for institutions that

capture loss events that are treated differently for regulatory capital

and management purposes to demonstrate that (1) loss events are being

captured consistently across the institution; (2) the data systems are

sufficiently advanced to allow for this differential treatment of loss

events; and (3) credit, market, and operational risk losses are being

appropriated in the correct manner for regulatory capital purposes.

The Agencies have established a clear boundary between credit and

operational risks for regulatory capital purposes. If a loss event has

any element of credit risk, it must be treated as credit risk for

regulatory capital purposes. This would include all credit-related

fraud losses. In addition, operational risk losses with credit risk

characteristics that have historically been included in institutions'

credit risk databases will continue to be treated as credit risk for

the purposes of calculating minimum regulatory capital.

The accounting guidance for credit losses provides that creditors

recognize credit losses when it is probable that they will be unable to

collect all amounts due according to the contractual terms of a loan

agreement. Credit losses may result from the creditor's own

underwriting, processing, servicing or administrative activities along

with the borrower's failure to pay according to the terms of the loan

agreement. While the creditor's personnel, systems, policies or

procedures may affect the timing or magnitude of a credit loss, they do

not change its character from credit to operational risk loss for

regulatory capital purposes. Losses that arise from a contractual

relationship between a creditor and a borrower are credit losses

whereas losses that arise outside of a relationship between a creditor

and a borrower are operational losses.

B. External Data

Supervisory Standards

S 20. The institution must have policies and procedures that

provide for the use of external loss data in the operational risk

framework.

S 21. Management must systematically review external data to ensure

an understanding of industry experience.

External data may serve a number of different purposes in the

operational risk framework. Where internal loss data is limited,

external data may be a useful input in determining the institution's

level of operational risk exposure. Even where external loss data is

not an explicit input to an institution's data set, such data provides

a means for the institution to understand industry experience, and in

turn, provides a means for assessing the adequacy of its internal data.

External data may also prove useful to inform scenario analysis, fit

severity distributions, or benchmark the overall operational risk

exposure results.

To incorporate external loss information into an institution's

framework, the institution should collect the following information:

[sbull] External loss amount;

[sbull] External loss description;

[sbull] Loss event type category;

[sbull] External loss event date;

[sbull] Adjustments to the loss amount (i.e., recoveries, insurance

settlements,

[[Page 45984]]

etc) to the extent that they are known; and

[sbull] Sufficient information about the reporting institution to

facilitate comparison to its own organization.

Institutions may obtain external loss data in any reasonable

manner. There are many ways to do so; some institutions are using data

acquired through membership with industry consortia while other

institutions are using data obtained from vendor databases or public

sources such as court records or media reports. In all cases,

management will need to carefully evaluate the data source to ensure

that they are comfortable that the information being reported is

relevant and reasonably accurate.

C. Business Environment and Internal Control Factor Assessments

Supervisory Standards

S 22. The institution must have a system to identify and assess

business environment and internal control factors.

S 23. Management must periodically compare the results of their

business environment and internal control factor assessments against

actual operational risk loss experience.

While internal and external loss data provide a historical

perspective on operational risk, it is also important that institutions

incorporate a forward-looking element to the operational risk measure.

In principle, an institution with strong internal controls in a stable

business environment will have less exposure to operational risk than

an institution with internal control weaknesses that is growing rapidly

or introducing new products. In this regard, institutions will be

required to identify the level and trends in operational risk in the

institution. These assessments must be current, comprehensive across

the institution, and identify the critical operational risks facing the

institution.

The business environment and internal control factor assessments

should reflect both the positive and negative trends in risk management

within the institution as well as changes in an institution's business

activities that increase or decrease risk. Because the results of the

risk assessment are part of the capital methodology, management must

ensure that the risk assessments are done appropriately and reflect the

risks of the institution. Periodic comparisons should be made between

actual loss exposure and the assessment results.

The framework established to maintain the risk assessments must be

sufficiently flexible to encompass an institution's increased

complexity of activities, new activities, changes in internal control

systems, or an increased volume of information.

D. Scenario Analysis

Supervisory Standards

S 24. Management must have policies and procedures that identify

how scenario analysis will be incorporated into the operational risk

framework.

Scenario analysis is a systematic process of obtaining expert

opinions from business managers and risk management experts to derive

reasoned assessments of the likelihood and impact of plausible

operational losses consistent with the regulatory soundness standard.

Within an institution's operational risk framework, scenario analysis

may be used as an input or may, as discussed below, form the basis of

an operational risk analytical framework.

As an input to the institution's framework, scenario analysis is

especially relevant for business lines or loss event types where

internal data, external data, and assessments of the business

environment and internal control factors do not provide a sufficiently

robust estimate of the institution's exposure to operational risk. In

some cases, an institution's internal loss history may be sufficient to

provide a reasonable estimate of exposure to future operational losses.

In other cases, the use of well-reasoned, scaled external data may

itself be a form of scenario analysis.

The institution must have policies and procedures that define

scenario analysis and identify its role in the operational risk

framework. The policy should cover key elements of scenario analysis,

such as the manner in which the scenarios are generated, the frequency

with which they are updated, and the scope and coverage of operational

loss events they are intended to reflect.

VIII. Risk Quantification

A. Analytical Framework

Supervisory Standards

S 25. The institution must have a comprehensive operational risk

analytical framework that provides an estimate of the institution's

operational risk exposure, which is the aggregate operational loss that

it faces over a one-year period at a soundness standard consistent with

a 99.9 per cent confidence level.

S 26. Management must document the rationale for all assumptions

underpinning its chosen analytical framework, including the choice of

inputs, distributional assumptions, and the weighting across

qualitative and quantitative elements. Management must also document

and justify any subsequent changes to these assumptions.

S 27. The institution's operational risk analytical framework must

use a combination of internal operational loss event data, relevant

external operational loss event data, business environment and internal

control factor assessments, and scenario analysis. The institution must

combine these elements in a manner that most effectively enables it to

quantify its operational risk exposure. The institution can choose the

analytical framework that is most appropriate to its business model.

S 28. The institution's capital requirement for operational risk

will be the sum of expected and unexpected losses unless the

institution can demonstrate, consistent with supervisory standards, the

expected loss offset.

The industry has made significant progress in recent years in

developing analytical frameworks to quantify operational risk. The

analytical frameworks, which are a part of the overall operational risk

framework, are based on various combinations of an institution's own

operational loss experience, the industry's operational loss

experience, the size and scope of the institution's activities, the

quality of the institution's control environment, and management's

expert judgment. Because these models capture specific characteristics

of each institution, such models yield unique risk-sensitive estimates

of the institutions' operational risk exposures.

While the Agencies are not specifying the exact methodology that an

institution should use to determine its operational risk exposure,

minimum supervisory standards for acceptable approaches have been

developed. These standards have been set so as to assure that the

regulation can accommodate continued evolution of operational risk

quantification techniques, yet remain amenable to consistent

application and enforcement across institutions. The Agencies will

require that the institution have a comprehensive analytical framework

that provides an estimate of the aggregate operational loss that it

faces over a one-year period at a soundness standard consistent with a

99.9 percent confidence level, referred to as the institution's

operational risk exposure. The institution will multiply the exposure

estimate by 12.5 to obtain risk weighted assets for operational risk,

[[Page 45985]]

and add this figure to risk-weighted assets for credit and market risk

to obtain total risk-weighted assets. The final minimum regulatory

capital number will be 8 percent of total risk-weighted assets.

The Agencies expect that there will be significant variation in

analytical frameworks across institutions, with each institution

tailoring its framework to leverage existing technology platforms and

risk management procedures. These approaches may only be used, provided

they meet the supervisory standards and include, as inputs, internal

operational loss event data, relevant external operational loss event

data, assessments of business environment and internal control factors,

and scenario analysis. The Agencies do expect that there will be some

uncertainty and potential error in the analytical frameworks because of

the evolving nature of operational risk measurement and data capture.

Therefore, a degree of conservatism will need to be built into the

analytical frameworks to reflect the evolutionary status of operational

risk and its impact on data capture and analytical modeling.

A diversity of analytical approaches is emerging in the industry,

combining and weighting these inputs in different ways. Most current

approaches seek to estimate loss frequency and loss severity to arrive

at an aggregate loss distribution. Institutions then use the aggregate

loss distribution to determine the appropriate amount of capital to

hold for a given soundness standard. Scenario analysis is also being

used by many institutions, albeit to significantly varying degrees.

Some institutions are using scenario analysis as the basis for their

analytical framework, while others are incorporating scenarios as a

means for considering the possible impact of significant operational

losses on their overall operational risk exposure.

The primary differences among approaches being used today relate to

the weight that institutions place on each input. For example,

institutions with comprehensive internal data may place less emphasis

on external data or scenario analysis. Another example is that some

institutions estimate a unique loss distribution for each business

line/loss type combination (bottom-up approach) while others estimate a

loss distribution on a firm-wide basis and then use an allocation

methodology to assign capital to business lines (top-down approach).

The Agencies expect internal loss event data to play an important

role in the institution's analytical framework, hence the requirement

for five years of internal operational risk loss data. However, as

footnote 5 makes clear, five years of data is not always required for

the analytical framework. For example, if a bank exited a business

line, the institution would not be expected to make use of that

business unit's loss experience unless it had relevance for other

activities of the institution. Another example would be where a bank

has made a recent acquisition where the acquired firm does not have

internal loss event data. In these cases, the Agencies expect the

institution to make use of the loss data available at the acquired

institution and any internal loss data from operations similar to that

of the acquired firm, but the institution will likely have to place

more weight relevant external loss event data, results from scenario

analysis, and factors reflecting assessments of the business

environment and internal controls.

Whatever analytical approach an institution chooses, it must

document and provide the rationale for all assumptions embedded in its

chosen analytical framework, including the choice of inputs,

distributional assumptions, and the weighting of qualitative and

quantitative elements. Management must also document and justify any

subsequent changes to these assumptions. This documentation should:

[sbull] Clearly identify how the different inputs are combined and

weighted to arrive at the overall operational risk exposure so that the

analytical framework is transparent. The documentation should

demonstrate that the analytical framework is comprehensive and

internally consistent. Comprehensiveness means that all required inputs

are incorporated and appropriately weighted. At the same time, there

should not be overlaps or double counting.

[sbull] Clearly identify the quantitative assumptions embedded in

the methodology and provide explanation for the choice of these

assumptions. Examples of quantitative assumptions include

distributional assumptions about frequency and severity, the

methodology for combining frequency and severity to arrive at the

overall loss distribution, and dependence assumptions between

operational losses across and within business lines.

[sbull] Clearly identify the qualitative assumptions embedded in

the methodology and provide explanations for the choice of these

assumptions. Examples of qualitative assumptions include the use of

business environment and control factors as well as scenario analysis

in the approach.

[sbull] Where feasible, provide results based purely on

quantitative methods separately from results that incorporate

qualitative factors. This will provide a transparent means of

determining the relative importance of quantitative versus qualitative

inputs.

[sbull] Where feasible, provide results based on alternative

quantitative and qualitative assumptions to gauge the overall model's

sensitivity to these assumptions.

[sbull] Provide a comparison of the operational risk exposure

estimate generated by the analytical framework with actual loss

experience over time, to assess the reasonable of the framework's

outputs.

[sbull] Clearly identify all changes to assumptions, and provide

explanations for such changes.

[sbull] Clearly identify the results of an independent verification

of the analytical framework.

The regulatory capital charge for operational risk will include

both expected losses (EL) and unexpected losses (UL). The Agencies have

considered two approaches that might allow for some recognition of EL;

these approaches are reserving and budgeting. However, both approaches

raise questions about their ability to act as an EL offset for

regulatory capital purposes. The current U.S. GAAP treatment for

reserves (or liabilities) is based on an incurred-loss (liability)

model. Given that EL is looking beyond current losses to losses that

will be incurred in the future, establishing a reserve for operational

risk EL is not likely to meet U.S. accounting standards. While reserves

are specific allocations for incurred losses, budgeting is a process of

generally allocating future income for loss contingencies, including

losses resulting from operational risk. Institutions will be required

to demonstrate that budgeted funds are sufficiently capital-like and

remain available to cover EL over the next year. In addition, an

institution will not be permitted to recognize EL offsets on budgeted

loss contingencies that fall below the established data thresholds;

this is relevant as many institutions currently budget for low

severity, high frequency events that are more likely to fall below most

institutions' thresholds.

An institution's analytical framework complements but does not

substitute for prudent controls. Rather, with improved risk

measurement, institutions are finding that they can make better-

informed strategic decisions regarding enhancements to controls and

processes, the desired scale and scope of the operations, and how

insurance and

[[Page 45986]]

other risk mitigation tools can be used to offset operational risk

exposure.

B. Accounting for Dependence

Supervisory Standards

S 29. Management must document how its chosen analytical framework

accounts for dependence (e.g., correlations) among operational losses

across and within business lines. The institution must demonstrate that

its explicit and embedded dependence assumptions are appropriate, and

where dependence assumptions are uncertain, the institution must use

conservative estimates.

Management must document how its chosen analytical framework

accounts for dependence (e.g., correlation) between operational losses

across and within business lines. The issue of dependence is closely

related to the choice between a bottom-up or a top-down modeling

approach. Under a bottom-up approach, explicit assumptions regarding

cross-event dependence are required to estimate operational risk

exposure at the firm-wide level. Management must demonstrate that these

assumptions are appropriate and reflect the institution's current

environment. If the dependence assumptions are uncertain, the

institution must choose conservative estimates. In so doing, the

institution should consider the possibility that cross-event dependence

may not be constant, and may increase during stress environments.

Under a top-down approach, an explicit assumption regarding

dependence is not required. However, a parametric distribution for loss

severity may be more difficult to specify under the top-down approach,

as it is a statistical mixture of (potentially) heterogeneous business

line and event type distributions. Institutions must carefully consider

the conditions necessary for the validity of top-down approaches, and

whether these conditions are met in their particular circumstances.

Similar to bottom-up approaches, institutions using top-down approaches

must ensure that implicit dependence assumptions are appropriate and

reflect the institution's current environment. If historic dependence

assumptions embedded in top-down approaches are uncertain, the

institution must be conservative and implement a qualitative adjustment

to the analysis.

IX. Risk Mitigation

Supervisory Standards

S 30. Institutions may reduce their operational risk exposure

results by no more than 20% to reflect the impact of risk mitigants.

Institutions must demonstrate that mitigation products are sufficiently

capital-like to warrant inclusion in the adjustment to the operational

risk exposure.

There are many mechanisms to manage operational risk, including

risk transfer through risk mitigation products. Because risk mitigation

can be an important element in limiting or reducing operational risk

exposure in an institution, an adjustment is being permitted that will

directly impact the amount of regulatory capital that is held for

operational risk. The adjustment is limited to 20% of the overall

operational risk exposure result determined by the institution using

its loss data, qualitative factors, and quantitative framework.

Currently, the primary risk mitigant used for operational risk is

insurance. There has been discussion that some securities products may

be developed to provide risk mitigation benefits; however, to date, no

specific products have emerged that have characteristics sufficient to

be considered capital-replacement for operational risk. As a result,

securities products and other capital market instruments may not be

factored in to the regulatory capital risk mitigation adjustment at

this time.

For an institution that wishes to adjust its regulatory capital

requirement as a result of the risk mitigating impact of insurance,

management must demonstrate that the insurance policy is sufficiently

capital-like to provide the cushion that is necessary. A product that

would fall in this category must have the following characteristics:

[sbull] The policy is provided through a third party \13\ that has

a minimum claims paying ability rating of A; \14\

---------------------------------------------------------------------------

\13\ Where operational risk is transferred to a captive or an

affiliated insurer such that risk is retained within the group

structure, recognition of such risk transfer will only be allowed

for regulatory capital purposes where the risk has been transferred

to a third party (e.g., an unaffiliated reinsurer) that meets the

standards set forth in this section.

\14\ Rating agencies may use slightly different rating

scales.For the purpose of this supervisory guidance, the insurer

must have a rating that is at least the equivalent of A under

Standard and Poor's Insurer Financial Strength Ratings or an A2

under Moody's Insurance Financial Strength Ratings.

---------------------------------------------------------------------------

[sbull] The policy has an initial term of one year; \15\

---------------------------------------------------------------------------

\15\ Institutions must decrease the amount of the adjustment if

the remaining term is less than one year. The institution must have

a clear policy in place that links the remaining term to the

adjustment factor.

---------------------------------------------------------------------------

[sbull] The policy has no exclusions or limitations based upon

regulatory action or for the receiver or liquidator of a failed bank;

[sbull] The policy has clear cancellation and non-renewal notice

periods; and

[sbull] The policy coverage has been explicitly mapped to actual

operational risk exposure of the institution.

Insurance policies that meet these standards may be incorporated

into an institution's adjustment for risk mitigation. An institution

should be conservative in its recognition of such policies, for

example, the institution must also demonstrate that insurance policies

used as the basis for the adjustment have a history of timely payouts.

If claims have not been paid on a timely basis, the institution must

exclude that policy from the operational risk capital adjustment. In

addition, the institution must be able to show that the policy would

actually be used in the event of a loss situation; that is, the

deductible may not be set so high that no loss would ever conceivably

exceed the deductible threshold.

The Agencies will not specify how institutions should calculate the

risk mitigation adjustment. Nevertheless, institutions are expected to

use conservative assumptions when calculating adjustments. An

institution should discount (i.e., apply its own estimates of haircuts)

the impact of insurance coverage to take into account factors, which

may limit the likelihood or size of claims payouts. Among these factors

are the remaining terms of a policy, especially when it is less than a

year, the willingness and ability of the insurer to pay on a claim in a

timely manner, the legal risk that a claim may be disputed, and the

possibility that a policy can be cancelled before the contractual

expiration.

X. Data Maintenance

Supervisory Standards

S 31. Institutions using the AMA approach for regulatory capital

purposes must use advanced data management practices to produce

credible and reliable operational risk estimates.

Data maintenance is a critical factor in an institution's

operational risk framework. Institutions with advanced data management

practices should be able to track operational risk loss events from

initial discovery through final resolution. These institutions should

also be able to make appropriate adjustments to the data and use the

data to identify trends, track problem areas, and identify areas of

future risk. Such data should include not only operational risk loss

event information, but also information on risk assessments, which are

factored into the operational risk exposure calculation. In general,

institutions using the AMA

[[Page 45987]]

should have the same data maintenance standards for operational risk as

those set forth for A-IRB institutions under the credit risk guidance.

Operational risk data elements captured by the institution must be

of sufficient depth, scope, and reliability to:

[sbull] Track and identify operational risk loss events across all

business lines, including when a loss event impacts multiple business

lines.

[sbull] Calculate capital ratios based on operational risk exposure

results. The institution must also be able to factor in adjustments

related to risk mitigation, correlations, and risk assessments.

[sbull] Produce internal and public reports on operational risk

measurement and management results, including trends revealed by loss

data and/or risk assessments. The institution must also have sufficient

data to produce exception reports for management.

[sbull] Support risk management activities.

The data warehouse \16\ 16 must contain the key data elements

needed for operational risk measurement, management, and verification.

The precise data elements may vary by institution and also among

business lines within an institution. An important element of ensuring

consistent reporting of the data elements is to develop comprehensive

definitions for each data element used by the institution for reporting

operational risk loss events or for the risk assessment inputs. The

data must be stored in an electronic format to allow for timely

retrieval for analysis, verification and testing of the operational

risk framework, and required disclosures.

---------------------------------------------------------------------------

\16\ In this document, the terms ``database'' and ``data

warehouse'' are used interchangeably to refer to a collection of

data arranged for easy retrieval using computer technology.

---------------------------------------------------------------------------

Management will need to identify those responsible for maintaining

the data warehouse. In particular, policies and processes will need to

be developed for delivering, storing, retaining, and updating the data

warehouse. Policies and procedures must also cover the edit checks for

data input functions, as well as the requirements for the testing and

verification function to verify data integrity. Like other areas of the

operational risk framework, it is critical that management ensure

accountability for ongoing data maintenance, as this will impact

operational risk management and measurement efforts.

XI. Testing and Verification

Supervisory Standards

S 32. The institution must test and verify the accuracy and

appropriateness of the operational risk framework and results.

S 33. Testing and verification must be done independently of the

firm-wide operational risk management function and the institution's

lines of business.

The operational risk framework must provide for regular and

independent testing and verification of operational risk management

policies, processes and measurement systems, as well as operational

risk data capture systems. For most institutions, operational risk

verification and testing will primarily be done by the audit function.

Internal and external audits can provide an independent assessment of

the quality and effectiveness of the control systems' design and

performance. However, institutions may use other independent internal

units (e.g. quality assurance) or third parties. The testing and

verification function, whether internally or externally performed,

should be staffed by qualified individuals who are independent from the

firm-wide operational risk management function and the institution's

lines of business.

The verification of the operational risk measurement system should

include the testing of:

[sbull] Key operational risk processes and systems;

[sbull] Data feeds and processes associated with the operational

risk measurement system;

[sbull] Adjustments to empirical operational risk capital

estimates, including operational risk exposure;

[sbull] Periodic certification of operational risk models used and

their underlying assumptions; and

[sbull] Assumptions underlying operational risk exposure, data

decision models, and operational risk capital charge.

The operational risk reporting processes should be periodically

reviewed for scope and effectiveness. The institution should have

independent verification processes to ensure the timeliness, accuracy,

and comprehensiveness of operational risk reporting systems, both at

the firm-wide and the line of business levels.

Independent verification and testing should be done to ensure the

integrity and applicability of the operational risk framework,

operational risk exposure/loss data, and the underlying assumptions

driving the regulatory capital measurement process. Appropriate

reports, summarizing operational risk verification and testing findings

for both the independent firm-wide risk management function and lines

of business should be provided to appropriate management and the board

of directors or a designated board committee.

Appendix A: Supervisory Standards for the AMA

S 1. The institution's operational risk framework must include

an independent firm-wide operational risk management function, line

of business management oversight, and independent testing and

verification functions.

S 2. The board of directors must oversee the development of the

firm-wide operational risk framework, as well as major changes to

the framework. Management roles and accountability must be clearly

established.

S 3. The board of directors and management must ensure that

appropriate resources are allocated to support the operational risk

framework.

S 4. The institution must have an independent operational risk

management function that is responsible for overseeing the

operational risk framework at the firm level to ensure the

development and consistent application of operational risk policies,

processes, and procedures throughout the institution.

S 5. The firm-wide operational risk management function must

ensure appropriate reporting of operational risk exposures and loss

data to the board of directors and senior management.

S 6. Line of business management is responsible for the day-to-

day management of operational risk within each business unit.

S 7. Line of business management must ensure that internal

controls and practices within their line of business are consistent

with firm-wide policies and procedures to support the management and

measurement of the institution's operational risk.

S 8. The institution must have policies and procedures that

clearly describe the major elements of the operational risk

management framework, including identifying, measuring, monitoring,

and controlling operational risk.

S 9. Operational risk management reports must address both firm-

wide and line of business results. These reports must summarize

operational risk exposure, loss experience, relevant business

environment and internal control assessments, and must be produced

no less often than quarterly.

S 10. Operational risk reports must also be provided

periodically to senior management and the board of directors,

summarizing relevant firm-wide operational risk information.

S 11. An institution's internal control structure must meet or

exceed minimum regulatory standards established by the Agencies.

S 12. The institution must demonstrate that it has appropriate

internal loss event data, relevant external loss event data,

assessments of business environment and internal controls factors,

and results from scenario analysis to support its operational risk

management and measurement framework.

S 13. The institution must include the regulatory definition of

operational risk as the baseline for capturing the elements of the

[[Page 45988]]

AMA framework and determining its operational risk exposure.

S 14. The institution must have clear standards for the

collection and modification of the elements of the operational risk

AMA framework.

S 15. The institution must have at least five years of internal

operational risk loss data \17\ captured across all material

business lines, events, product types, and geographic locations.

---------------------------------------------------------------------------

\17\ With supervisory approval, a shorter initial historical

observation period is acceptable for banks newly authorized to use

an AMA methodology.

---------------------------------------------------------------------------

S 16. The institution must be able to map internal operational

risk losses to the seven loss-event type categories.

S 17. The institution must have a policy that identifies when an

operational risk loss becomes a loss event and must be added to the

loss event database. The policy must provide for consistent

treatment across the institution.

S 18. The institution must establish appropriate operational

risk data thresholds.

S 19. Losses that have any characteristics of credit risk,

including fraud-related credit losses, must be treated as credit

risk for regulatory capital purposes. The institution must have a

clear policy that allows for the consistent treatment of loss event

classifications (e.g., credit, market, or operational risk) across

the organization.

S 20. The institution must have policies and procedures that

provide for the use of external loss data in the operational risk

framework.

S 21. Management must systematically review external data to

ensure an understanding of industry experience.

S 22. The institution must have a system to identify and assess

business environment and internal control factors.

S 23. Management must periodically compare the results of their

business environment and internal control factor assessments against

actual operational risk loss experience.

S 24. Management must have policies and procedures that identify

how scenario analysis will be incorporated into the operational risk

framework.

S 25. The institution must have a comprehensive operational risk

analytical framework that provides an estimate of the institution's

operational risk exposure, which is the aggregate operational loss

that it faces over a one-year period at a soundness standard

consistent with a 99.9 per cent confidence level.

S 26. Management must document the rationale for all assumptions

underpinning its chosen analytical framework, including the choice

of inputs, distributional assumptions, and the weighting across

qualitative and quantitative elements. Management must also document

and justify any subsequent changes to these assumptions.

S 27. The institution's operational risk analytical framework

must use a combination of internal operational loss event data,

relevant external operational loss event data, business environment

and internal control factor assessments, and scenario analysis. The

institution must combine these elements in a manner that most

effectively enables it to quantify its operational risk exposure.

The institution can choose the analytical framework that is most

appropriate to its business model.

S 28. The institution's capital requirement for operational risk

will be the sum of expected and unexpected losses unless the

institution can demonstrate, consistent with supervisory standards,

the expected loss offset.

S 29. Management must document how its chosen analytical

framework accounts for dependence (e.g., correlations) among

operational losses across and within business lines. The institution

must demonstrate that its explicit and embedded dependence

assumptions are appropriate, and where dependence assumptions are

uncertain, the institution must use conservative estimates.

S 30. Institutions may reduce their operational risk exposure

results by no more than 20% to reflect the impact of risk mitigants.

Institutions must demonstrate that mitigation products are

sufficiently capital-like to warrant inclusion in the adjustment to

the operational risk exposure.

S 31. Institutions using the AMA approach for regulatory capital

purposes must use advanced data management practices to produce

credible and reliable operational risk estimates.

S 32. The institution must test and verify the accuracy and

appropriateness of the operational risk framework and results.

S 33. Testing and verification must be done independently of the

firm-wide operational risk management function and the institution's

lines of business.

Dated: July 17, 2003.

John D. Hawke, Jr.,

Comptroller of the Currency.

By order of the Board of Governors of the Federal Reserve

System, July 21, 2003.

Jennifer J. Johnson,

Secretary of the Board.

Dated at Washington, DC, this 11th day of July, 2003.

By order of the Board of Directors.

Federal Deposit Insurance Corporation.

Robert E. Feldman,

Executive Secretary.

Dated: July 18, 2003.

By the Office of Thrift Supervision.

James E. Gilleran,

Director.

[FR Doc. 03-18976 Filed 8-1-03; 8:45 am]

BILLING CODE 4810-33-P

Last Updated 07/07/2003 regs@fdic.gov

Last Updated: August 4, 2024