This is the accessible text file for GAO report number GAO-07-791 
entitled 'Reexamining Regulations: Opportunities Exist to Improve 
Effectiveness and Transparency of Retrospective Reviews' which was 
released on August 15, 2007. 

This text file was formatted by the U.S. Government Accountability 
Office (GAO) to be accessible to users with visual impairments, as part 
of a longer term project to improve GAO products' accessibility. Every 
attempt has been made to maintain the structural and data integrity of 
the original printed product. Accessibility features, such as text 
descriptions of tables, consecutively numbered footnotes placed at the 
end of the file, and the text of agency comment letters, are provided 
but may not exactly duplicate the presentation or format of the printed 
version. The portable document format (PDF) file is an exact electronic 
replica of the printed version. We welcome your feedback. Please E-mail 
your comments regarding the contents or accessibility features of this 
document to Webmaster@gao.gov. 

This is a work of the U.S. government and is not subject to copyright 
protection in the United States. It may be reproduced and distributed 
in its entirety without further permission from GAO. Because this work 
may contain copyrighted images or other material, permission from the 
copyright holder may be necessary if you wish to reproduce this 
material separately. 

Report to Congressional Requesters: 

United States Government Accountability Office: 

GAO: 

July 2007: 

Reexamining Regulations: 

Opportunities Exist to Improve Effectiveness and Transparency of 
Retrospective Reviews: 

GAO-07-791: 

GAO Highlights: 

Highlights of GAO-07-791, a report to congressional requesters 

Why GAO Did This Study: 

Congress and presidents require agencies to review existing regulations 
to determine whether they should be retained, amended, or rescinded, 
among other things. GAO was asked to report the following for agency 
reviews: (1) numbers and types completed from 2001 through 2006; (2) 
processes and standards that guided planning, conducting, and 
reporting; (3) outcomes; and (4) factors that helped or impeded in 
conducting and using them. GAO evaluated the activities of nine 
agencies covering health, safety, environmental, financial, and 
economic regulations and accounting for almost 60 percent of all final 
regulations issued within the review period. GAO also reviewed 
available documentation, assessed a sample of completed reviews, and 
solicited perspectives on the conduct and usefulness of reviews from 
agency officials and knowledgeable nonfederal parties. 

What GAO Found: 

From 2001 through 2006, the selected agencies completed over 1,300 
reviews of existing regulations. The mix of reviews conducted, in terms 
of impetus (mandatory or discretionary) and purpose, varied among 
agencies. Mandatory requirements were sometimes the impetus for 
reviews, but agencies more often exercised their own discretionary 
authorities to review regulations. The main purpose of most reviews was 
to examine the effectiveness of the implementation of regulations, but 
agencies also conducted reviews to identify ways to reduce regulatory 
burdens and to validate the original estimates of benefits and costs. 

The processes and standards guiding reviews varied across agencies and 
the impetus and phase of the review process. They varied by the extent 
to which agencies applied a standards-based approach, incorporated 
public participation, and provided complete and transparent 
documentation. For example, while almost all agencies had standards for 
conducting mandatory reviews, only about half of the agencies had such 
standards for conducting discretionary views. The extent of public 
involvement varied across review phases, with relatively more in the 
selection process for discretionary reviews. Agencies more often 
documented all phases of mandatory reviews compared to discretionary 
reviews. 

The outcomes of reviews included amendments to regulations, changes to 
guidance and related documents, decisions to conduct additional 
studies, and confirmation that existing rules achieved the intended 
results. Mandated reviews, in particular, most often resulted in no 
changes. Agencies noted that discretionary reviews generated additional 
action more often than mandatory reviews. Agencies and nonfederal 
parties generally considered all of the various review outcomes useful. 

Multiple factors helped or impeded the conduct and usefulness of 
retrospective reviews. Agencies identified time and resources as the 
most critical barriers, but also cited factors such as data limitations 
and overlapping or duplicative review requirements. Nonfederal parties 
said that the lack of transparency was a barrier; they were rarely 
aware of the agencies’ reviews. Both agencies and nonfederal parties 
identified limited public participation as a barrier. To help improve 
the conduct and usefulness of reviews, agencies and nonfederal parties 
suggested practices such as pre-planning to identify data needed to 
conduct effective reviews, a prioritization process to address time and 
resource barriers, high-level management support, grouping related 
regulations together when conducting reviews, and making greater use of 
diverse communication technologies and venues to promote public 
participation. 

What GAO Recommends: 

GAO recommends that agencies incorporate various elements into their 
policies and procedures to improve the effectiveness and transparency 
of retrospective regulatory reviews and that they identify 
opportunities for Congress to revise and consolidate existing 
requirements. In commenting on a draft of this report, SBA’s Office of 
Advocacy agreed with and provided updated guidance in response to the 
recommendations. OMB reported having no comments on the draft. All 
others provided technical comments. 

[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-07-791]. 

To view the full product, including the scope and methodology, click on 
the link above. For more information, contact Mathew J. Scire at (202) 
512-6806 or sciremj@gao.gov. 

[End of section] 

Contents: 

Letter: 

Results in Brief: 

Background: 

All Selected Agencies Reviewed Existing Regulations, but the Purposes 
of the Reviews Varied: 

Agencies Varied in the Extent to Which They Used Formal Processes and 
Standards for Selecting, Conducting, and Reporting on Retrospective 
Reviews: 

Reviews Resulted in a Variety of Outcomes that Were Considered Useful, 
but Mandatory Reviews Most Often Resulted in No Change: 

Agencies and Nonfederal Parties Identified Barriers and Facilitators to 
the Conduct and Usefulness of Retrospective Reviews: 

Conclusions: 

Recommendations for Executive Action: 

Matters for Congressional Consideration: 

Agency Comments: 

Appendix I: Objectives, Scope and Methodology: 

Appendix II: Department of Agriculture Retrospective Reviews: 

Appendix III: Department of Justice Retrospective Reviews: 

Appendix IV: Department of Labor Retrospective Reviews: 

Appendix V: Department of Transportation Retrospective Reviews: 

Appendix VI: Consumer Product Safety Commission Retrospective Reviews: 

Appendix VII: Environmental Protection Agency Retrospective Reviews: 

Appendix VIII: Federal Communications Commission Retrospective Reviews: 

Appendix IX: Federal Deposit Insurance Corporation Retrospective 
Reviews: 

Appendix X: Small Business Administration Retrospective Reviews: 

Appendix XI: Department of Labor's Employee Benefit Security 
Administration Example of a Documented Formal Review: 

Appendix XII: Comments from the Small Business Administration Office of 
Advocacy: 

Appendix XIII: GAO Contact and Acknowledgments: 

Tables: 

Table 1: Impetus for Mandatory and Discretionary Reviews Conducted 
Between 2001 and 2006: 

Table 2: Retrospective Reviews for which Results Were Reported 2001- 
2006, for Selected Agencies: 

Table 3: Examples of Overlapping Timing and Review Factors in 
Retrospective Review Requirements: 

Table 4: Description of USDA Retrospective Reviews: 

Table 5: Description of DOJ Retrospective Reviews: 

Table 6: Description of DOL Retrospective Reviews: 

Table 7: Description of DOT Retrospective Reviews: 

Table 8: Description of CPSC Retrospective Reviews: 

Table 9: Description of EPA Retrospective Reviews: 

Table 10: Description of FCC Retrospective Reviews: 

Table 11: Description of FDIC Retrospective Reviews: 

Table 12: Description of SBA Retrospective Reviews: 

Figures: 

Figure 1: Agencies' Use of Standards-based Approaches in the Review 
Process, for Discretionary and Mandatory Reviews: 

Figure 2: Agencies' Incorporation of Public Involvement in the Review 
Process, for Discretionary and Mandatory Reviews: 

Figure 3: Agencies' Documentation of the Review Process, for 
Discretionary and Mandatory Reviews: 

Figure 4: Modifications to Rules Listed in the December 2006 Unified 
Agenda by Type of Review for Selected Agencies: 

Figure 5: USDA Review Process: 

Figure 6: ATF Section 610 Review, Explosive Materials in the Fireworks 
Industry: 

Figure 7: EBSA Retrospective Review Process: 

Figure 8: OSHA Retrospective Review Process: 

Figure 9: Illustration of DOT's Review Program: 

Figure 10: CPSC Retrospective Review Process: 

Figure 11: EPA Retrospective Review Process: 

Figure 12: FCC Retrospective Review Processes: 

Figure 13: FDIC Retrospective Review Process: 

Figure 14: SBA's General Retrospective Review Process: 

Abbreviations: 

AMS: Agricultural Marketing Service: 
APHIS: Animal and Plant Health Inspection Service: 
ATF: Bureau of Alcohol, Tobacco, Firearms, and Explosives: 
CFR: Code of Federal Regulations: 
CPSC: Consumer Product Safety Commission: 
CRA: Congressional Review Act: 
DEA: Drug Enforcement Administration: 
DOJ: Department of Justice: 
DOL: Department of Labor: 
DOT: Department of Transportation: 
EBSA: Employee Benefits Security Administration: 
EGRPRA: Economic Growth and Regulatory Paperwork Reduction Act: 
EPA: Environmental Protection Agency: 
ETA: Employment and Training Administration: 
FAA: Federal Aviation Administration: 
FCC: Federal Communications Commission: 
FDIC: Federal Deposit Insurance Corporation: 
FFIEC: Federal Financial Institutions Examination Council: 
FSIS: Food Safety and Inspection Service: 
FTE: full-time equivalent: 
GPRA: Government Performance and Results Act of 1993: 
MSHA: Mine Safety and Health Administration: 
NHTSA: National Highway Traffic Safety Administration: 
OIRA: Office of Information and Regulatory Affairs: 
OMB: Office of Management and Budget: 
OSHA: Occupational Safety and Health Administration: 
PART: Program Assessment Rating Tool: 
PRA: Paperwork Reduction Act: 
RFA: Regulatory Flexibility Act: 
SBA: Small Business Administration: 
SEISNOSE: significant economic impact upon a substantial number of 
small entities: 
USDA: Department of Agriculture: 

United States Government Accountability Office: 
Washington, DC 20548: 

July 16, 2007: 

The Honorable Joe Barton: 
Ranking Member:
Committee on Energy and Commerce: 
House of Representatives: 

The Honorable Ed Whitfield: 
Ranking Member: 
Subcommittee on Oversight and Investigations: 
Committee on Energy and Commerce: 
House of Representatives: 

Regulation is a basic tool of government. Each year, federal agencies 
issue thousands of regulations to ensure public health and safety, 
protect the environment, and facilitate the effective functioning of 
financial markets, among other goals. The total costs of these 
regulations are estimated to be in the hundreds of billions of dollars, 
and the estimated benefits are even higher. Congress and presidents 
have imposed many procedural and analytical requirements on the 
regulatory process, in part, because of the substantial costs and 
benefits of regulations. The requirements focus predominately on 
agencies' development of new rules, but some, such as Section 610 
reviews required by the Regulatory Flexibility Act (RFA),[Footnote 1] 
call for the evaluation of existing regulations. Such evaluations can 
be done through a variety of activities that may be considered 
retrospective regulatory reviews.[Footnote 2] For purposes of this 
report, we generally use the term retrospective review to mean any 
assessment of an existing regulation, primarily for purposes of 
determining whether (1) the expected outcomes of the regulation have 
been achieved; (2) the agency should retain, amend, or rescind the 
regulation; and/or (3) the actual benefits and costs of the implemented 
regulation correspond with estimates prepared at the time the 
regulation was issued. We have reported that federal and nonfederal 
stakeholders and experts identified such evaluations as highly 
desirable and potentially useful but, at the same time, recognized that 
they may be difficult to do.[Footnote 3] 

To provide insights concerning how agencies evaluate existing 
regulations, you requested that we examine agencies' implementation of 
retrospective regulatory reviews and the results of such reviews. For 
selected agencies, we are reporting on: 

1) the magnitude of retrospective review activity and type of 
retrospective reviews agencies completed from calendar year 2001 
through 2006, including the frequency, impetus (mandatory or 
discretionary), and purposes of the reviews; 

2) the processes and standards that guide agencies' planning, 
conducting, and reporting on reviews, and the strengths and limitations 
of the various review processes and requirements;[Footnote 4] 

3) the outcomes of reviews, including the perceived usefulness of the 
reviews to agencies and the public and how they affected subsequent 
regulatory activities; and: 

4) the factors that appear to help or impede agencies in conducting or 
using retrospective reviews, including which methods, if any, that 
agencies and we identified as most cost-effective for conducting 
reviews. 

We assessed the activities of nine agencies and their relevant 
subagencies from 2001 through 2006, including the Departments of 
Agriculture (USDA), Justice (DOJ), Labor (DOL), and 
Transportation(DOT); Consumer Product Safety Commission (CPSC); 
Environmental Protection Agency (EPA); Federal Communications 
Commission (FCC); Federal Deposit Insurance Corporation (FDIC); and the 
Small Business Administration (SBA).[Footnote 5] We selected these 
agencies because they include Cabinet departments, independent 
agencies, and independent regulatory agencies covering a wide variety 
of regulatory activities in areas such as health, safety, 
environmental, financial, and economic regulation.[Footnote 6] We were 
not able to assess the activities of all regulatory agencies, due to 
time and resource constraints, but, given the diversity and volume of 
federal regulation conducted by the nine selected agencies, the results 
of our assessment should provide a reasonable characterization of the 
variety of retrospective regulatory reviews and the issues associated 
with their implementation. Further, our Federal Rules Database showed 
that the nine agencies accounted for almost 60 percent of all final 
rules published from 2001 through 2006. The nine agencies accounted for 
almost 36 percent of all major rules (for example, rules with at least 
a $100 million impact on the economy) published during that time 
period.[Footnote 7] 

To address our four objectives, we reviewed documentation from the 
selected agencies and interviewed agency officials. We administered and 
collected responses to a structured data collection instrument that 
solicited information on agencies' retrospective review activities and 
lessons learned. We supplemented this data collection by obtaining 
information from the Federal Register, Unified Agenda, and published 
dockets and reports listed on agency Web sites, and by conducting a 
more detailed assessment of a sample of studies completed between 2001 
and 2006. We also solicited perspectives on the usefulness of agency 
reviews and factors that help or impede the conduct of reviews from 
regulatory oversight entities, such as the Office of Information and 
Regulatory Affairs (OIRA) within the Office of Management and Budget 
(OMB), the Office of Advocacy within SBA, and knowledgeable nonfederal 
parties from a variety of sectors (academia, business, public advocacy, 
and state government). We reviewed agency policies, executive orders, 
and statutory requirements to identify what policies and procedures 
agencies have in place to guide planning, conducting, and reporting on 
reviews. Further, to identify the strengths and limitations of the 
processes, we assessed agencies' use of three systematic review 
practices that are important to the effectiveness and transparency of 
agency reviews, including the (1) use of a standards-based approach, 
(2) incorporation of public involvement, and (3) documentation of 
review processes and results.[Footnote 8] In our more detailed 
assessment of a limited sample of retrospective reviews completed 
between 2001 and 2006, we also assessed the agencies' application of 
research and economic standards and practices. The sample that we 
assessed was too small to be generalizable to all agency retrospective 
reviews, but this assessment illustrated some of the strengths and 
limitations that exist in agency reviews. We conducted our work in 
accordance with generally accepted government auditing standards from 
May 2006 through April 2007. (See app. I for a more detailed 
description of our objectives, scope, and methodology.) 

Results in Brief: 

Since 2001, all of the selected agencies conducted multiple 
retrospective reviews of existing regulations. The nine agencies 
reported completing over 1,300 reviews from 2001 through 2006. However, 
it is not possible to compile a complete tally of all reviews that the 
agencies completed, primarily because agencies reported that they did 
not always document reviews that may have followed more informal review 
processes. The mix of reviews conducted--in terms of impetus (mandatory 
or discretionary) and purpose--varied among agencies. The impetus for 
initiating reviews sometimes reflected various governmentwide or 
agency/program-specific mandatory requirements. More often it reflected 
how an agency exercises its own discretionary authorities, such as 
responding to petitions or concerns raised by regulated entities or to 
changes over time in an industry sector or technology. Agency officials 
said that the primary purpose of most reviews was to examine the 
effectiveness of the implementation of regulations, but they also 
conducted reviews to identify opportunities to reduce regulatory 
burdens and to validate the original estimates of benefits and costs 
associated with regulations. 

The processes and standards guiding retrospective regulatory reviews, 
and the strengths and limitations of these processes, varied widely 
across their three major phases, including the selection of regulations 
to review, conduct of the review, and reporting review results. We 
identified three important practices that can have an impact on the 
effectiveness and transparency of each phase of the review process and 
assessed the extent to which they existed within agencies' processes 
and standards, including: (1) use of a standards-based approach, (2) 
incorporation of public involvement, and (3) documentation of review 
processes and results. We found that, while some agencies have not 
developed formal processes or standards that include these practices, 
others have developed detailed, systematic processes. Further, whether 
agencies used these practices often depended on whether they conducted 
discretionary or mandatory reviews. Even for reviews that had 
standards, agencies did not always apply them to their analysis. For 
example, in a Section 610 review, an agency relied only on public 
comments and did not assess the other four factors identified in the 
mandatory review requirements.[Footnote 9] The extent to which agencies 
involved the public varied across review phases, with relatively more 
public involvement in the selection process for discretionary reviews. 
Similarly, the level of documentation of agency reviews varied 
depending on the impetus of the review and the phase of the review 
process. For example, agencies appeared to document the selection, 
conduct, and results of reviews that they initiated in response to 
mandatory requirements more often than they did for reviews conducted 
at their own discretion. 

Agencies reported that their reviews of existing regulations resulted 
in various outcomes, including changes to regulations, changes to 
guidance and related documents, decisions to conduct additional 
studies, and confirmation that existing rules achieved the intended 
results. Agency officials and nonfederal parties who we interviewed 
reported that each of these outcomes could be useful to the agency and 
the public. Agency officials reported that some types of reviews more 
often resulted in subsequent action, such as prompting the agencies to 
complete additional studies or to initiate rulemaking to amend the 
existing rule. In particular, they reported that discretionary reviews 
more often generated additional action than mandatory reviews. For the 
mandatory reviews completed within our time frame, the most common 
result was a decision by the agency that no changes were needed to the 
regulation.[Footnote 10] Nonfederal parties who we interviewed 
generally considered reviews that addressed multiple purposes (such as 
validation of prior estimates of benefits and costs, burden reduction, 
and efficiency and effectiveness) more useful than reviews that focused 
on a single purpose. One of the major findings in our review was the 
difference in the perceived usefulness of mandatory versus 
discretionary regulatory reviews. Generally, the agencies characterized 
the results of discretionary reviews as more productive and more likely 
to generate further action. A primary reason for this appears to be 
that discretionary reviews may better be suited to address emerging 
issues than mandatory reviews with a predetermined time frame. For 
example, USDA's AMS reported completing 11 mandated Section 610 reviews 
since 2001, which resulted in no regulatory changes. For 9 of these 
reviews, the related published Section 610 reports stated that AMS made 
no changes to the regulations because they were modified "numerous 
times" in advance of the 10-year Section 610 review to respond to 
changes in economic and other conditions affecting the industry. 

Agencies and others reported multiple factors that impeded the conduct 
and usefulness of retrospective reviews. Agencies reported that the 
most critical barrier to their ability to conduct reviews was the 
difficulty in devoting the time and staff resources required for 
reviews while also carrying out other mission activities. Another 
barrier included limits on their ability to obtain the information and 
data needed to conduct reviews. Agencies and others also identified 
factors that impeded the usefulness of retrospective reviews. Among 
these impediments were overlapping and duplicative review schedules and 
evaluation factors. While the predetermined review schedules prompt 
agencies to periodically reexamine certain regulations, their timing 
may either be too short to produce meaningful analysis or too long to 
account for ongoing changes that agencies make in their discretionary 
reviews. For example, both EPA and FCC officials reported that their 
annual and/or biannual review requirements do not allow enough time to 
most effectively complete reviews and/or observe new changes before 
starting a subsequent review. Further, our examination of the timing 
and evaluative factors in mandatory review requirements revealed that 
there is considerable overlap, which agencies reported creates 
challenges to efficiently allocating limited resources. Agencies and 
nonfederal parties identified other factors that impeded the usefulness 
of retrospective reviews, including difficulties in limiting reviews to 
a manageable scope, and constraints in agencies' abilities to modify 
some regulations without additional legislative action. Further, they 
identified the lack of public participation as a barrier to the 
usefulness of reviews, but differed in their explanations for the lack 
of public participation. The nonfederal parties also identified the 
lack of transparency in agency review and reporting processes as a 
barrier to the usefulness of reviews to the public; they were rarely 
aware of the retrospective review activities reported to us by the 
agencies. This lack of awareness may be, in part, because of 
limitations in agencies' documentation and reporting of discretionary 
reviews. 

Agency officials and nonfederal parties suggested a number of practices 
to help the conduct and usefulness of regulatory reviews, such as pre- 
planning reviews to identify the data and analysis needed to conduct 
effective reviews, developing a prioritization process to address time 
and resource barriers, ensuring an appropriate level of independence in 
the conduct of reviews to enhance credibility and effectiveness, 
obtaining high-level management support to sustain commitment to a 
review program and follow up on the review results, grouping related 
regulations together when conducting reviews, making greater use of 
informal networks to promote public participation, and targeting the 
level of detail and type of product used to report results to meet the 
needs of different audiences. Nonfederal parties also suggested that it 
would be more useful if agencies selected a few (two to five) high- 
priority regulations to substantively review each year rather than 
conducting a cursory review of many regulations. We found that only a 
few agencies tracked the costs associated with conducting their 
reviews, so we were unable to identify the most cost-effective 
approaches. However, agency officials told us that reviews have 
resulted in cost savings to their agencies and to regulated parties. 
For example, DOL's MSHA reported that it has a policy to review and 
modify regulations that regulated parties repeatedly petition for 
waivers. As a result, MSHA officials believe that they save the time 
and resources that would otherwise be spent on repetitively reviewing 
and responding to similar petitions. 

We are making recommendations to executive agencies to improve the 
effectiveness and transparency of reviews through incorporating, where 
appropriate, various elements into the policies and procedures that 
govern their retrospective review activities, such as establishing 
plans for measuring the performance of regulations, a prioritization 
mechanism for review activities based upon defined evaluation criteria, 
and minimum standards for documenting and reporting review results, 
among other things. We also recommend that the Administrator of OIRA 
and the Chief Counsel for Advocacy work with regulatory agencies to 
identify opportunities for Congress to revise the timing and scope of 
existing requirements and/or consolidate existing requirements. In 
addition, we suggest that Congress authorize a pilot program with 
selected agencies to test the effectiveness of allowing agencies to 
satisfy various retrospective review requirements with similar 
objectives and evaluation factors that can be consolidated into one 
review that is reported to all of the appropriate relevant parties and 
oversight bodies. In formal comments on the draft report, the SBA 
Office of Advocacy concurred with the recommendations and, as an 
attachment, provided a copy of draft guidance that they developed in 
response to our recommendations. OMB told us that they reviewed our 
draft report and had no comments. All other agencies provided technical 
and editorial comments, which we incorporated as appropriate. In its 
technical comments, one agency suggested that one recommendation to 
identify opportunities to consolidate existing retrospective reviews 
for statutory requirements could be broadened to include executive 
requirements. However, we limit our recommendation to statutory 
retrospective review requirements because they tend to be recurring 
and/or on a predetermined review schedule. 

Background: 

The performance and accountability of government agencies and programs 
have attracted substantial attention by Congress, the executive branch, 
and others, including GAO. For example, the Government Performance and 
Results Act of 1993 (GPRA) established a statutory framework designed 
to provide congressional and executive decision makers with objective 
information on the relative effectiveness and efficiency of federal 
programs and spending.[Footnote 11] A central element of the current 
administration's Presidential Management Agenda is the Program 
Assessment Rating Tool (PART), designed by OMB to provide a consistent 
approach to assessing federal programs in the executive budget 
formation process.[Footnote 12] Over the past 2 years, we have 
emphasized that the long-term fiscal imbalance facing the United States 
and other significant trends and challenges establish the need to 
reexamine the base of the federal government and its existing programs, 
policies, functions, and activities.[Footnote 13] We noted that a top- 
to-bottom review of federal programs and policies is needed to 
determine if they are meeting their objectives.[Footnote 14] To support 
this reexamination, the policy process must have the capacity to 
provide policymakers not only with information to analyze the 
performance and results achieved by specific agencies and programs, but 
that of broad portfolios of programs and tools (including regulation) 
contributing to specific policy goals. 

While initiatives such as GPRA and PART can evaluate regulatory 
performance at the agency or program level, Congress and presidents 
also have instituted requirements that focus on a more basic element, 
the agencies' existing regulations.[Footnote 15] For example, through 
Section 610, Congress requires agencies to review all regulations that 
have or will have a "significant economic impact upon a substantial 
number of small entities" (generally referred to as SEISNOSE) within 10 
years of their adoption as final rules. The purpose of these reviews is 
to determine whether such rules should be continued without change, or 
should be amended or rescinded, consistent with the stated objectives 
of applicable statutes, to minimize impacts on small entities. As 
discussed later in this report, Congress also established other 
requirements for agencies to review the effects of regulations issued 
under specific statutes, such as the Clean Air Act.[Footnote 16] Every 
president since President Carter has directed agencies to evaluate or 
reconsider existing regulations. For example, President Carter's 
Executive Order 12044 required agencies to periodically review existing 
rules; one charge of President Reagan's task force on regulatory relief 
was to recommend changes to existing regulations; President George H.W. 
Bush instructed agencies to identify existing regulations to eliminate 
unnecessary regulatory burden; and President Clinton, under section 5 
of Executive Order 12866, required agencies to develop a program to 
"periodically review" existing significant regulations.[Footnote 17] In 
2001, 2002, and 2004, the administration of President George W. Bush 
asked the public to suggest reforms of existing regulations. 

The Office of Advocacy within SBA and OIRA within OMB have issued 
guidance to federal agencies on the implementation of, respectively, 
the RFA and Executive Order 12866.[Footnote 18] The available guidance 
documents, including OMB Circular A-4 on regulatory analysis, focus 
primarily on the procedural and analytical steps required for reviewing 
draft regulations, but they are also generally applicable whenever 
agencies analyze the benefits and costs of regulations, including those 
of existing regulations. However, the documents provide limited 
guidance focused specifically on retrospective reviews. In a short 
discussion on periodic reviews of existing rules pursuant to Section 
610, the Office of Advocacy's RFA guidance briefly touches on planning, 
conducting, and reporting the results of reviews, but the OMB/OIRA 
guidance does not specifically address the executive order's periodic 
review requirement.[Footnote 19] 

In our prior work on this subject, we found that agencies infrequently 
performed certain types of reviews and identified potential challenges 
and benefits of conducting retrospective reviews.[Footnote 20] In 1999, 
we reported that assessments of the costs and benefits of EPA's 
regulations after they had been issued had rarely been done.[Footnote 
21] In a series of reports on agencies' compliance with Section 610 
requirements, we again noted that reviews were not being 
conducted.[Footnote 22] We identified a number of challenges to 
conducting retrospective reviews. In general, these included the 
difficulties that regulatory agencies face in demonstrating the results 
of their work, such as identifying and collecting the data needed to 
demonstrate results, the diverse and complex factors that affect 
agencies' results (for example, the need to achieve results through the 
actions of third parties), and the long time period required to see 
results in some areas of federal regulation.[Footnote 23] We also 
identified concerns about the balance of regulatory analyses, because 
it may be more difficult to estimate the benefits of regulations than 
it is to estimate the costs. Our report on EPA's retrospective studies 
noted that such studies were viewed as hard to do because of the 
difficulty in obtaining valid cost data from regulated entities and 
quantifying actual benefits, among other reasons. Our work on agencies' 
implementation of Section 610 requirements revealed that there was 
confusion among the agencies regarding the meaning of key terms such as 
SEISNOSE, what RFA considers to be a "rule" that must be reviewed, and 
whether amending a rule within the 10-year period provided in Section 
610 would "restart the clock," among other things. 

However, our prior work also pointed out that retrospective evaluation 
could help inform Congress and other policymakers about ways to improve 
the design of regulations and regulatory programs, as well as play a 
part in the overall reexamination of the base of the federal 
government.[Footnote 24] For example, we reported that retrospective 
studies provided insights on a market-based regulatory approach to 
reduce emissions that cause acid rain and that the studies found that 
the actual costs of reducing emissions were lower than initially 
estimated. Experts and stakeholders whom we consulted during work on 
federal mandates (including regulations) and economic and regulatory 
analyses told us that they believe more retrospective analysis is 
needed and, further, that there are ways to improve the quality and 
credibility of the analyses that are done.[Footnote 25] One particular 
reason cited for the usefulness of retrospective reviews was that 
regulations can change behavior of regulated entities, and the public 
in general, in ways that cannot be predicted prior to implementation. 

All Selected Agencies Reviewed Existing Regulations, but the Purposes 
of the Reviews Varied: 

Since 2001, the nine selected agencies conducted multiple retrospective 
reviews of their existing regulations to respond to mandatory and 
discretionary authorities. Between 2001 and 2006, the nine agencies 
reported completing over 1,300 mandatory or discretionary reviews, but 
many could not tally the total number of reviews that they have 
conducted. The reviews addressed multiple purposes, such as examining 
the efficiency and effectiveness of regulations and identifying 
opportunities to reduce regulatory burden. The mix of reviews 
conducted--in terms of the impetus to start a review and the purpose of 
the review--varied not only across but also within the agencies that we 
examined. 

Agencies Conducted Most Reviews on Their Own Discretion Rather than in 
Response to Mandatory Requirements, but the Total Number Is Unknown: 

Agencies conducted reviews in response to various mandatory 
requirements, but most agencies indicated that they conducted the 
majority of reviews based on their own discretion. Agencies reported 
conducting reviews, at their own discretion, in response to their 
agencies' own formal policies and procedures to conduct reviews or to 
respond to accidents or similar events, changes in technology and 
market conditions, advances in science, informal agency feedback, and 
petitions, among other things. Among the main mandatory sources of 
reviews were governmentwide statutory requirements (such as Section 610 
of the RFA or the Paperwork Reduction Act (PRA)), agency-or program- 
specific statutory requirements (such as Section 402 of the 
Telecommunications Act of 1996 that requires FCC to review the 
regulations promulgated under the act every 2 years),[Footnote 26] 
Executive Order 12866 on Regulatory Planning and Review (which requires 
agencies to periodically conduct reviews of their significant 
regulations),[Footnote 27] and other executive branch directives (such 
as the memorandum on plain language). In addition, agencies conducted 
reviews in response to OMB initiatives to solicit nominations for 
regulatory reexamination, which were not statutorily mandated reviews 
or required by a specific executive order, but were a part of executive 
branch regulatory reform efforts. The frequency of agency reviews 
varied based on review requirements. In some cases, agencies were 
required to conduct reviews every 2 years to 10 years. Available 
information on completed reviews indicated that the numbers of reviews 
completed by individual agencies in any given year varied from a few to 
more than a hundred. 

Agencies' officials reported they conducted discretionary reviews more 
often than mandated studies. These discretionary reviews were often 
prompted by drivers such as informal suggestions or formal petitions 
from regulated entities seeking regulatory changes, suggestions from 
the agency personnel who implement and enforce the regulations, 
departmentwide initiatives to review regulations, or changes in 
particular technologies, industries, or markets. Although these 
discretionary reviews were often undocumented, our review of publicly 
available results for retrospective reviews confirmed that, in some 
instances, agencies like those within DOL, DOT, and DOJ cited 
discretionary drivers as the motivation for their reviews. 

Among the various reasons for agency-initiated reviews, petitions were 
a major review driver for almost all of the agencies. Agency officials 
cited three types of petition activities that largely influenced them 
to conduct regulatory reviews. These petition activities included: (1) 
petitions for rulemaking where the regulated entities or public 
requested a modified rule that includes new or updated information, (2) 
petitions for reconsideration where regulated entities or the public 
requested that the agency revisit an aspect of the rule, and (3) 
petitions for waivers where the regulated entities or public requested 
waivers from regulatory requirements. Some agencies, such as DOT and 
MSHA, have policies to begin a more formal review of the entire 
regulation when there are a substantial number of petitions. 

Agencies also reported conducting reviews in response to various 
mandatory requirements. However, whether they conducted them more often 
in response to governmentwide requirements or requirements specific to 
the policy area or sector that they regulate varied among and within 
agencies. For example, DOT, USDA's APHIS, and SBA conducted many 
mandatory reviews in response to Section 610 requirements, while 
others, such as EPA and FCC, conducted most mandatory reviews in 
response to statutes that applied specifically to areas that they 
regulate. Specifically, all of the mandatory reviews completed by APHIS 
and SBA since 2001 were conducted in response to Section 610 
requirements. Similarly, DOT initiated a systematic 10-year plan for 
reviewing all of its sections of the CFR in order to satisfy the 
provisions of Section 610, along with other mandatory review 
requirements.[Footnote 28] DOT reported completing over 400 reviews 
between 2001 and 2006 under this 10-year plan. However, even within 
DOT, variation existed with regard to which review requirements more 
often prompted reviews. For example, unlike some other DOT agencies, 
FAA conducted the majority of its mandatory reviews in response to 
industry-specific review requirements rather than governmentwide 
retrospective review mandates. 

While EPA, FCC, and FDIC also conduct some reviews in response to 
governmentwide requirements, such as Section 610, they most often 
conducted mandatory reviews to comply with statutes that apply 
specifically to areas that they regulate. EPA officials provided a list 
of seven statutes that require the agency to conduct mandatory 
retrospective reviews, including requirements in the Safe Drinking 
Water Act, Clean Air Act, Clean Water Act, and Federal Food, Drug and 
Cosmetic Act, among others. Similarly, FCC conducts many of its 
mandatory retrospective reviews to comply with the biennial and 
quadrennial regulatory review requirements under the Communications 
Act, as amended. One agency in our review, FDIC, conducted most of its 
mandatory reviews in response to the financial-sector-specific Economic 
Growth and Regulatory Paperwork Reduction Act of 1996 (EGRPRA), which 
requires federal financial regulatory agencies to identify outdated, 
unnecessary, or unduly burdensome statutory or regulatory requirements 
every 10 years.[Footnote 29] Finally, agencies also conducted single 
comprehensive reviews of a regulation as a method to satisfy multiple 
review requirements. For example, DOL's EBSA and OSHA, DOT, and FDIC 
conducted reviews that incorporated Section 610 and other review 
requirements into broader reviews that the agency initiated as part 
either of their regular review program or in response to industry 
feedback and petitions, among other things. (Table 1 illustrates the 
range of mandatory and discretionary reviews conducted by selected 
agencies included in our scope.) 

Table 1: Impetus for Mandatory and Discretionary Reviews Conducted 
Between 2001 and 2006: 

Impetus of review: Agency: USDA;
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: Check; 
Impetus of review: Mandatory reviews: Sector-specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: Check; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: DOJ; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: [Empty]; 
Impetus of review: Mandatory reviews: Sector- specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: [Empty]; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: DOL; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: [Empty]; 
Impetus of review: Mandatory reviews: Sector- specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: Check; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: DOT; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: Check; 
Impetus of review: Mandatory reviews: Sector-specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: Check; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: CPSC; 
Impetus of review: Mandatory reviews: Governmentwide: [Empty]; 
Impetus of review: Mandatory reviews: Agency- specific: [Empty]; 
Impetus of review: Mandatory reviews: Sector- specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: Check; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: EPA; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: Check; 
Impetus of review: Mandatory reviews: Sector-specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: [Empty]; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: FCC; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: Check; 
Impetus of review: Mandatory reviews: Sector-specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: [Empty]; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: FDIC; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: [Empty]; 
Impetus of review: Mandatory reviews: Sector- specific: Check; 
Impetus of review: Discretionary reviews: Agency policy: Check; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Impetus of review: Agency: SBA; 
Impetus of review: Mandatory reviews: Governmentwide: Check; 
Impetus of review: Mandatory reviews: Agency- specific: [Empty]; 
Impetus of review: Mandatory reviews: Sector- specific: [Empty]; 
Impetus of review: Discretionary reviews: Agency policy: Check; 
Impetus of review: Discretionary reviews: Petitions or other forms of 
industry and consumer feedback: Check; 
Impetus of review: Discretionary reviews: Sector changes (e.g., new 
science, technology, or market trends): Check; 
Impetus of review: Discretionary reviews: Internal drivers (e.g., 
improving enforcement or compliance): Check. 

Source: GAO analysis of agency data and regulatory statutes. 

Note: We were not able to verify the conduct of some discretionary 
reviews reported by agencies because they were not documented. 

[End of table] 

The frequency with which agencies were required to conduct certain 
reviews varied depending on the statutory requirement. For example, 
under some statutory requirements, agencies must review certain 
regulations every 2 or 3 years. Other requirements cover a longer 
period, such as Section 610's requirement to revisit certain rules 10 
years after their issuance, or specify no particular time. In addition, 
agency policies on the frequency with which to conduct discretionary 
reviews varied. For example, USDA has a departmentwide requirement 
generally to review economically significant and major regulations 
every 5 years, while FAA conducts its reviews on a 3-year cycle. Some 
agencies, such as DOT, had a departmentwide requirement to review all 
regulations in its CFR within 10 years of the creation of its Unified 
Agenda Regulatory Plan, but provided discretion to its agencies on when 
to conduct reviews during that period. 

Despite a perception expressed by some federal and nonfederal parties 
that agencies are not actively engaged in reviewing their existing 
regulations, we found that agencies reported results for over 1,300 
reviews completed between 2001 through 2006. However, even this number 
may understate the total because it does not account for all of the 
undocumented discretionary reviews conducted by agencies. In addition, 
the available information reported by the agencies (and others, such as 
OMB) may include some duplication or fail to capture additional follow- 
up reviews by the agencies. It is also important to note that the units 
of analysis that agencies used in their reviews, and the scope of 
individual reviews, varied widely. Therefore, the level of agency 
review activity cannot be compared by only assessing the number of 
reviews completed. For example, a review might be as narrowly focused 
as the review that DOT's NHTSA completed in 2005 on 49 CFR part 571.111 
(regarding rearview mirrors) or as broad as a FDIC review that included 
analyses of 131 regulations within a single review. DOT also pointed 
out that because rules vary widely in complexity, even a narrowly 
focused review can be a major undertaking. For example, NHTSA may 
review a one-sentence rule that says, "all motor vehicles sold in the 
U.S. must be equipped with airbags." In another review, they may look 
at 131 regs that impose very minor requirements. It may well be a far 
more time and resource intensive effort for NHTSA to review the effect 
of the one-sentence airbag requirement. Further, because some agencies 
produce many more regulations than others do, the number of reviews 
that agencies reported conducting also should be considered within the 
context of their volume of regulations and rulemaking 
activity.[Footnote 30] 

Table 2 lists the number of reviews that agencies reported completing 
between 2001 and 2006. 

Table 2: Retrospective Reviews for which Results Were Reported 2001- 
2006, for Selected Agencies: 

Agency: USDA; 
Numbers and types of completed reviews that were reported: At least 531 
reviews: 
* APHIS completed 9 Section 610 reviews and numerous other reviews 
resulting in 139 rule modifications; 
* AMS completed 11 Section 610 and 300 discretionary reviews; 
* FSIS completed 1 Section 610 review and 36 discretionary reviews in 
response to petitions; 
* USDA completed 17 reviews in response to OMB regulatory reform 
nominations from 2001 to 2004. 

Agency: DOJ; 
Numbers and types of completed reviews that were reported: At least 20 
reviews: 
* Completed 1 Section 610 review; 
* Completed 10 discretionary reviews and 2 other reviews for agency-
specific requirements; 
* Completed 7 reviews in response to OMB regulatory reform nominations 
from 2001 to 2004. 

Agency: DOL; 
Numbers and types of completed reviews that were reported: At least 60 
reviews: 
* EBSA completed 7 reviews, including 4 Section 610 reviews; 
* OSHA completed 4 Section 610 reviews; 
* MSHA completed 4 discretionary reviews; 
* ETA completed 3 discretionary reviews; 
* All DOL agencies conducted reviews under departmentwide initiatives 
in 2005 and 2006, including those listed for the agencies above; 
* Completed 42 reviews in response to OMB regulatory reform nominations 
from 2001 to 2004. 

Agency: DOT; 
Numbers and types of completed reviews that were reported: At least 488 
reviews: 
* Completed 406 reviews per the departmentwide 10-year regulatory 
review plan, which addresses mandatory requirements such as Section 
610; 
* Completed a stand-alone regulatory review initiative in 2005 (that 
generated 120 public comments to which DOT responded); 
* FAA conducted reviews of 3 regulations per requirements in the 
Federal Aviation Authorization Act of 1996; 
* NHTSA conducted 17 other reviews of existing regulations; 
* Completed 61 reviews in response to OMB regulatory reform nominations 
from 2001 to 2004. 

Agency: CPSC; 
Numbers and types of completed reviews that were reported: At least 4 
reviews: 
* Conducted as part of the agency's regulatory review program. 

Agency: EPA; 
Numbers and types of completed reviews that were reported: At least 
156: 
* Completed 14 Section 610 reviews; 
* Completed 20 reviews in response to agency specific mandates; 
* Completed 6 discretionary reviews; 
* Completed 116 reviews in response to OMB regulatory reform 
nominations from 2001 to 2004. 

Agency: FCC; 
Numbers and types of completed reviews that were reported: At least 47 
reviews: 
* Completed 1 Section 610 review (2006); 
* Completed 15 biennial reviews of all telecommunications rules (2002, 
2004, 2006); 
* Completed 2 quadrennial reviews of all broadcast ownership rules 
(2002, 2006); 
* Completed 1 review of rules regarding the pricing of unbundled 
network elements (2003); 
* Completed 2 periodic reviews of digital television conversion 
process; 
* Completed 26 reviews in response to OMB regulatory reform nominations 
from 2001 to 2004. 

Agency: FDIC; 
Numbers and types of completed reviews that were reported: At least 4 
reviews: 
* Completed 2 Section 610 reviews; 
* Completed 1 EGRPRA review, in conjunction with the members of the 
Federal Financial Institutions Examination Council (FFIEC); 
* Completed 1 discretionary review. 

Agency: SBA; 
Numbers and types of completed reviews that were reported: At least 32 
reviews; 
* Completed 4 Section 610 reviews; 
* Completed 27 reviews in response to petitions; 
* Completed 1 review in response to OMB regulatory reform nominations 
from 2001 to 2004. 

Agency: Total; 
Numbers and types of completed reviews that were reported: At least 
1,300 reviews conducted in response to various requirements and 
drivers. 

Source: GAO analysis of agency and other publicly available data 
sources. 

[A] The numbers in this table represent individual reviews conducted 
and not the number of regulations reviewed because, in some cases, one 
review covered multiple regulations. 

[B] The information presented in this table also only reflects reviews 
reported by the agencies as having been completed. The agencies also 
reported initiating other reviews during this time period that are not 
yet complete. 

[End of table] 

Agency Officials Reported that They Most Often Conducted Reviews to 
Assess the Effectiveness of Their Regulations: 

According to agency officials, they conducted reviews for various 
purposes but most often focused on assessing the effectiveness of 
regulations. Agency officials reported conducting reviews to evaluate 
or identify (1) the results produced by existing regulations, including 
assessments to validate the original estimates of projected benefits 
and costs associated with the regulation; (2) ways to improve the 
efficiency or effectiveness of regulatory compliance and enforcement; 
and (3) options for reducing regulatory burdens on regulated 
entities.[Footnote 31] Overall, agency officials reported that their 
reviews more often focused on improving effectiveness, with burden 
reduction as a secondary consideration, even for review requirements 
that are geared toward identifying opportunities for burden reduction, 
such as those within Section 610. The approaches that agencies reported 
taking to assess "effectiveness" varied, including measuring whether 
the regulations were producing positive outcomes, facilitated industry 
compliance with relevant statutes, and/or were assisting the agency in 
accomplishing its goals. For example, DOL officials reported that the 
agency attempts to assess "effectiveness" by measuring improvements 
resulting from the regulation and by analyzing factors required by 
Section 610 of RFA and Section 5 of Executive Order 12866, such as 
whether there is new technology, excessive complexity, conflict with 
other regulations, or whether cost-effectiveness can be improved. 
However, the agency does not conduct what would be considered a 
traditional benefit-cost economic analysis. Other agencies, such as 
EPA, reported assessing "effectiveness" by determining whether the 
regulation is achieving its intended goal as identified in related 
statutes. For example, officials from EPA reported that their 
retrospective review of the Clean Water Act helped them estimate the 
extent to which toxic pollutants remained, thereby aiding them to 
assess the effectiveness of each existing regulation in various 
sections of the act. Since the goal of the Clean Water Act is zero 
discharge of pollutants, the review was an indicator of the progress 
the agency had made toward the statutory goals. Our limited review of 
agency summaries and reports on completed retrospective reviews 
revealed that agencies' reviews more often attempted to assess the 
effectiveness of their implementation of the regulation rather than the 
effectiveness of the regulation in achieving its goal. 

Agencies Varied in the Extent to Which They Used Formal Processes and 
Standards for Selecting, Conducting, and Reporting on Retrospective 
Reviews: 

The use of systematic evaluation practices and the development of 
formal retrospective regulatory review processes varied among and even 
within the agencies.[Footnote 32] To assess the strengths and 
limitations of various agency review processes, we examined the three 
phases of the process: the selection of regulations to review, conduct 
of reviews, and reporting of review results. We identified three 
practices that are important for credibility in all phases of the 
review, including the extent to which agencies (1) employed a standards-
based approach, (2) involved the public, and (3) documented the process 
and results of each phase of the review. The use of these evaluation 
practices in agency review processes, and the development of formal 
policies and procedures to guide their review processes, varied among 
the agencies. Furthermore, whether agencies used these practices often 
varied according to whether they conducted discretionary or mandatory 
reviews.[Footnote 33] Fewer agencies used standards-based approaches or 
documented the selection, conduct, or reporting of reviews when they 
were discretionary. While more agencies incorporated public involvement 
in the selection of regulations for review in discretionary reviews, 
fewer included public involvement in the conduct of those reviews. 
Generally, agencies did not consistently incorporate the three 
practices we identified into their discretionary reviews as often as 
they did for their mandatory reviews. However, it is important to note 
that some agencies have recently developed their review programs and 
others are attempting to find ways to further develop and improve their 
retrospective review processes (i.e., establishing units that focus on 
retrospective reviews and seeking assistance with establishing 
prioritization systems). For example, CPSC and EBSA recently 
established their review programs, and ETA recently established a 
centralized unit to develop retrospective review processes for the 
agency. Furthermore, although the process has been delayed because of 
other regulatory priorities, MSHA recently sought contractor assistance 
with developing a more standard selection process for its reviews. 
(Additional details on each agency's review process can be seen in 
apps. II through XI). 

Agencies Employed Standards-Based Approaches More Often in Mandatory 
Reviews: 

All of the agencies in our review reported that they have practices in 
place to help them select which of their existing regulations to 
review. Some of these agencies have established or are establishing 
standard policies and procedures, and guidance for this selection. 
While almost all of the agencies reported having standards to select 
regulations for their mandated reviews because the mandates identified 
which regulations agencies must review or prescribed standards for 
selecting regulations to review, fewer agencies had yet developed 
formal standards for selecting regulations to review under their own 
discretion. Agencies that had established such processes reported that 
they were useful in determining how to prioritize agency review 
activities. For example, DOL's EBSA and CPSC established detailed 
standards for the selection of their discretionary reviews, which they 
used to prioritize their retrospective review activities and the 
corresponding use of agency resources. The officials reported that 
their standards-based selection processes allowed them to identify 
which regulations were most in need of review and to plan for 
conducting those reviews. Furthermore, officials from both agencies 
reported that their prioritization processes allowed them to focus on 
more useful retrospective review activities, which resulted in 
identifying important regulatory changes. We observed that this 
standards-based approach to selecting regulations also increased the 
transparency of this phase of the review process. We were better able 
to determine how these agencies selected which regulations to review. 
Further, as identified by CPSC officials and others, applying a 
standards-based approach to selecting regulations to review can provide 
a systematic method for agencies to assess which regulations they 
should devote their resources toward, as they balance retrospective 
review activities with other mission-critical priorities. The 
consequence of not using a standards-based approach could result in 
diverting attention from regulations that, through criteria, agencies 
could identify as needing the most consideration. Otherwise, agencies 
may instead focus their reviews on regulations that would warrant less 
attention. In addition--for each phase of the review process--using a 
standards-based approach can allow agencies to justify the 
appropriateness of the criteria that they use (either because they are 
objective, or at least mutually agreed upon), and thus gain credibility 
for their review. Selecting a different criterion or set of standards 
for each review could imply a subjective evaluation process and 
possibly an arbitrary treatment of one regulation versus another. 

Similarly, agencies varied in their use of a standards-based approach 
when analyzing regulations in reviews. While most of the mandatory 
review requirements identified by the selected agencies establish 
standard review factors that agencies should consider when conducting 
their reviews--which agencies reported following--about half of the 
agencies within our review have formal policies and procedures that 
establish a standards-based approach for conducting discretionary 
reviews. Specifically, the five agencies that had formal procedures 
that established standards defined the steps needed to conduct reviews 
and review factors used to assess regulations. Other agencies had not 
yet established written guidance or policies to guide their conduct of 
reviews or define the analytical methods and standards they should use 
to assess regulations. When we identified agencies that did specify 
objectives and review factors that they used when conducting reviews, 
they ranged from general (such as to identify if they were still needed 
or could be simplified) to specific (such as to account for new 
developments in practices, processes, and control technologies when 
assessing emission standards). We observed that the more specific sets 
of evaluative factors that agencies considered in different types of 
reviews shared several common elements, such as prompting agencies to 
consider comments and complaints received by the public and the 
regulation's impact on small entities. 

Our assessment of a small sample of agency reviews revealed that, even 
when relevant standards were available to agencies for conducting 
reviews, they did not always apply them. In one case, the economic 
analyses conducted in the agency review did not employ some of the 
relevant best practices identified for these types of analyses in OMB's 
guidance. In another case, in conducting a mandated Section 610 review, 
the agency relied only on public comments to evaluate the regulation 
and did not provide any additional assessment on the other four factors 
identified in the mandatory review requirements.[Footnote 34] According 
to agency documentation, it (1) provided no further assessment of the 
factors and (2) concluded that the regulation did not need changes, 
because the agency received no public comments. Conversely, our review 
of another agency's Section 610 assessment of regulations provided an 
example of how agencies that apply standards can use the reviews to 
produce substantive results. Because the agency relied both on public 
comments and its own assessment of the five Section 610 review factors/ 
standards, the agency identified additional changes that would not have 
been identified if it had relied on only one of these standards. When 
we assessed whether agencies applied other generally accepted review 
standards, such as identifying how well they implemented the 
regulation, whether there was a pressing need for the regulation, or 
whether intended or unintended effects resulted from the regulation, we 
observed that some agencies' analyses did not. 

Most of the agencies within our review had established policies on 
reporting the results of their mandatory retrospective reviews to 
senior management officials and to the public, although in some cases, 
there was no requirement to do so. For example, Section 610 requires 
federal agencies to report on the initiation of a review, but does not 
require agencies to report the findings or policy decisions resulting 
from the review. Nevertheless, as a matter of practice, all of the 
agencies within our review reported information on the results of their 
Section 610 reviews, though the content and level of detail varied. 
Conversely, about half of the agencies in our review had not 
established written policies or procedures for reporting the results of 
their discretionary retrospective reviews to the public. As a result, 
we found inconsistencies in practices within agencies on how and 
whether they reported the results of these reviews to the public, 
resulting in less transparency in this process. For example, agencies 
within DOT, USDA, and some within DOL indicated that, at times, they 
report review results of discretionary reviews in the preambles of 
proposed rule changes or through other mechanisms, but they did not 
consistently do so. Agencies also reported that they often do not 
report the results of discretionary reviews at all, if they did not 
result in a regulatory change. This lack of transparency may be the 
cause for assertions from nonfederal parties that they were unaware 
that agencies conducted discretionary reviews of their existing 
regulations. Figure 1 illustrates the differences in agencies' use of a 
standards-based approach in the three phases of the review process, for 
discretionary and mandatory reviews. 

Figure 1: Agencies' Use of Standards-based Approaches in the Review 
Process, for Discretionary and Mandatory Reviews: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[A] One of the agencies within our review did not have any regulations 
that were subject to mandatory review requirements. 

[End of figure] 

Level of Public Participation Varied by Impetus and Phase of the Review 
Process: 

Agency practices varied for soliciting and incorporating public input 
during the selection of regulations to review. For example, in 2005 DOT 
formally requested nominations from the public on which regulations it 
should review. Further, in the 2006 semi-annual Unified Agenda, the 
agency sought public suggestions for which regulations it should 
review. However, agencies in our review more often reported that they 
solicit public input on which regulations to review during informal 
meetings with their regulated entities for their discretionary reviews. 
Techniques used by some of the agencies to obtain public input were 
informal networks of regulated entities, agency-sponsored listening 
sessions, and participation in relevant conferences. For example, USDA 
officials reported that they regularly meet with industry committees 
and boards and hold industry listening sessions and public meetings to 
obtain feedback on their regulations. DOJ's DEA reported holding yearly 
conferences with industry representatives to obtain their feedback on 
regulations. While almost all of the agencies in our review reported 
soliciting public input in a variety of ways for their discretionary 
reviews, SBA relied primarily on the Federal Register's Unified Agenda 
to inform the public about their reviews. For mandatory reviews, 
agencies appeared to do less outreach to obtain public input on the 
selection of regulations to review. However, it is important to note 
that such public input into the selection phase of mandated reviews may 
not be as appropriate because agencies have less discretion in choosing 
which regulations to review under specific mandates. 

Almost all agencies within our review reported soliciting public input 
into the conduct of their mandatory reviews, either through the notices 
in the Federal Register, or in more informal settings, such as 
roundtable discussions with industries, or both. For these reviews, 
agencies appeared to almost always place notices in the Federal 
Register, including the semiannual Unified Agenda, soliciting public 
comments, but nonfederal parties cited these tools as being ineffective 
in communicating with the public because these sources are too 
complicated and difficult to navigate.[Footnote 35] Our review of the 
Federal Register confirmed that agencies often provided such notice and 
opportunity for public comment on these regulatory reviews. In 
addition, some agencies, such as DOJ's ATF, USDA's AMS, and FCC, 
reported on the agencies' analysis of the public comments and their 
effect on the outcome of the review. However, we were not always able 
to track such notices or discussions of public input into the conduct 
of discretionary reviews. Officials from DOL's MSHA and ETA stated 
that, if internal staff generated the review and it was technical in 
nature, the agency might not include the public when conducting a 
review. However, we were able to observe that some agencies, such as 
FCC, DOT, and EPA, did post comments received from petitioners and 
solicited public comments on these types of discretionary reviews in 
order to inform their analyses. 

Most agencies within our review did not solicit or incorporate public 
input on the reported results of their reviews. We are only aware of a 
few agencies (NHTSA, FCC, and EPA) that provided opportunities for 
additional public feedback on the analysis of their regulations before 
making final policy decisions. Figure 2 illustrates the difference in 
agencies' incorporation of public involvement in the three phases of 
the review process, for discretionary and mandatory reviews. 

Figure 2: Agencies' Incorporation of Public Involvement in the Review 
Process, for Discretionary and Mandatory Reviews: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[A] One of the agencies within our review did not have any regulations 
that were subject to mandatory review requirements. 

[End of figure] 

The Level of Documentation in Agencies' Reviews Varied Depending on the 
Impetus and Phase of the Review, but Overall Could Be Improved: 

Agency documentation for selecting regulations to review varied from 
detailed documentation of selection criteria considered to no 
documentation of the selection process. For example, DOL's EBSA 
documented its selection process, including selection criteria used, in 
detail in its Regulatory Review Program. However, agencies did not 
always have written procedures for how they selected regulations for 
discretionary reviews. SBA officials, for example, were not able to 
verify the factors they considered during the selection of some 
regulations that they reviewed because employees who conducted the 
reviews were no longer with the agency and they did not always document 
their review process. The officials indicated that the agency 
considered whether the regulations that they reviewed under Section 610 
requirements were related to or complement each other, but did not 
always document selection factors for discretionary reviews. This lack 
of documentation was particularly important for SBA because the agency 
reported having high staff turnover that affected maintaining 
institutional knowledge about retrospective regulatory review plans. 
For example, officials reported that within the 8(a) Business 
Development program, there is a new Director almost every 4 years who 
sets a new agenda for retrospective regulatory review needs. However, 
because of other pressing factors, these reviews are often not 
conducted. Consequently, we conclude that this lack of documentation 
may result in duplicative agency efforts to identify rules for review 
or not including rules that the agency previously identified as needing 
review. 

Agency documentation of the analyses conducted in reviews ranged from 
no documentation to detailed documentation of their analysis steps in 
agency review reports. While some agencies reported the analysis 
conducted in great detail in review reports, others summarized review 
analysis in a paragraph or provided no documentation of review analysis 
at all. Some agencies did not provide detailed reports because they did 
not conduct detailed analyses. For example, SBA officials reported 
that, for Section 610 reviews, they do not conduct any additional 
analysis of the regulation if the public does not comment on the 
regulation. Our assessment of a sample of agency review reports 
revealed that, even for some reviews that provided a summary of their 
analysis, we could not completely determine what information was used 
and what analysis the agency conducted to form its conclusions. 
Further, agencies in our reviews reported that they less often 
documented the analysis of those reviews conducted on a discretionary 
basis. One SBA official acknowledged that it would be helpful if the 
agency better documented its reviews. 

For each of the agencies we reviewed, we were able to find reports on 
the results of some or all of their completed reviews. Nonetheless, the 
content and detail of agency reporting varied, ranging from detailed 
reporting to only one-sentence summaries of results. Some agencies told 
us that they typically only document and report the results if their 
reviews result in a regulatory change. Further, officials from many 
agencies primarily reported conveying the results of only mandatory 
reviews to the public. Agencies employed a variety of methods to report 
review results to the public, but more often used the Federal Register 
and Unified Agenda. Although agencies in our review often reported the 
results of their mandatory reviews by posting them in the Federal 
Register, agencies like OSHA, CPSC, FCC, and those within DOT also made 
some or all their review reports available on their Web sites. During 
our joint agency exit conference, officials indicated that agencies 
could do more to report their review analysis and results to a wider 
population of the public by using the latest information technology 
tools. Specifically, they said that agencies could: (1) use listserves 
to provide reports to identified interested parties, (2) make review 
analysis and results more accessible on agency Web sites, and (3) share 
results in Web-based forums, among other things.[Footnote 36] 
Nonfederal parties also reported that agencies could improve their 
efforts to report review results to the public and cited similar 
communication techniques. Additionally, nonfederal parties reported 
that agencies could improve communication by conducting more outreach 
to broad networking groups that represent various stakeholders, such as 
the Chamber of Commerce, the National Council of State Legislators, and 
the Environmental Council of States, and tailoring their summary of the 
review results to accommodate various audiences. Figure 3 illustrates 
the differences in agencies' use of documentation in the three phases 
of the review process, for discretionary and mandatory reviews. 

Figure 3: Agencies' Documentation of the Review Process, for 
Discretionary and Mandatory Reviews: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[A] One of the agencies within our review did not have any regulations 
that were subject to mandatory review requirements. 

[End of figure] 

Reviews Resulted in a Variety of Outcomes that Were Considered Useful, 
but Mandatory Reviews Most Often Resulted in No Change: 

Agency reviews of existing regulations resulted in various outcomes-- 
from amending regulations to no change at all--that agencies and 
knowledgeable nonfederal parties reported were useful. Mandatory 
reviews most often resulted in no changes to regulations. Conversely, 
agency officials reported that their discretionary reviews more often 
generated additional action. Both agency officials and nonfederal 
parties generally considered reviews that addressed multiple purposes 
more useful than reviews that focused on a single purpose. 

Agency reviews of existing regulations resulted in various outcomes 
including: changes to regulations, changes or additions to guidance and 
other related documents, decisions to conduct additional studies, and 
validation that existing rules were working as planned. Agencies and 
nonfederal parties that we interviewed reported that each of the 
outcomes could be valuable to the agency and the public. However, for 
the mandatory reviews completed within our time frame, the most common 
result was a decision by the agency that no changes were needed to the 
regulation. There was a general consensus among officials across the 
agencies that the reviews were sometimes useful, even if no subsequent 
actions resulted, because they helped to confirm that existing 
regulations were working as intended. Officials of some agencies 
further noted that, even when mandatory reviews do not result in 
changes, they might have already made modifications to the regulations. 
Our examinations of selected completed reviews confirmed that this is 
sometimes the case. 

Among the various outcomes of retrospective reviews were changes to 
regulations, changes or additions to guidance and other related 
documents, decisions to conduct additional studies, and validation that 
existing rules were working as planned. Agencies and nonfederal parties 
that we interviewed reported that each of the outcomes could be 
valuable to the agency and the public. In our review of agency 
documentation, we confirmed that some reviews resulted in regulatory 
actions that appeared useful. Our review of agency documentation 
confirmed that some reviews can prompt potentially beneficial 
regulatory changes. For example, OSHA's review of its mechanical press 
standard revealed that the standard had not been implemented since its 
promulgation in 1988 because it required a validation that was not 
available to companies. Consequently, OSHA is currently exploring ways 
to revise its regulation to rely upon a technology standard that 
industries can utilize and that will provide for additional 
improvements in safety and productivity.[Footnote 37] Although some 
reviews appeared to result in useful changes, the most common result 
for mandatory reviews was a decision by the agency that no changes were 
needed to the regulation. However, there was a general consensus among 
officials across the agencies that such decisions are still sometimes 
useful because they helped to confirm that existing regulations were 
working as intended. Officials of some agencies further noted that, 
even when mandatory reviews do not result in changes, they might have 
already made modifications to the regulations. Our examinations of 
selected completed reviews confirmed that this is sometimes the case. 

Agency officials reported that their discretionary reviews resulted in 
additional action--such as prompting the agencies to complete 
additional studies or to initiate rulemaking to amend the existing 
rule--more often than mandatory reviews. In particular, officials from 
USDA's AMS and FSIS, FCC, SBA, EPA, DOJ, and DOT reported that Section 
610 reviews rarely resulted in a change to regulations. Although AMS 
has initiated 19 Section 610 reviews since 2001, AMS officials reported 
that, because of their ongoing engagement with the regulated community, 
these reviews did not identify any issues that the agency did not 
previously know, and therefore resulted in no regulatory changes. 
Similarly, none of the Section 610 reviews conducted by SBA and DOL's 
EBSA resulted in changes to regulations, and few changes resulted from 
Section 610 reviews conducted by EPA, FCC, DOJ, and DOT. The one 
apparent outlier in our analysis was FDIC, which conducted many of its 
reviews in response to the financial-sector-specific burden reduction 
requirement in EGRPRA. According to FDIC officials and the agency's 
2005 annual report, reviews conducted in response to this mandate 
resulted in at least four regulatory changes by the agency since 2001 
and over 180 legislative proposals for regulatory relief that FDIC and 
other members of the FFIEC presented to Congress. The legislative 
proposals led to the passage of the Financial Services Regulatory 
Relief Act, which reduced excessive burden in nine areas in the 
financial sector. In addition, our analyses of the December 2006 
Unified Agenda revealed that FDIC attributed four of its nine proposed 
or initiated modifications to existing regulations to statutory 
mandates. 

Most agencies' officials reported that reviews they conduct at their 
own discretion--in response to technology and science changes, industry 
feedback, and petitions--more often resulted in changes to regulations. 
As one of many examples, EBSA officials reported that because the 
reviews initiated and conducted by the agency to date have been 
precipitated by areas for improvement identified by the regulated 
community or the agency, virtually all the reviews have resulted in 
changes to the reviewed rules. They reported that, in general, these 
changes have tended to provide greater flexibility (e.g., the use of 
new technologies to satisfy certain disclosure and recordkeeping 
requirements) or the streamlining and/or simplifying of requirements 
(e.g., reducing the amount of information required to be reported). 
Similarly, DOT officials and other agencies' officials reported that 
reviews that they conduct in response to industry and consumer feedback 
and harmonization efforts also resulted in changes to regulations more 
often than mandated reviews. 

In addition, some agencies also reported that reviews that incorporated 
review factors from both their mandatory requirements and factors 
identified by the agency in response to informal feedback often 
resulted in useful regulatory changes. These agencies' reviews 
incorporated factors identified by the agency as well as ones that were 
requirements in mandatory reviews. For example, DOL's OSHA and EBSA 
selected regulations for review based upon criteria that they 
independently identified and selection criteria identified by Section 
610 requirements. They also incorporated review factors listed in 
Section 610 requirements into a broader set of evaluative factors 
considered during their discretionary reviews, including assessing: (1) 
whether the regulation overlaps, duplicates, or conflicts with other 
federal statutes or rule; and (2) the nature of complaints against the 
regulation. As a result, they reported that these reviews generated 
useful outcomes. Nonfederal parties also indicated that reviews that 
focus on multiple review factors and purposes are more useful than 
reviews that focus only one purpose, such as only burden reduction or 
only enforcement and compliance or only one factor, such as public 
comments. 

Because agencies did not always document discretionary reviews that 
they conducted, it is not possible to measure the actual frequency with 
which they resulted in regulatory change. However, we observed that, 
for cases where agencies reported modifications to regulations, these 
actions were most often attributed to factors that agencies addressed 
at their own discretion, such as technology changes, harmonization 
efforts, informal public feedback, and petitions. For example, although 
EPA officials reported that they have many mandatory regulatory review 
requirements, our review of proposed or completed modifications to 
existing regulations reported in the December 2006 Unified Agenda 
showed that 63 of the 64 modifications reported were attributed to 
reasons associated with agencies' own discretion. As illustrated in 
figure 4, other agencies within our review had similar results. 

Figure 4: Modifications to Rules Listed in the December 2006 Unified 
Agenda by Type of Review for Selected Agencies: 

[See PDF for image] 

Source: GAO analysis of December 2006 Unified Agenda entries. 

[End of figure] 

Although agencies reported, and our analysis of the Unified Agenda 
indicated, that agencies more often modify existing regulations for 
reasons attributed to their own discretion, it is important to note 
that mandatory reviews may serve other valuable purposes for Congress. 
Such reviews may provide Congress with a means for ensuring that 
agencies conduct reviews of regulations in policy areas that are 
affected by rapidly changing science and technology and that agencies 
practice due diligence in reviewing and addressing outdated, 
duplicative, or inconsistent regulations. For example, Congress 
required FCC to conduct reviews of its regulations that apply to the 
operation or activity of telecommunication service providers to 
"determine whether any such regulation is no longer necessary in the 
public interest as the result of meaningful economic competition 
between providers of such service."[Footnote 38] 

Agencies' officials reported that reviews often had useful outcomes 
other than changes to regulations, such as changes or additions to 
guidance and other related documents, decisions to conduct additional 
studies, and validation that existing rules were working as planned. 
For example, OSHA officials reported that, outside of regulatory 
changes, their reviews have resulted in recommended changes to guidance 
and outreach materials and/or the development of new materials or 
validation of the effectiveness of existing rules. Our review of OMB's 
regulatory reform nominations process confirmed that at least four of 
OSHA's reviews conducted in response to OMB's manufacturing reform 
initiative resulted in changes to or implementation of final guidance 
or the development of a regulatory report. We observed similar results 
from OMB's regulatory reform process for EPA. Similarly, DOT officials 
reported that their reviews also often led to changes in guidance or in 
further studies, and our examination of review results reported by DOT 
confirmed that this was often the case. Moreover, all of the agencies 
within our review reported that reviews have resulted in validating 
that specific regulations produced the intended results. 

Agencies and Nonfederal Parties Identified Barriers and Facilitators to 
the Conduct and Usefulness of Retrospective Reviews: 

Agencies' officials reported that barriers to their ability to conduct 
and use reviews included: (1) difficulty in devoting the time and staff 
resources required for retrospective review requirements, (2) 
limitations on their ability to obtain the information and data needed 
to conduct reviews, and (3) constraints in their ability to modify some 
regulations without additional legislative action, among other 
important factors. Both agencies and nonfederal parties identified the 
lack of public participation in the review process as a barrier to the 
usefulness of reviews. The nonfederal parties also identified the lack 
of transparency in agency review processes as a barrier to the 
usefulness of reviews. Agency officials and nonfederal parties also 
suggested a number of practices that could facilitate conducting and 
improving the usefulness of regulatory reviews, including: (1) 
development of a prioritization process to facilitate agencies' ability 
to address time and resource barriers and allow them to target their 
efforts at reviews of regulations that are more likely to need 
modifications, (2) pre-planning for regulatory reviews to aid agencies 
in identifying the data and analysis methodology that they will need to 
conduct effective reviews, and (3) utilizing independent parties to 
conduct the reviews to enhance the review's credibility and 
effectiveness, among other things. While there was general consensus 
among federal and nonfederal parties on the major facilitators and 
barriers, there were a few clear differences of opinions between them 
regarding public participation and the extent to which reviews should 
be conducted by independent parties. Because only a few agencies track 
the costs associated with conducting their reviews, one cannot identify 
the type and approach to retrospective review that may be most cost 
effective. However, agency officials told us that the reviews have 
resulted in cost savings to their agencies and to regulated parties, 
for example by saving both the agency and the public the costs of 
repeatedly dealing with petitions for change or waivers in response to 
difficulties implementing particular regulatory provisions. 

Time, Resources, and Information Were Among the Critical Barriers to 
Conducting Reviews: 

All of the agencies in our review reported that the lack of time and 
resources are the most critical barriers to their ability to conduct 
reviews. Specifically, they said that it is difficult to devote the 
time and staff resources required to fulfill various retrospective 
review requirements while carrying out other mission-critical 
activities. Agencies' officials reported that, consequently, they had 
to limit their retrospective review activities during times when they 
were required to respond to other legislative priorities. For example, 
officials from MSHA reported that they conducted fewer reviews in 2006 
because they were heavily engaged in trying to implement the Mine 
Improvement and New Emergency Response Act of 2006 (MINER Act), which 
Congress passed in response to mining accidents that occurred in 
2006.[Footnote 39] Prior to these events, MSHA was engaged in 
soliciting a contractor to assist the agency in prioritizing its 
retrospective review efforts. The officials reported that, because of 
the need to develop regulations pursuant to the act, they stopped the 
process of looking for a contractor, and conducted fewer reviews. 
Officials from various agencies reported that retrospective reviews are 
the first activities cut when agencies have to reprioritize based upon 
budget shortfalls. A DOT official reported that, despite having high- 
level management support for retrospective review activities, the 
department has still experienced funding limitations that have affected 
their ability to conduct retrospective review activities. Our 
examination of agency documents confirmed that several agencies 
indicated that they did not complete all of the reviews that they 
planned and scheduled for some years within the scope of our review 
because sufficient resources were not available. In one example, we 
found that FAA delayed conducting any planned reviews for an extended 
period because, as reported in the Unified Agenda, they did not have 
the resources to conduct them. Many of the agencies in our review did 
not track the costs (usually identified in terms of full-time 
equivalent (FTE) staff resources) associated with their reviews; 
therefore, they could not quantify the costs of conducting reviews. 

Information and Data Limitations Present Challenges to the Conduct of 
Retrospective Reviews: 

Most agencies' officials reported that they lack the information and 
data needed to conduct reviews. Officials reported that a major data 
barrier to conducting effective reviews is the lack of baseline data 
for assessing regulations that they promulgated many years ago. Because 
of this lack of data, agencies are unable to accurately measure the 
progress or true effect of those regulations. Similar data collection 
issues were also identified by agencies in the Eisner and Kaleta study 
published in 1996, which concluded that, in order to improve reviews 
for the future, agencies should collect data to establish a baseline 
for measuring whether a regulation is achieving its goal, and identify 
sources for obtaining data on ongoing performance.[Footnote 40] 

Agencies and nonfederal parties also considered PRA requirements to be 
a potential limiting factor in agencies' ability to collect sufficient 
data to assess their regulations. For example, EPA officials reported 
that obtaining data was one of the biggest challenges the Office of 
Water faced in conducting its reviews of the effluent guideline and 
pretreatment standard under the Clean Water Act, and that as a result 
the Office of Water was hindered or unable to perform some analyses. 
According to the officials, while EPA has the authority to collect such 
data, the PRA requirements and associated information collection review 
approval process take more time to complete than the Office of Water's 
mandated schedule for annual reviews of the effluent guideline and 
pretreatment standard allows. While one nonfederal party did not agree 
that PRA restrictions posed a significant barrier to conducting 
reviews, agencies and nonfederal parties generally agreed that the act 
was an important consideration in agency data collection. However, 
while agencies identified the potential limitations of PRA, it is 
important to recognize that PRA established standards and an approval 
process to ensure that agencies' information collections minimize the 
federal paperwork burden on the public, among other purposes.[Footnote 
41] 

In general, data collection appeared to be an important factor that 
either hindered or facilitated reviews. Some of the agencies in our 
review that promulgate safety regulations, such as CPSC, NHTSA, and 
those within DOJ, reported that having sufficient access to established 
sources of safety data, such as death certificates or hospital 
databases on deaths and injuries related to products, greatly 
facilitated their ability to conduct retrospective reviews of their 
regulations. Finally, agencies also reported facing limits on their 
ability to obtain data on their regulations because of the length of 
time it takes to see the impact of some regulations and the scarcity of 
data related to areas that they regulate. Nonfederal parties also cited 
this data limitation as a challenge to agency reviews. 

Barriers to the Usefulness of Reviews Included Overlapping and Generic 
Schedules and Review Factors, Scope, Statutory Limitations, Limited 
Public Participation, and Transparency: 

Overlapping Schedules and Review Factors: 

To make efficient use of their time and resources, various agency 
officials said that they consider all relevant factors, including 
effectiveness and burden reduction, whenever they review an existing 
regulation. Therefore, when reviews that have predetermined or generic 
schedules and review factors (such as 10-year Section 610 reviews) 
arise, the agency might have already reviewed and potentially modified 
the regulation one or more times, based upon the same factors outlined 
in Section 610. The officials reported that, although the subsequent 
predetermined reviews are often duplicative and less productive, they 
nevertheless expend the time and resources needed to conduct the 
reviews in order to comply with statutory requirements. However, they 
reported that these reviews were generally less useful than reviews 
that were prompted because of informal industry and public feedback, 
petitions, changes in the market or technology, and other reasons. 
Furthermore, agencies expressed concerns about whether predetermined 
schedules may conflict with other priorities. DOT acknowledged this 
issue even as it was establishing an agency policy to require 
retrospective reviews. In response to a public suggestion that DOT 
conduct reviews based upon a regular predetermined schedule, the agency 
cautioned that arbitrary schedules might mean delaying other, more 
important regulatory activities. 

As examples of predetermined reviews that may be duplicative or 
unproductive, officials from agencies within DOT, USDA, and DOL 
reported that the regulations that most often apply to their industries 
may need review sooner than the 10-year mark prescribed by Section 610. 
To be responsive to the regulated community, the agencies regularly 
review their regulations in response to public feedback, industry and 
technology changes, and petitions, among other things, and make 
necessary changes before a Section 610 review would be required. Our 
assessment of reviews listed in the Unified Agenda confirmed that 
agencies often noted that they had not made changes because of their 
Section 610 reviews, but had previously made changes to these 
regulations because of factors that previously emerged. For example, 
USDA's AMS reported completing 11 mandated Section 610 reviews since 
2001, which resulted in no regulatory changes. For 9 of these reviews, 
the related published Section 610 reports stated that AMS made no 
changes to the regulations because they were modified "numerous times" 
in advance of the 10-year Section 610 review to respond to changes in 
economic and other emerging conditions affecting the industry. 

Similar to agency views on timing, views by an OMB official and some 
nonfederal parties indicated that the period immediately after an 
agency promulgates a rule may be a critical time for agencies to review 
certain types of regulations, in part because once the regulated 
community invests the resources to comply with the regulations and 
integrates them into their operations, they are less likely to support 
subsequent changes to the regulation. In addition, the immediate 
effects of certain types of regulations, such as economic incentive 
regulations, may be more apparent and changes, if needed, can be 
brought about sooner.[Footnote 42] Nonfederal parties reported that 
this may be especially important during the time that regulated 
entities are facing challenges with the implementation of a regulation. 
Some of these commenters noted that such immediate reviews might be 
especially appropriate for rules that have a high profile, are 
controversial, or involve a higher degree of uncertainty than usual. 

Two agencies within our review that had predetermined deadlines that 
are set only a few years apart also reported that these schedules 
affected their ability to produce more useful reviews. The officials 
reported that they do not have enough time to effectively complete the 
reviews prior to beginning another review. For example, EPA and FCC 
both stated that agency-specific review requirements to conduct reviews 
of their regulations every few years make it difficult for the agencies 
because either the agencies do not have enough time to effectively 
gather data for the reviews or do not have enough time to observe new 
effects of the regulation between reviews. As a result, the agencies 
may be doing a less comprehensive job in conducting the reviews and 
have more difficulty in meeting their review deadlines. 

For requirements that specify a predetermined schedule for conducting 
reviews, agencies also identified, as a potential barrier, the lack of 
clarity on when to "start the clock" for regulations that have been 
amended over time. For example, as previously mentioned, in order to 
satisfy Section 610 requirements, DOT initiated an extensive process 
for reviewing its sections of the CFR every year. The agency's 
officials reported that they adopted this extensive approach because 
they were unable to determine whether to review a regulation 10 years 
after its promulgation or 10 years after its last modification. Other 
agencies included in our review did not take this approach to meeting 
Section 610 requirements. Similarly, in our 1999 report on RFA, we 
reported that agencies' varying interpretations of Section 610 
requirements affected when they conducted reviews.[Footnote 43] 

While agencies' officials reported that predetermined schedules can 
sometimes be ineffective, it is important to note that such schedules 
can also help ensure that reviews occur. Specifically, some parties 
have noted that a benefit of prespecifying the timing of reviews is 
that this provides Congress with a way to force agencies to 
periodically reexamine certain regulations. 

In general, as illustrated in table 3, our review of the timing of 
reviews and the evaluative factors that agencies are supposed to assess 
in those reviews revealed that there is considerable overlap in the 
various mandatory and discretionary review requirements. 

Table 3: Examples of Overlapping Timing and Review Factors in 
Retrospective Review Requirements: 

Impetus for review: Governmentwide RFA, Section 610 review 
requirements; 
Regulations covered by the review: All regulations that are at least 10 
years old and have a significant economic impact on a substantial 
number of small entities; 
Frequency: Every 10 years; 
Evaluation factors to be considered: Identify burden reduction 
opportunities by assessing the: 
* continued need for the rule; 
* nature of complaints or comments received concerning the rule from 
the public; 
* complexity of the rule; 
* the extent to which the rule overlaps, duplicates or conflicts with 
other federal rules, and, to the extent feasible, with state and local 
governmental rules;[A]; 
* length of time since the rule has been evaluated or the degree to 
which technology, economic conditions, or other factors have changed in 
the area affected by the rule. 

Impetus for review: EGRPRA review for Financial and Banking Regulatory 
agencies; 
Regulations covered by the review: All regulations prescribed by 
financial regulatory agencies within FFIEC or by any such appropriate 
federal banking agency; 
Frequency: Every 10 years; 
Evaluation factors to be considered: Identify and assess public 
comments on areas of regulations that are: 
* outdated,; 
* unnecessary,; 
* unduly burdensome. 

Impetus for review: EBSA Regulatory Review Program; 
Regulations covered by the review: Substantively review any EBSA 
regulations that are selected as a priority, based upon 16 factors, 
including regulations that are subject to the RFA, Section 610 
requirements, and E.O. 12866; 
Frequency: Annually; 
Evaluation factors to be considered: Assess rules to determine whether: 
* there is a continued need for the regulation; 
* the regulation has been the subject of complaints or comments from 
the public and the nature of those complaints; 
* the regulation overlaps, duplicates, or conflicts with other federal 
statutes or rules or with nonpreempted state or local statutes or 
rules; 
* the regulation is overly complex and could be simplified without 
impairing its effectiveness; 
* the regulation may be based upon outdated or superseded employment, 
industrial, or economic practices or assumptions and whether 
participants and/or beneficiaries of employee benefit plans may be 
exposed to harm as a result; 
* the regulation may impose significant economic costs on regulated 
entities and whether the benefit(s) or purpose(s) of the regulation 
could be achieved as effectively through an alternative regulatory 
approach that would impose less economic burden on regulated 
industries. 

Impetus for review: FCC Communications Act Biennial Review; 
Regulations covered by the review: All rules applicable to providers of 
telecommunications service; 
Frequency: Every 2 years; 
Evaluation factors to be considered: Identify rules that are: 
* no longer necessary in the public interest as a result of meaningful 
competition. 

Impetus for review: FCC Communication Act, Quadrennial Review; 
Regulations covered by the review: Almost all broadcast ownership 
rules; 
Frequency: Every 4 years; 
Evaluation factors to be considered: Determine whether any rules: 
* are necessary in the public interest as the result of competition, 
or; 
* are no longer in the public interest. 

Impetus for review: EPA 's Drinking Water Program /Review of National 
Drinking Water Regulation; 
Regulations covered by the review: All drinking water regulations that 
were promulgated under Section 1412(b)(9); 
Frequency: Every 6 years; 
Evaluation factors to be considered: To assess: whether regulations 
should be modified based upon contaminate occurrences in finished 
drinking water. 

Impetus for review: EPA Section 405(d)(2)(C) Clean Water Act Review; 
Regulations covered by the review: All regulations on the Use and 
Disposal of Sewage Sludge; 
Frequency: Every 2 years; 
Evaluation factors to be considered: To identify: 
* whether the regulation should be updated to include additional toxic 
pollutants. 

Source: GAO analysis of agency data and regulatory statutes. 

[A] Similar to this criterion, PRA requires agencies to consider 
whether information that they propose to collect from public--including 
information requests per regulatory requirements--overlaps, duplicates, 
or conflicts with other federal requests for information. All proposed 
collections from 10 or more persons, other than agencies, 
instrumentalities, or employees of the United States must be approved 
by OMB, and those approvals expire after 3 years, so must be reviewed 
if the agency wishes to continue the collection. 

[End of table] 

Scope of Reviews: 

Various agencies identified scoping issues as a barrier to the 
usefulness of reviews. Agencies' officials reported significant delays 
in completing reviews and making timely modifications, as well as 
obtaining meaningful input in reviews that involved multiple 
regulations as the unit of analysis. Some agencies, such as DOL's MSHA, 
reported experiencing delays up to 16 years in completing a review 
because they scoped their review too broadly. Specifically, MSHA 
officials reported that, during a comprehensive review of their 
ventilation standards, the scope of the review increased due to input 
from other departmental agencies. Because of this input and the 
complexity of the rule itself, it took 16 years to complete the 
modifications, resulting in a major rewrite of the ventilation 
standards. In our assessment of this review, the resulting information 
was not as timely as it otherwise could have been, and therefore may 
have been less useful.[Footnote 44] Similarly, officials from other 
agencies reported that scoping reviews too broadly also affected their 
ability to conduct expedient reviews. Agencies' officials suggested 
that having a narrow and focused unit of analysis, such as a specific 
standard or regulation, is a more effective approach to conducting 
reviews. Specifically, officials from DOT and FDIC reported that, when 
they conducted narrowly defined reviews, the public provided more 
meaningful input on their regulations. Furthermore, one nonfederal 
party emphasized that, when agencies choose to analyze a broad unit of 
analysis, such as an act, it is difficult for the public to discern 
which regulations are doing well and which are not. The positive 
effects of one regulation under the legislation can overshadow the 
negative effects of other regulations. Therefore, the performance 
assessment of the relevant regulations is less transparent and, 
consequently, less useful. 

Statutory Provisions: 

Agencies' officials reported that statutory requirements are a major 
barrier to modifying or eliminating regulations in response to 
retrospective regulatory reviews because some regulations are aligned 
so closely with specific statutory provisions. Therefore, the agencies 
may be constrained in the extent to which they can modify such 
regulations without legislative action. For example, officials from 
MSHA, FDIC, and SBA reported that many of their regulations mirror 
their underlying statutes and cannot be modified without statutory 
changes. During its retrospective reviews to reduce burden, FDIC along 
with other banking agencies within the FFIEC, identified 180 financial 
regulations that would require legislative action to revise. Similarly, 
in our 1999 report on regulatory burden, we found that agencies often 
had no discretion, because of statutory provisions, when they imposed 
requirements that businesses reported as most burdensome.[Footnote 45] 
One approach taken by FDIC to address this issue was to identify 
regulations that required legislative action in their review process 
and to coordinate with Congress to address these potential regulatory 
changes. Because of this approach, Congress is actively involved in 
FDIC's regulatory burden relief efforts and has passed changes in 
legislation to provide various forms of burden relief to the financial 
sector. 

Limited Public Participation: 

Agencies and nonfederal parties identified the lack of public 
participation in the review process as a barrier to the usefulness of 
reviews. Agencies stated that despite extensive outreach efforts to 
solicit public input, they receive very little participation from the 
public in the review process, which hinders the quality of the reviews. 
Almost all of the agencies in our review reported actively soliciting 
public input into their formal and informal review processes. They 
reported using public forums, and industry meetings, among other things 
for soliciting input into their discretionary reviews, and primarily 
using the Federal Register and Unified Agenda for soliciting public 
input for their mandatory reviews. For example, USDA officials reported 
conducting referenda of growers to establish or amend AMS marketing 
orders, and CPSC officials reported regularly meeting with standard- 
setting consensus bodies, consumer groups, and regulated entities to 
obtain feedback on their regulations. Other agencies reported holding 
regular conferences, a forum, or other public meetings. However, most 
agencies reported primarily using the Unified Agenda and Federal 
Register to solicit public comments on mandatory reviews, such as 
Section 610 reviews. Despite these efforts, agency officials reported 
receiving very little public input on their mandatory reviews. 

Nonfederal parties we interviewed were also concerned about the lack of 
public participation in the retrospective review process and its impact 
on the quality of agency data used in reviews. However, these 
nonfederal parties questioned the adequacy and effectiveness of 
agencies' outreach efforts. Specifically, 7 of the 11 nonfederal 
parties cautioned that the Federal Register and Unified Agenda are not 
sufficiently effective tools for informing the public about agency 
retrospective review activities. In addition, most of the nonfederal 
parties we interviewed were unaware of the extent to which agencies 
conducted reviews under their own discretion, and most of those parties 
reported that they were not aware of the outreach efforts agencies are 
making to obtain input for these reviews. Limited public participation 
in some review activities was cited by both agencies and nonfederal 
parties as a barrier to producing quality reviews, in part because 
agencies need the public to provide information on the regulations' 
effects. Both agency officials and nonfederal parties identified 
methods for improving communication, including using agency Web sites, 
e-mail listserves, or other Web-based technologies (such as Web 
forums), among other things. 

Shortcomings in Transparency: 

Nonfederal parties identified the lack of transparency in agency review 
processes, results, and related follow-up activities as a barrier to 
the usefulness of reviews to the public. Nonfederal parties were rarely 
aware of the retrospective review activities reported to us by the 
agencies in our review. Similarly, in our review of the Federal 
Register and Unified Agenda, we were not always able to track 
retrospective review activities, identify the outcome of the review, or 
link review results to subsequent follow-up activities, including 
initiation of rulemaking to modify the rule. As stated earlier, some 
mandatory reviews do not require public reporting and many agencies did 
not consistently report the results of their discretionary reviews, 
especially if the reviews resulted in no changes to regulations. Some 
nonfederal parties told us that lack of transparency was the primary 
reason for the lack of public participation in agencies' review 
processes. 

Agencies and Nonfederal Parties Identified Lessons Learned and 
Practices for Improving Retrospective Reviews: 

Agencies and nonfederal parties identified pre-planning for regulatory 
reviews as a practice that aids agencies in identifying the data and 
analysis methodology that they need to conduct effective outcome-based 
performance reviews. Some agencies within our review planned how they 
would collect performance data on their regulations before or during 
the promulgation of the relevant regulations or prior to the review. 
They cited this technique as a method for reducing data collection 
barriers. For example, DOT's NHTSA was an agency that OMB officials and 
nonfederal parties identified as appearing to conduct effective 
retrospective reviews of its regulations. NHTSA officials reported to 
us that, to conduct effective reviews, they plan for how they will 
review their regulations even before they issue them. Prior research on 
regulatory reviews also cited the need for agencies to set a baseline 
for their data analysis, in order to conduct effective reviews. In 
addition, we have long advocated that agencies take an active approach 
to measuring the performance of agency activities. Furthermore, we 
observed that pre-planning for data collection could address some 
challenges that agencies reported facing with PRA data collection 
requirements, such as the length of time required to obtain approval. 

Agencies reported that prioritizing which regulations to review 
facilitated the conduct of and improved usefulness of their reviews. 
Agencies that developed review programs with detailed processes for 
prioritizing which regulations to review reported that this 
prioritization facilitated their ability to address time and resource 
barriers to conducting reviews and allowed them to target their efforts 
at more useful reviews of regulations that were likely to need 
modifications. As previously mentioned, DOL's EBSA and CPSC developed 
detailed prioritization processes that allowed officials to identify 
which regulations were most in need of review and to plan for 
conducting those reviews. Furthermore, this process allowed CPSC to 
prospectively budget for its reviews and to identify the number of 
substantive reviews per year that the agency could effectively conduct, 
while meeting its other agency priorities. Officials from both agencies 
reported that their prioritization processes allowed them to focus on 
the most useful retrospective review activities, which identified 
important regulatory changes. Nonfederal parties that we interviewed 
also asserted that it is not necessary or even desirable for agencies 
to expend their time and resources reviewing all of their regulations. 
Instead, they reported that it would be more efficient and valuable to 
both agencies and the public for agencies to conduct substantive 
reviews of a small number of regulations that agencies and the public 
identify as needing attention. Nonfederal parties and agency officials 
suggested that factors that agencies should consider when prioritizing 
their review activities could include economic impact, risk, public 
feedback, and length of time since the last review of the regulation, 
among other things. 

Nonfederal regulatory parties believed that reviews would be more 
credible and effective if the parties that conduct them were 
independent. For example, two different parties who we interviewed said 
that EPA's first report in response to Section 812 under the Clean Air 
Act could have been improved by involving independent analysts. 
However, they recognized that it is important to include input from 
those who were involved in the day-to-day implementation of the 
regulation and were responsible for producing initial benefit-cost 
estimates for the regulations. Almost all of the nonfederal parties 
that we interviewed expressed concern that agency officials who 
promulgated and implemented regulations may be the same officials who 
are responsible for evaluating the performance of these regulations. 
Although the nonfederal parties acknowledged that it is important for 
officials with critical knowledge about the program to be involved with 
providing input into the review, they were concerned that officials 
placed in this position may not be as objective as others may be. 
Nonfederal parties also expressed concerns about agencies' capacity to 
conduct certain types of analyses for their reviews, such as benefit- 
cost assessments. The nonfederal parties suggested that agencies could 
consider having an independent body like another agency, Inspector 
General, or a centralized office within the agency conduct the reviews. 
During our review, agencies' officials reported that they sometimes 
contract out their reviews if they do not have the expertise needed to 
conduct the analyses. However, during a discussion of this issue at our 
joint agency exit meeting, agency officials pointed out the difficulty 
in finding a knowledgeable independent review body to conduct 
retrospective reviews, and they noted that even contracted reviewers 
may be considered less independent, because they are paid by the agency 
to conduct the study. 

Agencies and nonfederal regulatory parties agreed that high-level 
management support in the review process is important to the successful 
implementation of not only individual reviews but also to sustaining 
the agency's commitment to a review program and following up on review 
results. As an example, officials from FDIC credited the 
accomplishments of their review program largely to the support of high- 
level managers who headed the FFIEC effort to reduce regulatory burden 
on financial institutions. Officials reported that the leadership of 
the Director of the Office of Thrift Supervision, who chaired the FFIEC 
effort, helped to catapult support for reviews at all of the FFIEC 
agencies, including FDIC, and helped to free up resources to conduct 
reviews at these agencies. Almost all of the selected agencies reported 
involving some high-level management attention in their reviews, but 
where and how they used this involvement varied. For example, while 
almost all of the agencies reported involving high-level management 
attention in decision-making processes that resulted from reviews, CPSC 
and EBSA's review programs also involved high-level managers early in 
their processes, in order to determine which regulations to review. 
Overall, agencies and nonfederal parties indicated that having high- 
level management attention is important to obtaining and sustaining the 
resources needed to conduct reviews and the credibility of agency 
reviews. 

According to agency officials from DOT, DOL, SBA, and FDIC, they 
learned that grouping related regulations together when conducting 
reviews is a technique that more often generated meaningful comments 
and suggestions from the public. For example, officials from FDIC 
stated that categorizing regulations for review and soliciting input 
over an extended time period proved to be a more effective way of 
receiving public input. They reported that placing regulations into 
smaller groups and soliciting feedback on these categories separately 
over a 3-year period helped the members of the FFIEC to avoid 
overwhelming the public with the regulatory reviews, and allowed the 
agencies to receive more thoughtful participation and input. SBA 
officials reported reviewing related regulations together because a 
change to one rule can have an impact on the related rules. Similarly, 
a DOT official reported that grouping similar regulations together to 
solicit public input was an effective technique for FAA because the 
agency regulates a broad policy area. FAA received 1800 suggestions for 
regulatory changes based upon one such review. However, the official 
cautioned that while grouping regulations is an effective technique to 
obtaining useful public input, defining the categories too broadly can 
lead to an effort that is too intensive. In addition, the practice may 
be less convenient and practical for agencies that write very specific 
standards, such as NHTSA. For these agencies it may be more effective 
to pick related characteristics of rules in order to group regulations 
to review. 

Nonfederal parties suggested that agencies need to be more aware of the 
different audiences that might be interested in their reviews, and 
target the level of detail and type of product used to report the 
results to meet the needs of these various audiences. For example, a 
product that focuses on specific details of implementing a regulation 
may be less useful to those interested in the policy effects of a 
regulation, and vice versa. Further, both agency officials and 
nonfederal parties identified methods for improving communication, 
including better use of information technology tools, such as agency 
Web sites, electronic docket systems, e-mail listserves, Web-based 
forums, or other Web-based technologies. 

Some Agencies Believe Retrospective Reviews Can Result in Cost Savings: 

Agencies have not estimated all of the costs and benefits associated 
with conducting retrospective reviews, but they believe that 
retrospective reviews have resulted in cost savings to their agencies. 
For example, MSHA officials reported that their retrospective 
regulatory reviews related to petitions for modification produce 
savings for the agency because the reviews prompt the agency to review 
and modify regulations that are heavily petitioned, which reduces costs 
associated with reviewing similar petitions.[Footnote 46] They reported 
that these reviews also save the mining industry from the costs 
associated with repeatedly filing petitions. In addition to petition- 
related cost savings, agencies could save costs by reviewing and 
eliminating regulations that are no longer useful. Therefore, agencies 
could reduce costs associated with implementing and enforcing outdated 
or unproductive regulations. 

We found that only a few agencies track the costs associated with 
conducting their reviews, so we were unable to identify which methods 
are most cost effective. Some agency officials, such as those in MSHA, 
reported that tracking direct costs associated with reviews is 
difficult because reviews are conducted as part of the normal operation 
of the agencies and in concert with other actions to fulfill the 
agencies' missions. However, some agencies like CPSC establish budgets 
for their reviews, and track the associated costs. As a result, CPSC 
determined that conducting about four regulatory reviews per year was a 
reasonable effort for the associated expense to the agency. OSHA also 
tracks the costs associated with its reviews. The agency's officials 
told us that each of its reviews typically requires 2/3 of a program 
analyst FTE in the Office of Evaluations and Audit Analysis, about 1/5 
of an attorney FTE in the Office of the Solicitor, 1/2 FTE for the 
involvement of staff from other directorates, and approximately $75,000 
to $100,000 of contractor support per review. 

Although agencies did not always track the cost of their reviews, 
officials reported that they know some reviews are not cost effective. 
For example, a USDA official reported that, by nature, some regulations 
are set up by the agency to be reviewed regularly. Therefore, 
externally imposed reviews only duplicate this effort. An example of 
such reviews would be those conducted for regulations that are 
consistently reviewed by industry committees that are appointed by the 
Secretary of an agency. AMS officials reported that industry committees 
appointed by the Secretary of Agriculture oversee many of the agency's 
regulations and, as one of their main functions, regularly review AMS 
regulations to identify needed changes. Therefore, regulations under 
the purview of these committees are already constantly being reviewed 
and updated, and thus may benefit less from a Section 610 review than 
other regulations. 

Conclusions: 

Our review revealed that agencies are conducting more reviews, and a 
greater variety of reviews, than is readily apparent, especially to the 
public. To facilitate their reviews, agencies, to greater and lesser 
extents, have been developing written procedures, processes, and 
standards to guide how they select which rules to review, conduct 
analyses of those rules, and report the results. Given the multiple 
purposes and uses of reviews, we recognize that there is no "one size 
fits all" approach. However, there are lessons to be learned from 
ongoing regulatory reviews that could benefit both the agencies in our 
scope and others that conduct retrospective regulatory reviews. Because 
agencies are attempting to find ways to further develop and improve 
their retrospective review processes (for example, establishing units 
that focus on retrospective reviews and seeking assistance with 
establishing prioritization systems), identifying ways to share 
promising practices could collectively improve agency review 
activities. Feedback from agency officials and nonfederal parties, as 
well as our own analysis, indicate that there are procedures and 
practices that may be particularly helpful for improving the 
effectiveness and transparency of retrospective review processes. For 
example, agencies can be better prepared to undertake reviews if they 
have identified what data will be needed to assess the effectiveness of 
a rule before they start a review and, indeed, before they promulgate 
the rule. If agencies fail to plan for how they will measure the 
performance of their regulations, and what data they will need to do 
so, they may continue to be limited in their ability to assess the 
effects of their regulations. 

Given increasing budgetary constraints, both agency officials and 
nonfederal parties emphasized the need to better prioritize agency 
review activities, when possible, to more effectively use their limited 
resources. Agency officials and nonfederal parties recognize that time 
and resources are too limited to allow for a regular, systematic review 
of all of their regulations, and that devoting excessive time and 
scarce resources to a formal review of all of their regulations could 
result in insufficient attention to other regulatory needs or statutory 
mandates. As we have observed, some agencies are already using such 
prioritization processes. Without a detailed prioritization system, 
agencies may not be able to effectively target their reviews so that 
they devote resources to conducting substantive and useful reviews of 
the regulations that need the most attention. 

Agencies and nonfederal parties also reported that reviews are more 
credible and useful to all parties if agencies have assessed multiple 
review factors in their analyses of the regulations, rather than 
relying on a single factor, such as public comments. The failure of 
agencies to do this could result in reviews that miss assessing crucial 
information that could provide context to the results of the analysis, 
such as weighing the benefits against the burdens of the regulation. 
Further, our assessment of the strengths and limitations of agency 
reviews revealed that agencies could improve their efforts to employ a 
standards-based approach to conducting discretionary reviews. Agencies 
are inconsistently applying a standards-based approach to conducting 
discretionary reviews. Applying a standards-based approach could 
enhance the transparency and consistency of reviews. 

Agencies' reporting of reviews appears largely ineffective. None of the 
nonfederal parties we contacted were aware of the extent of agency 
retrospective review activities. This lack of awareness might be 
attributable to two reasons. First, agencies typically did not report 
results for discretionary reviews, which account for most of agencies' 
review activities. Therefore, the public cannot be expected to know 
about these reviews. Second, when agencies do report on their 
activities, the mode and content of these communications may not be 
effective. For example, although we found that some agencies used 
multiple modes of communication, for the most part agencies reported 
that they rely heavily on the Federal Register. However, nonfederal 
parties indicated that reliance on the Federal Register is not 
sufficient. Further, the content that agencies do publish does not 
always provide adequate information about the analysis and results of 
the reviews. Our own assessment showed that it was sometimes difficult 
to determine the outcomes of the reviews or the bases for the agencies' 
conclusions. Some agencies have employed multiple communication modes 
and provided detailed content in their reports, but still report 
disappointing levels of public participation. Therefore, it is clear 
that agencies need to continue to explore methods to more effectively 
communicate and document information about their reviews and the 
underlying analyses. According to agency officials and nonfederal 
parties, such methods could include using agency Web sites, e-mail 
listserves, or other Web-based technologies (such as Web forums). When 
agencies do not effectively communicate the analysis and results of 
their reviews, they miss the opportunity to obtain meaningful comments 
that could affect the outcome of their reviews. Further, without 
showing the underlying analysis of reviews, the agencies' conclusions 
may lack credibility. 

Agencies and nonfederal parties also emphasized the importance of 
having high-level support for sustaining agency retrospective review 
activities, and increasing their credibility with the public. Without 
such attention, agencies will face difficulties in making retrospective 
review a priority that receives the resources necessary for conducting 
successful reviews. Agencies provided specific examples that 
illustrated how high-level management support helped to ensure that 
they followed through on the results of regulatory reviews. Although 
agency officials cautioned that even high-level management support 
might not be sufficient to overcome all budgetary constraints, having 
such support may ensure that some retrospective review activity will be 
sustained. 

One of the most striking findings during our review was the disparity 
in the perceived usefulness of mandatory versus discretionary 
regulatory reviews. The agencies characterized the results of their 
discretionary reviews as more productive and more likely to generate 
further action. A primary reason for this appears to be that 
discretionary reviews that address changes in technology, advances in 
science, informal agency feedback, harmonization efforts, and 
petitions, among other things, may be more closely attuned to 
addressing issues as they emerge. While agencies' officials reported 
that their discretionary reviews might be more useful than the 
mandatory reviews, we can not definitively conclude which reviews are 
most valuable. We did not assess the content and quality of 
discretionary reviews, and could not have done so because they often 
were not documented. Although the officials reported that the bulk of 
their review activity is associated with discretionary reviews, they 
could not provide evidence to show definitively that this was so or 
that discretionary reviews more often generated useful outcomes. 
Further, one cannot dismiss the value that Congress anticipated when 
establishing the mandatory requirements for agencies to conduct reviews 
for particular purposes and on particular schedules. 

The predetermined time frames of mandatory reviews can both help and 
hinder. On one hand, predetermined schedules are one means by which 
Congress can force agencies to periodically reexamine certain 
regulations. However, the timing for some mandatory reviews may either 
be too short or overlap with other review requirements, making it more 
difficult for agencies to produce meaningful analysis from their 
reviews. Conversely, from the cursory information that agencies 
reported for some mandatory reviews that have review periods as long as 
10 years, it appears that agencies may devote limited time and 
resources to conducting these reviews, perhaps partly because the 
required timelines do not recognize ongoing changes to regulations. 

Further, the criteria used in mandatory and discretionary reviews may 
be duplicative. In general, our review of the timing of reviews and the 
evaluative factors that agencies are supposed to assess in those 
reviews revealed that there is considerable overlap in the various 
mandatory and discretionary review requirements. To make efficient use 
of their time and resources, agency officials said that they consider 
all relevant factors, including effectiveness and burden reduction, 
whenever they review an existing regulation. Therefore, when there are 
duplicative review factors (such as assessing whether the rule is still 
needed, overly burdensome, or overlaps with other regulations), the 
agency might have already reviewed and potentially modified the 
regulation one or more times based upon the same factors. The officials 
reported that, although the subsequent reviews are often duplicative 
and less productive, they nevertheless expend the time and resources 
needed to conduct the reviews in order to comply with statutory 
requirements. 

Given the long-term fiscal imbalance facing the United States and other 
significant trends and challenges, Congress and the executive branch 
need to carefully consider how agencies use existing resources. In 
particular, overlapping or duplicative reviews may strain limited 
agency resources. As agencies face trade-offs in allocating these 
limited resources to conducting mandatory and discretionary reviews, as 
well as conducting other mission-critical activities, they have to make 
decisions about what activities will produce the most benefit. In some 
cases, we observed that agencies like FAA delayed conducting any 
planned reviews for an extended period because they reported that they 
did not have the resources to conduct them. Given the trade-offs that 
agencies face, it makes sense to consider the appropriate mix of 
mandatory and discretionary reviews, and other mission-critical 
activities, that agencies can and should conduct. More specifically, 
our findings and analysis suggest that it may be useful to revisit the 
scope and timing of some review requirements to see whether there are 
opportunities to consolidate multiple requirements to enhance their 
usefulness and make them more cost effective and easier to implement. 
If the current state of review requirements remains unchanged, agencies 
may continue to expend their limited time and resources on conducting 
pro forma reviews that appear to produce less useful results. Further, 
agencies may also continue to produce less useful results for reviews 
that they rush to complete, as identified by EPA and FCC officials who 
reported that their annual and/or biannual review requirements do not 
provide enough time for them to most effectively complete their reviews 
and/or observe new changes before starting a subsequent review. 

While we believe that employing the lessons learned by agencies may 
improve the effectiveness of their retrospective reviews, we 
acknowledge that the review of regulations is only one of the tools 
that agencies will need to fully understand the implications of their 
regulatory activities. In order to fully assess the performance of 
regulatory activities, agencies will need to consider the performance 
of the programs that implement their regulations and the statutes that 
underlie the regulations. Considering any of these elements in 
isolation will provide an incomplete picture of the impact of 
regulations on the public. However, neglecting any of these elements 
will have the same effect. 

Recommendations for Executive Action: 

In order to ensure that agencies conduct effective and transparent 
reviews, we recommend that both the Director of the Office of 
Management and Budget, through the Administrator of the Office of 
Information and Regulatory Affairs, and the Chief Counsel for Advocacy 
take the following seven actions. 

Specifically, we recommend that they develop guidance to regulatory 
agencies to consider or incorporate into their policies, procedures, or 
agency guidance documents that govern regulatory review activities the 
following elements, where appropriate: 

1. Consideration, during the promulgation of certain new rules, of 
whether and how they will measure the performance of the regulation, 
including how and when they will collect, analyze, and report the data 
needed to conduct a retrospective review. Such rules may include 
significant rules, regulations that the agencies know will be subject 
to mandatory review requirements, and any other regulations for which 
the agency believes retrospective reviews may be appropriate. 

2. Prioritization of review activities based upon defined selection 
criteria. These criteria could take into account factors such as the 
impact of the rule; the length of time since its last review; whether 
changes to technology, science, or the market have affected the rule; 
and whether the agency has received substantial feedback regarding 
improvements to the rule, among other factors relevant to the 
particular mission of the agency. 

3. Specific review factors to be applied to the conduct of agencies' 
analyses that include, but are not limited to, public input to 
regulatory review decisions. 

4. Minimum standards for documenting and reporting all completed review 
results. For reviews that included analysis, these minimal standards 
should include making the analysis publicly available. 

5. Mechanisms to assess their current means of communicating review 
results to the public and identify steps that could improve this 
communication. Such steps could include considering whether the agency 
could make better use of its agency Web site to communicate reviews and 
results, establishing an e-mail listserve that alerts interested 
parties about regulatory reviews and their results, or using other Web- 
based technologies (such as Web forums) to solicit input from 
stakeholders across the country. 

6. Steps to promote sustained management attention and support to help 
ensure progress in institutionalizing agency regulatory review 
initiatives. 

We further recommend that, in light of overlapping and duplicative 
review factors in statutorily mandated reviews and the difficulties 
identified by agencies in their ability to conduct useful reviews with 
predetermined time frames, the Administrator of OIRA and Chief Counsel 
for Advocacy take the following step. 

7. Work with regulatory agencies to identify opportunities for Congress 
to revise the timing and scope of existing regulatory review 
requirements and/or consolidate existing requirements. 

Matters for Congressional Consideration: 

In order to facilitate agencies' conduct of effective and transparent 
reviews, while maximizing their limited time and resources, Congress 
may wish to consider authorizing a pilot program with selected agencies 
that would allow the agencies to satisfy various retrospective review 
requirements with similar review factors that apply to the same 
regulations by conducting one review that is reported to all of the 
appropriate relevant parties and oversight bodies. 

Agency Comments: 

We provided a draft of this report to the Secretary of Agriculture, the 
Attorney General, the Secretary of Labor, the Secretary of 
Transportation, the Administrator of EPA, the Administrator of SBA, the 
Acting Chairman of CPSC, the Chairman of FCC, the Chairman of FDIC, the 
Director of OMB, and the Chief Counsel for Advocacy for their review 
and comment. 

We received formal comments from the SBA Office of Advocacy; they 
concurred with the recommendations and, as an attachment, provided a 
copy of draft guidance that they developed in response to our 
recommendations (see app. XII). The Office of Advocacy also suggested 
that it would be more appropriate to direct the recommendations to the 
Chief Counsel of Advocacy rather than the Administrator of SBA. Because 
the Chief Counsel of Advocacy is the official who would need to act 
upon these recommendations, we made the change. 

OMB told us that they reviewed our draft report and had no comments. 
All other agencies provided technical and editorial comments, which we 
incorporated as appropriate. In its technical comments, DOT suggested 
that we expand the recommendation for agencies to identify 
opportunities for Congress to examine the timing and scope of existing 
requirements and/or consolidate existing requirements, to include 
executive agency mandated reviews. However, the focus of the 
recommendation is on statutory requirements because they tended to have 
recurring and/or predetermined review schedules. Therefore, we did not 
expand the recommendation. 

As we agreed with your office, unless you publicly announce the 
contents of this report earlier, we plan no further distribution of it 
until 30 days from the date of this letter. We will then send copies of 
this report to interested congressional committees, the Secretary of 
Agriculture, the Attorney General, the Secretary of Labor, the 
Secretary of Transportation, the Administrator of EPA, the 
Administrator of SBA, the Acting Chairman of CPSC, the Chairman of FCC, 
the Chairman of FDIC, the Director of OMB, the Administrator of OIRA, 
and the Chief Counsel for Advocacy. Copies of this report will also be 
available at no charge on our Web site at http://www.gao.gov. 

If you or your staff have any questions about this report, please 
contact me at (202) 512-6806 or sciremj@gao.gov. Contact points for our 
Offices of Congressional Relations and Public Affairs may be found on 
the last page of this report. Key contributors to this report are 
listed in appendix XIII. 

Signed by: 

Mathew J. Scire: 
Director: 
Strategic Issues: 

[End of section] 

Appendix I: Objectives, Scope and Methodology: 

Objectives: 

To provide insights concerning how agencies assess existing 
regulations, congressional requesters asked us to examine agencies' 
implementation of retrospective regulatory reviews and the results of 
such reviews. Accordingly, for selected agencies, we are reporting on: 

1. the magnitude of retrospective review activity and type of 
retrospective reviews agencies completed from calendar year 2001 
through 2006, including the frequency, impetus (mandatory or 
discretionary), and purposes of the reviews; 

2. the processes and standards that guide agencies' planning, conduct, 
and reporting on reviews, and the strengths and limitations of the 
various review processes and requirements; 

3. the outcomes of reviews, including the perceived usefulness of the 
reviews and how they affected subsequent regulatory activities; and: 

4. the factors that appear to help or impede agencies in conducting or 
using retrospective reviews, including which methods, if any, that 
agencies and we identified as most cost-effective for conducting 
reviews. 

For purposes of this report, we generally use the term retrospective 
reviews to mean any assessment of an existing regulation, primarily for 
purposes of determining whether (1) the expected outcomes of the 
regulation have been achieved; (2) the agency should retain, amend, or 
rescind the regulation; and/or (3) the actual benefits and costs of the 
implemented regulation correspond with estimates prepared at the time 
the regulation was issued. We defined mandatory reviews as 
retrospective reviews that agencies conducted in response to 
requirements in statutes, executive orders, or executive branch 
directives. We defined discretionary reviews as reviews that agencies 
undertook based upon their own initiative. 

Scope: 

For calendar years 2001 through 2006, we assessed the retrospective 
review activities of nine agencies and their relevant subagencies. The 
nine agencies included the Departments of Agriculture, Justice, Labor, 
and Transportation; Consumer Product Safety Commission (CPSC); 
Environmental Protection Agency (EPA); Federal Communications 
Commission (FCC); Federal Deposit Insurance Corporation (FDIC); and the 
Small Business Administration(SBA). The subagencies covered in detail 
by our review included USDA's Animal and Plant Health Inspection 
Service, Agricultural Marketing Service, and Food Safety and Inspection 
Service; Department of Justice's Bureau of Alcohol, Tobacco, Firearms, 
and Explosives; Department of Labor's Employee Benefits Security 
Administration, Occupational Safety and Health Administration, Mine 
Safety and Health Administration, and Employment and Training 
Administration; and the Department of Transportation's Federal Aviation 
Administration and National Highway Traffic Safety Administration. 

We selected these agencies because they include Cabinet departments, 
independent agencies, and independent regulatory agencies covering a 
wide variety of regulatory activities in areas such as health, safety, 
environmental, financial, and economic regulation.[Footnote 47] 
Further, we selected these agencies because they were actively 
conducting regulatory reviews or were responsible for responding to 
multiple review requirements. We were not able to assess the activities 
of all regulatory agencies, due to time and resource constraints, but 
given the diversity and volume of federal regulation conducted by the 
nine selected agencies, we believe that the results of our assessment 
should provide a reasonable characterization of the variety of 
retrospective regulatory reviews and the issues associated with their 
implementation. GAO's Federal Rules Database, which is used to compile 
information on all final rules, showed that the nine agencies accounted 
for almost 60 percent of all final regulations published from 2001 
through 2006.[Footnote 48] However, the volume and distribution of 
reviews covered in this report are not generalizable to all regulatory 
reviews governmentwide. To supplement our assessment of these agencies' 
activities, we also solicited the perspectives of regulatory oversight 
entities and nonfederal parties knowledgeable about regulatory issues, 
such as the Office of Information and Regulatory Affairs within the 
Office of Management and Budget, the Office of Advocacy within SBA, and 
11 nonfederal parties that represented a variety of sectors (academia, 
business, public advocacy, and state government). 

Methodology: 

To address our first objective, we interviewed and obtained 
documentation from agency officials as well as other knowledgeable 
regulatory parties on agency retrospective reviews. We administered and 
collected responses to a structured data collection instrument that 
solicited information on agencies' retrospective review activities and 
lessons learned. We supplemented this data collection by obtaining 
information from the Federal Register, Unified Agenda, and published 
dockets and agency reports. We used information obtained to describe 
the "types" of reviews that agencies conducted--in terms of impetus 
(mandatory or discretionary) and purpose (for example, burden reduction 
or effectiveness). We compared agency review activities in terms of 
impetus and purpose because important differences can be seen in the 
processes used, outcomes derived, and lessons learned, based upon these 
characteristics, which we further analyze in objectives two through 
four. Although we note that reviews can be described and compared using 
other characteristics, such as policy area assessed in the review (such 
as health, safety, or economic) or type of analyses conducted (such as 
economic benefit-cost analysis, other quantitative, and qualitative), 
we believe our selection of characteristics in describing the types of 
reviews conducted was most useful and relevant for addressing our 
objectives. 

To address our second objective, we interviewed and obtained 
documentation from agency officials as well as other knowledgeable 
regulatory parties on agency retrospective reviews. We collected 
responses to the aforementioned structured data collection instrument 
that solicited information on agencies' retrospective review activities 
and lessons learned. We supplemented this data collection by obtaining 
information from the Federal Register, Unified Agenda, and published 
dockets and agency reports. We also reviewed agency policies, executive 
orders, and statutory requirements to identify policies and procedures 
that guide the planning, conduct, and reporting of agencies' reviews. 
Further, to identify the strengths and limitations of agency review 
processes, we assessed agencies' use of three review and economic 
practices and standards that are important to the effectiveness and 
transparency of agency reviews, including the (1) use of a standards- 
based approach, (2) incorporation of public involvement, and (3) 
documentation of review processes and results. In prior work, we 
identified some overall strengths or benefits associated with 
regulatory process initiatives, including: increasing expectations 
regarding the analytical support for proposed rules, encouraging and 
facilitating greater public participation in rulemaking, and improving 
the transparency of the rulemaking process. Because these strengths or 
benefits are also relevant and useful for assessing agency 
retrospective review initiatives, we considered them in our selection 
of assessment criteria for this review. Other practices that could 
improve the effectiveness and transparency of reviews may exist and 
could be considered when developing retrospective review processes. 
However, we believe that the three practices that we assessed are among 
the most important. While we did not assess whether agencies employed 
these practices all the time, to the extent possible we did seek 
documentation and evidence that they were applied. Further, while we 
assessed whether agencies employed standards-based approaches in their 
retrospective review processes--within the scope of our review--we did 
not attempt to assess the quality of such standards. We compared the 
strengths and limitations of review processes across agencies, types of 
reviews, and phases of the review process. In our more detailed 
assessment of a limited sample of retrospective reviews completed 
between 2001 and 2006, we also evaluated the use of research and 
economic practices and standards. The sample that we assessed was too 
small to generalize to all agency retrospective reviews, but this 
assessment illustrated some of the strengths and limitations that exist 
in the agencies we reviewed. 

To address the third objective, we interviewed and obtained 
documentation from agency officials and collected responses on the 
usefulness of various types of retrospective reviews using the 
structured data collection instrument identified in objective one. To 
obtain the perspectives of nonfederal parties on the usefulness of 
agency reviews, we identified and interviewed 11 parties that represent 
a variety of sectors (academic, business, public, advocacy, and state 
government) and points of view.[Footnote 49] The parties were selected 
based on their contributions to prior GAO work on regulatory issues and 
our assessment of their recent publications on regulatory issues. The 
opinions expressed by agency officials and these nonfederal parties may 
be subjective and may not capture the views of all regulatory agencies, 
experts, and stakeholders on the usefulness of reviews. However, we 
believe that our selection represents a reasonable range of 
knowledgeable perspectives on retrospective reviews. We supplemented 
our data collection on the outcomes of agency reviews by reviewing the 
Federal Register, Unified Agenda, and published dockets and reports. 
For mandatory and discretionary reviews, we identified the reported 
results of reviews, including whether the review prompted any change to 
existing regulations. We also synthesized and described the usefulness 
of different types of reviews, as determined by agency officials and 
nonfederal parties knowledgeable about regulatory issues. 

To address the fourth objective, we interviewed and obtained 
documentation from agency officials, collected responses to a 
structured data collection instrument, and reviewed existing research 
on agency regulatory review initiatives. Further, we solicited 
perspectives of the selected oversight and nonfederal parties on the 
facilitating factors and barriers to the usefulness of agency reviews. 
Based on our analysis of agency responses and documentation, we 
described the lessons learned from the different agencies, and the 
views of oversight and nonfederal parties on facilitating and impeding 
practices. To supplement the lessons identified and to identify the 
most prevalent and/or critical facilitators or barriers for the conduct 
and usefulness of reviews, as well as options to overcome any barriers 
identified, we hosted a joint agency exit conference. During this joint 
exit conference, we discussed the collective responses of agencies and 
nonfederal parties, and similarities and differences in experiences and 
views.[Footnote 50] We had general consensus among federal agencies on 
the points discussed during this exit conference and report on areas 
where there was not consensus in agency and nonfederal parties' views. 
We conducted our work from May 2006 through April 2007 in accordance 
with generally accepted government auditing standards. 

[End of section] 

Appendix II: Department of Agriculture Retrospective Reviews: 

The three Department of Agriculture (USDA) agencies examined in this 
study actively reviewed their existing regulations under both mandatory 
and discretionary authorities. The Animal and Plant Health Inspection 
Service (APHIS), the Agricultural Marketing Service (AMS), and the Food 
Safety and Inspection Service (FSIS) conducted reviews to reduce burden 
on small entities under Section 610. USDA conducted discretionary 
reviews to respond to industry petitions or informal feedback, to meet 
recommendations from committees, to address new risks in regulated 
environments, or to update rules due to advances in technology or 
scientific knowledge. The agencies use both centralized and 
decentralized review processes that rely on the input of outside 
parties to inform their reviews. 

USDA Retrospective Reviews: 

The three USDA agencies examined in this study actively reviewed their 
existing regulations under both mandatory and discretionary 
authorities, with reviews conducted at their own discretion more common 
than mandated reviews. For example, during the 2001 through 2006 period 
covered in our review, APHIS reported conducting 18 Section 610 reviews 
and completing rulemakings for 9, with 8 others currently in progress. 
APHIS also reported that since 2001, it has completed a total of 139 
regulatory reviews, which resulted in 139 rule modifications across 12 
broad content areas. AMS officials reported initiating 19 and 
completing 11 Section 610 reviews since 2001. However, AMS also 
reported that it has issued approximately 300 modifications to 30 
regulations based on interaction with Industry Committees between 
fiscal years 2002 and 2006. In addition, AMS also reported that since 
2001, the agency has conducted 18 independent assessments of its 
commodity promotion programs, as required of AMS under 7 U.S.C. § 7401. 
FSIS reported initiating 1 Section 610 review since 2001; however 
during the same time period, the agency has conducted 36 reviews of its 
rules as a result of industry petitions. The agencies' officials 
reported that discretionary reviews more often resulted in regulatory 
changes. Our analysis of the December 2006 Unified Agenda confirmed 
that most modifications to the department's regulations were attributed 
to reasons under USDA's own discretion rather than because of mandates. 
Of the 132 rule changes listed in the Unified Agenda, 113 resulted from 
decisions made at agency discretion while 19 of those changes were the 
result of mandated actions. 

Table 4: Description of USDA Retrospective Reviews: 

Review authorities; 
Mandatory reviews: 
* Section 610; 
* AMS: Commodity Promotion and Evaluation under 7 U.S.C. § 7401; 
Discretionary reviews: 
* Agency discretion; 
* USDA internal policy to review significant rules every 5 years. 

Frequency; 
Mandatory reviews: 
* 10 years (Section 610); 
* 5 years (AMS Commodity Promotion and Evaluation); 
Discretionary reviews: 
* As needed based on industry feedback or petition; 
* 5 years (USDA internal policy). 

Purposes; 
Mandatory reviews: 
* Reduce economic impact on small entities; 
* Reduce complexity of regulation; 
* Consolidate duplicative rules; 
* Respond to changes over time in technology, economic conditions; 
* Evaluate program effectiveness and ensure that the objectives of the 
program are being met (AMS Commodity Promotion and Evaluation); 
Discretionary reviews: 
* Respond to industry petitions, recommendations from committees; 
* Advances in technology, science; 
* Meet emerging risks (disease, pests); 
* Align with international standards; 
* Requests from trading partners, states; 
* Feedback from regulated communities; 
* Respond to new, revised legislation; 
* Reduce complexity of regulation; 
* Consolidate duplicative rules; 
* Respond to changes over time in technology, economic conditions. 

General outcomes; 
Mandatory reviews: 
* APHIS: 9 of 18 reviews have resulted in completed rulemaking 
proceedings. Five other proceedings are pending completion; 
* AMS: All 11 completed Section 610 reviews have resulted in no 
changes; 
* FSIS: The only Section 610 review conducted has yet to be reported; 
Discretionary reviews: 
* APHIS: Rule changes vary in terms of less/more stringent; 
* FSIS: According to officials, 8 of the 36 reviews from petitions 
resulted in changes, and generally less stringent rules; 
* 113 of the 132 rule changes listed in the December 2006 Unified 
Agenda resulted from decisions made under agency discretion. 

Agency views on; 
usefulness of reviews; 
Mandatory reviews: 
* APHIS officials reported that periodic reviews are less useful, as 
small modifications are made to regulations on ongoing basis; 
* AMS officials reported that Section 610 reviews have not identified 
issues that were not previously known through other oversight 
mechanisms; 
Discretionary reviews: 
* FSIS officials reported that reviews focus the agency's efforts on 
measuring a regulation's effectiveness and determining whether a 
particular policy is still consistent with the agency's overall policy 
direction. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

USDA Retrospective Review Processes: 

The processes employed for review varied by agency, with AMS program 
offices conducting reviews of their own regulations, APHIS program 
offices conducting reviews in concert with centralized offices within 
the agency, and centralized offices within the agency conducting FSIS 
reviews. However, all three agencies relied on the input of regulated 
communities to inform their processes. As an example of a centralized 
approach: APHIS' technical and policy program staff work with the 
agency's Policy and Program Development (PPD) unit to conduct reviews, 
and PPD works with the Deputy Administrators for each regulatory 
program to set regulatory priorities. The program staff that oversees 
the regulation, on the other hand, conducts AMS reviews, in-house. All 
three agencies reported that they rely on outside parties to inform 
their review process. For example, AMS reported that the agency 
conducts periodic referenda of regulated growers of fruit and 
vegetables to amend agency marketing orders and to identify programs 
for discontinuance. APHIS reported that its review decisions are 
influenced by ongoing discussions with industry, state and tribal 
authorities, and foreign governments regarding setting international 
trade standards. APHIS also reported that it has acted on 
recommendations made by outside reviews of its programs conducted by 
the National Plant Board and the National Association of State 
Departments of Agriculture. FSIS reported that it holds industry 
listening sessions and public meeting to inform its rulemaking and 
affect the day-to-day implementation of regulations. Figure 5 depicts 
USDA's general process for regulatory review. 

Figure 5: USDA Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix III: Department of Justice Retrospective Reviews: 

While the Department of Justice (DOJ) is not primarily a regulatory 
agency, during the 2001 through 2006 period covered in our review, DOJ 
component agencies have conducted reviews of their existing regulations 
under both mandatory review requirements and under their own 
discretionary authorities.[Footnote 51] Most DOJ reviews were 
discretionary and in response to such drivers as changes in technology 
or feedback back from regulated entities, among other factors. The 
three mandatory reviews conducted by DOJ since 2001 were driven by 
separate statutory requirements to review regulations or set 
enforceable standards for others to follow. While DOJ has few formal 
processes or standards to guide the planning, conduct, and reporting of 
its internally conducted discretionary reviews, in the conduct of the 
one Section 610 review conducted by DOJ and evaluated by GAO, statutory 
standards were followed. 

DOJ Retrospective Review Activity: 

DOJ is not primarily a regulatory agency and officials reported that 
most of its primary activities, including antiterrorism, investigation, 
and law enforcement do not involve the department's regulatory process. 
Officials reported that few of the department's regulations are subject 
to Section 610 review, and one official reported that regulatory 
review, as a whole, is not a major priority within the agency, compared 
to its other functions. However, since 2001 DOJ agencies reported 
completing at least 13 reviews of existing regulations. Based on 
published documents in the Federal Register or Unified Agenda, 10 of 
these reviews were conducted under DOJ's own discretion, while 3 
reviews were in response to mandatory review requirements or to comply 
with statutory requirements to revise regulations. 

The drivers for the discretionary reviews conducted by DOJ included 
responding to changes in technology or feedback from regulated 
entities, among other factors. For example, FBI officials reported that 
the Bureau has reviewed and is revising a rule preventing the FBI from 
retaining or exchanging the fingerprints and criminal history record 
information of nonserious offenses in the FBI's Fingerprint 
Identification Records System. According to the proposed rule change 
resulting from this review, the existing regulations were originally 
implemented in 1974 and based on the data-processing capabilities of a 
manual record-keeping environment. Officials reported that advances in 
information technology precipitated a review of these regulations, 
which once revised, will enhance the FBI's search capability for 
fingerprint and criminal history background checks. DOJ also cited 
feedback from regulated entities as an important driver of 
discretionary reviews. DEA, for example, reported that the controlled 
substance manufacturer and distributor industries requested that DEA 
provide an electronic method to satisfy the legal requirements for 
ordering Schedule I and II controlled substances, which previously 
could only be ordered through a triplicate form issued by DEA. 
According to officials, DEA reviewed its regulations and worked with 
industry to develop a pilot program to update its system. After notice- 
and-comment rulemaking, DEA published a Final Rule revising its 
regulations on April 1, 2005.[Footnote 52] 

In addition to these reviews, ATF conducted five discretionary reviews 
since 2001, including a reorganization of Title 27 in the transition of 
ATF functions from the Department of the Treasury to DOJ after the 
creation of the Department of Homeland Security.[Footnote 53] 
Additionally, OJP conducted two discretionary reviews since 2001 and 
the BOP reported that it conducts annual, ongoing reviews of its Policy 
Statements, many of which correspond with its regulations in the CFR, 
to ensure that they are current. 

GAO was able to identify three mandatory regulatory reviews completed 
by DOJ since 2001, and the impetuses for these reviews varied. For 
example, ATF in 1997 initiated a Section 610 review evaluating the 
impact of changes to its fireworks storage and record-keeping 
requirements on small entities.[Footnote 54] This review, concluded in 
a January 29, 2003, Federal Register notice, certified that the revised 
rule will have a minimal economic impact on the explosives industry, 
and will no longer have a significant economic impact on a substantial 
number of small entities.[Footnote 55] The review also identified other 
areas of concern to the public, precipitating further actions. CRT 
conducted a review pursuant to Executive Order 12250, which requires 
the Attorney General to establish and implement a schedule for the 
review of executive branch agencies' regulations implementing various 
federal nondiscrimination laws, including the Civil Rights Act of 1964, 
among others.[Footnote 56] According to officials, this "Cureton Review 
Project" included an evaluation of the regulations of 23 agencies, 
including DOJ, which resulted in clarified statutory language to 
promote consistent compliance with the various nondiscrimination 
statutes.[Footnote 57] In a third review, CRT published an Advanced 
Notice of Proposed Rulemaking (ANPRM) to update regulations 
implementing Title II and Title III of the Americans with Disabilities 
Act of 1990 (ADA), including the ADA Standards for Accessible 
Design.[Footnote 58] According to the ANPRM, the ADA requires DOJ to 
adopt accessibility standards that are ''consistent with the minimum 
guidelines and requirements issued by the Architectural and 
Transportation Barriers Compliance Board," which were revised in July 
2004. DOJ has also reported that it may conduct a Regulatory Impact 
Analysis on the revised ADA standards, including a benefit-cost 
analysis pursuant to Executive Order 12866, OMB Circular A-4, and the 
Regulatory Flexibility Act. 

Table 5: Description of DOJ Retrospective Reviews: 

Review authorities; 
Mandatory reviews: 
* Section 610; 
* Executive Order 12250 (implementation of federal nondiscrimination 
laws); 
* Revision of ADA Accessible Design Standards; 
Discretionary reviews: 
* BOP internal policy to annually review its Program Statements. 

Frequency; 
Mandatory reviews: 
* 10 years (Section 610); 
* Executive Order 12250 (Attorney General sets evaluation schedule); 
* ADA Standards are revised as needed to enforce standards revised by 
the Architectural and Transportation Barriers Compliance Board; 
Discretionary reviews: 
* As needed, based on the feedback of regulated communities; 
* Annually (BOP). 

Purposes; 
Mandatory reviews: 
* Reduce economic impact on small entities; 
* Reduce complexity of regulation; 
* Consolidate duplicative rules; 
* Respond to changes over time in technology, economic conditions; 
Discretionary reviews: 
* Respond to petitions, feedback from regulated communities, law 
enforcement agencies, advisory boards; 
* Changes in technology, statutes, standards; 
* Ensure that regulations are current (BOP). 

General outcomes; 
Mandatory reviews: 
* In the ATF Section 610 review, ATF certified that the revised rule 
will have a minimal economic impact on the explosives industry, and 
will no longer have a SIESNOSE. However, the review identified 
additional issues for further action; 
* CRT's Executive Order 12250 review resulted in clarification of 
regulatory language within federal agency programs; 
Discretionary reviews: 
* ATF reported four discretionary reviews since 2001 which resulted in 
rulemaking proceedings. 

Agency views on usefulness of reviews; 
Mandatory reviews: ATF officials reported that reviews ensure that ATF 
is meeting the needs of government, industry, and the law enforcement 
community while meeting legislative mandates. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

DOJ Review Process: 

Department officials stated that much of DOJ's regulatory review 
process was "informally" structured, and without formal procedures and 
standards. Professional judgment, officials stated, was used in some 
cases in lieu of documented practices. However, a GAO evaluation of the 
recent ATF Explosive Materials in the Fireworks Industry review 
indicates that DOJ followed the statutorily defined process for its 
completion. As required by Section 610, the review must describe (a) 
the continued need for the rule; (b) the nature of complaints or 
comments received concerning the rule from the public; (c) the 
complexity of the rule; (d) the extent to which the rule overlaps, 
duplicates, or conflicts with other federal rules and, to the extent 
feasible, with state and local governmental rules; and (e) the length 
of time since the rule has been evaluated or the degree to which 
technology, economic conditions, or other factors have changed in the 
area affected by the rule. GAO's evaluation of this proceeding 
concluded that ATF addressed the requirements for responding to public 
comments, complaints, and the rule's complexity. ATF's analysis was 
primarily in response to public comments and a review of its own 
experience implementing the rule. In a few cases, ATF responded to 
comments by referencing published experts' opinions and scientific 
tests. However, ATF provided no overall analysis of the cost of these 
storage regulations or of their effectiveness in promoting public 
safety, or law enforcement's ability to trace fireworks to their 
manufacturer--a specific desired outcome referred to in the 
notice.[Footnote 59] Figure 6 depicts the general process for 
regulatory review in one ATF Section 610 review. 

Figure 6: ATF Section 610 Review, Explosive Materials in the Fireworks 
Industry: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix IV: Department of Labor Retrospective Reviews: 

During the 2001 through 2006 period covered in our review, agencies 
within the Department of Labor (DOL) have actively reviewed their 
existing regulations in response to both mandatory and discretionary 
drivers. Specifically, the Employee Benefits Security Administration 
(EBSA), Occupational Safety and Health Administration (OSHA), Mine 
Safety and Health Administration (MSHA), and Employment and Training 
Administration (ETA) have conducted various retrospective reviews of 
their regulations. The types of reviews--in terms of impetus and 
purpose--outcomes of reviews and processes used to conduct the reviews 
varied among the agencies. Specifically, while EBSA has established a 
formal and documented regulatory review program, OSHA, MSHA, and ETA 
have somewhat less formal review programs, but MSHA and ETA were in the 
process of developing more standardized processes. Furthermore, while 
all of the agencies reported that their discretionary reviews more 
often resulted in subsequent regulatory action, the outcomes of 
mandatory reviews varied slightly among the agencies. 

DOL Retrospective Review Activity: 

All of the DOL agencies within our review reported actively conducting 
reviews of their regulations. However, the types of reviews--in terms 
of impetus and purpose--and outcomes of reviews varied slightly among 
the agencies. All of the DOL agencies reported that they conducted 
ongoing reviews of their regulations, at their own discretion. However, 
two of the agencies--OSHA and EBSA--also incorporated requirements from 
mandatory reviews into these discretionary reviews. Furthermore, EBSA 
conducts its discretionary reviews more formally as part of its 
Regulatory Review Program. According to documentation that we reviewed 
on this program, EBSA formally conducted reviews of its existing 
regulations in response to specific developments and/or changes in the 
administration of group health, pension, or other employee benefit 
programs, changes in technology and industries, and legislation. EBSA 
also reviewed regulations in response to identified enforcement 
problems or the need to further the agency's compliance assistance 
efforts through improved guidance. Furthermore, the review program 
incorporates Section 610 reviews as part of the program. While OSHA did 
not have a program that was as formalized and documented as EBSA, the 
officials reported and our review of their analyses confirmed that the 
agency also incorporated Section 610 criteria into broader review 
initiatives that the agency undertook to address informal feedback from 
industry, stakeholders, and staff. MSHA and ETA also reported 
initiating reviews in response to either stakeholder input, technology 
or policy updates, petitions, or internal identification of needed rule 
changes. However, the agencies' officials reported that they have not 
conducted any Section 610 reviews (which focus on burden reduction) 
during the period covered in our review because they have not had any 
regulations within the last 10 years that had a SEISNOSE effect. 

Outcomes of reviews varied slightly among the agencies. While it was 
not possible to account for all of the reviews conducted by all of the 
agencies because the agencies did not document some informal reviews, 
collectively the agencies reported completing at least 60 reviews since 
2001. According to EBSA documentation, the agency completed at least 7 
of its 13 formal retrospective reviews, including 4 Section 610 
reviews. All of the discretionary reviews resulted in subsequent 
regulatory changes, including changes to the regulation, guidance, or 
related materials. None of EBSA's Section 610 reviews resulted in 
regulatory changes. OSHA completed 4 reviews in response to both 
discretionary and Section 610 requirements which resulted in regulatory 
changes or changes to guidance documents or related materials. 
According to OSHA documentation, 2 of their completed Section 610 
reviews and 2 of their Standards Improvement Project Reviews 
recommended regulatory changes, including clarifications to standards 
or additional outreach or compliance assistance materials. MSHA 
officials reported engaging in a 2004 MSHA Strategic Initiative Review 
(a review of all Title 30 CFR regulations) and a review conducted 
according to an MSHA initiative to improve and eliminate regulations 
that were frequently the subject of petitions for modification. Both of 
these reviews resulted in changes to regulations. ETA officials 
reported that, in 2002, the agency conducted a regulatory cleanup 
initiative that resulted in updates to individual regulations and that 
ETA has updated individual regulations when the agency's program 
offices identified a need to do so through their course of business. 
The agencies also reported making regulatory changes based upon 
departmentwide regulatory cleanup initiatives in 2002, and 2005/2006, 
which the department's Office of the Assistant Secretary for Policy 
spearheaded. Additionally, the department completed 42 reviews in 
response to Office of Management and Budget (OMB) regulatory reform 
nominations from 2001 to 2004, which resulted in subsequent regulatory 
action.[Footnote 60] 

Table 6: Description of DOL Retrospective Reviews: 

Review Authorities; 
Mandatory reviews: 
* Section 610; 
Discretionary reviews: 
* Agency initiatives in response to technology changes, petitions, 
enforcement and compliance issues; 
* Department Regulatory Clean Up Initiative (all agencies, including 
OSHA, EBSA, and ETA); 
* EBSA: Reviews also conducted under the agency's formal Regulatory 
Review Program. 

Frequency; 
Mandatory reviews: 
* Every 10 years; 
* MSHA and ETA: No reviews conducted under this authority; 
Discretionary reviews: 
* Discretionary reviews are completed as needed; 
* Departmentwide Review Initiatives occurred at least twice; 
* EBSA: Regulatory Review Program reviews are conducted annually. 

Purposes; 
Mandatory reviews: 
* Reduce burden on small entities; 
* Reduce complexity of regulation; 
* Consolidate duplicative rules; 
* Respond to changes over time in technology, economic conditions; 
Discretionary reviews: 
* Effectiveness (for example, performance measurement or other 
measurements of efficiency); 
* Improved enforcement and compliance; 
* Burden reduction. 

General outcomes; 
Mandatory reviews: 
* EBSA: No changes to regulations; 
* OSHA: When Section 610 and discretionary reviews were combined, 
reviews resulted in some changes; 
Discretionary reviews: 
* Subsequent regulatory action (e.g., changes to regulations, guidance, 
or related materials). 

Agency views on usefulness of reviews; 
Mandatory reviews: 
* EBSA officials reported Section 610 mandatory review requirements are 
less comprehensive than their discretionary reviews because they are 
limited to regulations with SEISNOSE and focus primarily on 
deregulation; 
* OSHA officials reported Section 610 mandatory review criteria were 
reasonably clear and relevant, while Executive Order 12866 criteria are 
less concise; 
Discretionary reviews: 
* Burden reduction; 
* Officials reported that reviews are useful to identifying whether: 
(1) the agency is achieving its goals, or (2) regulation or guidance 
changes are needed. MSHA also reported reviews are useful to saving 
costs associated with reviewing several petitions for a rule that needs 
modification. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

DOL Retrospective Review Processes: 

The development of review processes for DOL agencies ranged from 
processes that were documented and formal with established review 
structures and procedures, to informal undocumented review processes 
with structures and procedures that were still developing. For example, 
EBSA established a formal review program that established a formal 
structure for reviews, including identification of what resources 
(staff) would be involved, criteria that the agency would use to select 
and assess regulations, and the method for reporting results. While 
OSHA did not have a documented formal review program, the agency 
described a somewhat formal structure that it uses to conduct its 
reviews. Similarly, ETA officials reported that they just recently 
established a more formal structure for their review process, including 
the creation of a Regulations Unit that will coordinate the development 
of regulations for ETA legislative responsibilities and formalize 
regulatory procedures within the agency. According to the officials, 
the Regulations Unit will establish time frames and/or internal 
triggers for reviews to ensure the agency properly reviews and updates 
regulations. However, they noted that, given the recent establishment 
of this unit, it might take some time to implement these procedures. 
MSHA did not appear to have a documented formal review process or 
structure for its discretionary and mandatory reviews. However, the 
agency reported that it had been engaged in soliciting contractors to 
develop a more formal process for how to prioritize what regulations 
that agency would review. Figures 7 and 8 illustrate an example of the 
variation in the agencies' review processes. To facilitate sharing 
practices, in appendix XI we provide a more detailed description of 
practices within EBSA's review process, which was the most formalized 
and documented review process that we examined within the scope of our 
review. 

Figure 7: EBSA Retrospective Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

Figure 8: OSHA Retrospective Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix V: Department of Transportation Retrospective Reviews: 

Between 2001 and 2006, Department of Transportation (DOT) agencies 
within the scope of our evaluation actively reviewed their existing 
regulations under both mandatory and discretionary 
authorities.[Footnote 61] The mandatory reviews conducted by DOT 
agencies addressed governmentwide, departmentwide, and agency-specific 
review requirements. DOT conducted discretionary reviews in response to 
formal petitions and informal feedback from the public and in response 
to accidents or similar events and changes in specific industries, 
technologies, or underlying standards. Additionally, DOT conducted 
reviews in response to Office of Management Budget (OMB) regulatory 
reform initiatives as well as a stand-alone initiative to review all 
rules under the department's authority. DOT has written policies and 
procedures guiding the planning, conduct, and reporting of reviews. 
While review processes may vary somewhat within DOT agencies, overall 
these agencies follow DOT guidelines in the conduct of their reviews. 

DOT Retrospective Review Activity: 

DOT has conducted a number of initiatives to systematically review 
existing regulations to comply with federal mandates and DOT's own 
policies and procedures for regulatory review. In order to satisfy 
Section 610 and other review requirements, DOT initiated a 10-year plan 
in 1998 to systematically review some of its sections of the Code of 
Federal Regulations every year, with the objective of reviewing all of 
its regulations over a 10-year cycle. DOT also maintains a 
departmentwide review requirement, instituted in 1979, to periodically 
review existing regulations to determine whether they continue to meet 
the needs for which they originally were designed or whether reviewed 
rules should be revised or revoked. More recently, in 2005, acting 
under its own discretion, DOT initiated and completed a special stand- 
alone regulatory review in which the department sought public comment 
on all rules and regulations under DOT's authority. DOT also reviewed 
regulations in response to OMB initiatives in 2001, 2002, and 2004, 
which solicited nominations from the general public for federal 
regulations and guidance documents for reform. The agency completed 61 
reviews in response to these reform initiatives, and the department 
took subsequent action on 43 of the regulations it reviewed. Overall, 
during the 2001 through 2006 period covered in our review, DOT has 
reported conducting over 400 reviews of existing regulations to meet 
governmentwide review requirements, including those under Executive 
Order 12866 on Regulatory Planning and Review, Section 610, and the 
Executive Memorandum of June 1, 1998, on Plain Language in Government 
Writing.[Footnote 62] 

In addition to reviews conducted under departmentwide requirements, 
various agencies within DOT have reviewed regulations within the 
specific statutes under their purview. For example, since 2001 FAA has 
reviewed three regulations pursuant to requirements in the Federal 
Aviation Reauthorization Act of 1996. According to agency officials, 
these reviews included post implementation cost-benefit assessments of 
three, high-cost FAA rules. FMCSA reported that it also reviews any 
regulation impacted by the Motor Carrier Act of 1980; the Motor Carrier 
Safety Improvement Act; and the Safe, Accountable, Flexible, Efficient 
Transportation Equity Act: A Legacy for Users (SAFETEA-LU). Although 
not within the time frame for this study, FTA announced in the December 
2006 Unified Agenda that it will undertake a review of its regulations 
to bring them into conformity with the SAFETEA-LU statute. 

In addition to these more formal regulatory review efforts, DOT 
officials reported that the department also reviews its existing 
regulations at its own discretion as a function of its daily, ongoing 
activities. According to officials, such reviews are often the result 
of petitions from or consultations with parties affected by DOT 
regulations or based on the experience of agency staff members in light 
of changes in specific industries, technologies, or underlying 
standards. DOT officials said that, for some of their agencies, 
reviewing petitions for rulemaking or regulatory waivers is the most 
productive way to obtain public input on a review of that rule. An 
evaluation of NHTSA's entries in DOT's December 2005 Unified Agenda 
indicated 10 rule change proceedings in various stages of completion 
which were the result of petitions from regulatory stakeholders. NHTSA 
also reported that, since 2001, it has conducted 17 reviews of Federal 
Motor Vehicle Safety Standards (FMVSS), including a few studies 
evaluating the benefits and costs of various standards. PHMSA reported 
that the granting of numerous waivers of a regulation is a particular 
signal that new technology or conditions may render that regulation 
obsolete or in need of amendment. 

Table 7: Description of DOT Retrospective Reviews: 

Review authorities; 
Mandatory reviews: 
* Section 610; 
* Executive Order 12866; 
Discretionary reviews: 
* Responses to petitions, requests for interpretations; 
* Responses to changes in specific industries, technologies, or 
underlying standards. 

Frequency; 
Mandatory reviews: 
* DOT has a formal program for reviewing all of its regulations over a 
10-year cycle. Every year, DOT reviews some of its sections of the CFR 
in that 10-year effort; 
Discretionary reviews: 
* As needed, as a function of daily activities; 
* As determined by the office initiating the review, under DOT 
departmentwide policy to periodically review existing regulations. 

Purposes; 
Mandatory reviews: 
* Reduce burden on small entities; 
* Reduce complexity of regulation; 
* Consolidate duplicative rules; 
* Respond to changes over time in technology, economic conditions; 
Discretionary reviews: 
* Meet complaints or suggestions of public; 
* Simplify or clarify language, particularly in response to petition, 
problems evidenced in the enforcement of the regulation;; 
* Eliminate overlapping, duplicative regulations;; 
* Eliminate conflicts and inconsistencies in its own regulations or 
those of other initiating offices or other agencies;; 
* Insure continued relevance of the problem the regulations were 
originally intended to solve;; 
* Address changes in technology, economic conditions, or other factors; 
and; 
* Monitor rules receiving a number of requests for exemption from a 
regulation. 

General outcomes; 
Mandatory reviews: 
* According to DOT review summary reports, 102 of the 406 reviews 
conducted between 2001 and 2006 resulted in further action, such as 
making substantive or plain language amendments as well as identifying 
regulations for further study; 
* DOT reported 5 rulemakings resulting from Section 610 reviews; 
* FAA reported that none of the agency's Section 610 reviews have 
resulted in any changes to the regulations; 
Discretionary reviews: 
* 43 of the 61 DOT rules nominated in the OMB reform initiatives 
resulted in subsequent actions; 
* According to NHTSA officials, the majority of NHTSA reviews do not 
result in changes to regulations because they confirmed that the 
regulations were effective; 
* According to FAA officials, no changes have been promulgated or 
proposed for reviews conducted in response to the Federal Aviation 
Reauthorization Act of 1996; 
* DOT received and responded to 120 public comments based on 2005 stand-
alone review process. DOT identified 36 comments as warranting further 
action and 21 as warranting further consideration. 

Agency views on usefulness of reviews; 
Mandatory reviews: 
* DOT officials characterized regulatory reviews as generally valuable 
because they ensure the agency focuses on whether the rule is meeting 
its objectives, whether changed circumstances warrant revisions or 
revocation, or whether the rule is as cost effective as originally 
thought; 
* DOT officials reported that when formal reviews do not result in 
changes, it may be because changes were already made as a result of 
informal or discretionary reviews.[A]. 

Source: GAO analysis of agency data and regulatory statutes. 

[A] Informal reviews generally refer to those conducted as a routine 
occurrence during daily general operations of an agency, when problems 
are identified with existing rules that might warrant further action. 
See Eisner and Kaleta at page 7. 

[End of table] 

DOT Retrospective Review Process: 

DOT has written policies and procedures guiding the planning, conduct, 
and reporting of reviews. While the processes employed by DOT agencies 
may vary somewhat, overall these agencies follow DOT guidelines in the 
conduct of their reviews. For example, DOT's Policies and Procedures 
provide guidance for prioritizing regulations for review, including the 
extent of complaints or suggestions received from the public; the 
degree to which technology or economic factors have changed; and the 
length of time since the regulations were last reviewed, among other 
factors.[Footnote 63] DOT's procedures also provide agencies with 
discretion in applying the procedures. For example, NHTSA reported that 
it gives highest priority to the regulations with the highest costs, 
potential benefits, and public interest, while PHMSA reported that it 
gives highest priority to initiatives it deems most likely to reduce 
risk and improve safety. Additionally, while DOT officials reported 
that DOT considers OMB Circular A-4 on "Regulatory Analysis" as a guide 
for cost/benefit analysis of regulatory outcomes, FAA reported that it 
uses a set of flexible procedures recommended by an outside consultant 
to conduct ex post evaluations of some rules. With regard to public 
participation in the review process, Appendix D to the department's 
Unified Agenda announces the complete schedule for all reviews, 
requests public comments for reviews in progress, and reports the 
results of completed reviews and their results. DOT agencies also 
pointed out that they regularly interact with stakeholders, such as 
regulated industries, consumers, and other interested parties to obtain 
feedback on regulations. For example, FAA officials stated that the 
agency holds conferences with industry and consumer groups to identify 
regulatory issues for review. In terms of the reporting of review 
results, DOT publishes brief summaries of completed reviews in Appendix 
D of its Unified Agenda. However, agencies may report review results in 
other ways. For example, FMCSA documents the results of its Section 610 
reviews in an annual internal report, while NHTSA publishes the 
technical reports of its reviews in the Federal Register, requesting 
public comments on its determinations. Figure 9 depicts DOT's general 
process for regulatory review. 

Figure 9: Illustration of DOT's Review ProgramA: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[A] This flow chart presents a summary of what DOT describes as its 
formal review process, which includes reviews conducted in its 10-year 
review plan. This illustration may not convey the actual process used 
in a given situation, which may vary to accommodate complexities not 
included in this depiction. 

[End of figure] 

[End of section] 

Appendix VI: Consumer Product Safety Commission Retrospective Reviews: 

Since 2001, the Consumer Product Safety Commission (CPSC or the 
Commission) systematically reviewed its regulations under its own 
discretion, but has not conducted any mandatory reviews because none of 
its rules triggered Section 610 or other mandatory review requirements. 
Moreover, agency officials noted that because of its reliance on 
voluntary consensus standards, the agency does not promulgate as many 
rules as other regulatory agencies. However, the primary purpose of 
CPSC discretionary reviews is to assess whether the regulations that 
CPSC promulgates remain consistent with the objectives of the 
Commission. In performing its reviews, CPSC has created systematic 
processes for the planning, conduct, and reporting of its reviews. 
Through this process, the Commission prospectively budgets for its 
reviews. Because CPSC's review program is so new, the agency has not 
completed most of the reviews that it has initiated, but the Commission 
has proposed changes to at least two existing regulations. In addition, 
the officials reported that their review program has been useful to the 
Commission. 

CPSC Retrospective Review Activity: 

CPSC actively conducted reviews of its existing regulations under its 
own discretion. Specifically, the Commission implemented a pilot review 
program in 2004, with annual follow-up efforts in 2005 and 2006, which 
resulted in the initiation of 14 retrospective reviews. CPSC initiated 
this review process partly because of an Office of Management and 
Budget (OMB) Program Assessment Rating Tool (PART) recommendation that 
the agency develop a plan to systematically review its current 
regulations to ensure consistency among them in accomplishing program 
goals. The primary purpose of CPSC reviews is to assess the degree to 
which the regulations under review remain consistent with the 
Commission's program policies and program goals. CPSC also assesses 
whether it can streamline regulations to minimize regulatory burdens, 
especially on small entities. The officials reported that their review 
process is so new that they have not yet fully completed it for all of 
the reviews that they have initiated. However, they have completed at 
least 3 of their 14 initiated reviews. 

Officials reported that, while some of the regulations they reviewed 
did not need a revision, they have proposed regulatory changes for two 
regulations, including standards for flammability of clothing textiles 
and surface flammability of carpets and rugs. They reported that their 
reviews could focus on opportunities to either expand or streamline 
existing regulations. Thus, their reviews could lead to increases or 
decreases in the scope of CPSC regulations. As examples, CPSC officials 
reported that during their review of their existing bicycle regulation 
they identified that the regulation did not reflect new technology and 
materials, and therefore needed to be modified and updated. Conversely, 
their review of their cigarette lighter rule revealed that the agency 
needed to promote greater compliance and more effective enforcement, 
which increased the agency's regulatory oversight. Table 8 provides 
additional detail on the CPSC retrospective reviews. 

Table 8: Description of CPSC Retrospective Reviews: 

Review authorities; 
Discretionary reviews: 
* Agency-initiated reviews conducted under the agency's formal 
regulatory review program, in response to OMB PART recommendation; 
* Technology changes, petitions, test lab input, new industry 
standards, obsolescence of test equipment. 

Frequency; 
Discretionary reviews: 
* Under new program review about 4 regulations per year; 
* Other discretionary reviews conducted as needed. 

Purposes; 
Discretionary reviews: 
* Effectiveness in meeting agency goals (such as performance 
measurement); 
* Burden reduction. 

General outcomes; 
Discretionary reviews: 
* Process new, but completed reviews have led to subsequent regulatory 
action (e.g., proposed changes to regulations). 

Agency views on usefulness of reviews; 
Discretionary reviews: 
* Officials reported that reviews are generally useful to identifying 
and responding to needed regulatory changes, confirming that some rules 
are producing intended results and improving enforcement. However, 
review of regulations not as prominent for CPSC because of reliance on 
voluntary standards. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

CPSC Review Process: 

CPSC established a formal review program that prospectively budgets for 
the substantive reviews that the agency will conduct. Officials 
reported that they have conducted about four substantive reviews per 
year using this process, while still managing other agency priorities. 
The process consists of three phases, including: (1) prioritization and 
selection of regulations to substantively review, (2) substantive 
review of the selected regulations, and (3) reporting results to the 
Commissioners and the public, for certain reviews. As part of this 
process, CPSC staff prioritize which regulations need review by 
considering: (1) which rules have the oldest effective dates, (2) which 
rules were adopted under various statutes under CPSC's authority, and 
(3) which rules staff identified as good candidates for change (from 
their experience working with the regulation). As resources allow, the 
agency selects one substantive regulation from each of their statutes' 
areas (with the exception of the Refrigerator Safety Act), starting 
with their earliest regulations.[Footnote 64] As part of this 
prioritization process, the agency considers input from CPSC's 
technical staff and outside groups. CPSC staff initiates substantive 
review of regulations that the Commission chooses for review. In this 
process, the agency solicits public comments using the Federal 
Register, assesses the comments received, conducts an internal 
technical review of the regulation, and reports the results to the 
Commissioners. The Commissioners make a policy decision on actions the 
agency will take based upon staff recommendations. If the agency 
decides to conduct a follow-on activity to update a rule, it 
subsequently notifies the public via the Federal Register. For rule 
reviews that result in Commission-approved projects for certain 
rulemaking activities (such as developing revisions to a rule for 
Commission consideration), CPSC makes the briefing packages available 
on its Web site. Other rule reviews (such as reviews for which staff 
suggests no action) are given to the Commissioners, but are not posted 
on the Web site. Figure 10 illustrates the general review process. 

Figure 10: CPSC Retrospective Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix VII: Environmental Protection Agency Retrospective Reviews: 

During the 2001 through 2006 period covered in our review, program 
offices within Environmental Protection Agency (EPA) have conducted 
numerous retrospective reviews of EPA existing regulations and 
standards. The mix of reviews conducted by EPA--in terms of 
authorities--varied across the agency, but the purposes of these 
reviews--effectiveness, efficiency, and burden reduction--were similar 
across the agency. EPA's retrospective review results provided three 
distinctive outcomes. While the agency conducts many reviews under its 
mandates, reviews conducted on its own discretion yielded more changes 
to existing regulations than mandated reviews. The review processes 
within EPA's program offices, though different, typically shared 
similar elements in the planning, conduct, and reporting of results. 
Overall, EPA reported that its retrospective reviews have proven to be 
useful to the agency. 

EPA Retrospective Review Activities: 

The Office of Air and Radiation (OAR), the Office of Prevention, 
Pesticides, and Toxic Substance (OPPTS), the Office of Solid Waste and 
Emergency Response (OSWER), and the Office of Water within EPA each 
conduct mandatory retrospective reviews under their guiding statutes 
and focus the reviews on what is stated in statute or developed by the 
agency. Thus, the frequency of mandated reviews varies within EPA as 
well as the program offices. For instance, the frequency of reviews 
required by the Safe Drinking Water Act (SDWA), conducted by the Office 
of Water, ranges from every 3 years to every 7 years, depending on the 
review requirement, while the OAR is required to conduct reviews by the 
Clean Air Act ranging from every 3 years to every 8 years. Mandated 
reviews, such as those required by agency-specific statutes, mainly 
focused on effectiveness, while Section 610 reviews and Office of 
Management and Budget (OMB) Regulatory Reform Nominations were focused 
on burden reduction. According to EPA officials, mandatory 
retrospective reviews have generally resulted in limited or no changes 
to regulations, while reviews conducted under discretionary authority 
usually resulted in more changes.[Footnote 65] For instance, of the 14 
Section 610 reviews conducted by the program offices since 2001, only 1 
resulted in change.[Footnote 66] Moreover, OAR noted that most of its 
reviews validated the need for the regulation or standard. However, 
EPA's review of regulations in response to OMB's Manufacturing 
Regulatory Reform initiative resulted in 19 regulatory changes and 19 
nonregulatory actions, including the development of regulatory guidance 
and reports. In addition, GAO's review of EPA's December 2006 Unified 
Agenda entries also revealed that 63 out of 64 rules identified as 
changed or proposed for changes were the result of decisions made under 
EPA's discretionary authority. Though the use of discretionary 
authority produced more rules changes, officials reported that 
retrospective reviews, in general, were valuable in (1) determining 
whether new information exists which indicates the need for revisions 
and (2) enabling the agency to gain new insights about its analytical 
methods. In addition, officials noted that retrospective reviews were 
useful in determining whether the rule was working as intended and 
helping to achieve the agency's or statute's goals. 

Table 9: Description of EPA Retrospective Reviews: 

Review Authorities; 
Mandatory reviews: In response to: 
* Clean Air Act; 
* Clean Water Act; 
* Federal Food, Drug, & Cosmetic Act; 
* Federal Insecticide, Fungicide, and Rodenticide Act; 
* Comprehensive Environmental Response, Compensations and Liability 
Act; 
* Food Quality Act; 
* Safe Drinking Water Act; 
* Section 610 of the Regulatory Flexibility Act; 
* Court decisions; 
Discretionary reviews: In response to: 
* OMB's Manufacturing Regulatory Reform Nominations; 
* Formal petitions; 
* Industry changes; 
* New scientific studies; 
* Informal comments from the public; 
* Emergencies or disasters. 

Frequency; 
Mandatory reviews: Ranges from every 2 years to every 10 years 
depending on the review requirement; 
Discretionary reviews: Depends largely on the nature of the source. 

Purposes; 
Mandatory reviews: Effectiveness, burden reduction, and efficiency; 
Discretionary reviews: Depends largely on the nature of the source. 

General outcomes; 
Mandatory reviews: Reviews generally resulted in no changes, a 
validation that existing regulations are working. Sometimes reviews 
resulted in the development of guidance materials in lieu of a rule 
modification; 
Discretionary reviews: Reviews generally resulted in changes, 
particularly in response to OMB nominations. 

Agency views on usefulness of reviews; 
Mandatory reviews: Officials reported that the frequency of some 
mandated reviews limits the usefulness of the reviews. However, reviews 
were useful to determine whether the regulations were working as 
intended; 
Discretionary reviews: Officials did not differentiate between the 
usefulness of discretionary reviews versus mandatory reviews. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

EPA Review Process: 

EPA's review process varied by program office and by review 
requirement; however, most mandatory and discretionary reviews 
contained consistent elements. The four EPA program offices included in 
our review perform various functions of the agency that rarely overlap 
into other program offices duties. For example, OAR exclusively 
oversees the air and radiation protection activities of the agency, 
while the Office of Water solely manages the agency's water quality 
activities. These two offices have different guiding statutes that 
require them to conduct reviews and, within those statutes, processes 
are sometimes outlined for how the agency should conduct the reviews. 
Therefore, the processes for these program offices varied. However, 
three elements were similar across the offices: these included formal 
or informal notification of the public; involvement of the public in 
the conduct of the review mainly through the request of public 
comments, science, risk, or policy assessments of the regulation; and 
release of the results to the public primarily through the Federal 
Register and the EPA Web site. In addition, mandatory and discretionary 
regulatory reviews that were high profile in nature (e.g., because they 
were conducted in response to emergencies, were contentious, or 
received heavy attention from the public, Congress, or regulatory 
experts) had the aforementioned elements as well as high-level 
management attention from the Assistant Administrator of the program 
office or the EPA Administrator. For example, the review processes of 
the mandatory National Ambient Air Quality Standards and the Lead and 
Copper review, which was initiated after elevated levels of lead were 
found in the District of Columbia, were defined, documented, and 
included extensive public involvement and high-level management 
attention. Figure 11 illustrates the general review process for the 
different review drivers. 

Figure 11: EPA Retrospective Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix VIII: Federal Communications Commission Retrospective Reviews: 

The Federal Communications Commission (FCC or the Commission) actively 
reviews its existing regulations to meet congressionally mandated 
review requirements, and to respond to petitions from regulated 
entities and changes in technology and market conditions, under its own 
discretionary authority. While FCC's retrospective review processes 
vary depending on the review requirement the agency is addressing, 
FCC's biennial and quadrennial review processes provide opportunities 
for public participation and transparency. According to FCC officials, 
the frequency of the biennial review requirement presents staffing 
challenges to the agency, while the 10-year requirement for the Section 
610 review presents a challenge to the usefulness of this review, as 
regulations may have been previously modified under other requirements 
prior to the review. 

FCC Retrospective Review Activity: 

FCC actively reviews its existing regulations to meet congressionally 
mandated review requirements and to respond to petitions from regulated 
entities and changes in technology and market conditions under its own 
discretionary authority. Under the Communications Act, as amended, the 
Commission is subject to two agency-specific mandated reviews: (1) the 
biennial regulatory review of FCC telecommunications rules,[Footnote 
67] and (2) the quadrennial regulatory review of the broadcast and 
media ownership rules.[Footnote 68] FCC officials reported that these 
reviews are guided by the deregulatory tenor of the Telecommunications 
Act, which instructed the Commission to promote competition and reduce 
regulation in the telecommunications and broadcast industries. The 
purpose of these reviews is to identify rules no longer necessary in 
the public interest so that they may be modified or repealed. In the 
2002 biennial review, the Commission conducted and reported 89 separate 
review analyses of its telecommunications regulations and made more 
than 35 recommendations to open proceedings to consider modifying or 
eliminating rules. The Commission is also subject to the governmentwide 
review requirement to minimize significant economic impact on small 
entities under Section 610. FCC has initiated 3 multiyear Section 610 
review projects (1999, 2002, 2005), plus 1 single-year review (2006), 
issuing public notices listing all rules subject to Section 610 review. 
Officials pointed out that these reviews rarely result in rulemaking 
proceedings and cited only one proceeding which resulted in the 
elimination of obsolete rules as a result of the Section 610 
process.[Footnote 69] 

In addition to these mandatory requirements, FCC officials reported 
that the Commission reviews existing regulations at its own discretion 
in response to rapid changes in technology and market conditions and to 
petitions from regulated entities. A GAO analysis of the December 2006 
Unified Agenda indicated that most of FCC's proposed and final rule 
changes for that year were the result of decisions made under FCC's 
discretionary authority. Of the 39 rule changes listed in the Unified 
Agenda, 33 were the result of decisions made at the Commission's own 
discretion, while 6 of those changes were the results of mandated 
actions. This informal analysis indicates that, in addition to its 
mandatory review requirements, FCC does make efforts to review and 
amend regulations under its own discretion. 

Table 10: Description of FCC Retrospective Reviews: 

Review authorities; 
Mandatory reviews: 
* Section 610; 
* Section 11 of Communications Act; 
* Section 202(h) of Telecommunications Act; 
Discretionary reviews: 
* Agency discretion. 

Frequency; 
Mandatory reviews: 
* 10 years (Section 610); 
* 2 years (Section 11); 
* 4 years (Section 202(h)); 
Discretionary reviews: 
* As needed, based on industry developments, issues of interest to the 
general public. 

Purposes; 
Mandatory reviews: Section 610 Review: 
* Reduce burden on small entities; 
* Reduce complexity of regulation; 
* Consolidate duplicative rules; 
* Respond to changes over time in technology, economic conditions; 
Biennial review (Section 11): 
* Repeal or modify telecommunications rules that are no longer 
necessary in the public interest as a result of meaningful economic 
competition; 
Quadrennial review (Section 202(h)): 
* Repeal or modify broadcast ownership rules determined to be no longer 
in the public interest; 
Discretionary reviews: 
* Respond to petitions for reconsideration, rulemaking, waiver, or 
forbearance; 
* Changes in technology or economic conditions in regulated sectors; 
* Respond to interactions with regulated entities, general public. 

General outcomes; 
Mandatory reviews: Section 610 Review: 
* FCC reported that most reviews result in no changes, and cited only 
one proceeding which resulted in the elimination of rules; 
Biennial review: 
* Most biennial reviews result in some changes to the Commission's 
regulations, either to modify or remove rules that are no longer 
necessary; 
Quadrennial review: 
* The 2002 review order would have relaxed most rules under review; 
Discretionary reviews: 
* Officials stated that the outcomes of reviews varied based upon the 
regulation reviewed, purpose of the review, and impetus of the review. 

Agency views on usefulness of reviews; 
Mandatory reviews: 
* Officials reported that review requirements enable FCC to conduct 
notice-and- comment reviews of existing regulations more often than it 
might otherwise; 
Discretionary reviews: 
* Officials reported that discretionary reviews enable FCC to 
prioritize reviews and provide timely response to industry 
developments. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

FCC Review Process: 

FCC's retrospective review processes vary somewhat depending on the 
review requirement the agency is addressing. However, in its conduct of 
the biennial and quadrennial review, FCC follows a notice-and-comment 
review process which provides opportunities for public participation 
and transparency. For example, in the 2002 biennial review, FCC Bureaus 
and Offices issued public notices listing rules for review under their 
purview and requesting comments regarding the continued necessity of 
rule parts under review. The Bureaus and Offices published Staff 
Reports on the FCC Web site summarizing public comments and making 
determinations as to whether the Commission should open proceedings to 
modify or eliminate any of the reviewed rules. The Commission released 
Notices of Proposed Rulemaking, seeking further public comments. 
Officials reported that if the Commission modifies or eliminates any 
regulations as a result of its proceeding, that decision is announced 
in a rulemaking order, which is published in the Federal Register. 
Similarly, in the 2006 quadrennial review (which was in process at the 
time this report was written) the Commission released a Further Notice 
of Proposed Rulemaking (FNPR) and posted a Web page providing 
background information and hyperlinks to FCC documents relevant to the 
review. The FNPR requests public comment on the media ownership rules 
and factual data about their impact on competition, localism, and 
diversity. The Commission reported that it will hold six public 
hearings in locations around the country and make available for public 
comment 10 economic studies commissioned by FCC on issues related to 
the media ownership rules. 

Despite the opportunities for public participation in these regulatory 
reviews, the mandated structure of some review processes presents a 
challenge to the usefulness of FCC reviews. For example, according to 
an FCC official, the requirement to review the Commission's 
telecommunications rules every 2 years forces Bureau staff to be 
constantly reviewing regulations. This official reported that the 
quadrennial requirement is a more appropriate time period for review, 
as it provides greater opportunity for regulatory changes to take hold. 
Additionally, an official reported that too much time between reviews 
can be problematic. For example, rules that require Section 610 review 
every 10 years may have been modified or previously reviewed as part of 
an overlapping review requirement or as part of a discretionary review 
occurring prior to the 10-year review requirement. 

Figure 12: FCC Retrospective Review Processes: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix IX: Federal Deposit Insurance Corporation Retrospective 
Reviews: 

During the 2001 through 2006 period covered in our review, the Federal 
Deposit Insurance Corporation (FDIC), has performed numerous 
retrospective reviews of its existing regulations in response to 
mandatory authorities such as Section 610 of the Regulatory Flexibility 
Act and the Economic Growth Regulatory and Paperwork Reduction Act of 
1996 (EGRPRA) and at its own discretion. The focus of FDIC's reviews 
has been on burden reduction, which is part of the agency's strategic 
goals. The process that FDIC used to plan, conduct, and report its 
reviews was coordinated by a larger organizational body. The 
centralized review effort helped to leverage the agencies' resources 
and facilitate the regulatory changes recommended as a result of the 
EGRPRA reviews. 

FDIC Retrospective Reviews: 

FDIC, along with members of the Federal Financial Institutions 
Examination Council (FFIEC) has examined 131 regulations under 
EGRPRA.[Footnote 70] FDIC conducted two Section 610 reviews after 2001, 
but before the initiation of the EGRPRA reviews in 2003. Because the 
EGRPRA review affected almost all of FDIC's regulations, the agency 
subsequently included Section 610 reviews within the EGRPRA review 
effort. Also, the agency has conducted discretionary reviews in 
response to petitions and external emergencies, such as natural 
disasters. For instance, the agency reported reviewing its regulations 
to reduce burden for businesses affected by Hurricane Katrina. In doing 
so, the agency made 8 temporary regulatory changes to ease the burden 
on affected entities. FDIC also made changes to 4 regulatory areas, 
which included changes to 3 regulations, as a result of the EGRPRA 
reviews. Additionally, GAO's review of the December 2006 Unified 
Agenda, indicated FDIC made changes to 5 regulations as a result of 
decisions under its own discretion and 4 changes as result of mandates. 
FDIC and the other banking agencies also worked with Congressional 
staff regarding legislative action as a result of the EGRPRA reviews. 
For example, the agencies reviewed over 180 legislative initiatives for 
burden relief in 2005. Furthermore, the agencies testified before the 
Senate Banking Committee and House Financial Services Committee on a 
variety of burden reduction measures and upon request, agency 
representatives offered technical assistance in connection with the 
development of legislation to reduce burden. Congress later passed and 
the President signed the Financial Services Regulatory Relief Act of 
2006 on October 13, 2006. 

Table 11: Description of FDIC Retrospective Reviews: 

Review authorities; 
Mandatory reviews: 
* EGRPRA; 
* Section 610; 
Discretionary reviews: 
* Formal petitions; 
* Emergencies or disasters. 

Frequency; 
Mandatory reviews: 
* Every 10 years for both EGRPRA and Section 610 after adoption; 
Discretionary reviews: 
* Depends largely on the nature of the source. 

Purposes; 
Mandatory reviews: 
* Both acts focus on reducing burden; 
Discretionary reviews: 
* Depends largely on the nature of the source. 

General outcomes; 
Mandatory reviews: 
* Reviews generally resulted in changes and recommendations for 
legislative action; 
Discretionary reviews: 
* Reviews generally resulted in changes. 

Agency views on usefulness of reviews; 
Mandatory reviews: 
* Officials reported that these reviews help the agency to accomplish 
its goals to reduce burden while also being valuable because there is a 
need for regulatory decisions to reflect current realties; 
Discretionary reviews: 
* Officials did not differentiate between the usefulness of 
discretionary reviews versus mandatory reviews. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

FDIC Retrospective Review Process: 

FDIC and other financial regulatory agencies that are members of the 
FFIEC decided to use the FFIEC as the coordinating body for the EGRPRA 
review process because the act affected all of the agencies and the 
agencies wanted to: (1) establish a centralized process for selecting, 
conducting, and reporting its reviews; and (2) leverage the expertise 
and resources of all of the member agencies. EGRPRA required the 
agencies to categorize their rules, solicit public comment, and publish 
the comments in the Federal Register. The act also required the 
agencies to report to Congress no later than 30 days after publishing 
the final summarized comments in the Federal Register. The FFIEC 
established additional processes for planning, conducting, and 
reporting of retrospective reviews conducted under EGRPRA outside of 
these specified requirements, such as providing 90 public comment 
periods, holding outreach meetings with regulated entities as well as 
consumer groups across the United States, and establishing a Web site 
dedicated to the EGRPRA reviews. Within all of the processes developed 
by the FFIEC, a high level of management attention was maintained. For 
instance, the Director of the Office of Thrift Savings, who is also a 
member of FDIC's Board of Directors, headed the interagency effort. In 
this capacity, a political appointee was involved in planning, 
conducting, and reporting the reviews. As illustrated by figure 13, the 
process involved interagency coordination and review activities within 
each individual agency, including FDIC. 

Figure 13: FDIC Retrospective Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix X: Small Business Administration Retrospective Reviews: 

During the 2001 through 2006 period covered in our review, the Small 
Business Administration (SBA) has reviewed its existing regulations in 
accordance with Section 610 of the Regulatory Flexibility Act and on it 
own discretion. While the purpose of the Section 610 reviews was to 
reduce burden, the purpose of every discretionary review was to 
increase effectiveness. SBA had written procedures to plan, conduct, 
and report its Section 610 reviews. However, the agency did not have 
written processes to guide planning, conduct, and reporting of 
discretionary reviews. Overall, SBA's discretionary reviews have 
resulted more often in regulatory changes than reviews mandated by 
statute. 

SBA Retrospective Review Activities: 

Officials reported that SBA has conducted discretionary reviews based 
on congressional interest or industry petitions. Specifically, 
officials from the HUBZONE program indicated that their office receives 
attention from congress about the workings of their regulations, 
thereby prompting them to review their existing regulations. In 
addition, SBA's Division of Size Standards completed 27 reviews in 
response to industry petitions and congressional requests. SBA also 
completed 4 Section 610 reviews in 2005. While the purpose of the 
Section 610 reviews was to reduce burden, officials from one division 
in SBA said that they focused many of their retrospective reviews on 
examining the effectiveness of their regulations by evaluating their 
progress on outcomes. However, they stated that because some of their 
regulations are linked to the regulatory activity of other agencies, 
they are not always able to achieve the intended outcome of the 
regulation. Of the reviews conducted by SBA, discretionary reviews 
yielded more changes to existing regulations than mandated reviews. For 
instance, there were no changes made to the 4 Section 610 reviews 
completed but there were 23 final or proposed changes to regulations in 
response to industry petitions. In addition, GAO's examination of SBA's 
December 2006 Unified Agenda entries indicated that 22 rule changes 
were the result of the agency's discretionary authority rather than 
statutory mandates. 

Table 12: Description of SBA Retrospective Reviews: 

Review authorities; 
Mandatory reviews: 
* Section 610; 
Discretionary reviews: 
* Formal petitions; 
* Congressional interest. 

Frequency; 
Mandatory reviews: 
* Every 10 years; 
Discretionary reviews: 
* Depends largely on the nature of the source. 

Purposes; 
Mandatory reviews: 
* Burden reduction; 
Discretionary reviews: 
* Effectiveness. 

General outcomes of reviews; 
Mandatory reviews: 
* Reviews generally resulted in limited or no changes. According to the 
December 2006 Unified Agenda, SBA made changes to two regulations as a 
result of an agency-specific mandate; 
Discretionary reviews: 
* SBA made changes to 23 out of 27 regulations resulting from 
petitions. 

Agency views on usefulness of reviews; 
Mandatory reviews: 
* Officials stated that retrospective reviews are useful because they 
provide an opportunity for the public and the agency to revisit rules 
to ensure continued effectiveness and achievement of goals and 
objectives; 
Discretionary reviews: 
* Officials did not differentiate between the usefulness of 
discretionary reviews versus mandatory reviews. 

Source: GAO analysis of agency data and regulatory statutes. 

[End of table] 

SBA Retrospective Review Process: 

SBA's Section 610 Plan in the May 2006 Unified Agenda described 
procedures for conducting Section 610 reviews. The plan specifies that 
SBA will consider the factors identified in Section 610. The plans also 
specifies that the conduct of the review will be performed by the 
program office of jurisdiction, which entails reviewing any comments 
received from the public, in consultation with the Office of General 
Counsel (OGC) and the Office of Advocacy. The document notes that the 
program office may contact associations that represent affected small 
entities in order to obtain information on impacts of the rules. 
Although Section 610 does not require agencies to report the results of 
the reviews, SBA reported its results in the Unified Agenda. 

Under SBA's standard operating procedures each program office is 
responsible for evaluating the adequacy and sufficiency of existing 
regulations that fall within its assigned responsibilities. However, 
according to the officials, the agency does not have a uniform way to 
plan, conduct, and report these discretionary reviews. In general, the 
agency conducts reviews in an informal manner; therefore, documentation 
does not exist for the procedures or standards used to conduct these 
reviews. However, officials described considering these factors to 
prioritize their reviews: (1) the level of congressional interest in a 
specific review, (2) OGC's input on which rules should be reviewed, and 
(3) the number of petitions and appeals SBA has received regarding a 
particular rule. Reviews are conducted differently in the various 
program offices within SBA. Moreover, the agency described a high 
turnover of employees, which makes it important to document SBA reviews 
and processes. Currently, it does not appear that the agency documents 
its review and processes. 

Figure 14: SBA's General Retrospective Review Process: 

[See PDF for image] 

Source: GAO analysis of agency data. 

[End of figure] 

[End of section] 

Appendix XI: Department of Labor's Employee Benefit Security 
Administration: Example of a Documented Formal Review: 

Employee Benefit Security Administration's (EBSA) retrospective 
regulatory review process was the most documented and detailed formal 
review process included in our review. According to EBSA officials and 
our review of EBSA documentation on its Regulatory Review Program (the 
Program), the agency established its program as a continuing and 
systematic process that allows the agency to periodically reviews its 
regulations to determine whether they need to be modified or updated. 
The Program takes into account technology, industry, economic, 
compliance and other factors that may adversely affect a rule's 
continued usefulness, viewed with respect to either costs or benefits. 
According to program documentation, through the integration of 
prescribed review criteria, regulatory reviews conducted under the 
Program would also help EBSA to satisfy the Section 610 requirement for 
periodic reviews of agency regulations. In addition, the Program 
provides information and data that assists EBSA in conducting 
regulatory reviews of EBSA regulations in accordance with the 
requirements of Executive Order 12866. 

EBSA's regulatory review process is conducted annually by a Regulatory 
Review Committee (RRC) composed of the Counsel for Regulation, Office 
of the Solicitor's Plan Benefits and Security Division (or his 
delegate), and the Directors of the following offices (or their 
respective delegates): Office of Regulations and Interpretations, 
Office of Policy and Research, Office of Enforcement, Office of Health 
Plan Standards and Compliance Assistance, and Office of Exemption 
Determinations. The Director of Regulations and Interpretations (or his 
delegate) chairs the RRC. EBSA's review process consists of three 
formal phases: (1) selection of regulations for further review, (2) 
substantive review of the selected regulations, and (3) reporting 
review results to high-level management and the public. 

During phase 1 of EBSA's review process, the selection and 
recommendation of regulations for substantive review, the RRC conducts 
a preliminary review of EBSA's regulations and assigns a value for the 
need for additional review on a scale of 1 (lowest value) to 5 (highest 
value), based upon several factors (listed below). This ranking is 
based on the RRC's "expert" judgment and knowledge of the regulated 
community. For purposes of this type of review, the RRC may consider a 
series of closely related regulations as one regulation. The RRC 
subsequently compiles the results of the review in chart or outline 
form in order to compare the regulations and recommends to the 
Assistant Secretary at least three regulations as candidates for review 
under the Program. At least one of the three regulations recommended is 
subject to review under Section 610. The RRC presents its 
recommendations for regulations that need further review in a written 
report to the Assistant Secretary, including an explanation of the 
reasons for its recommendations. The factors that the RRC considers 
when preliminarily reviewing the regulations are: 

* whether the regulation is subject to review under the RFA; 

* whether the regulation is subject to review under statutory or 
Executive Order requirements other than the RFA; 

* absolute age of regulation (time elapsed since promulgation); 

* time elapsed since regulation was last amended and nature of 
amendment (major/minor); 

* court opinions adjudicating issues arising under regulation; 

* number of EBSA investigations that have found violations of 
regulation; 

* number of public requests received for interpretation of regulation; 

* type(s) of plans affected by the regulation; 

* number of plans affected; 

* cumulative number of participants and beneficiaries affected by 
regulation; 

* cumulative amount of plan assets affected by regulation; 

* relative difficulty of compliance with regulation for the regulated 
entities (complexity, understandability); 

* potential for cost burden as compared with intended benefits of the 
regulation; 

* extent to which development of new technology or industry practice 
since promulgation may reduce effectiveness of regulation; 

* extent to which legal changes (statutory, regulatory, executive 
order) since promulgation of the regulation may affect its validity; 

* significance of the regulation with respect to EBSA's goals; 

* significance of the regulation with respect to enforcement, 
compliance assistance and voluntary compliance efforts; and: 

* availability of information pertinent to evaluating the regulation. 

EBSA's review program also outlines detailed criteria for the selection 
of regulations to review under Section 610. The program sets a clear 
definition of what constitutes a "significant impact on a substantial 
number of small entities" for its regulations, a step which, as we have 
noted in the past, many agencies have not yet taken.[Footnote 71] 
Specifically, the program sets threshold criteria for what constitutes 
"significant impact" and "substantial number of entities." GAO has 
reported on numerous occasions that the lack of clarity about these 
terms is a barrier to agency conduct of reviews and has resulted in 
fewer reviews being conducted. Therefore, this step in the review 
program appears to be a very useful factor. Under EBSA's approach for 
measuring these thresholds, the rules to be reviewed each year are 
first subjected to quantitative analysis to determine whether they are 
considered to have a significant economic impact on a substantial 
number of small entities. For its initial Section 610 reviews, EBSA has 
adopted a uniform standard of $25 per plan participant to measure the 
discretionary impact of regulations reviewed under Section 610, and 
whether it constitutes a "significant economic impact." EBSA's 
definition of a small entity as an employee pension or welfare plan 
with fewer than 100 participants is grounded in sections 104(a)(2) and 
(3) of the Employee Retirement Income Security Act (ERISA),[Footnote 
72] which permit the Secretary to prescribe simplified annual reports 
for pension and welfare plans with fewer than 100 participants. 
Additional details on these definitions and how they were derived can 
be found in the agency's Regulatory Review Program guidance. 

During phase two of EBSA's review process, the substantive review of 
regulations, the RRC substantively analyzes each regulation selected by 
the Assistant Secretary, and considers the following factors, among 
others, in determining whether the agency should amend, rescind, or 
retain a regulation: 

* whether there appears to be a continued need for the regulation; 

* whether the regulation has been the subject of complaints or comments 
from the public and the nature of those complaints; 

* whether the regulation overlaps, duplicates, or conflicts with other 
federal statutes or rules or with nonpreempted state or local statutes 
or rules; 

* whether the regulation is overly complex and could be simplified 
without impairing its effectiveness; 

* whether the regulation may be based on outdated or superseded 
employment, industrial, or economic practices or assumptions and 
whether participants and/or beneficiaries of employee benefit plans may 
be exposed to harm as a result; 

* whether the regulation may impose significant economic costs on 
regulated entities and whether the benefit(s) or purpose(s) of the 
regulation could be achieved as effectively through an alternative 
regulatory approach that would impose less economic burden on regulated 
industries; 

* whether an alternative regulatory approach that does not increase the 
compliance burden for regulated industries could better serve the 
purpose(s) of the regulation or provide better protection(s) to 
participants and beneficiaries of employee benefit plans; and: 

* whether it would be in the public interest to initiate particular 
actions (e.g., contracting a research study, promulgating a Request for 
Information, conducting a public hearing) within the authority of EBSA 
to develop information or expertise pertinent to the regulation and 
relevant to consideration of the above issues. 

In phase 3 of the review process, reporting review results, the RRC 
provides the Assistant Secretary with a report that includes (1) a 
summary of the RCC's conclusions concerning its evaluation of the 
regulation, based upon the factors described above; and (2) 
recommendations to the Assistant Secretary concerning regulatory 
actions that could be taken regarding the regulation. According to 
Program documentation, the recommendations could include, but are not 
limited to, issuing a Request for Information to supplement EBSA's 
knowledge concerning the effect of the regulation, developing a 
proposal to amend or withdraw a regulation, or retaining the regulation 
in its current form. If the RRC recommends changes to a rule, the 
agency analyzes this options in accordance with Executive Order 12866, 
RFA, or other requirements, and publishes determined changes for public 
comment in the Federal Register, typically as a Notice of Proposed 
Rulemaking. If the RCC recommends issuing a Request for Information to 
supplement EBSA's knowledge concerning the effect of the regulation, 
the agency publishes this Request for Information in the Federal 
Register, before it issues a notice for proposed rulemaking.[Footnote 
73] (For an illustration of this process, see fig 7 in app. IV.) 

[End of section] 

Appendix XII: Comments from the Small Business Administration Office of 
Advocacy: 

SBA: 
Office of Advocacy: 
www.sba.gov/advo: 
Advocacy: the voice of small business in government: 

By Electronic Mail: 

June 21, 2007: 

Mr. Mathew Scire: 
Director, Strategic Issues: 
U.S. Government Accountability Office: 
Mail Stop 2440C: 
Room 2912: 
441 G Street, NW: 
Washington, D.C. 20548-0001: 

RE: GAO's Examination of Retrospective regulatory reviews: 

Dear Mr. Scire: 

The Office of Advocacy (Advocacy) has reviewed the Government 
Accountability Office's draft report, Reexamining Regulations: 
Opportunities Exist to Improve Effectiveness and Transparency of 
Retrospective Reviews-We believe that the draft report is well- 
researched and accurately portrays federal agencies' experiences with 
retrospective rule reviews. 

Our only substantive comment on the draft report relates to the final 
section, "Recommendations for Executive Action," in which GAO 
recommends on page 55 that "the Administrator of the U.S. Small 
Business Administration instruct the Chief Counsel for Advocacy to take 
the following sec en actions." Because the Office of Advocacy is an 
independent office, it would he inappropriate for the SBA Administrator 
to instruct the Chief Counsel to take a particular action. Instead, we 
recommend that GAO simply "recommend that the Administrator of the 
Office of Management and Budget instruct the Administrator of the 
Office of Information and Regulatory Affairs, together with the Chief 
Counsel for Advocacy, to take the following seven actions." 

Advocacy previously communicated to GAO that we are in the process of 
developing supplemental guidance to federal agencies on how to conduct 
retrospective reviews of existing regulations under section 610 of the 
Regulatory Flexibility Act. An initial draft of this guidance, which 
hich is being circulated for informal interagency review and comment, 
is attached. This guidance is intended to supplement our May 2003 
guidance, "How to Comply with the Regulatory Flexibility Act,"[Footnote 
74] and covers several aspects of section 610's requirements, such as 
the timing, scope, and purpose of a periodic rule review, and when 
other retrospective reviews can be considered functionally equivalent 
to a section 610 review. 

Thank you for giving us the opportunity to review the draft report and 
share our views. Please do not hesitate to call me or Keith Holman 
(keith.holman@sBa.gov or (202) 205-6936) if we can be of further 
assistance. 

Sincerely, 

Signed by: 

Shawne C. McGibbon: 
Deputy Chief Counsel for Advocacy: 

Enclosure: 

[This is the initial draft of supplemental guidance on compliance with 
section 610 of the regulatory flexibility act. This guidance will be 
issued by the office of advocacy following interagency review and 
receipt of comments. Input from other agencies will help advocacy in 
developing this guidance.] 

How to Comply with Section 610 of the Regulatory Flexibility Act: A 
Guide for Federal Agencies: 

Introduction: 

Section 610 of the Regulatory Flexibility Act (RFA)[Footnote 75] 
requires federal agencies to review regulations that have a significant 
impact on a substantial number of small entities within 10 years of 
their adoption as final rules. This review is intended to assess the 
impact of existing rules on small entities and to determine whether the 
rules should be continued without change, or should be amended or 
rescinded (consistent with the objectives of applicable statutes). 

In practice, compliance with section 610's retrospective review 
requirement has varied substantially from agency to agency.[Footnote 
76] Although the Office of Advocacy of the U.S. Small Business 
Administration (Advocacy) issued guidance in 2003 on how to comply with 
the RFA, including section 610,[Footnote 77] further clarification is 
warranted. The following provides Advocacy's interpretation of section 
610 of the RFA and answers common questions about conducting 
retrospective reviews of existing regulations. 

The text of 5 U.S.C. § 610, Pub. L 96-354,94 Stat. 1164 (1981): 

§ 610. Periodic review of rules: 

(a) Within one hundred and eighty days after the effective date of this 
chapter, each agency shall publish in the Federal Register a plan for 
the periodic review of the rules issued by the agency which will have a 
significant economic impact upon a substantial number of small 
entities. Such a plan may be amended by the agency at any time by 
publishing the revision in the Federal Register. The purpose of the 
review shall be to determine whether such rules should be continued 
without change, or should be amended or rescinded, consistent with the 
stated objectives of applicable statutes, to minimize any significant 
economic impact of the rules upon a substantial number of such small 
entities. The plan shall provide for the review of all such agency 
rules existing on the effective date of this chapter within ten years 
of the publication of such rules as the final rule. If the head of the 
agency determines that completion of the review of existing rules is 
not feasible by the established date, he shall so certify in a 
statement published in the Federal Register and may extend the 
completion date by one year at a time for a total of not more than five 
years. 

(b) In reviewing rules to minimize any significant economic impact of 
the rule on a substantial number of small entities in a manner 
consistent with the stated objectives of applicable statutes, the 
agency shall consider the following factors -: 

(1) the continued need for the rule; 

(2) the nature of complaints or comments received concerning the rule 
from the public; 

(3) the complexity of the rule; 

(4) the extent to which the rule overlaps, duplicates or conflicts with 
other Federal rules, and, to the extent feasible, with State and local 
governmental rules: and: 

(5) the length of time since the rule has been evaluated or the degree 
to which technology, economic conditions, or other factors have changed 
in the area affected by the rule. 

(c) Each year, each agency shall publish in the Federal Register a list 
of the rules which have a significant economic impact on a substantial 
number of small entities, which are to be reviewed pursuant to this 
section during the succeeding twelve months. The list shall include a 
brief description of each rule and the need for and legal basis of such 
rule and shall invite public comment upon the rule. 

Legislative history relating to section 610. 

Statements made during floor debate on the Regulatory Flexibility Act 
in 1980 suggest that Congress meant section 610 to be a tool for 
agencies to periodically re-examine their rules in light of changing 
circumstances and address increased regulatory burdens vis-a-vis small 
entities.[Footnote 78] Similarly, the section-by-section analysis of 
the periodic review provision of S. 299, which became the RFA, notes 
that the elements of a section 610 review mirror the evaluative factors 
in President Carter's Executive Order 12,044, Improving Government 
Regulations.[Footnote 79] Pursuant to that Executive Order, President 
Carter issued a Memorandum to the Heads of Executive Departments and 
Agencies in 1979, further instructing federal agencies: 

As you review existing regulatory and reporting requirements, take 
particular care to determine where, within statutory limits, it is 
possible to tailor those requirements to fit the size and nature of the 
businesses and organizations subject to them.[Footnote 80]  

This view was also reflected in Advocacy's 1982 guidance explaining the 
RFA, which stated that: 

The RFA requires agencies to review all existing regulations to 
determine whether maximum flexibility is being provided to accommodate 
the unique needs of small businesses and small entities. Because 
society is not static, changing environments and technology may 
necessitate modifications of existing, anachronistic regulations to 
assure that they do not unnecessarily impede the growth and development 
of small entities.[Footnote 81] 

Put simply, the objective of a section 610 review is like the goal of 
most other retrospective rule reviews[Footnote82]: to determine whether 
an existing rule is actually working as it was originally intended and 
whether revisions are needed. Has the problem the rule was designed to 
address been solved? Are regulated entities (particularly small 
entities) able to comply with the rule as expected? Are the costs of 
compliance in line with the agency's initial estimates? Are small 
businesses voicing continuing concerns about their inability to comply? 
The section 610 review is an excellent way to address these questions. 

Scope of the review: What should be included? 

At a minimum, once an agency has determined that an existing rule now 
has a significant economic impact on a substantial number of small 
entities, the agency's section 610 review should address each of the 
five elements listed in section 610(b)(1)-(5): 

* Whether or not there a continuing need for this rule; 

* Whether the public has ever submitted comments or complaints about 
this rule; 

* The degree of complexity of this rule; 

* Whether some other federal or state requirement accomplishes the same 
regulatory objective as this rule; and: 

* The length of time since the agency has reviewed this rule, and/or 
the extent to which circumstances have changed which may affect 
regulated entities. 

Particular attention should be paid to changes in technology, economic 
circumstances, competitive forces, and the cumulative burden faced by 
regulated entities. Has the impact of the rule on small entities 
remained the same? Even if an agency was originally able to properly 
certify under section 605 of the RFA that a rule would not have a 
significant economic impact on a substantial number of small 
entities,[Footnote 83] changed conditions may mean that the rule now 
has a significant impact and therefore should be reviewed under section 
610. If there is evidence (such as new cost or burden data) that a rule 
is now having a significant economic impact on a substantial number of 
small entities, including small communities or small non-profit 
associations, the RFA requires the agency to conduct the section 610 
review. 

Section 610(b) requires an agency to evaluate and minimize "any 
significant economic impact of a rule on a substantial number of small 
entities." To accomplish this economic analysis, the analysis best 
suited for a section 610 review is similar to the analysis required in 
an Initial Regulatory Flexibility Analysis (IRFA). [Footnote 
84]Agencies certainly have the discretion to place significant weight 
on other relevant factors besides the types of economic data required 
by an IRFA. These other factors include an agency's experience in 
implementing the rule, as well as the views expressed over time by the 
public, regulated entities, and Congress. With the benefit of actual 
experience with a rule, the agency and other interested parties are in 
a good position to evaluate potential improvements to the rule. 
Particular attention should be paid to such factors as unintended 
market effects and market distortions, unusually high firm mortality 
rates in specific industry sub-sectors, and widespread noncompliance 
with reporting and other "paperwork" requirements. Thus, the focus of 
the review should go beyond merely ensuring that regulatory 
requirements are expressed in plain language and that paperwork can be 
filed electronically. The analysis should be aimed at understanding and 
reducing unnecessary burdens that impact small entities. 

The section 610 analysis must be based on facts. To the extent that an 
agency relies on specific data to reach a conclusion about the 
continuing efficacy of a rule, the agency must be able to provide that 
data. Similarly, the agency must be prepared to explain any assumptions 
that it makes in the analysis. The review process must not be a "black 
box" that stakeholders cannot see or understand. 

Timing of the review: When does the agency have to start and finish? 

The language of section 610 specifies that the review should take place 
no more than 10 years after a rule is promulgated, but an agency can 
choose to review a rule earlier, before 10 years have clapsed. While it 
is recommended that agencies gain some experience with a rule before 
undertaking a retrospective review, the review may take place at any 
time during the 10-year period. If an agency substantially revises a 
rule after its initial promulgation, the 10 year trigger for the 
section 610 review may properly be delayed to correspond to the 
revision date. However, minor revisions to a rule would be inadequate 
as a basis for delaying the review of the rule. If there is confusion 
regarding whether a revision is minor or major, an agency should seek 
input form Advocacy. The RFA allows an agency to delay the initiation 
of a section 610 review by one-year increments, up to a total of five 
years. 

Section 610 does not specifically set a limit on the amount of time for 
a rule review. Some agencies have reported that they spend more than a 
year on each section 610 review. It is within an agency's discretion to 
determine how much time it needs to spend on retrospective rule 
reviews. Advocacy does not anticipate, however, that a section 610 
review should require substantially more time to complete than an 
Initial Regulatory Flexibility Analysis. As long as an agency addresses 
the required elements of a section 610 review and provides an adequate 
factual basis for its conclusions, Advocacy believes that many section 
610 analyses are capable of completion within a shorter timeframe. 

Agencies may also wish to take advantage of the opportunity afforded in 
section 605(c) of the RFA to consider a series of "closely related 
rules" as one rule for retrospective review purposes. An agency can 
accomplish a programmatic or comprehensive section 610 review of 
closely related rules, satisfying the requirements of the RFA while 
minimizing the resources needed for the review. 

How should agencies communicate with interested entities about section 
610 reviews they are conducting? 

Section 610(c) of the RFA requires agencies to publish in the Federal 
Register a list of the rules which they plan to review in the upcoming 
year. This listing requirement is intended to give small entities early 
notice of the section 610 reviews so that they will be ready and able 
to provide the agency with comments about the rule under review. As a 
practical matter, however, agencies often give stakeholders no other 
information about the ongoing status of a section 610 review, what 
factors an agency is considering in conducting the review, how comments 
can be submitted to the agency, or the factual basis on which the 
agency made its section 610 review findings. 

At a minimum, agencies should develop a mechanism for communicating 
with interested entities about the section 610 reviews that they are 
conducting, as well as those they have completed. Experience has shown 
that the Federal Register is not necessarily the most effective tool 
for conveying up-to-date information about the status of ongoing 
retrospective reviews. This information may be better communicated via 
an agency website or other electronic media, and should inform 
interested parties of their ability to submit comments and the agency's 
commitment to consider those comments. Several agencies already utilize 
web-based communications as an outreach tool during section 610 
reviews; [Footnote 85]every agency should achieve at least this level 
of accessibility. 

Insights about an existing regulation received from regulated entities 
and other interested parties who live with the rule should be a key 
component of a retrospective rule review. By making the review process 
transparent and accessible, agencies are more likely to identify 
improvements that will benefit all parties at the conclusion of the 
review. 

Can other agency retrospective rule reviews satisfy the section 610 
requirement? 

Agencies that undertake retrospective rule reviews for other reasons 
may satisfy the section 610 requirement, as long as the rule reviews is 
functionally equivalent. For example, agencies that evaluated a 
regulation pursuant to the Office of Management and Budget's 2002 
publicly-nominated rule reform process[Footnote 86] or OMB's 
manufacturing rule reform process[Footnote 87] would qualify as section 
610 reviews. Similarly, agencies that undertook retrospective reviews 
of their regulatory programs because of complaints from regulated 
entities would likely qualify as section 610 reviews - as long as the 
review includes the minimum elements required by section 610 and the 
agency adequately communicates with stakeholders. 

Examples. In Advocacy's view, what recent retrospective rule reviews 
have been successful? 

* Federal Railway Administration's Section 610 Review of Railroad 
Workplace Safety - On December 1, 2003, the Department of 
Transportation's Federal Railroad Administration completed a section 
610 review of its railroad workplace safety regulations. After 
determining that the workplace safety regulations had a significant 
economic impact on a substantial number of small entities, the FRA 
examined the rules in light of section 610's review elements. The FRA 
provided a good description of its analysis of the workplace safety 
regulations under each review element and the agency's conclusions. See 
http://www.fra.dot.gov/downloads/safety/railroad_workplace_safety.pdf: 

* EPA's RCRA Burden Reduction Initiative - As a result of public 
nominations for reforms to the Environmental Protection Agency's waste 
management program under the Resource Conservation and Recovery Act 
(RCRA), EPA evaluated the program and identified duplicative 
requirements, such as forcing filers to submit reports to multiple 
locations when one location is adequate. By reducing or eliminating 
these procedures after public notice and comment, EPA enabled regulated 
entities to save up to $3 million per year while preserving the 
protections of the RCRA program. The retrospective review was 
successful because it involved a detailed review of the program's 
requirements and their costs, based on years of practical experience. 
The agency considered technical changes such as computerization that 
have made some of the older paperwork requirements redundant, and found 
ways to modernize the program to reflect current realities. See 71 Fed. 
Reg. 16,862 (April 4, 2006). 

* OSHA Excavations Standard - In March 2007, the Occupational Safety 
and Health Administration (OSHA) completed a section 610 review of its 
rules governing excavations and trenches. These standards had been in 
place since 1989, and were designed to ensure that trenches do not 
collapse on workers and that excavated material does not fall back into 
a trench and bury workers. In the review, OSHA did a good job of 
seeking public input on how and whether the rule should be changed. 
While the agency ultimately decided that no regulatory changes to the 
standard were warranted, it did determine that additional guidance and 
worker training would help continue the downward trend of fewer deaths 
and injuries from trench and excavation work. OSHA had concluded that 
its current Excavations standard has reduced deaths from approximately 
90 per year to about 70 per year. See 72 Fed. Reg. 14,727 (March 29, 
2007). 

* FCC Section 610 Review of 1993-1995 Rules - In May 2005, the Federal 
Communications Commission undertook a section 610 review of rules the 
Commission adopted in 1993, 1994, and 1995 which have, or might have, a 
significant economic impact on a substantial number of small entities. 
The FCC solicited public comment on the rules under review, explained 
the criteria it was using to review the rules, and gave instructions 
where to file comments. This approach was transparent because the 
agency allowed adequate time for comments (three months) and gave 
interested parties sufficient information to prepare useful comments. 
See 70 Fed. Reg. 33,416 (June 8, 2005). 

The Office of Advocacy is ready to assist agencies that are planning a 
retrospective review of their regulations to ensure that the review 
fully meets the requirements of section 610. Discussions with the 
Office of Advocacy are confidential interagency communications, and the 
Advocacy staff is ready to assist you. For more information about this 
guidance, or for other questions about compliance with section 610, 
please contact Advocacy at (202) 205-6533. 

[End of section] 

Appendix XIII: GAO Contact and Acknowledgments: 

[End of section] 

GAO Contact: 

Mathew J. Scire, Director, Strategic Issues (202) 512-6806, 
sciremj@gao.gov. 

Acknowledgments: 

Tim Bober, Assistant Director, and Latesha Love, Analyst-in-Charge, 
managed this assignment. Other staff who made key contributions to this 
assignment were Matt Barranca, Jason Dorn, Tim Guinane, Andrea Levine, 
Shawn Mongin, Bintou Njie, Joe Santiago, Stephanie Shipman, Michael 
Volpe, and Greg Wilmoth. 

FOOTNOTES 

[1] Pub. L. No. 96-354, 94 Stat. 1164, 1169 (Sept. 16, 1980) codified 
at 5 U.S.C. § 610. These reviews are referred to as Section 610 
reviews. 

[2] There is no one standard term or definition for the variety of 
activities that might be considered retrospective regulatory reviews. 
In different contexts, these have been referred to as look-backs, ex 
post (postregulation) studies, retrospective studies, validation 
studies, and sometimes just as "reviews." 

[3] See, for example, GAO, Regulatory Reform: Prior Reviews of Federal 
Regulatory Process Initiatives Reveal Opportunities for Improvements, 
GAO-05-939T (Washington, D.C.: July 27, 2005), and Environmental 
Protection: Assessing the Impacts of EPA's Regulations Through 
Retrospective Studies, GAO/RCED-99-250 (Washington, D.C.: Sept. 14, 
1999). 

[4] For purposes of this report, we define mandatory reviews as 
retrospective reviews that agencies conducted in response to 
requirements in statutes, executive orders, or executive branch 
directives. We define discretionary reviews as reviews that agencies 
undertook based upon their own initiative. 

[5] We report findings from the following sub-agencies in the body of 
our report: USDA's Agricultural Marketing Service (AMS), Animal and 
Plant Health Inspection Service (APHIS), and Food Safety and Inspection 
Service (FSIS); Department of Justice's Bureau of Alcohol, Tobacco, 
Firearms, and Explosives (ATF) and Drug Enforcement Administration 
(DEA); Department of Labor's Employee Benefits Security Administration 
(EBSA), Occupational Safety and Health Administration (OSHA), Mine 
Safety and Health Administration, and Employment and Training 
Administration (ETA); and the Department of Transportation's Federal 
Aviation Administration (FAA) and National Highway Traffic Safety 
Administration (NHTSA). Additional information on these and other 
subagencies included in our review are reported in appendices II-XI. 

[6] The term "independent regulatory agencies" refers to the boards and 
commissions identified as such in the Paperwork Reduction Act (44 
U.S.C. § 3502(5)), including CPSC, FCC, and FDIC. "Independent 
agencies" refers to agencies that answer directly to the President but 
are not part of Cabinet departments. Some regulatory statutory 
requirements apply to Cabinet departments and independent agencies, 
while others apply to those agencies as well as the independent 
regulatory agencies. Unless otherwise indicated, executive orders in 
this report apply to agencies other than those considered to be 
independent regulatory agencies, although the independent regulatory 
agencies may be asked by the Office of Management and Budget (OMB) to 
voluntarily comply. 

[7] The Congressional Review Act (CRA) (5 U.S.C. § 801) requires 
agencies to file final rules with both Congress and GAO before the 
rules can become effective. To compile information on all the rules 
submitted to us under CRA, GAO established a database and created a 
standardized submission form to allow more consistent information 
collection. The Federal Rules Database is publicly available at 
http://www.gao.gov under Legal Products. 

[8] In prior work, we identified overall strengths or benefits 
associated with regulatory process initiatives, including: (1) 
increasing expectations regarding the analytical support for proposed 
rules, (2) encouraging and facilitating greater public participation in 
rulemaking, and (3) improving the transparency of the rulemaking 
process. Because these strengths or benefits are also relevant and 
useful for assessing agency retrospective review initiatives, we 
considered them in our selection of evaluation criteria. Other 
practices that could improve the effectiveness and transparency of 
reviews may exist and could be considered when developing retrospective 
review processes. However, we believe that the three practices that we 
assessed are among the most important. 

[9] Section 610 requires agencies to consider: (1) the continued need 
for the rule; (2) the nature of complaints or comments received 
concerning the rule from the public; (3) the complexity of the rule; 
(4) the extent to which the rule overlaps, duplicates, or conflicts 
with other federal rules, and, to the extent feasible, with state and 
local government rules; and (5) the length of time since the rule has 
been evaluated or the degree to which technology, economic conditions, 
or other factors have changed in the area affected by the rule. 

[10] Outside of DOT, we identified 88 mandatory reviews completed by 
our selected agencies. DOT has an agency policy that requires its 
agencies to conduct reviews of all its regulations in the Code of 
Federal Regulations (CFR) to ensure the department's compliance with 
Section 610 and other mandatory requirements. DOT completed 406 reviews 
under this policy between 2001 and 2006. 

[11] GPRA (Pub. L. No. 103-62, 107 Stat. 285 (Aug. 3, 1993)) required 
federal agencies to develop strategic plans with long-term, outcome- 
oriented goals and objectives, annual goals linked to achieving the 
long-term goals, and annual reports on the results achieved. See GAO, 
Results-Oriented Government: GPRA Has Established a Solid Foundation 
for Achieving Greater Results, GAO-04-38 (Washington, D.C.: Mar. 10, 
2004). 

[12] PART is a standard series of questions meant to serve as a 
diagnostic tool, drawing on available program performance and 
evaluation information to form conclusions about program benefits and 
recommend adjustments that may improve results. See GAO, Program 
Evaluation: OMB's PART Reviews Increased Agencies' Attention to 
Improving Evidence of Program Results, GAO-06-67 (Washington, D.C.: 
Oct. 28, 2005). 

[13] See, for example, GAO, United States Government Accountability 
Office: Supporting the Congress through Oversight, Insight, and 
Foresight, GAO-07-644T (Washington, D.C.: Mar. 21, 2007), and 21st 
Century Challenges: Reexamining the Base of the Federal Government, GAO-
05-325SP (Washington, D.C.: February 2005). 

[14] See GAO, Forces That Shape America's Future: Themes from GAO's 
Strategic Plan 2007-2012, GAO-07-467SP (Washington, D.C.: March 2007). 

[15] The underlying statutes also play an important role and contribute 
to differences in agencies' regulatory activities. See, for example, 
GAO, Chemical Risk Assessment: Selected Federal Agencies' Procedures, 
Assumptions, and Policies, GAO-01-810 (Washington, D.C.: Aug. 6, 2001). 

[16] See, for example, 42 U.S.C. §§ 7411(b)(1)(B), 7412(d)(6). 

[17] The Executive Order 12866 of September 30, 1993, as amended by 
Executive Order 13258 of February 26, 2002, and Executive Order 13422 
of January 18, 2007, is still in effect. 

[18] OIRA has primary responsibility for implementing Executive Order 
12866, reviewing and coordinating regulatory activities of Cabinet 
departments and independent agencies, and issuing guidance on the 
regulatory review process. SBA's Office of Advocacy monitors and 
provides guidance regarding the compliance of all agencies, including 
independent regulatory agencies, with the RFA. 

[19] See U.S. Small Business Administration, Office of Advocacy, A 
Guide for Government Agencies: How to Comply with the Regulatory 
Flexibility Act (Washington, D.C.: May 2003). 

[20] Other studies about retrospective reviews of federal regulations 
have also recognized the potential usefulness of such evaluations and 
identified problems associated with effectively reviewing existing 
regulations. See, for example: Neil R. Eisner and Judith S. Kaleta, 
Federal Agency Reviews of Existing Regulations, 48 Admin. L. Rev. 
(1996) pp. 139-174; Office of Management and Budget, Validating 
Regulatory Analysis: 2005 Report to Congress on the Costs and Benefits 
of Federal Regulations and Unfunded Mandates on State, Local, and 
Tribal Entities (Washington, D.C.: 2005); and Michael R. See, Willful 
Blindness: Federal Agencies' Failure to Comply with the Regulatory 
Flexibility Act's Periodic Review Requirement--and Current Proposals to 
Invigorative the Act, 33 Fordham Urb. L.J. (2006) pp. 1199-1255. 

[21] See GAO, Environmental Protection: Assessing the Impacts of EPA's 
Regulations Through Retrospective Studies, GAO/RCED-99-250 (Washington, 
D.C.: Sept. 14, 1999). 

[22] See, for example, GAO, Regulatory Flexibility Act: Congress Should 
Revisit and Clarify Elements of the Act to Improve Its Effectiveness, 
GAO-06-998T (Washington, D.C.: July 20, 2006). 

[23] GAO, Managing for Results: Regulatory Agencies Identified 
Significant Barriers to Focusing on Results, GAO/GGD-97-83 (Washington, 
D.C.: June 24, 1997). 

[24] See GAO-05-939T. 

[25] See GAO, Economic Performance: Highlights of a Workshop on 
Economic Performance Measures, GAO-05-796SP (Washington, D.C.: July 18, 
2005); Unfunded Mandates: Views Vary About Reform Act's Strengths, 
Weaknesses, and Options for Improvement, GAO-05-454 (Washington, D.C.: 
Mar. 31, 2005); GAO-05-939T ; and GAO/RCED-99-250. 

[26] Section 402 amended the Telecommunications Act to add a new 
Section 11. These reviews are referred to as Section 11 reviews. See 47 
U.S.C.§161. 

[27] A significant regulatory action under the executive order is 
defined as any regulatory action that, for example, may have an annual 
effect on the economy of $100 million or more, or create a serious 
inconsistency or otherwise interfere with an action taken or planned by 
another agency, among other things. 

[28] Each year of DOT's plan coincides with the fall-to-fall schedule 
for publication of the department's regulatory agenda. Year 1 of this 
process began in the fall of 1998. 

[29] 12 U.S.C. § 3311. 

[30] For example, during the 2001 through 2006 time period covered by 
our report, USDA published 1,132 final rules, DOJ 199, DOL 151, DOT 
6,415, CPSC 12, EPA 2,948, FCC 1,044, FDIC 40, and SBA 47. 

[31] OMB requested input on guidance and regulations, respectively. For 
this review we only count agencies' reforms of their existing 
regulations. 

[32] Appendix IV provides an illustration of the variation in review 
processes even within agencies, as demonstrated by agencies within DOL. 

[33] To illustrate how an agency can apply these characteristics in a 
detailed formal program that defines key terms and identifies important 
evaluation factors that agencies can consider in each phase, we provide 
a detailed example of EBSA's regulatory review program in app. XI. 
EBSA's program was the most comprehensive documented review process 
that we examined within the scope of our review. 

[34] Section 610 requires agencies to assess: (1) the continued need 
for the rule; (2) the nature of complaints or comments received 
concerning the rule from the public; (3) the complexity of the rule; 
(4) the extent to which the rule overlaps, duplicates, or conflicts 
with other federal rules, and, to the extent feasible, with state and 
local government rules; and (5) the length of time since the rule has 
been evaluated or the degree to which technology, economic conditions, 
or other factors have changed in the area affected by the rule. 

[35] Some statutory mandates, such as Section 610, require agencies to 
provide notice of retrospective reviews and solicit public comments on 
the relevant regulations. 

[36] A listserve is a program used to manage e-mails for members of a 
discussion group. For example, a listerve program can be used as an 
automatic mailing list server. When e-mail is addressed to a listserve 
mailing list, it is automatically broadcast to everyone on the list. 
However, listserves can take multiple forms. For example, DOT pointed 
out that within their Docket Management System, the listserve users 
sign up to be notified when the agency places something in a docket of 
interest. 

[37] When a press is equipped with presence sensing device initiation 
(PSDI), the press cycle will not initiate until the PSDI system senses 
that the danger zone is clear. OSHA reviewed its PSDI standard for 
mechanical power presses and found that companies were not using PSDI 
systems because the standard requires an OSHA-approved third party to 
validate the systems at installation and annually thereafter, but no 
party has pursued OSHA approval since the standard was implemented in 
1988. OSHA is currently attempting to revise the regulation to rely 
upon a technology standard that industries could utilize and that would 
provide for additional improvements in safety and productivity. 

[38] 47 U.S.C. § 161(a)(2). 

[39] Pub. L. No. 109-231, 120 Stat. 493 (June 15, 2006). 

[40] See Eisner and Kaleta at pages 9 and 14. 

[41] For a discussion of the purposes, standards, and approval process 
established by PRA see GAO, Paperwork Reduction Act: New Approach May 
Be Needed to Reduce Government Burden on Public, GAO-05-424 
(Washington, D.C.: May 20, 2005). 

[42] Economic incentive regulations are rules that encourage behavior 
through price signals rather than through explicit instructions on how 
to meet standards, such as pollution control levels. An example of such 
a regulation would be an EPA rule that uses tradable permits and 
pollution charges administered by the agency to achieve pollution 
control goals. 

[43] GAO, Regulatory Flexibility Act: Agencies' Interpretations of 
Review Requirements Vary, GAO/GGD-99-55 (Washington, D.C.: Apr. 2, 
1999). 

[44] This review was conducted between 1980 and 1996--outside of the 
time period of our report--so we did not include it in our summary data 
on the number of reviews conducted by MSHA. However, agency officials 
identified it as a relevant example of a "lesson learned" about 
conducting retrospective reviews. 

[45] GAO, Regulatory Burden: Some Agencies' Claims Regarding Lack of 
Rulemaking Discretion Have Merit, GAO/GGD-99-20 (Washington, D.C.: Jan. 
8, 1999). 

[46] MSHA officials pointed out that, when modifying regulations, the 
agency is prohibited from reducing the protection afforded miners by an 
existing mandatory health or safety standard. See Section 30 U.S.C. § 
811(a)(9). 

[47] The term "independent regulatory agencies" refers to the boards 
and commissions identified as such in the Paperwork Reduction Act (44 
U.S.C. § 3502(5)), including CPSC, FCC, and FDIC. "Independent 
agencies" refers to agencies that answer directly to the President but 
are not part of Cabinet departments. 

[48] The Congressional Review Act (CRA) requires agencies to file final 
rules with both Congress and GAO before the rules can take effect. To 
compile information on all the rules submitted to us under CRA, GAO 
established a database and created a standardized submission form to 
allow more consistent information collection. We determined that the 
data in the database were sufficiently reliable for the purpose for 
which it was used. The Federal Rules Database is publicly available at 
www.gao.gov under Legal Products. 

[49] We interviewed 15 individuals from 11 separate groups that 
represented the various sectors listed. Specifically, we interviewed 3 
groups from academia, 3 groups that represent public advocacy, 3 groups 
that represent business advocacy, and 2 groups that represent state and 
local government. 

[50] We held a separate exit meeting with the Department of Justice and 
discussed the same information presented at the joint agency exit 
meeting. 

[51] The DOJ agencies included in our review were the Bureau of 
Alcohol, Tobacco, Firearms, and Explosives (ATF); Civil Rights Division 
(CRT); Federal Bureau of Investigations (FBI); Drug Enforcement 
Administration (DEA); Bureau of Prisons (BOP); Executive Office for 
Immigration Review; and the Office of Justice Programs (OJP). 

[52] 70 Fed. Reg. 16,902 (Apr. 1, 2005). 

[53] 68 Fed. Reg. 3744 (Jan. 24, 2003). 

[54] 62 Fed. Reg. 1386 (Jan. 10, 1997). 

[55] 68 Fed. Reg. 4406 (Jan. 29, 2003). 

[56] 45 Fed. Reg. 72,995 (Nov. 2, 1980). 

[57] 68 Fed. Reg. 51,334 (Aug. 26, 2003). 

[58] 69 Fed. Reg. 58,768 (Sept. 30, 2004). 

[59] 68 Fed. Reg. 4406 (Jan. 29, 2003). 

[60] OSHA completed 24 of the 42 DOL reviews conducted in response to 
the OMB initiatives. 

[61] The DOT includes the Office of the Secretary (OST) and the 
following operating administrations: Federal Aviation Administration 
(FAA); Federal Highway Administration (FHWA); Federal Motor Carrier 
Safety Administration (FMCSA); Federal Railroad Administration (FRA); 
Federal Transit Administration (FTA); Maritime Administration (MARAD); 
National Highway Traffic Safety Administration (NHTSA); Pipeline and 
Hazardous Materials Safety Administration (PHMSA); Research and 
Innovative Technology Administration (RITA); and St. Lawrence Seaway 
Development Corporation (SLSDC). The components of DOT have changed 
since 2001. For example, with the establishment of the Department of 
Homeland Security (DHS) in 2003, the U.S. Coast Guard and the 
Transportation Security Administration were transferred from DOT to 
DHS. 

[62] 63 Fed. Reg. 31,883 (June 10, 1998). The Executive Memorandum 
directed agencies to consider rewriting existing rules in plain 
language when the opportunity and resources permit. 

[63] 44 Fed. Reg. 11,033 (Feb. 26, 1979). 

[64] CPSC officials reported that their review process typically 
involves a modest amount of staff resources per year (less than 10 
staff months), and involves work by staff who specialize in a variety 
of consumer product hazard categories. They pointed out that the costs 
of their review program are built into CPSC's operating plan and budget 
justification. Each portion of the plan and activities must be 
accounted for in this budget plan. If resources or agency priorities do 
not permit commitment of resources to revise regulations that CPSC 
staff identified for revision, CPSC officials may forego initiating a 
Phase 2 update until a later date. 

[65] The Office of Prevention, Pesticides and Toxic Substances and the 
Office of Solid Waste and Emergency Response reported that their 
reviews in response to the OMB Manufacturing Regulatory Reform 
initiative all resulted in the changes. 

[66] OPPTS reported two changes as a result of its Section 610 reviews; 
however, one change may have been conducted prior to years included 
within the scope of our review. 

[67] The biennial review is conducted under Section 11 of the 
Communications Act, as amended by Section 402 of the Telecommunications 
Act of 1996, which created the requirement for the biennial review. 

[68] Section 202(h) of the Communications Act required a biennial 
review of broadcast ownership rules. Congress, through the Consolidated 
Appropriations Act of 2004, changed this review requirement to every 4 
years (it is now referred to as the quadrennial review). 

[69] See FCC 06-129, In the Matter of Amendment of Part 13 and 80 of 
the Commission's Rules, PR Docket No. 92-257, Memorandum Opinion and 
Order, 2006. 

[70] The rule changes from the 2002 quadrennial review never went into 
effect. In 2004, the U.S. Court of Appeals remanded back to FCC for 
further review its rules for cross-media ownership, local television 
multiple ownership, and local radio multiple ownership (Prometheus 
Radio vs. F.C.C., 373 F.3d 372 (3rd Cir. 2004)). Additionally, Congress 
overturned FCC's national television ownership rule which would have 
allowed a broadcast network to own and operate local broadcast stations 
reaching 45 percent of U.S. television households. Through the 
Consolidated Appropriations Act of 2004, Congress set the national 
television ownership limit at 39 percent (Pub. L. No. 108-199, 118 
Stat. 3, 100 (Jan. 23, 2004)). 

[71] EBSA's document "Regulatory Review Program" notes that Congress 
did not prescribe specific standards for defining "substantial numbers" 
or "significance of impact" to be used in threshold testing for Section 
610 purposes. However, the document outlines, in great detail, guidance 
for determining whether regulations meet this threshold. The document 
also acknowledges that, while the agency considers its threshold 
standards appropriate for purposes of RFA Section 610 reviews, the 
agency is also considers it advisable to maintain some flexibility as 
the standard is applied going forward, should application of the 
threshold test produce a result that appears inconsistent with the 
intent of RFA. 

[72] The Employee Retirement Income Security Act of 1974 (ERISA) is a 
federal law that sets minimum standards for most voluntarily 
established pension and health plans in private industry to provide 
protection for individuals in these plans. See, e.g., 29 U.S.C. § 1001. 

[73] In the event that the RRC is unable to complete the review of a 
regulation before the end of the year following the year in which it is 
selected for review, it reports this to the Assistant Secretary in its 
next report and provides the reason(s) for the delay and an estimate of 
when the RRC expects to complete the review. 

[74] Office of Advocacy, A Guide for Government Agencies, How to Comply 
with the Regulatory Flexibility Act (May 2003) available at [hyperlink, 
http://www.sba.gov/advolaws/rfaguide.pdf].

[75] Pub. L. 96-354,94 Stat. 1164 (1981) codified as 5 U.S.C. § 610 
(2000). 

[76] See, for example, Michael R. See, Willful Blindness: Federal 
Agencies' Failure to Comply with the Regulatory Flexibility Act's 
Periodic Review Requirement-and Current Proposals to Invigorate the 
Act, 33 Fordham Urb. L.J. 1199-1255 (2006). 

[77] Office of Advocacy, How to Comply with the Regulatory Flexibility 
Act: A Guide for Government Agencies (May 2003) available at 
http://www.sba.gov/advo/laws/rfaguide.pdf]. 

[78] House Debate on the Regulatory Flexibility Act, 142 Cong. Rec. 
H24,575, H24,583-585 (daily ed. Sept. 8, 1980) ("At least once every 10 
years, agencies must assess regulations currently on the books, with a 
view toward modification of those which unduly impact on small 
entities." (Statement of Rep. McDade)) ("[Agencies must review all 
regulations currently on the books and determine the continued need for 
any rules which have a substantial impact on small business." 
(Statement of Rep. Ireland)). 

[79] Exec. Order No. 12,044,43 Fed. Reg. 12,661 (March 24, 1978). 

[80] President Jimmy Carter, Memorandum to the Heads of Executive 
Departments and Agencies, November 16, 1979. 

[81] Office of Advocacy, The Regulatory Flexibility Act (October 1982). 

[82] Common retrospective regulatory reviews include post-hoc 
validation studies, reviews initiated pursuant to petitions for 
rulemaking or reconsideration, paperwork burden reviews, and reviews 
conducted pursuant to agency policy. 

[83] See 5 U.S.C. § 605(b). 

[84] See 5 U.S.C. § 603. In the context of a section 610 review, the 
elements of an IRFA analysis that should be present include: a 
discussion of the number and types of small entities affected by the 
rule, a description of the compliance requirements of the rule and an 
estimate of their costs, identification of any duplicative or 
overlapping requirements, and a description of possible alternative 
regulatory approaches. 

[85] See, e.g., www.osha.gov, www.epa.gov, and www.dot.gov and search 
for "RFA section 610." 

[86] See Table 9, "New Reforms Planned or Underway - Regulations" and 
Table 10, "New Reforms Planned or Underway - Guidance Documents" in 
Informing Regulatory Decisions: 2003 Report to Congress on the Costs 
and Benefits of Federal Regulations and Unfunded Mandates on State, 
Local, and Tribal Entities (September 2003) at 26-34; available at 
http://www.whitehouse.gov/omb/inforeg/2003_cost_ben_final_rept.pdf. 

[87] See Regulatory Reform of the U.S. Manufacturing Sector (2005). 

GAO's Mission: 

The Government Accountability Office, the audit, evaluation and 
investigative arm of Congress, exists to support Congress in meeting 
its constitutional responsibilities and to help improve the performance 
and accountability of the federal government for the American people. 
GAO examines the use of public funds; evaluates federal programs and 
policies; and provides analyses, recommendations, and other assistance 
to help Congress make informed oversight, policy, and funding 
decisions. GAO's commitment to good government is reflected in its core 
values of accountability, integrity, and reliability. 

Obtaining Copies of GAO Reports and Testimony: 

The fastest and easiest way to obtain copies of GAO documents at no 
cost is through GAO's Web site (www.gao.gov). Each weekday, GAO posts 
newly released reports, testimony, and correspondence on its Web site. 
To have GAO e-mail you a list of newly posted products every afternoon, 
go to www.gao.gov and select "Subscribe to Updates." 

Order by Mail or Phone: 

The first copy of each printed report is free. Additional copies are $2 
each. A check or money order should be made out to the Superintendent 
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or 
more copies mailed to a single address are discounted 25 percent. 
Orders should be sent to: 

U.S. Government Accountability Office 441 G Street NW, Room LM 
Washington, D.C. 20548: 

To order by Phone: Voice: (202) 512-6000 TDD: (202) 512-2537 Fax: (202) 
512-6061: 

To Report Fraud, Waste, and Abuse in Federal Programs: 

Contact: 

Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov 
Automated answering system: (800) 424-5454 or (202) 512-7470: 

Congressional Relations: 

Gloria Jarmon, Managing Director, JarmonG@gao.gov (202) 512-4400 U.S. 
Government Accountability Office, 441 G Street NW, Room 7125 
Washington, D.C. 20548: 

Public Affairs: 

Paul Anderson, Managing Director, AndersonP1@gao.gov (202) 512-4800 
U.S. Government Accountability Office, 441 G Street NW, Room 7149 
Washington, D.C. 20548: