This is the accessible text file for GAO report number GAO-05-52 
entitled 'Environmental Indicators: Better Coordination Is Needed to 
Develop Environmental Indicator Sets That Inform Decisions' which was 
released on November 17, 2004.

This text file was formatted by the U.S. Government Accountability 
Office (GAO) to be accessible to users with visual impairments, as part 
of a longer term project to improve GAO products' accessibility. Every 
attempt has been made to maintain the structural and data integrity of 
the original printed product. Accessibility features, such as text 
descriptions of tables, consecutively numbered footnotes placed at the 
end of the file, and the text of agency comment letters, are provided 
but may not exactly duplicate the presentation or format of the printed 
version. The portable document format (PDF) file is an exact electronic 
replica of the printed version. We welcome your feedback. Please E-mail 
your comments regarding the contents or accessibility features of this 
document to Webmaster@gao.gov.

This is a work of the U.S. government and is not subject to copyright 
protection in the United States. It may be reproduced and distributed 
in its entirety without further permission from GAO. Because this work 
may contain copyrighted images or other material, permission from the 
copyright holder may be necessary if you wish to reproduce this 
material separately.

Report to Congressional Requesters:

November 2004:

ENVIRONMENTAL INDICATORS:

Better Coordination Is Needed to Develop Environmental Indicator Sets 
That Inform Decisions:

[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-05-52]: 

GAO Highlights:

Highlights of GAO-05-52, a report to congressional requesters:

Why GAO Did This Study:

Environmental indicator sets assemble quantitative measures of 
conditions and trends (known as indicators) to assess the state of the 
environment and natural resources and to gauge progress toward specific 
goals. Such sets are now being developed to bridge the gap between 
needed and available information and to prioritize further data 
collection. The widespread development and use of environmental 
indicator sets has led federal and nonfederal entities to consider the 
benefits such sets provide when measuring performance and improving 
oversight of environmental programs. In this context, GAO was asked to 
identify (1) the purposes for which federal and nonfederal 
organizations are developing and using environmental indicator sets, 
and how they are being used; and (2) the major challenges facing the 
development and use of environmental indicator sets. 

What GAO Found:

GAO identified the purposes for developing environmental indicator sets 
and major challenges facing their development and use to inform 
decisions by interviewing key experts, surveying developers and users, 
and studying eight major indicator sets. GAO found that federal and 
nonfederal organizations develop environmental indicator sets for 
several purposes, including assessing conditions and trends, 
communicating complex issues, and supporting performance management 
activities. Some environmental indicator sets are limited to use within 
specific political jurisdictional boundaries, while others are confined 
to specific natural areas, such as watersheds, lake basins, or 
ecosystems. Similarly, some sets address specific resources, such as 
water quality or land use, while others focus on quality of life issues 
or sustainable development. The indicator sets GAO reviewed are 
primarily used to assist in strategic planning efforts, communicate 
complex environmental issues, and track progress toward environmental 
goals.

Environmental indicator set developers, both federal and nonfederal, 
commonly face several major challenges. Such challenges include 
ensuring that a sound, balanced process is used to develop indicators, 
which can require a resource-intensive effort to address the needs of 
potential users. Similarly, obtaining sufficient data on environmental 
conditions and trends and their causes is particularly problematic. 
Another key challenge in developing useful environmental indicator sets 
involves coordinating and integrating the various related federal and 
other indicator sets in order to advance knowledge about the 
environment. In this regard, the efforts of the Council on 
Environmental Quality’s (CEQ) Interagency Working Group on Indicator 
Coordination are promising, but they lack the long-term, stable 
institutional arrangements needed to ensure continued guidance and 
coordination of federal activity in this area. Moreover, indicator sets 
designed to link management activities, environmental and natural 
resource conditions and trends, and human and ecological health have 
difficulty because many such relationships are not well understood. To 
that end, the Environmental Protection Agency’s (EPA) continuing work 
to develop indicators to assist the agency’s efforts to manage for 
results highlights this challenge. While EPA has made progress, its 
efforts to better understand such relationships over many years have 
been hampered not only by technical difficulties in establishing 
linkages between program activities and changes in the environment, but 
also by changes in leadership within the agency and the absence of a 
systematic approach, including clear expectations, milestones, and 
designated resources. Such institutional arrangements would enable the 
agency’s senior management, Congress, and other stakeholders to monitor 
and assist EPA’s efforts toward a complete and periodically updated 
Report on the Environment.

What GAO Recommends:

GAO recommends that the Chair of CEQ develop institutional arrangements 
needed to ensure a concerted, systematic, and stable approach to the 
development, coordination, and integration of environmental indicator 
sets. Moreover, GAO recommends that the EPA Administrator establish 
clear lines of responsibility and accountability and identify specific 
requirements for developing and using indicators. CEQ and EPA generally 
agreed with GAO’s recommendations.

www.gao.gov/cgi-bin/getrpt?GAO-05-52.

To view the full product, including the scope and methodology, click on 
the link above. For more information, contact John B. Stephenson at 
(202) 512-6225 or stephensonj@gao.gov.

[End of section]

Contents:

Letter:

Results in Brief:

Background:

Environmental Indicator Sets Are Developed for a Variety of Purposes, 
and Users Generally Report Positive Impacts:

Major Challenges Facing the Development and Use of Environmental 
Indicator Sets:

Conclusions:

Recommendations for Executive Action:

Agency Comments and Our Evaluation:

Appendixes:

Appendix I: Objectives, Scope, and Methodology:

Compendium of Environmental Indicator Sets:

Survey of Practitioners:

Case Study:

Meeting of Experts Convened by the National Academy of Sciences:

Appendix II: Key Environmental Indicator Initiatives Identified by 
Experts:

Appendix III: Environmental Indicator Set Case Study Profiles:

The Heinz Center's State of the Nation's Ecosystems:

EPA's National Coastal Assessment:

Chesapeake Bay Program:

Great Lakes State of the Lakes Ecosystem Conference:

Minnesota Department of Natural Resources' Strategic Conservation 
Agenda:

Environmental Protection Indicators for California:

Quality of Life Indicator Set, Jacksonville, Florida:

Environmental Indicators Project, West Oakland, California:

Appendix IV: Selected Activities Identifying Need for More 
Comprehensive Environmental Information:

Appendix V: Environmental Reporting by Private and Public 
Organizations:

Appendix VI: Accounting for the Environment:

Appendix VII: The Uncertain Cost of Environmental Information:

Appendix VIII: Selected Options:

Appendix IX: Comments from the Council on Environmental Quality:

Appendix X: Comments from the Department of the Interior:

Appendix XI: GAO Contact and Staff Acknowledgments:

GAO Contact:

Staff Acknowledgments:

Bibliography:

Related GAO Products:

Tables:

Table 1: Selected Major Environmental Research and Monitoring Networks 
and ProgramsA:

Table 2: Ten Challenges Most Frequently Cited as Major or Moderate by 
Survey Respondents:

Table 3: Ten Criteria Used to Select Indicators Most Frequently Cited 
by Survey Respondents:

Table 4: Sufficiency of Current Environmental Data to Support Three 
Major National Indicator Sets:

Table 5: Summary of Survey Participants:

Table 6: Environmental Indicator Sets Selected for Case Study Review:

Table 7: Major Pieces of Legislation to Address Federal Environmental 
Data and Indicator Issues, 1970-2004:

Table 8: Selected Congressional Hearings Addressing Federal 
Environmental Data and Indicator Management Issues, 1970-2004:

Table 9: Selected Academic Reports Addressing Federal Environmental 
Data and Indicator Management Issues:

Table 10: Direct Funding for Major Environment, Energy, and Natural 
Resources Statistical Programs:

Figures:

Figure 1: Nitrate Load Carried by Major Rivers:

Figure 2: Historical Wildfires in California, 1950 to 1997:

Figure 3: Ten Purposes for the Development of Environmental Indicator 
Sets Most Frequently Cited by Survey Respondents:

Figure 4: Selected Activities Identifying Need for More Comprehensive 
Environmental Information:

Figure 5: The State of the Nation's Ecosystems Report:

Figure 6: Draft National Coastal Condition Report II:

Figure 7: The State of the Chesapeake Bay Report:

Figure 8: State of the Great Lakes Report:

Figure 9: The Strategic Conservation Agenda Report:

Figure 10: Environmental Protection Indicators for California Report:

Figure 11: Jacksonville's 2003 Quality of Life Progress Report:

Figure 12: West Oakland's Neighborhood Knowledge for Change Report:

Abbreviations:

BEA: Bureau of Economic Analysis:

Cal/EPA: California Environmental Protection Agency:

CEQ: Council on Environmental Quality:

CSERA: Canadian System of Environmental and Resource Accounts:

EII: Environmental Indicators Initiative:

EIP: West Oakland Environmental Indicators Project:

EMAP: Environmental Monitoring and Assessment Program:

EPA: Environmental Protection Agency:

EPIC: Environmental Protection Indicators for California:

GEOSS: Global Earth Observation System of Systems:

GFT 250: Global Fortune Top 250 international companies:

GPRA: Government Performance and Results Act:

IEESA: Integrated Economic and Environmental Satellite Accounts:

JCCI: Jacksonville Community Council Inc.

NAMEA: National Accounting Matrix including Environmental Accounts:

NAS: National Academy of Sciences:

NCA: National Coastal Assessment:

NEPA: National Environmental Policy Act:

NOAA: National Oceanic and Atmospheric Administration:

NSTC: National Science and Technology Council:

OMB: Office of Management and Budget:

ORD: EPA's Office of Research and Development:

OSTP: Office of Science and Technology Policy:

PART: Performance Assessment Rating Tool:

PSR: pressure-state-response model:

SCA: Strategic Conservation Agenda:

SOLEC: Great Lakes State of the Lakes Ecosystem Conference:

TRI: Toxic Release Inventory:

USDA: U.S. Department of Agriculture:

USGS: U.S. Geological Survey:

Letter November 17, 2004:

The Honorable Sherwood L. Boehlert: 
Chairman, Committee on Science: 
House of Representatives:

The Honorable Vernon J. Ehlers: 
Chairman: 
The Honorable Mark Udall: 
Ranking Minority Member: 
Subcommittee on Environment, Technology and Standards: 
Committee on Science: 
House of Representatives:

Comprehensive and reliable information on the nation's environment and 
natural resources is a cornerstone of effective environmental 
management and an integral part of a national strategy to anticipate 
and address problems. Governments, businesses, and citizens depend on 
relevant, accurate, and timely federal data and statistics to make 
informed decisions about a range of environmental issues--including 
evaluating the performance of environmental programs, aligning the 
efficiency of markets with environmental protection, assessing the 
state of the environment and natural resources, and identifying 
emerging issues and options for action. Although data and statistics 
are rarely the sole factors that determine how society should address 
any given issue, reliable scientific information is essential to 
support the assessment of various alternatives and inform policy 
decisions.

Federal environmental monitoring and data collection activities provide 
critical inputs into the assessment process, and their planning and 
implementation must be linked to assessment and policy needs. The 
individual environmental problems that have been given much attention 
to date have given way to a growing realization of the overwhelming 
degree of interaction among the environmental, economic, and social 
sectors, and the degree to which the consequences of these interactions 
are cumulative, unpredictable, and--in many cases--difficult to repair. 
Developing an integrated understanding of such threats and the options 
for dealing with them is a central challenge for the nation. Moreover, 
the federal government relies on this information base to assess 
progress toward national goals as laid out in legislation and to 
improve and better account for its performance. In recent years, a 
general consensus has developed on the need to judge the success of the 
nation's environmental policies against environmental quality 
outcomes, rather than the number of management plans created, 
regulations or permits issued, or enforcement actions taken. The 
adoption of such a performance-based environmental policy, however, has 
been hampered by the lack of reliable scientific information on 
environmental conditions and trends.

Federal agencies collect and manage a tremendous volume of 
environmental data at a significant cost. The federal government spends 
at least $600 million each year on monitoring conditions and trends of 
the nation's natural and environmental resources.[Footnote 1] Numerous 
federal--and, in some cases, regional, state, or local--organizations 
conduct environmental research and monitoring programs using a variety 
of methods to address specific problems under different legislative 
authorizations, such as the Clean Air Act.[Footnote 2] Such activities 
can yield tangible and far-reaching benefits. For example, the National 
Oceanic and Atmospheric Administration (NOAA) produces climate 
forecasts based on data collected through satellites, ocean buoys, and 
other data collection activities that are often economically valuable 
because they give the public time and incentive to act to reduce 
weather-and climate-related losses. In one case, NOAA's forecast 
enabled residents of California to avoid an estimated $1.1 billion in 
damages during storms in the winter of 1997-'98, according to the 
agency.

However, adequate information is not always in place to help Congress 
or others determine how well the environment is doing, judge existing 
environmental policies, or develop sensible new ones. The nation's 
environmental data collection and monitoring systems were never 
intended to be comprehensive for all natural and environmental resource 
issues nationwide. A comprehensive picture of the nation's 
environmental and natural resources is not yet possible.

Numerous public and private initiatives are now developing sets of 
environmental indicators to bridge the gap between needed and available 
information and to prioritize further data collection. Environmental 
indicator sets assemble quantitative measures of conditions and trends 
to assess the state of the environment and natural resources and to 
gauge progress toward specific goals. In general, indicator sets are 
designed to provide environmental decision makers and the public with 
comprehensible information to assist developing strategic plans, 
setting priorities, and assessing which programs are, or are not, 
working well.

The widespread development and use of environmental indicator sets has 
led Congress, federal agencies, states, local communities, and 
corporations to consider the possible uses for sets of environmental 
indicators, such as for measuring performance and improving oversight 
of environmental programs. In this context, you asked us to examine (1) 
the purposes for which federal and nonfederal organizations are 
developing and using environmental indicator sets, and how they are 
being used; and (2) the major challenges facing the development and use 
of environmental indicator sets.

In addressing these objectives, we performed multiple lines of work:

* To identify a list of environmental indicator sets, we elicited the 
help of experts on environmental indicator set development. After 
conducting extensive Web and literature searches, conducting multiple 
background interviews, and following up with contacts made at 
professional conferences, we identified 48 experts. We then distributed 
a data collection instrument to these experts, asking them to specify 
(1) environmental indicator sets with which they were familiar that 
either were being developed or had been developed in the past 10 years, 
(2) states that had led or were leading the effort in developing and 
using environmental indicator sets, and (3) a contact person for each 
set. Twenty-three of the 48 experts responded. After combining 
duplicate responses and eliminating the responses that either did not 
meet our definition of an indicator set or could not be associated with 
enough information to locate a specific initiative, we developed a pool 
of 87 environmental indicator sets identified by experts that formed 
the basis for this review (see app. II for the complete list).

* To develop a list of environmental indicator set developers and users 
(or "practitioners") for GAO to survey, we spoke with representatives 
of each of the 87 indicator sets identified by experts and asked them 
to name a developer and a user to participate in our survey. This 
process yielded 87 practitioners, who collectively represented 52 of 
the indicator sets identified by experts. We then surveyed the 
practitioners, asking them to identify the primary characteristics of 
the indicator set, how it was developed, and how the set was being 
used. Forty-nine of the 87 practitioners responded to our survey.

* To gain a better understanding of the mechanics of environmental 
indicator set development and use, we conducted in-depth case studies 
on 8 of the 87 identified environmental indicator sets: two sets for 
each of four geographic scales--national, regional, state, and local. 
We selected the sets on the basis of their perceived level of maturity 
(current and active) and the availability and accessibility of 
individuals involved in their development and use. We conducted 
semistructured interviews that allowed practitioners the opportunity to 
supply information on a wide range of issues relating to their 
involvement with the development and use of the environmental indicator 
set. We also reviewed relevant documents that pertained to the 
development and use of each of the environmental indicator sets. Based 
on the information gathered, we then drafted case study profiles and 
provided them to the appropriate program manager for review and comment 
(see app. III).

* To assess the current status of environmental indicator sets and 
their impact on policy decisions, we contracted with the National 
Academy of Sciences (NAS) to convene a 2-day meeting of selected 
authorities with expertise in the interaction between science and 
policy making and who were familiar with indicator set development and 
use. NAS staff helped us identify a pool of authorities from which we 
selected 26 who collectively provided the meeting with a balance of 
expertise, interdisciplinary knowledge, and cross-jurisdictional 
representation. The meeting centered on discussions of three broad 
topics: (1) the organizations developing environmental indicator sets 
and the impact of these sets across the nation; (2) significant 
challenges facing the development and use of environmental indicator 
sets; and (3) what remedies, if any, existed to confront or mitigate 
these challenges.

In developing our findings, we corroborated the evidence gathered 
across these lines of work. A more detailed description of our scope 
and methodology is presented in appendix I. The findings in this report 
are not intended to apply to all environmental indicator sets. General 
references to indicator sets in this report refer to the 47 sets that 
we reviewed in detail--the 8 case studies and the 39 sets represented 
in our survey results. Furthermore, we did not independently assess the 
reliability of the data used in the 47 sets we reviewed because those 
data were not material to our findings. We conducted our work from June 
2003 to October 2004 in accordance with generally accepted government 
auditing standards.

Results in Brief:

Federal and nonfederal organizations are developing and using 
environmental indicator sets for assessing conditions and trends, 
communicating complex issues, and supporting performance management 
activities. Various organizations in the United States--including 
government agencies, nonprofit groups, universities, and corporations-
-have developed hundreds of environmental indicator sets. Some 
environmental indicator sets we reviewed are limited to political 
jurisdiction, such as county, state, or nation, while others are 
restricted to natural areas, such as watersheds, lake basins, or 
ecosystems. Some address specific resources, such as water quality or 
land use, while others focus on quality of life issues or sustainable 
development. For instance, the indicators reported through the Great 
Lakes State of the Lakes Ecosystem Conferences encompass the entire 
Great Lakes watershed--including aquatic, coastal, and terrestrial 
components, as well as human health and societal issues. In other 
cases, cities such as New Orleans, Pittsburgh, and Seattle have 
developed indicator sets that focus on broader issues that include 
economic prosperity, social equity, and environmental quality to 
measure and sustain the quality of life for the citizens in the 
community. The indicator sets we reviewed are primarily used to assist 
in strategic planning efforts, communicate complex environmental 
issues, and track progress toward environmental goals.

Whereas many challenges that inhibit the development of useful sets of 
environmental indicators are unique to the individual sets being 
developed, developers face several common challenges. Indicator set 
developers reported the following common challenges:

* Ensuring that a sound process is used to develop the indicator sets. 
Developers reported that creating an indicator set can be an intensely 
political process that challenges both the credibility and relevance of 
a set. Indicator sets we reviewed largely relied on collaborative 
processes to balance the various interests. Such processes define the 
purpose and intended use of the indicator set, determine the conceptual 
models--sets of qualitative assumptions to describe social, 
organizational, and natural systems--and criteria for selecting 
indicators, and selecting the indicators themselves. Such processes are 
difficult to manage, but essential to ensure that a set is ultimately 
accepted and used.

* Obtaining sufficient environmental data to report conditions and 
trends related to the indicators selected. Obtaining data for use in 
indicator sets can be difficult largely because long-standing 
limitations of federal environmental monitoring and data collection 
activities have not been resolved. Over half of the respondents to our 
survey identified obtaining data of sufficient quality as a major 
challenge to developing indicator sets. Indicator set developers and 
other experts noted that obtaining sufficient data on environmental 
conditions is difficult and costly because the many different 
organizations that collect data on the nation's environment and natural 
resources do so for specific purposes in different forms or on 
different geographic scales, and thus cannot be readily integrated to 
support indicators. Sharing such data can have significant, and 
sometimes prohibitive, costs because transforming the data to suit the 
needs of another user would require data managers to divert already 
limited resources from other projects. Moreover, past GAO work has 
emphasized that the federal government's current environmental 
information base suffers from data gaps between what is monitored and 
what needs to be monitored. Because of problems filling gaps in 
existing data and difficulties in integrating data from different 
databases, indicator set developers' efforts to identify data of 
sufficient quality from existing data sources has met with limited 
success.

* Coordinating and integrating various related indicator sets to 
develop a more comprehensive understanding of the environment. Experts 
we interviewed noted that the federal government lacks an 
organizational framework to provide a consistent basis for working with 
international, state, or nongovernmental indicator initiatives. 
Federal environmental indicator set developers employ a wide range of 
approaches. As a result, significant analytical and technical 
differences inhibit integration of related sets or synthesis of the 
diverse range of sets to draw a comprehensive picture of the nation's 
environment. Recognizing the need for coordination at the highest 
levels, the White House Council on Environmental Quality (CEQ) 
established an Interagency Working Group on Indicator Coordination in 
2002 to coordinate and integrate the federal investment in 
environmental indicator sets. According to officials, the Working Group 
was created as an ad hoc organization within the Executive Office of 
the President, operating without explicit responsibility and authority 
to ensure the continued and full involvement, cooperation, and 
resources from other federal agencies. Officials of agencies 
participating in the Working Group acknowledge the need for a more 
stable structure with the authority and resources necessary to achieve 
the Working Group's goals. On the basis of our discussions, we believe 
that a number of organizational options exist and should be studied to 
determine the most appropriate option or combination of options for 
implementing key functions, such as guiding and coordinating the 
development and use of environmental indicators.

* Linking specific environmental management actions and program 
activities to changes in environmental conditions and trends. 
Developers assembling environmental indicator sets to improve the 
performance of environmental management programs reported difficulty 
(1) accounting for relationships between management actions and other 
factors beyond the agency's control that can potentially affect 
environmental changes and (2) addressing the time lag between 
management actions and achieved results. Such problems are consistent 
with GAO's work on performance measurement in general, and the 
Environmental Protection Agency (EPA) in particular. Since our 1988 
report on EPA's management, GAO has stressed numerous times that EPA 
place priority on developing indicators to guide the agency's priority 
setting, strategic planning, and resource allocation. EPA's 
Environmental Indicators Initiative illustrates the difficulties in 
developing a set of national environmental indicators useful for 
establishing priorities, allocating resources, and assessing results. 
Past efforts to develop and use environmental indicators by the agency 
underscore both the importance and difficulty of doing so, and the need 
for a focused, long-term commitment as changes occur in the agency's 
senior management and priorities. These previous efforts have been 
hindered not only by technical difficulties in establishing linkages 
between program activities and changes in the environment, but also by 
changes in leadership within the agency and the lack of needed 
resources for monitoring the natural resources and the environment. 
Although a noteworthy step, EPA's effort thus far has not functioned as 
a key component of an agencywide comprehensive approach for managing 
EPA's work to achieve measurable results. EPA has not initiated or 
planned an institutional framework with clear lines of responsibility 
and accountability for developing and using environmental indicators, 
and no processes, procedures, or work plans exist to link the results 
of the initiative with EPA's strategic planning and performance 
reporting cycle.

In order to provide a comprehensive picture of environmental and 
natural resource conditions and trends to assess the nation's position 
and progress, we recommend that the Chairman of CEQ develop 
institutional arrangements needed to ensure a concerted, systematic, 
and stable approach to address the challenges associated with the 
development, coordination, and integration of environmental indicator 
sets.

Furthermore, building on EPA's initial efforts on indicators and to 
evaluate the purposes that indicators might serve, we recommend that 
the EPA Administrator establish clear lines of responsibility and 
accountability among EPA's various organizational components and 
identify specific requirements for developing and using environmental 
indicators.

Background:

Environmental indicators track changes to the quality and condition of 
the air, water, land, and ecosystems on various geographic scales, and 
related human health and economic conditions. Whereas definitions of 
"environmental indicator" vary, most of them emphasize that an 
environmental indicator is a selected quantifiable variable that 
describes, analyzes, and presents scientific information and its 
significance. Public and private initiatives assemble sets of 
indicators to address a variety of environmental issues. Federal 
agencies, private corporations, local communities, and others develop 
environmental indicator sets to condense complex topics or concepts, 
such as the health of ecosystems, into a manageable amount of 
meaningful information. Indicators are presented in statistical or 
graphical form, but are also referred to as concepts that have meaning 
beyond just the numeric value of the metric because of the importance 
of the phenomenon or element of a natural system being measured within 
the developers' worldview. For example, figure 1 presents the volume of 
nitrate carried by major rivers ("nitrate load") per year since the 
mid-1950s. Scientists generally accept this measure as an indicator of 
the condition of the nation's freshwater system, which, in turn, is a 
component of the health of ecosystems in the United States.

Figure 1: Nitrate Load Carried by Major Rivers:

[See PDF for image] 

[End of figure] 

Similarly, figure 2 shows an indicator drawn from a set of indicators 
addressing state-level environmental protection efforts. This 
indicator presents trend data on the extent of wildfires in California 
since 1950 as one measure to be used for gauging the performance of 
state programs to restore forest health.

Figure 2: Historical Wildfires in California, 1950 to 1997:

[See PDF for image] 

[End of figure] 

Organizations have developed and used indicator sets to address a broad 
array of economic, social, and environmental issues.[Footnote 3] For 
example, the Healthy People initiative, led by the Department of Health 
and Human Services, has worked since 1979 to develop a comprehensive 
set of national objectives for disease prevention and health promotion, 
and indicators with which to measure them. Healthy People has continued 
to be revised once every decade since 1980. Furthermore, economic 
indicator sets have been used to enhance understanding of economic 
phenomena, such as the business cycle. Economists generally agree that 
regular and consistent reporting of economic indicators such as 
unemployment, coupled with short explanations and extended discussion 
about the causes and consequences of the trends, has supported the 
development of economic theories and models and informed decision 
making in many institutions. However, as the National Research Council 
reported in 2000, while there are many well-known economic indicators, 
no current environmental indicators have achieved such status--although 
some environmental indicators, such as sea surface temperature as an 
indicator of global climate change, have begun to attract considerable 
attention. While much of the development of national indicators in the 
United States has focused on specific economic, social, and 
environmental concerns, the importance of interrelationships among 
these dimensions is growing. For example, there is a steady trend today 
to broaden and integrate various types of information used in 
decisionmaking contexts throughout society. The trend includes 
incorporating environmental and social measures into the regular 
reporting of economic measures by private corporations (see app. V) and 
linking environmental information to the information contained in the 
national economic accounts (see app. VI). Striving to understand the 
impact that human society has on the environment involves focusing on 
the interrelationships among economic, social, and environmental 
processes.

Environmental indicator sets are built upon a vast patchwork of 
environmental information. Federal agencies collect and manage a 
tremendous volume of environmental data at a cost of at least $600 
million each year. Across the United States, state, nonprofit, and 
private organizations also collect and manage research and monitoring 
data that feed into federal databases. Federal and nonfederal 
organizations collect such information to address specific problems 
under a variety of authorities using various research designs and 
methodologies, definitions, collection frequencies, and sites as 
determined by the collection agencies. As shown in table 1, numerous 
federal agencies are involved in key federal environmental research and 
monitoring programs, under a variety of legal authorities. Federal 
environmental monitoring and data collection activities provide 
critical feedback on the state of the nation's environment.

Table 1: Selected Major Environmental Research and Monitoring Networks 
and ProgramsA:

Program name: Coastal Change Analysis Program; 
Primary federal agencies: NOAA; 
Primary authority: NOAA Authorization Act of 1992.

Program name: Gap Analysis Program; 
Primary federal agencies: USGS; 
Primary authority: Fish and Wildlife Coordination Act.

Program name: National Wetlands Inventory; 
Primary federal agencies: FWS; 
Primary authority: Emergency Wetland Resources Act of 1986.

Program name: Breeding Bird Survey; 
Primary federal agencies: USGS; 
Primary authority: Migratory Bird Treaty Act.

Program name: Clean Air Status and Trends Network; 
Primary federal agencies: EPA; 
Primary authority: Clean Air Act Amendments of 1990.

Program name: Environmental Monitoring and Assessment Program; 
Primary federal agencies: EPA; 
Primary authority: Clean Water Act.

Program name: Forest Health Monitoring; 
Primary federal agencies: EPA and Forest Service; 
Primary authority: Forest Ecosystem and Atmospheric Pollution Research 
Act of 1988.

Program name: Forest Inventory Analysis; 
Primary federal agencies: Forest Service; 
Primary authority: Forest and Rangeland Renewable Resources Research 
Act of 1978, as amended by the Agricultural Research, Extension, and 
Education Reform Act of 1998.

Program name: National Atmospheric Deposition Program/National Trends 
Network; 
Primary federal agencies: USGS; 
Primary authority: Clean Air Act Amendments of 1990.

Program name: National Air Monitoring System/State and Local/
Photochemical Air Monitoring System; 
Primary federal agencies: EPA; 
Primary authority: Clean Air Act Amendments of 1977.

Program name: National Stream Quality Accounting Network; 
Primary federal agencies: USGS; 
Primary authority: USGS Organic Act.

Program name: National Stream Gauging Network; 
Primary federal agencies: USGS; 
Primary authority: USGS Organic Act.

Program name: National Resources Inventory; 
Primary federal agencies: NRCS; 
Primary authority: Rural Development Act of 1972.

Program name: National Status and Trends; 
Primary federal agencies: NOAA; 
Primary authority: Marine Protected Resources and Sanctuaries Act of 
1972, as amended.

Program name: NMFS marine mammal stock assessments; 
Primary federal agencies: NOAA and NMFS; 
Primary authority: Marine Mammal Protection Act Amendments of 1994.

Program name: Remote Automated Weather System; 
Primary federal agencies: Multiagency; 
Primary authority: Federal agency land management authorities.

Program name: Snowpack Telemetry; 
Primary federal agencies: NRCS; 
Primary authority: Pub. L. No. 74-46.

Program name: Agricultural Research Service; 
Primary federal agencies: USDA; 
Primary authority: USDA research authorities (e.g., 7 U.S.C. § 1010).

Program name: Forest and rangeland sites; 
Primary federal agencies: USDA; 
Primary authority: Forest and Rangeland Renewable Resources Research 
Act of 1978.

Program name: Long-term ecological research; 
Primary federal agencies: NSF; 
Primary authority: National Science Foundation Act of 1950, as amended.

Program name: National Park Ecological Monitoring Program; 
Primary federal agencies: NPS; 
Primary authority: NPS Organic Act.

Program name: Coastal Ocean Program; 
Primary federal agencies: NOAA; 
Primary authority: NOAA Authorization Act of 1992.

Program name: National Marine Sanctuary; 
Primary federal agencies: NOAA; 
Primary authority: National Marine Sanctuaries Amendments Act of 2000.

Program name: Hydrologic Benchmark Network; 
Primary federal agencies: USGS; 
Primary authority: USGS Organic Act.

Program name: National Water Quality Assessment; 
Primary federal agencies: USGS; 
Primary authority: USGS Organic Act.

Program name: Water, Energy, and Biogeochemical Budgets Program; 
Primary federal agencies: USGS; 
Primary authority: Global Change Research Act of 1990. 

Source: GAO analysis of National Science and Technology Council data.

[A] Networks and programs in this list were drawn from an inventory 
originally reported by the National Science and Technology Council in 
1997.

[End of table]

Although extensive, the environmental information base in the United 
States does not support comprehensive environmental and natural 
resource assessments. In 1997, the National Science and Technology 
Council (NSTC)--a Cabinet-level council that serves as the principal 
means for the president to coordinate research and development across 
federal agencies--evaluated the status of federal agency environmental 
monitoring and research activities and found that monitoring programs 
do not provide integrated data across multiple natural resources at the 
various scales needed to develop policies that take into account 
current scientific understanding. The NSTC called for a strategy for 
environmental monitoring and research to enable comprehensive 
assessments.[Footnote 4] More recently, the National Council for 
Science and the Environment--a nonprofit organization addressing the 
scientific basis for environmental decision making--convened a national 
conference of more than 450 scientists, policymakers, and academicians 
in December 2000 that underscored the need for comprehensive national 
assessments.[Footnote 5]

Until 2000, the White House Council on Environmental Quality (CEQ) was 
required to transmit an annual environmental quality report to 
Congress. Although the annual reporting requirement is no longer in 
effect, CEQ is still required to accumulate the necessary data and 
other information needed for a continuing analysis of changes and 
trends in the natural environment and an interpretation of their 
underlying causes.[Footnote 6] Whereas scientists, agency officials, 
and academicians generally agree on the need for periodic reporting of 
conditions and trends of environmental and natural resources, no 
consensus has been reached on who should be responsible for this task 
or how it would be best achieved.

The federal government relies on such trend information to assess 
progress toward national goals and to improve and better account for 
its performance, but credible and reliable information cannot always be 
obtained. In recent years, a general consensus has developed on the 
need to judge the success of the nation's environmental policies 
against environmental quality outcomes, rather than the number of 
management plans created, regulations or permits issued, or enforcement 
actions taken. The Government Performance and Results Act (GPRA)--the 
centerpiece of a statutory and management framework laid out in the 
1990s as the foundation for strengthening government performance and 
accountability--is designed to inform congressional and executive 
decision making by providing objective information on the relative 
effectiveness and efficiency of federal programs and spending. GPRA 
requires both a connection to the structures used in congressional 
budget presentations and consultation between the executive and 
legislative branches on agency strategic plans to ensure Congress an 
oversight stake in GPRA's success. The current administration has made 
the integration of performance and budget information one of five 
governmentwide management priorities under the President's Management 
Agenda. Central to this initiative is the Program Assessment Rating 
Tool (PART). The Office of Management and Budget developed PART as a 
diagnostic tool meant to provide a consistent approach to evaluating 
federal programs and as one tool applied it in formulating the 
executive branch's fiscal years 2004 and 2005 budget requests. The 
adoption of such a performance-based environmental policy, however, has 
been hampered by the lack of reliable scientific information on 
environmental conditions and trends.

Environmental Indicator Sets Are Developed for a Variety of Purposes, 
and Users Generally Report Positive Impacts:

Government agencies, universities, corporations, and other 
organizations have developed environmental indicator sets to address 
environmental issues on various geographic scales. Most of the 
environmental indicator sets we reviewed were developed for a myriad of 
purposes, including assessing environmental conditions and trends, 
raising public awareness, communicating complex issues, and tracking 
progress toward goals. Indicator set users reported that such sets 
generally had positive impacts, and were especially useful in assessing 
environmental conditions and trends, communicating complex 
environmental issues, and developing strategic plans. However, it is 
difficult to determine the benefits that arise from these impacts.

Organizations Develop Environmental Indicator Sets for Specific but 
Varied Purposes:

Various organizations throughout the United States--including 
government agencies at national, state, and local levels; nonprofit 
groups; universities; and corporations--have developed hundreds of 
environmental indicator sets in recent years to address environmental 
issues on a variety of geographic scales. Some environmental indicator 
sets are limited to political jurisdiction, such as county, state, or 
nation; others are limited to natural areas, such as watersheds, lake 
basins, or ecosystems. Many environmental indicator sets address 
complex, crosscutting issues--such as ecosystem health--that are 
affected by environmental, economic, and social factors. For instance, 
the Great Lakes Water Quality Agreement calls for the development of a 
set of about 80 ecosystem health indicators for the Great Lakes to 
inform the public and report progress toward achieving the objectives 
of the agreement. Indicators address specific geographic zones of the 
entire Great Lakes Basin ecosystem--such as offshore, nearshore, 
coastal wetlands, and shoreline--and other issues such as human health, 
land use, and societal well-being. The indicator list is continually 
evolving. Every 2 years, Environment Canada--the Canadian agency 
primarily responsible for the preservation and enhancement of the 
quality of the natural environment--and EPA host a review and 
discussion of the indicators as required under the agreement, either at 
the State of the Lakes Ecosystem Conference or through alternate 
processes. Moreover, some cities, such as New Orleans, Pittsburgh, and 
Seattle, have developed comprehensive indicator sets that focus on 
broader issues that incorporate such factors as economic prosperity, 
social equity, and environmental quality to measure and sustain the 
quality of life for the citizens in the community.

Respondents to our survey noted that the most common purposes for 
developing environmental indicator sets were to assess environmental 
conditions and trends, educate and raise awareness among the public, 
simplify and communicate complex issues, and track progress toward 
environmental goals (see fig. 3).[Footnote 7]

Figure 3: Ten Purposes for the Development of Environmental Indicator 
Sets Most Frequently Cited by Survey Respondents:

[See PDF for image] 

Note: Results out of a possible total of 42 responses.

[End of figure] 

Environmental indicator sets have been developed to serve multiple 
purposes and audiences. For example, the H. John Heinz III Center for 
Science, Economics, and the Environment (Heinz Center) developed The 
State of the Nation's Ecosystems indicator set, published in 2002, to 
identify a succinct set of indicators to report on the ecological 
condition of the nation, identify data gaps, and provide information to 
a broad audience. The intended audience of the indicator set 
encompassed members of Congress, executive branch agencies, business 
executives, environmental advocacy groups, state and local officials, 
and the general public.

Most environmental indicator sets are developed voluntarily. For 
example, the California Environmental Protection Agency (Cal/EPA) began 
developing the Environmental Protection Indicators for California 
(EPIC) in 2001 as part of the implementation plan for the agency's 2000 
Strategic Vision document. Cal/EPA made a commitment to focus more on 
measurable environmental results in assessing the effectiveness of its 
environmental programs, and in making program adjustments to better 
meet the state's environmental protection goals. EPIC developed about 
85 indicators based on categories that mirror the agency's areas of 
authority, and reported them in an April 2002 report. Similarly, 
Minnesota's Department of Natural Resources developed environmental 
indicators and targets for its Strategic Conservation Agenda. The 
department developed about 75 indicators in six performance areas to 
help the agency better define its priorities, communicate its progress, 
and manage for environmental results.

Other environmental indicator sets are developed in response to legal 
mandates. For example, the state of Michigan publishes a biennial 
report as required under the Michigan Natural Resources and Protection 
Act. The publication, prepared jointly by the Michigan Department of 
Environmental Quality and the Department of Natural Resources, reports 
on the conditions and trends of the environment, such as land use and 
cover, mammal and fish populations, and ambient air pollutant levels. 
At the federal level, the National Park Service created the Natural 
Resource Challenge in 1999 in response to the direction of the National 
Parks Omnibus Management Act of 1998 to enhance national parks 
management by using the highest quality science and information, and to 
create a resource inventory and monitoring program to establish 
baseline conditions and long-term trends.[Footnote 8] The Natural 
Resource Challenge includes indicators--referred to as vital signs--to 
identify ecosystem health status and trends and to determine compliance 
with laws and regulations. For example, park managers have used vital 
signs, such as the concentration of air pollutants in precipitation and 
its effects on water quality, to detect potential problems and identify 
steps to restore ecological health of park resources.

Environmental Indicator Set Users Generally Report Positive Impacts:

The use of environmental indicator sets has resulted in a variety of 
positive impacts. A majority of users of environmental indicator sets 
told us that the sets are either useful or very useful for their needs, 
especially in (1) assessing environmental conditions and trends, (2) 
communicating complex environmental issues, and (3) developing 
strategic plans. However, largely because indicator sets themselves do 
not create change--instead policymakers employ the information when 
making decisions--it is difficult to measure the benefits that accrue 
from these impacts.[Footnote 9]

The indicator sets we reviewed assess environmental and natural 
resource conditions and trends, and have been used to help identify 
data gaps and research needs, provide early warning of potential 
environmental problems, allocate resources, and analyze alternatives 
for environmental management. Several of the applications help 
demonstrate how environmental indicator sets had positive impacts:

* Experts called the Great Lakes State of the Lakes Ecosystem 
Conference (SOLEC) indicator set a key factor in identifying needed 
management approaches at the and served as a positive catalyst in 
promoting collaboration on key issues. In particular, the SOLEC 
indicator set helped influence the Fish and Wildlife Service decision 
to focus on the development of an ecosystem/watershed approach to 
environmental management for the Great Lakes that crosses multiple 
political boundaries.

* The ecological framework designed for the Heinz Center's State of the 
Nation's Ecosystems indicator set is used to inform the design of the 
ecological portion of the international Global Ocean Observing System-
-a major multinational initiative that is designed to observe, model, 
and analyze marine resources. In addition, the Heinz Center indicator 
set identified a number of missing or inadequate data needed to provide 
a complete picture of ecosystem condition, such as data to support an 
indicator measuring the biological condition of the soil in use as 
farmland. The center is working with federal, state, and local 
governmental, nongovernmental, and private organizations to call 
attention to the need for identifying priorities for filling data gaps 
and the need to fill these gaps.

* The National Coastal Assessment component of EPA's Environmental 
Monitoring and Assessment Program (EMAP) provides a more complete 
picture of the condition of the nation's estuaries. EPA's Office of 
Research and Development led the creation of the indicator set and 
monitoring program that constitute the assessment, which includes five 
aggregate indicators--water quality, sediment quality, coastal 
habitat, benthic community structure, and fish tissue contaminants. 
Three coastal states have fully implemented the monitoring and 
indicator approach to fulfill reporting requirements under the Clean 
Water Act,[Footnote 10] and 21 other states have begun to implement the 
approach or have used the approach to assess a part of their estuaries. 
Users reported that the indicator set and monitoring design provided a 
more effective approach to consistently measuring estuary conditions 
for coastal states.

* The development and use of an environmental indicator set for the 
Chesapeake Bay influenced the strategic allocation of approximately $18 
million of federal funds in fiscal year 2003 toward meeting restoration 
goals for the bay. The Chesapeake Bay Program--established by the 1983 
Chesapeake Bay Agreement, one of three overriding agreements aimed at 
restoring the health of the bay--began developing environmental 
indicators to support goal setting, to define targets and endpoints for 
restoration of the bay, and to make the program more accountable to the 
public by defining and communicating the bottom line environmental 
results achieved by the restoration program. The program distributes 
funds in the form of grants to state governments, local governments, 
interstate agencies, nonprofits, universities, and colleges to 
implement the restoration goals of the Chesapeake 2000 Agreement and to 
collect data and other information for use in the indicator set. The 
indicator set uses monitoring data and other information to measure 
environmental conditions of the Chesapeake Bay and progress in meeting 
goals.

Environmental indicator sets also serve as powerful tools for 
communicating information on complex environmental issues in a way that 
makes them more comprehensible and accessible. Two organizations in 
particular--the Pacific Institute for Studies in Development, 
Environment, and Security through its West Oakland Environmental 
Indicators Project and the Jacksonville Community Council Inc. (JCCI)-
-use their respective indicator sets to identify environmental issues, 
perform research to better understand the issues, and develop 
appropriate solutions. For example, West Oakland's indicator set helped 
decision makers identify and eventually close a major source of air 
pollution in the community, which likely would not have been 
accomplished without extensive public awareness and action galvanized 
by the indicator set. Similarly, JCCI uses its indicator set to 
identify issues for further study, such as ensuring an adequate water 
supply and reducing the municipal garbage burden, which the indicator 
set had shown to be areas of existing or emerging problems. At the 
culmination of each study, JCCI issues a report with recommendations to 
improve the situation and creates a task force to ensure implementation 
of the recommendations.[Footnote 11]

The process of developing an environmental indicator set enhances 
strategic planning by engaging a broad-based group of individuals in a 
structured, collaborative process. As we reported in March 2004, 
strategic planning for performance-based, results-oriented management 
requires transforming organizational cultures to improve decision 
making, maximize performance, and ensure accountability.[Footnote 12] 
Such a transformation requires investments of time and resources as 
well as sustained leadership, commitment, and attention. Throughout our 
review, indicator set developers and users emphasized the importance of 
broad collaboration in developing indicators as a way of strengthening 
their relevance and broad acceptance. The developers of some indicator 
sets use the indicator set development process to advance dialogue 
within their community or region by bringing together many different 
sectors, fostering new alliances and relationships, and providing a 
forum to discuss ways to better measure and manage environment issues. 
For instance, staff members of some organizations, such as the 
Minnesota Department of Natural Resources and the California 
Environmental Protection Agency, told us that the process of developing 
and refining their indicator sets helped staff identify and define 
environmental management goals to better manage for results. For 
instance, California EPA has traditionally assessed the success of its 
environmental programs based on measures of activities, such as the 
number of permits granted or notices of violations issued. The intent 
of developing environmental indicators at California EPA was to measure 
environmental results and to be able to use the indicators to support a 
results-based management system. The process of developing the 
indicators at California EPA brought various staff together to define 
issues and parameters to develop indicators that could be used to 
manage for results.

Nevertheless, it is not easy--or sometimes even possible--to measure 
the benefits of the sets that stem from these impacts. Developers 
reported that systematic monitoring of the effectiveness of 
environmental indicator sets and their benefits varies due in part to 
resource costs. Moreover, developers and users reported that 
environmental indicator sets themselves did not create change from 
which benefits could be measured; rather, they might influence 
environmental management activities and thus yield benefits from 
affecting the quality of a decision. However, such difficulties should 
not necessarily be seen as a precondition for developing and using 
indicator sets. Instead, these unanswered questions highlight the need 
for additional research on how to better gauge the return on the 
investment for organizations that have invested in indicator sets.

Major Challenges Facing the Development and Use of Environmental 
Indicator Sets:

A number of challenges face developers and users of environmental 
indicator sets. Selecting from a broad range of issues, survey 
respondents most frequently cited the 10 issues presented in table 2 as 
major or moderate challenges.

Table 2: Ten Challenges Most Frequently Cited as Major or Moderate by 
Survey Respondents:

Challenge: Obtaining data of sufficient quality; 
Number of responses: Major: 22; 
Number of responses: Moderate: 14; 
Number of responses: Total: 36.

Challenge: Obtaining data of appropriate geographic scope; 
Number of responses: Major: 19; 
Number of responses: Moderate: 13; 
Number of responses: Total: 32.

Challenge: Selecting sufficient indicators; 
Number of responses: Major: 15; 
Number of responses: Moderate: 16; 
Number of responses: Total: 31.

Challenge: Obtaining needed funds; 
Number of responses: Major: 15; 
Number of responses: Moderate: 13; 
Number of responses: Total: 28.

Challenge: Clearly defining the phenomena to be measured; 
Number of responses: Major: 12; 
Number of responses: Moderate: 12; 
Number of responses: Total: 24.

Challenge: Determining the criteria for selecting indicators; 
Number of responses: Major: 5; 
Number of responses: Moderate: 18; 
Number of responses: Total: 23.

Challenge: Staff with necessary expertise; 
Number of responses: Major: 7; 
Number of responses: Moderate: 15; 
Number of responses: Total: 22.

Challenge: Clearly defining the purpose of the indicator set; 
Number of responses: Major: 4; 
Number of responses: Moderate: 18; 
Number of responses: Total: 22.

Challenge: Clearly defining the intended use of the set; 
Number of responses: Major: 4; 
Number of responses: Moderate: 17; 
Number of responses: Total: 21.

Challenge: Determining the conceptual framework to use; 
Number of responses: Major: 5; 
Number of responses: Moderate: 15; 
Number of responses: Total: 20. 

Source: GAO.

Note: Results out of a possible total of 42 responses. Respondents 
chose from five response categories: Major, Moderate, Minor, Not a 
Challenge, or Don't Know.

[End of table]

Interviews with indicator set developers and other experts revealed 
that many challenges tended to revolve around the specific 
circumstances affecting the particular sets. However, we identified 
several categories of common challenges faced by indicator set 
developers and users on the basis of the survey responses and detailed 
interviews with developers and other experts:

* Ensuring that a sound process is used to develop the indicator sets. 
Developers reported that support for an indicator set can be undermined 
if it is viewed as biased because of its association with a particular 
political perspective or leader. The process of developing an indicator 
set can be an intensely political process that challenges both the 
credibility and relevance of a set. Developers of the sets we reviewed 
largely relied on collaborative processes to define the purpose and 
intended use of the indicator set, determine the conceptual model and 
criteria for select indicators, and selecting the indicators 
themselves. Such processes are difficult to manage to ensure a set's 
credibility and relevance.

* Obtaining sufficient environmental data to report conditions and 
trends related to the indicators selected. Over half of the respondents 
to our survey identified obtaining data of sufficient quality as a 
major challenge to developing indicator sets. Indicator set developers 
and other experts noted that the many different organizations that 
collect data on the nation's environment and natural resources do so 
for specific purposes in different forms or on different geographic 
scales.

* Coordinating and integrating various related indicator sets in order 
to obtain a better understanding of the environment. Experts that we 
interviewed noted the federal government lacks an organizational 
framework to provide a consistent basis for working with international, 
state, or nongovernmental indicator initiatives. Environmental 
indicator set developers employ a wide range of approaches. As a 
result, significant analytical and technical differences inhibit 
integration of related sets or synthesis of the diverse range of sets 
to draw a comprehensive picture of the nation's environment. The White 
House Council on Environmental Quality (CEQ) recognized the need for 
coordination and established an Interagency Working Group on Indicator 
Coordination (Working Group) in 2002 to coordinate and integrate the 
federal investment in environmental indicator sets.

* Linking specific environmental management actions and program 
activities to changes in environmental conditions and trends. 
Organizations that develop environmental indicator sets to improve the 
performance of environmental management programs can struggle with 
linking management actions and environmental conditions and trends and 
address the time lag between management actions and achieved results. 
EPA's past efforts to develop and use environmental indicators 
underscore both the importance and difficulty of doing so, and the need 
for a focused, long-term commitment as the agency undergoes changes in 
management and priorities.

Ensuring a Sound Process to Develop Indicator Sets:

Developers reported that support for an indicator set can be undermined 
if it is viewed as biased because of its association with a particular 
political perspective or leader. The process of developing an indicator 
set can be an intensely political process that challenges both the 
credibility and relevance of a set. When selecting one of the many 
indicators in a set, others are necessarily excluded because many 
indicator set developers strive to keep the number of indicators as 
small as possible. In some cases, that means an issue of interest to a 
particular stakeholder or user group does not get measured by the set. 
For example, the criteria used to select indicators for the Georgia 
Basin Puget Sound Ecosystem indicator set limited the number of 
indicators to only six, which led to gaps in the presentation of 
information on the complete state of the ecosystem.

The process used to select indicators can affect the usefulness of a 
set, producing a set of indicators of little or no relevance to the 
users' needs. Moreover, developers reported that support for an 
indicator set can be undermined if it is viewed as being biased or 
nonobjective because of its association with a particular political 
perspective or leader.

Indicator set developers stressed the need for a balanced process to 
manage such concerns. In particular, involving a set's varied users, 
developing and applying sound selection criteria, and identifying 
appropriate conceptual models were cited as important elements of the 
development process.

Many developers we interviewed noted the importance of--and 
difficulties in--incorporating users' needs when selecting indicators. 
Identifying, engaging, and balancing the information needs of the users 
can be a resource-intensive processes. For example, the Heinz Center 
spent significant time conducting outreach to each of four sectors--
businesses; environmental and conservation advocacy organizations; 
academia; and federal, state, and local governments--it identified as 
potential users of its indicator set on The State of the Nation's 
Ecosystems. The Heinz Center engaged about 150 representatives in a 3-
year consensus-building process, leading to the indicator set that was 
eventually adopted. Similarly, the Minnesota Department of Natural 
Resource's indicators supporting the state's Strategic Conservation 
Agenda were developed collaboratively by the department's Science 
Policy Unit--housed within the department's Office of Management and 
Budget Services--and departmental operations managers representing all 
divisions and regions. Developers stated that the process, although 
resource-intensive, ensured that the agency had support from users and 
other stakeholders of the indicator set. However, not all indicator 
sets have the resources to develop such a process or sustain it over 
time. As a result, indicator sets can have limited applicability to the 
users' needs. We found that some affected user groups were not 
identified, not effectively involved in the development of indicator 
sets, or both.

In many of the cases we reviewed, indicator set developers employed 
specific criteria to guide indicator selection. Such criteria describe 
desired characteristics, attributes, or standards--such as relevance to 
environmental policies or scientific soundness--that indicators must 
meet to be eligible for inclusion in a set (see table 3).

Table 3: Ten Criteria Used to Select Indicators Most Frequently Cited 
by Survey Respondents:

Criteria: Measurable; 
Number of responses: 35.

Criteria: Relevant; 
Number of responses: 35.

Criteria: Appropriate geographic scale; 
Number of responses: 34.

Criteria: Understandable; 
Number of responses: 34.

Criteria: Data available; 
Number of responses: 32.

Criteria: Data quality; 
Number of responses: 31.

Criteria: Importance; 
Number of responses: 28.

Criteria: Appropriate temporal scale; 
Number of responses: 28.

Criteria: Data comparability; 
Number of responses: 26.

Criteria: Trend data available; 
Number of responses: 24.

Source: GAO.

Note: Results out of a possible total of 42 responses.

[End of table]

In some cases, set developers engage users and other stakeholders in 
defining selection criteria early in the selection process to screen, 
rank, or otherwise prioritize the field of potential indicators before 
addressing and selecting the individual indicators. For example, the 
process for selecting indicators for the Environmental Protection 
Indicators for California indicator set involved developers first 
identifying environmental issues that are significant for the state--
such as air quality or human health--along with more specific 
components of such issues--such as criteria air pollutants such as 
ozone, carbon monoxide, and particulate matter. Developers then 
identified relevant, measurable parameters within each issue, such as 
vehicle miles traveled, to help derive candidate indicators. Candidate 
indicators were then subject to criteria, such as data quality, 
representativeness, sensitivity, and decision support to help select 
the final set of indicators.

In addition, many indicator set developers designed conceptual models 
to serve as foundations for structuring and selecting indicator sets. 
Conceptual models present the set developers' understanding of how 
systems operate, and help integrate the different fields of science 
relevant to an issue that cuts across environmental disciplines, such 
as ecosystem management. Such models can enhance the degree to which an 
indicator set incorporates the best available scientific knowledge and 
understanding, presents assumed causal relationships between different 
variables, and identifies different types of performance management 
indicators for assessing the results of specific environmental 
policies. For example, one common model is the pressure-state-response 
model. Such a model helps developers understand real and potential 
causal relationships between human actions, such as population growth 
and pollution, on the environment.

Obtaining Sufficient Environmental Data to Report Conditions and 
Trends:

Obtaining data for use in indicator sets can be difficult largely 
because longstanding limitations of federal environmental monitoring 
and data collection activities have not been resolved. Over half of the 
respondents to our survey identified obtaining data of sufficient 
quality as a major challenge to developing indicator sets. Indicator 
set developers and other experts noted that the many different 
organizations that collect data on the nation's environment and natural 
resources do so for specific purposes. To meet these purposes, these 
data are collected in different forms or on different geographic 
scales, and thus cannot be readily integrated to support indicators. 
Such limitations of federal environmental monitoring and data 
collection activities, however, are long-standing and, despite a number 
of attempts, have not been resolved.

Responsibility for research, monitoring, and assessment of various 
environmental and natural resources currently resides in various 
federal and other organizations whose activities focus on achieving 
specific programmatic objectives. Differences in definitions, study 
design and methodology; frequency of collection; site selection; 
quality assessment and control; and other technical issues compound the 
fragmentation of data collection activities. For example, our January 
2001 report detailed major management issues facing EPA, one of which 
was the agency's outmoded data management system that relies on 
separately designed, media-specific databases that are generally not 
technically compatible.[Footnote 13] Data generated through such 
disparate activities are not being integrated in common databases or 
otherwise being made accessible to potential users. Data sharing can 
have significant costs because environmental data are generally 
collected according to the specific needs or purposes of the collecting 
agency or organization, and transforming the data to suit the needs of 
another user would require data managers to divert already limited 
resources--staff time, computing resources, and money--from ongoing 
agency projects.

The recent commitment to develop a Global Earth Observation System of 
Systems (GEOSS) by the United States underscores the need for 
coordinated information about the environment. GEOSS is a 10-year 
international cooperative effort to make it possible for all existing 
and new earth-observing hardware and software around the globe to 
communicate so they can continuously monitor the land, sea, and air. 
GEOSS is built on the idea that the dozens of observational systems now 
generating reams of data around the world could be more powerful if 
they could be combined and widely disseminated. A completed 10-year 
implementation plan will be presented at the third Earth Observation 
Summit in February 2005. More than 15 federal agencies--including NOAA 
and EPA--and several White House offices are developing a draft 
strategic plan for the United States Integrated Earth Observation 
System, which will be a key component of the GEOSS 10-year plan.

Moreover, gaps in existing data also limit the usefulness of many 
federal environmental datasets to support the crosscutting issues 
addressed by indicator sets. Our past work has emphasized that the 
federal government's current environmental information base suffers 
from data gaps between what is monitored and what needs to be 
monitored. For example, we reported in July 1998 and again in December 
2002 on how the lack of consistent data on federal wetlands programs 
implemented by different agencies prevented the government from 
measuring progress toward achieving the governmentwide goal of no net 
loss of the nation's wetlands.[Footnote 14] Furthermore, we reported in 
June 2004 that hundreds of entities across the nation collect water 
quality data that provide a great deal of information about the 
condition of the nation's waters--however, the United States does not 
have enough information to provide a comprehensive picture at the 
national level because of the way in which these entities collect water 
quality data.[Footnote 15] This shortfall impairs its understanding of 
the state of its waters and complicates decision making on such 
critical issues as which waters should be targeted for cleanup and how 
such cleanups can best be achieved.

Problems with integrating databases and filling gaps in federal 
environmental data are long-standing issues that were recognized at 
least 3 decades ago. In 1970, the Council on Environmental Quality 
(CEQ) noted in its first report to Congress on the nation's environment 
that contemporary efforts did not provide the type of information or 
the geographic coverage needed to evaluate the condition of the 
nation's environment, track changes in its quality, or trace their 
causes.[Footnote 16] Moreover, academicians have found that nearly 
every comprehensive study during this period on national environmental 
protection has called for more coherent and comprehensive information 
on the state of our environment and natural resources.[Footnote 17] 
Congress has discussed federal environmental data and indicator issues 
many times since 1970. Figure 4 shows these efforts, as well as 
selected relevant scholarly reports issued during the same period.

Figure 4: Selected Activities Identifying Need for More Comprehensive 
Environmental Information:

[See PDF for image] 

Note: Refer to appendix IV for a description of legislation, hearings, 
and reports.

[End of figure] 

Although not intended to be exhaustive, this figure illustrates 
significant legislative and academic milestones in federal 
environmental data and indicator management over the last 35 years. As 
shown in the figure, both Congress and the academic community had 
already identified and analyzed, but not addressed, many of the 
fundamental issues confronting indicator development and data 
management by the close of the 1970s.

Because of problems filling gaps in existing data and difficulties in 
integrating data from different databases, indicator set developers' 
efforts to identify data of sufficient quality from existing data 
sources has met with limited success. For example, the developers of 
the Heinz Center's State of the Nation's Ecosystems report were unable 
to obtain sufficient data for reporting nationally 45 of 103 indicators 
included in the report. The report identified Total Impervious Area--a 
classification of urban and suburban areas according to the percentage 
of roads, parking lots, driveways, and rooftops that they contain--as 
an important measure of the degree of urbanization of the United 
States, and closely related to water quality in urban and suburban 
areas. However, the report explained that such data had not been 
compiled regionally or nationally and there were no standard methods 
for estimating this metric.[Footnote 18] As illustrated in table 4, 
other national indicator sets experienced a similar challenge.

Table 4: Sufficiency of Current Environmental Data to Support Three 
Major National Indicator Sets:

Indicator set: The State of the Nation's Ecosystems; 
Number of indicators: 103; 
Indicators with sufficient data: 58 (56%); 
Indicators with insufficient data: 45 (44%).

Indicator set: Draft Report on the Environment 2003; 
Number of indicators: 146; 
Indicators with sufficient data: 44 (30%); 
Indicators with insufficient data: 102 (70%).

Indicator set: National Report on Sustainable Forests--2003; 
Number of indicators: 67; 
Indicators with sufficient data: 8 (12%); 
Indicators with insufficient data: 59 (88%).

Sources: EPA, Forest Service, and the Heinz Center.

Note: GAO applied the various quality criteria developed and reported 
by each project. GAO did not independently evaluate these criteria or 
the project's application of the criteria.

[End of table]

Coordinating and Integrating Indicator Sets to Improve the Current 
Understanding of Environmental Conditions and Trends:

Experts we interviewed noted the federal government lacks an 
organizational framework or institutional arrangements to provide a 
consistent basis for working with international, state, or 
nongovernmental indicator initiatives. Currently these efforts are not 
coordinated, resulting in significant differences and 
incompatibilities between sets that inhibit integration and synthesis. 
For example, federal environmental indicator sets cannot always be 
integrated with each other, or with regional-or state-level indicator 
initiatives on similar topics, largely because the sets are based on 
different frameworks and include indicators relevant at different 
geographic scales. As a result, congressional, federal agency, and 
other users must reconcile information that seems to deliver 
inconsistent or conflicting messages. For example, both the Forest 
Service's National Report on Sustainable Forests--2003 and the Heinz 
Center's State of the Nation's Ecosystems include an indicator related 
to species rarity: the status (threatened, rare, vulnerable, 
endangered, or extinct) of forest-dependent species at risk of not 
maintaining viable breeding populations, as determined by legislation 
or scientific assessment and at-risk native forest species, 
respectively. However, though the datasets appear to be similar, the 
data in each set are presented in different ways and could appear 
confusing--even contradictory--to a reader unfamiliar with the 
different risk classification schemes used.

Moreover, even as federal activity developing indicator sets is 
increasing, developers at the various agencies may be missing 
opportunities to share knowledge and transfer experience. Federal 
developers have little to no access to best practices and lessons 
learned through others' experience with indicator sets needed to 
optimize the federal investment in this activity. Despite the extensive 
federal involvement in developing environmental indicators over the 
past decade, no clearinghouse has been established for collecting, 
classifying, and distributing information on best practices and lessons 
learned, either within or outside of the federal government. Experts 
involved in our meeting on environmental indicator sets said that such 
a clearinghouse could help developers avoid the sometimes duplicative 
time and resources currently devoted to identifying the elements of 
effective indicator sets. Several federal agencies have acknowledged 
the need for such and have begun taking initial actions to address this 
need. For example, the Forest Service's Northeastern Area State and 
Private Forestry unit recently developed a sourcebook and an Internet-
based clearinghouse to disseminate information for states and other 
organizations to use when attempting to use indicators for assessing 
forest sustainability.

Recognizing the need for improved coordination at the highest federal 
levels, the Interagency Working Group on Indicator Coordination was 
created at the request of the Chairman of CEQ in a December 31, 2002, 
memo. One purpose of the National Environmental Policy Act (NEPA) is to 
enrich the understanding of ecological systems and natural resources 
important to the nation.[Footnote 19] The act requires that CEQ review 
and appraise federal programs and activities to determine the extent to 
which these activities are achieving the purposes of NEPA and to make 
appropriate recommendations to the President. In addition, NEPA 
requires CEQ to document and define changes and trends in the natural 
environment, and accumulate the necessary data and other information 
for a continuing analysis of such changes and trends and an 
interpretation of their underlying causes.

The Interagency Working Group on Indicator Coordination is composed of 
representatives from the Departments of Agriculture, Commerce, Defense, 
Health and Human Services, the Interior, and Transportation, as well as 
EPA and the White House Offices of the Federal Environmental Executive, 
Management and Budget, and Science and Technology Policy. The Working 
Group first met in March 2003 to consider ways to enhance the nation's 
capacity to regularly report on natural and environmental resources, as 
well as related health, social, and economic factors, using a 
comprehensive set of indicators. It is currently considering a National 
System of Indicators on Natural and Environmental Resources, and is 
studying ways to improve institutional arrangements among the federal 
agencies for statistical reporting of such indicators.

The Working Group has developed an approach and policy framework for 
developing a national indicator system by building on existing federal 
and nonfederal efforts and has agreed that the system is a long-term 
goal. Furthermore, the Integration and Synthesis Group, an effort to 
coordinate several key federal "building block" indicator sets[Footnote 
20] under the leadership of the Working Group, has begun to develop a 
systems-based framework to organize environmental and natural resource 
indicators and provide a strong theoretical foundation for future 
integration work. The Working Group has also agreed on a general 
conceptual framework to guide the selection and use of indicators and 
is working to reach agreement on a detailed architecture to guide the 
management and use of data and information technology resources, and 
institutional arrangements to develop and operate a national system of 
indicators.

Officials of agencies participating in the Working Group acknowledge 
the need for a more stable structure with the authority and resources 
necessary to achieve the Working Group's goals. In this regard, as an 
ad hoc organization within the Executive Office of the President, the 
CEQ Working Group lacks a stable institutional arrangement with 
explicit responsibility and authority to ensure the continued and full 
involvement, cooperation, and resources from other federal agencies.

Experts participating in our two-day meeting on environmental indicator 
sets hosted by the National Academy of Sciences--including officials 
from CEQ, EPA, NOAA, the U.S. Geological Survey, and the Forest Service 
(within the Department of Agriculture)--discussed a number of different 
structures that could be employed to create a lead organization 
responsible for coordinating and integrating environmental indicator 
sets. Specifically, they discussed models ranging from using an 
executive order to build upon existing activity to creating a new 
quasi-governmental organization with the authority to oversee the 
development of a national environmental indicator system. In 
particular, the experts emphasized the importance of credibility and 
transparency as keys to the success of such an endeavor, in addition to 
authorities for addressing the widespread challenges of developing 
coordinated federal environmental indicator sets and ensuring the 
continued and full involvement, cooperation, and resources of the 
federal agencies. The experts did not settle on any particular 
approach, but instead noted that all of the options available should be 
studied to determine which option or combination of options is most 
appropriate. Furthermore, they generally agreed that whatever 
institutional arrangements are developed should be capable of 
performing the following functions:

* designing an information architecture using the best available 
information technology;

* providing leadership, vision, and overall scope;

* providing guidance and coordination with regard to environmental 
indicator development and use;

* assisting in environmental indicator selection, development, 
improvements, and evaluation;

* designing and managing data collection and monitoring, including 
consolidation and prioritization (identifying potential data sources, 
identifying areas where no data exist, and establishing ways to fill 
data gaps to support environmental indicators);

* organizing statistical compilation and reporting (connecting data to 
environmental indicator sets);

* identifying environmental research and development focus areas--
including environmental indicator methods--and developing and 
investigating conceptual frameworks, statistical methods, 
interpretation, assessment, diagnosis, and basis for interpretation;

* interpreting environmental indicators for planning, policy, 
management, and communication purposes; and:

* conducting audience analysis and public engagement to understand what 
information is needed to support outside entities.

Linking Environmental Management Actions and Program Activities to 
Changes in Environmental Conditions and Trends:

Environmental indicator sets are developed for many purposes, including 
tracking progress toward environmental goals and program performance. 
However, organizations that develop environmental indicator sets to 
improve the performance of environmental management programs can 
encounter challenges that inhibit the use of indicator sets in this 
context. Specifically, organizations encounter problems accounting for 
(1) causal relationships between management actions and other factors 
beyond the agency's control that can potentially affect environmental 
changes and (2) the delay between management actions and achieved 
results. Because complex webs of variables interact to determine 
ecological and human health outcomes, the role of a particular program 
in shaping environmental or natural resource conditions cannot always 
be determined. Organizations sometimes rely on indicator sets as 
diagnostic tools to highlight problem areas requiring further study, 
rather than as direct measures of performance, because indicator sets 
generally demonstrate a correlative--rather than causal--relationship 
between specific policies or programs and environmental conditions. 
Moreover, management actions can take many years to yield environmental 
results. A developer reported concern that the conditions and trends 
measured in their indicator sets would be used to determine funding 
allocations without regard to the long-term nature of environmental 
programs.

Such problems are consistent with our work on performance measurement 
in general. We reported in a June 1997 report on GPRA that the limited 
or indirect influence that the federal government sometimes has in 
determining whether a desired result is achieved complicates the effort 
to identify and measure the discrete contribution of the federal 
initiative to a specific program result.[Footnote 21] Our March 2004 
review of GPRA explained that this impediment occurs primarily because 
many federal programs' objectives are the result of complex systems or 
phenomena outside the program's control. In such cases, it is 
particularly challenging for agencies to confidently attribute changes 
in outcomes to their program--the central task of program impact 
evaluation.[Footnote 22] Our January 2001 report on management 
challenges at EPA noted that environmental programs may not yield 
measurable results for many years into the future.[Footnote 23] 
However, our prior work also discussed best practices for addressing 
challenges to measuring the results of such programs. In particular, to 
address the challenge of discerning the impact of a federal program, 
when other factors also affect results, we suggested agencies establish 
a rationale of how the program delivers results. Establishing such a 
rationale involves three related practices: (1) taking a holistic or 
"systems" approach to the problem being addressed, (2) building a 
program logic model that described how activities translated to 
outcomes, and (3) expanding program assessments and evaluations to 
validate the model linkages and rationale.

EPA's recent attempts to develop a set of environmental indicators 
illustrate the difficulties in linking management actions with the 
environmental results of such actions.[Footnote 24] In November 2001, 
at the direction of its Administrator, EPA embarked on a major effort-
-called the Environmental Indicators Initiative--to develop an 
assessment of the nation's environmental conditions and trends to 
enhance the agency's efforts to manage for environmental results, and 
to identify data gaps and the research and information collection 
efforts needed to fill those gaps. EPA's long-term goal for the 
initiative was to improve the data and indicators that are being used 
to guide its strategic plans, priorities, performance reports, and 
policy and management decisions.[Footnote 25] EPA's initiative, which 
resulted in the publication of its Draft Report on the Environment 
2003, seeks to provide a coherent picture of the nation's environment. 
This initiative is a major step toward developing indicators to provide 
a better understanding of the status and trends in human health and 
environmental conditions, as well as the more traditional measures of 
air, water, and land conditions. While EPA's two independent science 
advisory organizations--the Science Advisory Board and the National 
Advisory Council for Environmental Policy and Technology--have 
identified data limitations and other problems with the draft report, 
they commended EPA for its efforts and strongly recommended that EPA 
finalize the report after making needed revisions and improvements. 
According to EPA, work on EPA's next Report on the Environment--
scheduled for release in the summer of 2006--is currently under way. 
The next report will continue the efforts to develop a more 
comprehensive set of environmental indicators that could be used for a 
variety of purposes. EPA plans to include a set of regional 
environmental indicators in the next report that enhances the 
comprehensiveness of the indicators at multiple geographic scales. EPA 
is also working to integrate environmental information into a variety 
of planning processes. For example, the Office of Environmental 
Information and the Office of the Chief Financial Officer are currently 
working to link the forthcoming Report on the Environment 2006 to the 
agency's strategic planning effort.

EPA's recent actions represent noteworthy progress, but the agency 
still has considerable distance to travel and important challenges to 
overcome in developing a set of national environmental indicators 
useful for establishing priorities, allocating resources, and assessing 
environmental results. Since our 1998 report on EPA's management, GAO 
has stressed numerous times that EPA place priority on developing 
indicators to manage for results. In this regard, the few outcome 
measures in EPA's collection of performance metrics is largely a 
reflection that scientific knowledge essential to permit outcome 
measurement is often lacking, and that significant time lags often 
exist between actions taken to protect and improve the environment and 
demonstrable effects. In the absence of measures to detect and assess 
changes in the environment that could be supported with data, it 
becomes a matter of judgment as to how efficiently and effectively EPA 
is using its resources to address the nation's environmental problems.

Even with the agency's recent progress toward developing better outcome 
measures, EPA continues to face substantial challenges in understanding 
and describing the complex relationships among its programs, specific 
environmental pollutants, and human health and ecological conditions. 
EPA plans to continue developing and refining its indicator set as it 
seeks to clarify more fully the linkages between environmental 
pollution and other factors with human health and ecological 
conditions. To do so, it must continue to work to obtain credible and 
reliable environmental data from its own and other federal and 
nonfederal databases to support the indicators framework laid out in 
the Draft Report on the Environment. This task will involve continued 
collaborative effort with other federal, state, and tribal agencies.

As we reported in January 2003,[Footnote 26] EPA's progress in managing 
for results, particularly in describing current conditions and trends 
and identifying and filling research and data gaps, hinges on its 
efforts to translate its vision into specific actions. Such actions 
include establishing target dates for meeting specific milestones, 
identifying and obtaining sufficient staff and financial resources, and 
developing a structured approach for establishing direction, setting 
priorities, and measuring performance. Identifying and implementing 
specific actions aimed at better managing for results by developing and 
using environmental measures in planning, budgeting, and evaluating 
results continues to be difficult for EPA. The agency's earliest 
attempts to do so date back to 1974 and, in 1990, the agency made 
measuring changes in environmental conditions and trends a major policy 
and operational focus for the agency. These previous efforts to develop 
and use environmental indicators illustrate both the importance and 
difficulty of doing so, and the need for a focused, long-term 
commitment as changes occur in the agency's senior management and 
priorities. The previous EPA efforts have been hindered not only by 
technical difficulties in establishing linkages between program 
activities and changes in the environment, but also by changes in 
leadership within the agency and the lack of needed resources for 
monitoring environmental conditions. Monitoring activities have had 
trouble in competing for limited resources with EPA's regulatory 
programs and activities.

Recently, the Administrator of EPA has endorsed the continuation of the 
agency's indicators initiative in principle, and EPA has included the 
initiative as a performance measure in its annual performance plan for 
data quality activities. In addition, two of EPA's external scientific 
advisory organizations--the Science Advisory Board and the National 
Advisory Council for Environmental Policy and Technology--have lauded 
EPA's efforts thus far. Nonetheless, thus far the initiative--managed 
by EPA's Office of Information and Office of Research and Development-
-is not a key component of an agencywide comprehensive approach for 
identifying priorities, focusing resources on the areas of greatest 
concern, and managing EPA's work to achieve measurable results. For 
example, EPA has not initiated or planned an institutional framework 
with clear lines of responsibility and accountability among its various 
program offices and other organizational components for developing and 
using environmental indicators. Consequently, EPA has no systematic 
means to ensure that its efforts to identify environmental conditions 
and trends are used to inform priorities, strategic plans, allocation 
of resources, and agency reporting systems to establish accountability 
for EPA's efforts and determine whether programs and activities are 
having desired results, or need to be modified to better address the 
agency's priorities.

Conclusions:

Despite decades of activity and billions of dollars of investment, the 
nation is not yet capable of producing a comprehensive picture of 
environmental or natural resource conditions or trends. Federal and 
nonfederal organizations are developing and using environmental 
indicator sets to identify data gaps and bridge the gap between needed 
and available information. Despite several significant challenges, 
users of the indicator sets that we reviewed reported positive impacts 
in enhancing strategic planning efforts, communicating complex 
environmental issues, and tracking progress toward environmental goals. 
However, it is difficult to determine the benefits that arise from 
these impacts because environmental indicator sets themselves do not 
create change from which benefits can be measured. Rather, indicator 
sets might influence environmental management activities and thus yield 
benefits from affecting the quality of a decision. Much research 
remains to be done on how to better gauge returns on the investment 
made by organizations that have developed indicator sets. Nevertheless, 
the picture of the nation's environmental conditions and trends remains 
incomplete, as indicator set developers struggle to obtain sufficient 
data and coordinate their efforts with those of other set developers.

Federal agencies moving toward developing sets of environmental 
indicators face several major common challenges. These challenges 
include selecting the most appropriate indicators and sustaining a 
balanced process over time, linking the environmental outcomes 
represented by the indicators to steer specific environmental programs, 
enhancing the compatibility and coverage of environmental data, and 
overcoming obstacles to coordinating and integrating indicator sets to 
develop a comprehensive picture of the state of the nation's 
environment and natural resources. The refinement and usefulness of 
future sets of environmental indicators will largely depend on the 
extent to which these common challenges are resolved. Nonetheless, 
there is no entity with the authority, responsibilities, and resources 
to bring a concerted, focused, and systematic approach to addressing 
these common challenges and move toward a more fully systematic and 
integrated approach to developing federal sets of environmental 
indicators. Individual federal organizations may be missing 
opportunities to improve the quality of their indicator sets by not 
integrating their work with other similar efforts. Moreover, 
independently developing sets of indicators runs the risk of 
introducing increased possibilities of duplicating the activities of 
others.

Recognizing the need for a more coordinated approach to the federal 
investment in developing environmental indicator sets, CEQ's 
Interagency Working Group on Indicator Coordination is beginning to 
address challenges in developing environmental indicators sets. The 
Working Group is focused on developing institutional arrangements to 
provide the capacity and collaboration needed to produce and publish 
the indicator information, guide the selection and development of 
indicators and the organization of data for effective access and use, 
and develop processes for the coordination and integration of ongoing 
federal indicator development projects. However, the Working Group does 
not have a stable institutional arrangement with explicit 
responsibility and authority to ensure the continued and full 
involvement, cooperation, and resources from other federal agencies. 
Participants in our expert meeting convened by the National Academy of 
Sciences generally believed that the specific institutional 
arrangements utilized to coordinate and integrate federal environmental 
indicator projects should be carefully considered to ensure credibility 
of the outputs, both inside and outside the federal government. 
Moreover, they noted that specific key functions should be addressed, 
such as providing guidance for developing and using environmental 
indicators, designing an information architecture using the best 
available information technology, identifying the most crucial areas 
requiring environmental research, and assisting in environmental 
indicator selection, development, improvements, and evaluation.

We have long encouraged EPA to develop environmental indicators as a 
means to establish priorities, allocate resources, assess progress, 
and, in general, manage for environmental results. While we believe 
that EPA's Environmental Indicators Initiative and Draft Report on the 
Environment are a much-needed step in the right direction, this is not 
the first time the agency has tried to develop such environmental 
measures. The agency's successive efforts to develop and use 
environmental indicators since 1974 illustrate both the importance and 
difficulty of doing so and emphasize the need for dedicated, long-term 
commitment as changes occur in the agency's senior management and 
priorities. Given the complexity of the effort, a strong commitment to 
an institutional framework for developing and using indicators that 
emphasizes a systematic approach--including clear lines of 
responsibility and accountability among program offices and other 
organizational components and specific expectations, schedules, 
milestones, and resources--would better enable the agency's management 
to ensure that indicators of environmental conditions and trends are 
incorporated into EPA's efforts to plan strategically, allocate 
resources, and assess progress toward meeting environmental goals and 
objectives.

Recommendations for Executive Action:

To provide a comprehensive picture of environmental and natural 
resource conditions and trends to assess the nation's position and 
progress, we recommend that the Chairman of CEQ develop institutional 
arrangements needed to ensure a concerted, systematic, and stable 
approach to address the challenges associated with the development, 
coordination, and integration of environmental indicator sets. Such 
arrangements should be capable--either separately or jointly--of 
assisting in the development, selection, evaluation, and refinement of 
a national system of environmental indicators. The arrangements should 
provide for the coordination of federal data collection, monitoring, 
and statistical compilation activities, including consolidation and 
prioritization of data gaps, to support environmental indicators. 
Arrangements should also be capable of guiding and coordinating 
environmental indicator development and use, including creating a 
clearinghouse for best practices and lessons learned. The Chairman's 
strategy should incorporate the best available information technology 
to develop an information architecture for collecting, maintaining, and 
distributing environmental information. Moreover, the Chairman should 
provide for methods to identify environmental research and development 
focus areas. Finally, the system of arrangements should be designed to 
ensure the authority and credibility of its outputs.

Building on EPA's initial efforts on indicators and to evaluate the 
purposes that indicators might serve, we recommend that the EPA 
Administrator establish clear lines of responsibility and 
accountability among EPA's various organizational components and 
identify specific milestones, resources, and other requirements for 
developing and using environmental indicators to inform the agency's 
strategic systems for planning, budgeting, and reporting on progress.

Agency Comments and Our Evaluation:

We provided a draft of this report for review and comment to CEQ, the 
Departments of Agriculture and the Interior, EPA, and NOAA, all of 
which provided comment. Each of the agencies generally agreed with the 
report's findings and recommendations. Additional agency comments 
included the following:

* CEQ said that the report was a timely and comprehensive review of the 
many efforts underway, and that the report properly documents the many 
advancements and challenges recognized by experts. CEQ noted that the 
report should more clearly recognize that a comprehensive set of 
environmental indicators has the potential for benefiting environmental 
management governmentwide. We agree that environmental indicators stand 
to enhance management activities, such as strategic planning or 
resource allocation, across all federal agencies. Furthermore, CEQ 
commented that the report should make note of the Program Assessment 
Rating Tool, recently developed by the Office of Management and Budget, 
because it can enable both the executive and legislative branch of 
government to better understand program performance and identify 
opportunities for improvement. CEQ also noted that the report should 
make reference to the Global Earth Observation System of Systems--the 
international cooperative effort to bring together existing and new 
hardware and software to harmonize the supply data and information. We 
modified the report text as appropriate to incorporate these recent 
developments.

* The Department of Agriculture noted that the report effectively 
recognizes the need for better coordination of environmental indicator 
development and reporting among federal and nonfederal entities. Some 
Agriculture reviewers believed that, while the report emphasizes EPA's 
efforts in this area, many other agencies have authorities and 
responsibilities regarding environmental indicators. Additionally, 
Agriculture's Economic Research Service thought the report would have 
benefited from additional emphasis on the importance of coordinating 
behavioral and environmental data.

* The Department of the Interior noted that further efforts to identify 
institutional arrangements are essential given the unique 
characteristics and complex interrelationships among the range of 
agency programs noted in the report.

* EPA expressed some concern that the report implied that the Draft 
Report on the Environment 2003 was not successful in achieving its 
goals. We do not believe that the report makes such an implication, and 
we did not attempt to evaluate the success of the report in meeting its 
goals. Rather, we focused on the persistent need for the agency to 
provide clear lines of responsibility and accountability for meeting 
the goals of the Environmental Indicators Initiative--which produced 
the 2003 report--one of which was to improve the agency's ability to 
manage for results. EPA noted that it is currently working to link the 
planned Report on the Environment 2006 to the agency's strategic 
planning effort, and investigating other opportunities to link 
environmental information to management reporting and accountability 
systems. We modified the report text to better reflect these 
activities.

* NOAA questioned the practicality of coordinating the independent 
efforts of the many federal agencies currently collecting environmental 
monitoring data on coastal conditions. However, NOAA agreed that the 
report correctly characterizes the importance--as well as the 
difficulty--of doing so.

Finally, CEQ, the departments, and EPA recommended a number of 
technical changes to the report, which we incorporated as appropriate.

We are sending copies of this report to the Administrators of EPA and 
NOAA, the Chairman of the Council on Environmental Quality, the 
Secretaries of Agriculture and the Interior, and other interested 
parties. We also will make copies available to others upon request. In 
addition, the report will be available free of charge via the GAO Web 
site at [Hyperlink, http://www.gao.gov].

Should you or your respective staffs have any questions about this 
report, please contact me at (202) 512-6225, or Ed Kratzer, Assistant 
Director, at (202) 512-6553. Key contributors to this report are listed 
in appendix XI.

Signed by: 

John B. Stephenson: 
Director, Natural Resources and Environment:

[End of section]

Appendixes:

Appendix I: Objectives, Scope, and Methodology:

Specifically we were asked to report on the following questions: (1) 
How and for what purposes are federal and nonfederal organizations 
developing and using environmental indicator sets? And (2) What are the 
major challenges facing the development and use of environmental 
indicator sets?

For the purpose of this review, we defined an "environmental indicator 
set" as a selected group of quantifiable variables that shows a 
significant condition or trend of the state of the environment and 
natural resources, or related human activity. Our review focused 
primarily on the development and use of sets of environmental 
indicators, rather than on any single indicator. Our review included 
sets organized around environmental conditions and trends, ecological 
health, environmental performance, sustainable development, and 
corporate environmental information.

To meet our objectives, we performed multiple lines of work as detailed 
below, including reviewing literature on the development and use of 
environmental indicator sets; interviewing key experts from both the 
United States and abroad; developing a compendium of environmental 
indicator sets; surveying developers and users affiliated with 39 
environmental indicator sets at the national, state, regional, and 
local levels; conducting in-depth case studies of 8 indicator sets at 
the national, state, regional, and local levels; and contracting with 
the National Academy of Sciences to convene a meeting of experts. In 
developing our findings, we compiled evidence from across our lines of 
work to corroborate and "triangulate" salient themes. However, we did 
not intend to exhaustively catalog the universe of environmental 
indicator sets. General references to indicator sets in this report 
refer to the 47 sets we reviewed in detail--the 8 case studies and the 
39 sets represented in our survey results. Moreover, we did not 
evaluate the quality of data used in any of the indicator sets we 
reviewed, and we did not rely on these data for any of our findings. A 
thorough review of the data systems that support the indicator sets we 
reviewed was outside the scope of this project.

Compendium of Environmental Indicator Sets:

To identify a list of environmental indicator sets for review, we 
solicited input from experts in the field and asked them to identify 
indicator sets on four geographic scales--national, regional, state, 
and local. Forty-eight experts were selected from extensive Web and 
literature searches, background interviews, and contacts from 
professional conferences spanning our geographic scales. We distributed 
an electronic data collection instrument to each of the experts asking 
for information on environmental indicator sets with which they were 
familiar that either were being developed or had been developed in the 
past 10 years, states that have been or are currently leading the 
effort in developing and using environmental indicator sets, and a 
project contact person for each set. Twenty-three experts responded. We 
combined duplicate responses and eliminated responses that: (1) did not 
meet our definition of an indicator set or (2) could not be 
substantiated with enough information to locate a specific initiative. 
A pool of 87 environmental indicator sets was identified for review in 
detail (see app. II.)

Survey of Practitioners:

To develop a list of environmental indicator set developers and users-
-which we called practitioners--to survey, we contacted the points of 
contact at the 87 indicator sets identified by the experts and asked 
them to provide us with a developer and a user to receive our survey. 
This process yielded 87 practitioners to be surveyed, representing 52 
of the indicator sets. Forty-nine of the 87 practitioners responded to 
our survey for a 56 percent response rate. Table 5 provides summary 
information. The survey results are not necessarily representative of 
the entire population of environmental indicator set practitioners.

Table 5: Summary of Survey Participants:

Indicator sets: 

Indicator sets identified for survey; 
National: 17; 
Regional: 8; 
State: 14; 
Local: 13; 
Total: 52.

Indicator sets represented by a completed survey; 
National: 15; 
Regional: 4; 
State: 7; 
Local: 13; 
Total: 39.

Practitioners: 

Practitioners identified for survey; 
National: 28; 
Regional: 14; 
State: 23; 
Local: 22; 
Total: 87.

Practitioners that responded to the survey; 
National: 20; 
Regional: 6; 
State: 9; 
Local: 14; 
Total: 49.

Developer; 
National: 8; 
Regional: 3; 
State: 3; 
Local: 5; 
Total: 19.

User; 
National: 2; 
Regional: 1; 
State: 3; 
Local: 1; 
Total: 7.

Both; 
National: 10; 
Regional: 2; 
State: 3; 
Local: 8; 
Total: 23. 

Source: GAO.

[End of table]

We identified the areas to cover in the survey based on the assignment 
request, the Internet and literature searches, background interviews, 
and the professional conferences we had attended. The survey questions 
focused on the characteristics of the indicator set, how it was 
developed, and how the set is being used.

We pretested the survey with two developers and two users. We evaluated 
the appropriateness and quality of the survey questions and responses 
and tested the usability of the Internet-based survey. Based on the 
pretest results, we made the necessary changes to the survey prior to 
its implementation.

We administered the survey through the Internet. During our early 
efforts to determine whether we had accurate information on the survey 
population, we obtained their e-mail addresses. We used e-mail to 
inform the practitioners of the survey administration, and provided 
them with the Web link for the survey and their log-in name and 
password. To maximize the response rate, we sent an e-mail reminder and 
followed up by telephone to encourage survey participation.

The survey was structured in two separate sections: one for developers 
to complete and the other for users to complete. At least one developer 
or user from 39 of the 52 indicator sets completed our survey. However, 
some respondents answered the survey in a capacity other than how we 
originally classified them. The survey results for some indicator sets 
are represented with answers from two individuals. Given that the 
purpose of the survey was to gather general descriptive information on 
indicator sets and how they are developed and used, we do not believe 
that the multiple responses for some indicator sets greatly influence 
the survey results. Our survey of developers and users of environmental 
indicator sets and a more complete tabulation of the survey results 
(GAO-05-56SP) will also be available on the GAO Web site at 
www.gao.gov/cgi-bin/getrpt?GAO-05-56SP.

Case Study:

To contribute to our understanding of the development and use of 
environmental indicator sets, we reviewed 8 environmental indicator 
sets in-depth through case study. We selected two indicator sets for 
case study review at each of four geographic scales--national, 
regional, state, and local--from the pool of 87 indicator sets 
identified by experts. The selection of case studies for review was 
based on the level of maturity of the indicator set (current and 
active) and the availability and accessibility of individuals involved 
in the development and use of the indicator set. Table 6 provides a 
breakdown of the environmental indicator sets selected and the 
geographic scale that each set represents.

Table 6: Environmental Indicator Sets Selected for Case Study Review:

Case study name: The Heinz Center's State of the Nation's Ecosystems; 
Geographic scale: National.

Case study name: EPA's National Coastal Assessment; 
Geographic scale: National.

Case study name: Chesapeake Bay Program; 
Geographic scale: Regional.

Case study name: Great Lakes State of the Lakes Ecosystem Conference; 
Geographic scale: Regional.

Case study name: Minnesota's Department of Natural Resources Strategic 
Conservation Agenda; 
Geographic scale: State.

Case study name: Environmental Protection Indicators for California; 
Geographic scale: State.

Case study name: Quality of Life Indicator Set, Jacksonville, Florida; 
Geographic scale: Local.

Case study name: Environmental Indicators Project, West Oakland, 
California; 
Geographic scale: Local.

Source: GAO.

[End of table]

We conducted semistructured interviews with at least three individuals 
who were involved in the development, use, and data gathering 
activities of each environmental indicator set. An additional 
environmental indicator set was selected to test our interview 
questions. Semistructured interviews allowed interviewees the 
opportunity to openly and candidly supply information on a wide range 
of issues relating to their involvement with the development and use of 
the environmental indicator set. We also reviewed relevant documents 
pertaining to the development and use of each of the environmental 
indicator sets. In addition to providing evidence in the report, the 
case study information was used to construct case study profiles that 
were provided to the appropriate program manager for review. The 
profiles are in appendix III.

Meeting of Experts Convened by the National Academy of Sciences:

To assess the current state of environmental indicator set development 
and use, we contracted with the National Academy of Sciences (NAS) to 
host a 2-day meeting of experts. The selection of experts to 
participate in the meeting was a two-step process. First, we worked 
with the NAS staff to identify individuals with expertise in 
environmental indicator sets. After reviewing the background of each 
expert, we selected participants using the following criteria:

* balance of expertise (e.g., managers, data gathering, developers, 
users, scientists, researchers, and policymakers);

* balance of knowledge across various disciplines (e.g., natural 
resources, ecology, and agriculture); and:

* balance in representation (e.g., federal agencies, state agencies, 
academia, and nonprofit and private organizations).

Based on the availability of the selected participants, we invited 26 
experts--representing the geographic levels and sectors--to 
participate in the meeting held March 9-10, 2004, in Washington, D.C., 
all of whom attended. Prior to the meeting, we provided the selected 
experts with background materials that highlighted past reports written 
by GAO, the National Research Council, and other organizations 
addressing environmental indicator set issues. The following 26 experts 
participated in the meeting:

Albert Abee: 
Sustainable Development Coordinator: 
U.S. Forest Service:

James R. Bernard:
Environmental Management Consulting:

David Berry: 
Sustainable Water Resources Roundtable:

Zach Church: 
Pennsylvania Department of Environmental Protection, Policy Office:

J. Clarence Davies, Ph.D.: 
Senior Fellow: 
Resources for the Future:

Dennis Fenn, Ph.D.: 
Center Director: 
U.S. Geological Survey, Southwest Biological Science Center:

Keith G. Harrison, M.A., R.S., Certified Ecologist: 
Executive Director: 
Michigan Environmental Science Board: 
Special Projects Coordinator: 
Michigan Department of Environmental Quality:

R. Lee Hatcher: 
Managing Director:
AtKisson Inc.

Theodore Heintz: 
Indicator Coordinator: 
White House Council on Environmental Quality:

Rainer Hoenicke, Ph.D.: 
Environmental Scientist: 
San Francisco Estuary Institute:

Robert J. Huggett, Ph.D.: 
Professor of Zoology, Vice President for Research and Graduate Studies: 
Michigan State University:

Suellen Terrill Keiner, J.D.: 
Academy General Counsel and Vice President for Academy Programs: 
The National Academy of Public Administration:

Daniel Markowitz, Ph.D.: 
Associate: 
Malcolm Pirnie Inc.

Gary Matlock, Ph.D.: 
Director:
National Oceanic Atmospheric Administration, National Centers for 
Coastal Ocean Science:

Shelley Metzenbaum, Ph.D.: 
Executive Director: 
Environmental Compliance Consortium:

Visiting Professor: 
University of Maryland School of Public Affairs:

Patrick O'Brien, Ph.D.: 
Consulting Environmental Scientist:
Chevron-Texaco Energy Technology Company:

Robin O'Malley: 
Senior Fellow: 
The H. John Heinz III Center for Science, Economics and the 
Environment:

Gordon Orians, Ph.D.: 
Professor Emeritus: 
University of Washington: 
Department of Biology:

Duncan Patten, Ph.D.: 
Research Professor: 
Montana State University: 
Big Sky Institute:

Marcus Peacock: 
Associate Director: 
Office of Management and Budget, Natural Resources, Energy and Science:

Dee Peace Ragsdale: 
Performance and Recognition Manager: 
Washington Department of Ecology:

Mark Schaefer, Ph.D.: 
President and Chief Executive Officer: 
NatureServe:

Michael Slimak, Ph.D.: 
Associate Director for Environmental Ecology:
U.S. Environmental Protection Agency, National Center for Environmental 
Assessment:

Greg Wandrey, Ph.D.: 
Director of Product Stewardship: 
Pioneer Hi-Bred Inc.

John R. Wells: 
Sustainable Development Director:
Minnesota Environmental Quality Board:

Robin P. White, Ph.D.: 
Senior Associate: 
World Resources Institute:

During the meeting, experts participated in roundtable sessions and 
breakout groups to discuss the following:

* Why are organizations developing and using environmental indicator 
sets and what impacts are these sets having in the United States?

* What significant scientific, environmental data, communication, and 
institutional challenges hinder the development and use of 
environmental indicator sets?

* What actions could be taken to overcome the significant challenges to 
the development and use of environmental indicator sets?

The meeting was audio recorded to facilitate transcription. We reviewed 
the written transcript of the proceedings, the documents produced by 
experts, and other notes from the 2-day meeting to produce a summary 
document, which was provided to the experts for review. Their comments 
were incorporated into the summary, where appropriate. We used the 
summary document in preparing this report.

[End of section]

Appendix II: Key Environmental Indicator Initiatives Identified by 
Experts:

Indicator set initiative: Sustainable Development in the United States; 
Web site: http://clinton1.nara.gov/White_House/EOP/pcsd/ 
Scale: National.

Indicator set initiative: EPA--Draft Report on the Environment; 
Web site: http://www.epa.gov/indicators/ 
Scale: National.

Indicator set initiative: Sustainable Minerals Roundtable; 
Web site: http://www.unr.edu/mines/smr/ 
Scale: National.

Indicator set initiative: Sustainable Water Resources Roundtable; 
Web site: http://water.usgs.gov/wicp/acwi/swrr/ 
Scale: National.

Indicator set initiative: Roundtable on Sustainable Forests; 
Web site: http://www.sustainableforests.net/info.php; 
Scale: National.

Indicator set initiative: Sustainable Rangelands Roundtable; 
Web site: http://sustainablerangelands.cnr.colostate.edu/ 
Scale: National.

Indicator set initiative: State of the Nation's Ecosystems; 
Web site: http://www.heinzctr.org/ecosystems/ 
Scale: National.

Indicator set initiative: Ecological Monitoring and Assessment Program; 
Web site: http://www.epa.gov/emap/ 
Scale: National.

Indicator set initiative: Ecological Indicators for the Nation; 
Web site: http://books.nap.edu/catalog/9720.html; 
Scale: National.

Indicator set initiative: Index of Watershed Indicators; 
Web site: http://www.epa.gov/iwi/ 
Scale: National.

Indicator set initiative: Chemical and Pesticide Results Measures; 
Web site: http://www.pepps.fsu.edu/CAPRM/ 
Scale: National.

Indicator set initiative: Waste Indicator System for the Environment; 
Web site: http://www.pepps.fsu.edu/WISE/ 
Scale: National.

Indicator set initiative: America's Children and the Environment; 
Web site: http://www.epa.gov/envirohealth/children/ 
Scale: National.

Indicator set initiative: National Report on Human Exposure to 
Environmental Chemicals; 
Web site: http://www.cdc.gov/exposurereport/ 
Scale: National.

Indicator set initiative: Index of Leading Environmental Indicators; 
Web site: http://www.aei.org/publications/bookID.407/book_detail.asp; 
Scale: National.

Indicator set initiative: Agricultural Resource and Environmental 
Indicators; 
Web site: http://www.ers.usda.gov/publications/arei/ 
Scale: National.

Indicator set initiative: Environmental Public Health Indicators; 
Web site: http://www.cdc.gov/nceh/indicators/default.htm; 
Scale: National.

Indicator set initiative: Risk-Screening Environmental Indicators; 
Web site: http://www.epa.gov/opptintr/rsei/ 
Scale: National.

Indicator set initiative: The Status and Trends of Our Nation's 
Biological Resources; 
Web site: http://biology.usgs.gov/s+t/SNT/index.htm; 
Scale: National.

Indicator set initiative: National Coastal Condition Report; 
Web site: http://www.epa.gov/owow/oceans/nccr/ 
Scale: National.

Indicator set initiative: The Status of Biodiversity in the United 
States; 
Web site: http://www.natureserve.org; 
Scale: National.

Indicator set initiative: National Estuarine Reserves System Wide 
Monitoring Program; 
Web site: http://nerrs.noaa.gov/ 
Scale: National.

Indicator set initiative: National Coastal Management Performance 
Measurement System; 
Web site: http://www.ocrm.nos.noaa.gov/ 
Scale: National.

Indicator set initiative: National Park Service--Vital Signs Program; 
Web site: http://science.nature.nps.gov/im/monitor/index.htm; 
Scale: National.

Indicator set initiative: Relative Sea Level Trends; 
Web site: http://pubs.usgs.gov/of/2002/of02-233/ppvariables.htm; 
Scale: National.

Indicator set initiative: U.S. Land Cover Trends; 
Web site: http://gam.usgs.gov/LandUseDynamics/ludatacollection.shtml; 
Scale: National.

Indicator set initiative: Forest Health Monitoring Vegetation Indicator 
Pilot Program; 
Web site: http://www.fs.fed.us/na/briefs/fhm99/fhm99.htm; 
Scale: National.

Indicator set initiative: Chesapeake Bay Program; 
Web site: http://www.chesapeakebay.net; 
Scale: Regional.

Indicator set initiative: State of the Great Lakes Ecosystem 
Conference; 
Web site: http://www.epa.gov/glnpo/solec/ 
Scale: Regional.

Indicator set initiative: Environmental Indicators in the Estuarine 
Environment; 
Web site: http://www.aceinc.org/ 
Scale: Regional.

Indicator set initiative: Environmental Health Indicators for the U.S.-
-Mexico Border; 
Web site: http://www.fep.paho.org/english/env/Indicadores/IndSA.htm; 
Scale: Regional.

Indicator set initiative: New England Environmental Goals and 
Indicators Project; 
Web site: http://www.gmied.org; 
Scale: Regional.

Indicator set initiative: Western Regional Climate Center; 
Web site: http://www.wrcc.dri.edu/ 
Scale: Regional.

Indicator set initiative: Puget Sound/Georgia Basin Ecosystem 
Indicators; 
Web site: http://www.ecy.wa.gov/biblio/0201002.html; 
Scale: Regional.

Indicator set initiative: Southeastern Louisiana Top 10 by 2010 
Indicators Report; 
Web site: http://www.top10by2010.org/ 
Scale: Regional.

Indicator set initiative: North State (California) Vital Signs; 
Web site: http://www.mcconnellfoundation.org/ 
Scale: Regional.

Indicator set initiative: Mid-Atlantic Integrated Assessment; 
Web site: http://www.epa.gov/emap/maia/ 
Scale: Regional.

Indicator set initiative: South Florida/Everglades Comprehensive 
Ecosystem Restoration Plan; 
Web site: http://www.evergladesplan.org/ 
Scale: Regional.

Indicator set initiative: Tennessee Valley Authority Vital Signs 
Program; 
Web site: http://www.tva.gov/environment/reports/envreports/index.htm; 
Scale: Regional.

Indicator set initiative: Pacific Northwest Salmon Habitat Indicators; 
Web site: http://www.ecy.wa.gov/biblio/99301.html; 
Scale: Regional.

Indicator set initiative: Aquatic Habitat Indicators for the Pacific 
Northwest; 
Web site: http://yosemite.epa.gov/R10/ecocomm.nsf/0/
74476bae1ae7e9fb88256b5f00598b43?OpenDocument; 
Scale: Regional.

Indicator set initiative: Tampa Bay Estuary Program Baywide 
Environmental Monitoring Report; 
Web site: http://www.tbep.org/baystate/bemr.html; 
Scale: Regional.

Indicator set initiative: Ecosystem Indicators for the Lake Champlain 
Basin Program; 
Web site: http://www.uvm.edu/envnr/indicators/ 
Scale: Regional.

Indicator set initiative: Environmental Protection Indicators for 
California; 
Web site: http://www.oehha.ca.gov/multimedia/epic/ 
Scale: State.

Indicator set initiative: Minnesota Environmental Indicators; 
Web site: http://www.dnr.state.mn.us/eii/index.html; 
Scale: State.

Indicator set initiative: Minnesota Strategic Conservation Agenda; 
Web site: http://www.dnr.state.mn.us/conservationagenda/index.html; 
Scale: State.

Indicator set initiative: Central Texas Sustainability Indicators 
Initiative; 
Web site: http://www.centex-indicators.org/ 
Scale: State.

Indicator set initiative: Pennsylvania Environmental Futures Planning; 
Web site: http://www.dep.state.pa.us/hosting/efp2/PDF_ICF_EFP2X/
priorities.htm; 
Scale: State.

Indicator set initiative: State of the Texas Environment Report; 
Web site: http://www.tnrcc.state.tx.us/ 
Scale: State.

Indicator set initiative: Texas Index of Leading Environmental 
Indicators 2000; 
Web site: http://www.texaspolicy.com/
research_reports.php?report_id=143&loc_id=1; 
Scale: State.

Indicator set initiative: Texas Environmental Almanac; 
Web site: http://www.texascenter.org/almanac/ 
Scale: State.

Indicator set initiative: Water for Texas; 
Web site: http://www.twdb.state.tx.us/publications/reports/
State_Water_Plan/2002/FinalWaterPlan2002.htm; 
Scale: State.

Indicator set initiative: Utah Air Monitoring--Mobile Sources; 
Web site: http://www.airmonitoring.utah.gov/amc.htm; 
Scale: State.

Indicator set initiative: Ambient Air Monitoring Program; 
Web site: http://www.epa.gov/air/oaqps/qa/monprog.html; 
Scale: State.

Indicator set initiative: Minnesota Milestones; 
Web site: http://www.mnplan.state.mn.us/mm/ 
Scale: State.

Indicator set initiative: Oregon Shines; 
Web site: http://egov.oregon.gov/DAS/OPB/os.shtml; 
Scale: State.

Indicator set initiative: Florida Assessment of Coastal Trends; 
Web site: http://www.pepps.fsu.edu/FACT/ 
Scale: State.

Indicator set initiative: Washington Department of Ecology; 
Web site: http://www.ecy.wa.gov/ 
Scale: State.

Indicator set initiative: Oregon State of the Environment Report; 
Web site: http://egov.oregon.gov/DAS/OPB/soer2000index.shtml; 
Scale: State.

Indicator set initiative: State of Kentucky's Environment; 
Web site: http://www.eqc.ky.gov/pubs/soke/ 
Scale: State.

Indicator set initiative: Illinois Department of Environmental Quality 
indicators; 
Web site: http://www.dnr.state.il.us/orep/NRRC/balancedgrowth/
indicators.htm; 
Scale: State.

Indicator set initiative: Environmental Indicators for Delaware 
Estuary; 
Web site: http://www.epa.gov/owow/estuaries/coastlines/jan02/
envindicator.html; 
Scale: State.

Indicator set initiative: Indicators of Livable Communities; 
Web site: http://www.mdf.org/megc/pubs/livable_communities.htm; 
Scale: State.

Indicator set initiative: Oregon's First Approximation Report; 
Web site: http://www.oregonforestry.org/sustainability/
first_approximation_report.htm; 
Scale: State.

Indicator set initiative: Sustainable Development Indicators for 
Pennsylvania; 
Web site: http://www.paconsortium.state.pa.us/
pointing_pa_sustainable_future. htm; 
Scale: State.

Indicator set initiative: New Jersey Hudson Bay Environmental 
Indicators Initiatives; 
Web site: http://www.harborestuary.org/reports/harborh.htm; 
Scale: State.

Indicator set initiative: Everglades Comprehensive Annual Report; 
Web site: http://www.sfwmd.gov/org/ema/everglades/ 
Scale: Local.

Indicator set initiative: The State of the Bay--a Characterization of 
the Galveston Bay Ecosystem; 
Web site: http://www.tnrcc.state.tx.us/admin/topdoc/pd/020/02-04/
galvestonbay.html; 
Scale: Local.

Indicator set initiative: Index of Silicon Valley; 
Web site: http://www.jointventure.org/resources/2002Index/ 
Scale: Local.

Indicator set initiative: Santa Monica Sustainable City Plan; 
Web site: http://santa-monica.org/epd/scp/ 
Scale: Local.

Indicator set initiative: Current Status and Historical Trends of 
Selected Estuarine and Coastal Habitats in Corpus Christi Bay National 
Estuary Program Study Area; 
Web site: http://www.sci.tamucc.edu/ccs/ 
Scale: Local.

Indicator set initiative: Bay Area Alliance for Sustainable 
Communities; 
Web site: http://www.bayareaalliance.org/ 
Scale: Local.

Indicator set initiative: Bay Institute; 
Web site: http://www.bay.org/main.htm; 
Scale: Local.

Indicator set initiative: Bay Area EcoAtlas and Pulse of the Bay 
report; 
Web site: http://www.sfei.org/ 
Scale: Local.

Indicator set initiative: Mecklenburg County State of the Environment 
Report; 
Web site: http://www.charmeck.org/Departments/LUESA/
Water+and+Land+Resources/State+of+the+Environment+Report.htm; 
Scale: Local.

Indicator set initiative: Sustainable Seattle--Indicators of 
Sustainable Community; 
Web site: http://www.sustainableseattle.org/Publications/
40indicators.shtml; 
Scale: Local.

Indicator set initiative: Legacy 2002--Greater Orlando; 
Indicators Report; 
Web site: http://www.hcbs.org/moreInfo.php/source/62/sby/Author/doc/
251/Legacy_2002_-_Greater_Orlando_Indicator's_Report_-; 
Scale: Local.

Indicator set initiative: Sierra Nevada Wealth Index; 
Web site: http://www.sbcouncil.org/wealth.htm; 
Scale: Local.

Indicator set initiative: Sustainable Nantucket--a Compass for The 
Future; 
Web site: http://indicators.sustainablenantucket.org/intro.cfm; 
Scale: Local.

Indicator set initiative: Community-based Environmental Health 
Assessment Program; 
Web site: http://www.naccho.org/general955.cfm; 
Scale: Local.

Indicator set initiative: Multnomah County--Benchmarks; 
Web site: http://www.portlandonline.com/auditor/
index.cfm?&a=39665&c=27347; 
Scale: Local.

Indicator set initiative: Valley Vision (California); 
Web site: http://www.calregions.org/civic/partners/mid-vvr.html; 
Scale: Local.

Indicator set initiative: King County Benchmarks; 
Web site: http://www.metrokc.gov/budget/benchmrk/bench03/ 
Scale: Local.

Indicator set initiative: State of Boston Harbor; 
Web site: http://www.mwra.state.ma.us/harbor/html/2002-09.htm; 
Scale: Local.

Indicator set initiative: West Oakland--Environmental Indicators; 
Web site: http://www.neip.org/ 
Scale: Local.

Indicator set initiative: Jacksonville Community Council Inc. Quality 
of Life Indicators; 
Web site: http://www.jcci.org; 
Scale: Local. 

Source: GAO.

Note: Web addresses are current as of August 10, 2004.

[End of table]

[End of section]

Appendix III: Environmental Indicator Set Case Study Profiles:

We conducted eight in-depth case studies of environmental indicator 
sets over the course of the review. We reviewed two environmental 
indicator sets at each of the following geographic scales: national, 
regional, state, and local. The indicator sets profiled are:

1. The Heinz Center's State of the Nation's Ecosystems;

2. EPA's National Coastal Assessment;

3. Chesapeake Bay Program;

4. Great Lakes State of the Lakes Ecosystem Conference;

5. Minnesota's Department of Natural Resources Strategic Conservation 
Agenda;

6. Environmental Protection Indicators for California;

7. Quality of Life Indicator Set, Jacksonville, Florida; and:

8. Environmental Indicators Project, West Oakland, California.

Each profile contains a brief overview of the program, the process of 
development, the use and impact of the indicator set, and next steps 
planned for the indicator set.

The Heinz Center's State of the Nation's Ecosystems:

Overview: In early 1997, as a follow up to a major review of federal 
environmental monitoring efforts, the White House Office of Science and 
Technology Policy (OSTP) requested that the H. John Heinz III Center 
for Science, Economics, and the Environment (Heinz Center)--a nonprofit 
institution--develop a nonpartisan, science-based report on the state 
of the nation's environment.

The Heinz Center lists 103 indicators in the set, with approximately 15 
indicators for each of 6 major ecosystem types (Coasts and Oceans, 
Farmlands, Forests, Fresh Waters, Grasslands and Shrublands, Urban and 
Suburban Areas) and 10 additional core national indicators that provide 
a broad yet succinct view of the national ecosystem condition and use. 
The indicator set is national in scope with limited breakout by 
regions. The indicators focus on the condition of ecosystems that 
support policy debate and decision-making at the national scale. The 
environmental indicator set information was disseminated through a 
report in 2002 (see fig. 5) that was issued simultaneously in print and 
on the Web.

Figure 5: The State of the Nation's Ecosystems Report:

[See PDF for image]

[End of figure]

Development: The Heinz Center assembled a small in-house staff and a 
large team of part-time contributors drawn from government, the private 
sector, environmental organizations, and academia. A design committee 
oversaw the entire project and technical work groups, which provided 
expertise in particular ecosystems, identified the indicators and 
selected and assessed the data sources. Overall, nearly 150 individuals 
participated in the project as committee and group members, with many 
more participating as contributors, reviewers, and advisers.

The committee selected indicators that could provide a broad, balanced 
description of each ecosystem type based on 10 characteristics that 
covered the physical dimensions of the systems, their chemical and 
physical conditions, the status of their biological components, and the 
amounts of goods and services people receive from them. Once the 
committee chose an indicator and identified relevant sources of data, 
it reviewed the data based on the following three criteria: (1) Data 
had to be of sufficient quality to provide a scientifically credible 
description of actual ecosystem conditions; (2) data had to have 
adequate geographic coverage to represent the state of the nation's 
ecosystems; and (3) data had to be collected through an established 
monitoring program that offered a reasonable likelihood of future data 
availability.

Use: The indicator set highlights the need for a comprehensive view of 
ecosystem condition and change and the need for additional information 
to fill the gaps in data available to describe key aspects of the 
nation's ecosystems. The major use to date has been by managers of 
major monitoring systems who are using it in designing their collection 
and reporting systems.

Next steps: The 2002 report was the first in what is intended to be a 
regular series of reports on the state of the nation's ecosystems every 
5 years. The next edition in the series is planned for issuance in 
2007. Between the issuance of major editions, substantial revisions--
such as the incorporation of new data sets--will be issued in a 
periodic update on the Web. Before the next version is published, Heinz 
Center staff will fill data gaps and improve the consistency of both 
data and indicators; consult with key scientific communities in order 
to refine and clarify certain indicators; work with public and private 
agencies to regularly provide data in the form needed for national 
reporting; and strengthen the linkages between the Heinz Center project 
and other efforts related to ecosystem reporting.

EPA's National Coastal Assessment:

Overview: In 1988, the Environmental Protection Agency's (EPA) Science 
Advisory Board charged the Office of Research and Development (ORD) to 
develop a nationally consistent way to report on the condition of 
coasts for the purpose of Clean Water Act Section 305(b) reporting. 
ORD's Environmental Monitoring and Assessment Program (EMAP), which 
involved the efforts of several other federal agencies, developed the 
National Coastal Assessment (NCA) indicator set and monitoring program. 
The program was implemented in 2000 as a 5-year effort to evaluate the 
assessment methods and environmental indicators that ORD had developed 
to advance the science of ecosystem condition monitoring and 
evaluation. The program created an integrated, comprehensive coastal 
monitoring program and environmental indicator set among the coastal 
states to assess the condition of the nation's estuaries and offshore 
waters. Through strategic partnerships with 24 coastal states using a 
compatible, probabilistic design and a common set of survey indicators, 
each of the 24 states involved in the NCA program have conducted the 
survey and assessed the conditions of their respective coastal 
resources. These assessments in turn can be aggregated to assess 
conditions at the EPA regional, biogeographical, and national levels.

The NCA includes five aggregate indicators--water quality, sediment 
quality, coastal habitat, benthic community structure, and fish tissue 
contaminants--based on 200 to 250 separate measurements. The indicators 
cover a range of geographic scales--state, regional, biogeographical, 
and national. The indicators focus on showing the condition of 
estuaries and the association between condition and stressors 
(effects). As such, the indicators are based on science rather than on 
administrative policy performance. The states report the indicators 
through state Section 305(b) reports to EPA, which submits them to 
Congress. The indicators are also aggregated with other data collection 
efforts and reported through the National Coastal Condition Report (see 
fig. 6).

Figure 6: Draft National Coastal Condition Report II:

[See PDF for image]

[End of figure]

Development: A number of pilot projects held over a 10-year period at 
different geographic areas, helped identify and develop the indicators. 
The indicators were developed based on 15 guidelines organized around 
four evaluation phases: conceptual relevance, feasibility of 
implementation, response variability, and interpretation and utility.

Use: The NCA indicator set and monitoring program are used by 24 marine 
coastal states and Puerto Rico to provide an assessment of estuary 
conditions for the purposes of Clean Water Act Section 305(b) 
reporting. Before development of NCA, states or territories had little 
or no coastal monitoring in place and no mechanism to evaluate the 
condition of the resource. The NCA indicators provided states with a 
small set of indicators that are adaptable to the specific needs of the 
state utilizing them. Three coastal states have fully implemented the 
NCA monitoring and indicator approach to fulfill Section 305(b) 
reporting requirements; the other 21 states either are just beginning 
to implement the approach or have used the approach to assess a part of 
their estuaries.

Next steps: The 5-year NCA program is set to expire in the summer of 
2004, after which the EPA Office of Water may take over the program. At 
the end of the period, ORD officials will evaluate the effectiveness of 
the program and provide assistance to the Office of Water as needed. 
ORD is currently structuring monitoring programs and indicator 
development to provide tools to states to monitor and evaluate not only 
the conditions of waters for reporting purposes (Section 305(b)) but 
also for other provisions of the Clean Water Act, such as nonpoint 
source control (Section 319), Total Maximum Daily Loads allocation 
(Section 303(d)), and the National Pollutant Discharge Elimination 
System permitting program (Section 402).

Chesapeake Bay Program:

Overview: In 1991, the Chesapeake Bay Program Office, headed by EPA, 
began developing environmental indicators to support goal setting, to 
define targets and end points for restoration of the bay, and to make 
the program more accountable to the public by defining and 
communicating the bottom-line environmental results achieved by the 
restoration program. The EPA coordinates the development, revision, and 
updates of the environmental indicators with more than 50 federal, 
state, and local government agencies and nongovernmental organizations 
that participate as bay program partners. The bay program carries out 
its work through a series of committees, advisory committees, and 
subcommittees.

A basic tenet of the bay program's environmental indicators effort is 
that environmental indicators (outcome measures) need to be clearly 
associated with strategic goals for the program. As such, the bay 
program has developed a framework for linking environmental outcome 
measures to strategic program goals. The Chesapeake Bay Program 
currently utilizes nearly 90 environmental indicators to gauge the 
Chesapeake Bay's environmental condition and progress made in 
restoration. The Chesapeake Bay Program organizes the indicators into 
six levels that range from indicators that measure management actions-
-such as implementing advanced treatment of wastewater to reduce 
nutrient discharges--to those that are direct or indirect measures of 
ecological or human health. The indicators are further categorized into 
a performance measure; context indicator; emerging science indicator; 
or pressure, state, or response indicator (these indicators are based 
on a concept of causality, where human activities place pressures on 
the environment that cause a change in the state of the environment; 
these changes alert society, which then implements a response to reduce 
the pressures or to change the affected environment). Environmental 
indicator set information is reported by a variety of mechanisms, such 
as briefing packages, presentations, and fact sheets, and a triennial 
State of the Chesapeake Bay report (see fig. 7).

Figure 7: The State of the Chesapeake Bay Report:

[See PDF for image]

[End of figure]

Development: The process of developing and subsequently adding, 
deleting, or modifying indicators is collaborative and includes 
hundreds of individuals working through bay program committees, 
subcommittees, and work groups. The criteria for indicator selection 
are: (1) data availability; (2) environmental results measure; (3) 
management needs; (4) and request of the public. Indicators are 
developed to measure performance of restoration goals, which have been 
primarily established through three overriding Chesapeake Bay 
agreements. The most recent of the agreements--the Chesapeake 2000 
agreement--establishes many goals to be achieved by 2010.

Use: Goal setting through Chesapeake Bay agreements has given the 
Chesapeake Bay Program an important tool to develop and use indicators 
that improve its ability to garner and target resources and to evaluate 
the bay program's management strategies. The indicator set also 
presents information to the public on the condition of the Chesapeake 
Bay through various reporting mechanisms. The environmental indicator 
set has supported goal setting for the bay program both in longer-term 
strategic implementation plans and for annual planning and budgeting.

Next steps: The Chesapeake Bay Program office plans to develop more 
river-specific or subwatershed indicators in addition to baywide 
average indicators. They also plan to modify, replace or develop new 
indicators as necessary to measure goals in the Chesapeake 2000 
agreement, fill key gaps in the indicators hierarchy and continuum to 
complete the "cause and effect picture" for the watershed, and initiate 
the development of sustainable development indicators that reflect 
stewardship and land use.

Great Lakes State of the Lakes Ecosystem Conference:

Overview: The Great Lakes Water Quality Agreement of 1978, as amended, 
calls for the development of a set of comprehensive ecosystem health 
indicators for the Great Lakes. Accordingly, the indicator set is meant 
to be used to inform the public and report progress in achieving the 
objectives of the agreement. The indicators are reviewed and discussed 
every 2 years at the State of the Lakes Ecosystem Conference (SOLEC), 
hosted by the EPA and Environment Canada in response to a reporting 
requirement of the agreement. The two governments established SOLEC in 
1992 to report on the state of the Great Lakes ecosystem and the major 
factors impacting it, and to provide a forum for exchange of this 
information among Great Lakes decision makers. In the year following 
each conference, the governments prepare a report on the state of the 
lakes based in large part upon the conference process and environmental 
indicators discussed there. The first conference was held in 1994, and 
the first comprehensive basinwide set of indicators was developed after 
the 1996 conference. The 1998 SOLEC conference was the first to utilize 
a comprehensive set of indicators.

Approximately 80 indicators address specific geographic zones of the 
entire Great Lakes Basin ecosystem, such as offshore, nearshore, 
coastal wetlands, and shoreline and address issues, such as human 
health, land use, and societal. The indicators are based on a pressure-
state-response (PSR) model--a causality framework where human 
activities place pressures on the environment that cause a change in 
the state of the environment; these changes alert society, which then 
implements a response to reduce the pressures or to change the affected 
environment. These changes alert society, which then implements a 
response to reduce the pressures or to change the affected environment. 
The indicators are reported primarily through biennial State of the 
Great Lakes reports (see fig. 8).

Figure 8: State of the Great Lakes Report:

[See PDF for image] 

[End of figure] 

Development: Over 130 experts participated in the development and 
selection of indicators. Experts divided into seven core groups, which 
directly selected and developed indicators or reviewed draft products 
throughout the process for the more than 850 indicators they 
identified. Expert panels initially screened the indicators according 
to the criteria--necessary, sufficient, and feasible--and then analyzed 
them for validity, understandability, interpretability, information 
richness, data availability, timeliness, and cost considerations. This 
vetting process reduced the number of indicators to 80. The Great Lakes 
indicator set draws upon and complements indicators used for more 
specific purposes, such as management plans created for individual 
lakes.

Use: The indicator development and revision process has in itself 
proved beneficial by providing to scientists, resources managers, and 
the public a forum in which to discuss and better understand the 
conditions of the Great Lakes and the impacts affecting its quality. 
The SOLEC indicator set has also identified key data gaps and has 
spurred collaborative monitoring efforts between the United States and 
Canada.

Next steps: In order to establish a consistent, easily understood 
indicator set, EPA and Environment Canada will continue to review and 
refine the indicator set. Indicators are currently being grouped into 
bundles to reduce and organize essential information to a few 
understandable topics. EPA and Environment Canada also plan to build 
appropriate monitoring and reporting activities into existing Great 
Lakes programs at the federal, provincial, state, tribal, and industry 
levels to fully report on all of the approximately 80 indicators.

Minnesota Department of Natural Resources' Strategic Conservation 
Agenda:

Overview: In 2003, the Minnesota Department of Natural Resources (DNR) 
began the development of a Strategic Conservation Agenda (SCA) 
indicator set in response to a directive from the DNR Commissioner's 
office to strengthen accountability and public confidence by better 
communicating progress toward conservation results. The objective of 
the SCA was to provide internal management direction for defining 
agency-level performance goals, demonstrating accountability to 
citizens, and fulfilling the governor's expectations for agency 
accountability to results. The SCA is one piece in a larger policy 
hierarchy as it fits within a DNR mission statement and strategic plan, 
and the department's budgeting process.

The SCA indicator set includes about 75 indicators that target natural 
resource conditions, DNR management activities, and results toward 
which DNR will strive through management efforts. The indicator set 
does not represent all of the natural resources in Minnesota but the 
areas in which DNR will commit resources to achieve specific results. 
The SCA indicators measure natural resource trends or resource work 
performed. The SCA indicator set is defined by six key performance 
areas at DNR: Natural Lands, Fisheries and Wildlife, Healthy Waters and 
Watersheds, Forests, Outdoor Recreation, and Natural Resources 
Stewardship Education. Targets are assigned to each indicator to define 
expected results and serve as specific milestones that help DNR gauge 
progress toward long-term goals. Environmental indicator set 
information was presented in the first SCA report (see fig. 9), which 
was issued to the public through the DNR Web site in March 2004.

Figure 9: The Strategic Conservation Agenda Report:

[See PDF for image] 

[End of figure] 

Development: DNR developed the indicator set through a multistep, 
agencywide process under the direction of the DNR Commissioner's 
Office. The Science Policy Unit, housed in DNR's Office of Management 
and Budget Services, worked with DNR operations managers representing 
all DNR divisions and regions to develop the indicators. The model used 
by DNR for the selection of indicators was based on prior work through 
the Minnesota Environmental Indicators Initiative, which existed from 
1995 through 2000. The DNR relied on that past work to select 
indicators for its focused use. Indicators were selected within goal 
areas established in DNR's strategic planning process called 
Directions. Different DNR divisions provided a menu of existing and new 
indicators along with initial targets. The targets state strategic 
goals in specific and measurable terms where indicators track progress 
and document results. Senior management at DNR then reviewed, modified 
as needed, and approved a final set of indicators that were designed to 
be measurable, accurate, meaningful, and compelling.

Use: DNR uses the indicator set to assist in management decision 
making, to communicate how DNR programs are achieving results, and to 
provide accountability to citizens. For example, the indicator "number 
of cords of wood offered for sale on DNR lands" allows DNR to set 
targets to ensure a predictable, sustainable supply of quality wood. 
The indicator would be reported on and tracked by DNR as well as the 
public to evaluate management practices and be held accountable for 
sustaining timber supplies. DNR staff's involvement in the process of 
development has provided them an opportunity to think about natural 
resource management along the dimensions of performance measurement.

Next steps: DNR will update the indicators periodically. Existing 
indicators will be tracked over time to chart and report progress 
toward conservation targets. New indicators will be added to fill 
information gaps. DNR will work with the public to adjust targets as 
conditions change and develop new targets as opportunities arise to 
better conserve natural resources.

Environmental Protection Indicators for California:

Overview: The California Environmental Protection Agency (Cal/EPA) 
developed the Environmental Protection Indicators for California 
(EPIC), in response to the agency's July 2000 Strategic Vision document 
that committed the agency to manage for environmental results as well 
as to adopt environmental indicators as a priority. The environmental 
indicators in EPIC were developed for the purposes of strategic 
planning, policy formulation, resource allocation, and priority setting 
under a results-based management system.

The EPIC project developed an initial set of indicators based on issue 
categories that generally mirror Cal/EPA areas of authority. EPIC is 
designed to measure the pressures exerted on the environment in 
California by human activities and ambient environmental conditions, as 
well as the resulting effects on human and ecological health in 
California. Most of the indicators focus on environmental resources at 
the state level. Global or transboundary issues that affect the state, 
such as global climate change, are also included in EPIC. In total, 
Cal/EPA identified about 85 indicators for inclusion in EPIC. The 
indicators are organized into six levels that range from indicators 
measuring management actions to those that are direct or indirect 
measures of ecological or human health. The indicators were presented 
by Cal/EPA and the California Resources Agency in an April 2002 report 
(see fig. 10) and a shorter summary document created to provide a more 
general overview of the project and the indicators.

Figure 10: Environmental Protection Indicators for California Report:

[See PDF for image] 

[End of figure] 

Development: The EPIC project began in January 2001 with a conference 
designed to engage individuals other than those in the participating 
state agencies in discussions about the areas the indicators should 
address. Cal/EPA's Office of Environmental Health Hazard Assessment was 
designated by Cal/EPA to lead and oversee EPIC as a whole. The offices, 
departments and boards within Cal/EPA participated in the development 
of EPIC by identifying data sources and developing indicators. In 
addition, recognizing the need of EPIC to address environmental 
protection issues in tandem with resource management issues and the 
interplay between environment and human health, both the California 
Resources Agency and the California Department of Health Services 
collaborated in the development. Approximately 130 individuals 
representing various groups were involved in the selection and 
development of the indicators--an external advisory group, interagency 
advisory group, project staff, and seven work groups.

Within each issue area, work groups identified parameters that could be 
used to derive candidate indicators. The indicators they developed in 
the various issue areas were subject to criteria that included data 
quality, representativeness, sensitivity, and decision support. 
Indicators that met criteria were further evaluated as to whether data 
are available to present a condition or trend for the issue area. 
Indicators were then classified into three categories according to the 
availability of data that are collected on a systematic basis.

Use: Because EPIC's primary purpose is to evaluate Cal/EPA programs, 
Cal/EPA has begun to use the indicators in a pilot project to institute 
a performance management system. The project was scheduled for 
completion in June 2004. Participants in the indicator development 
process stated that EPIC helped to get the agency to initiate 
discussion between boards and departments on what indicators were 
available, and how the agency could begin to measure results. The 
process also helped to identify data gaps.

Next steps: California has suspended funding for the EPIC project. Cal/
EPA staff, however, will continue to evaluate the current set of 
indicators, identify new indicators, revise and replace existing 
indicators as appropriate, and publish a progress report outlining its 
activities on a regular basis.

Quality of Life Indicator Set, Jacksonville, Florida:

Overview: The Jacksonville Community Council Inc. (JCCI)--a nonprofit 
organization in Florida--started the Quality of Life indicator set in 
1985 to measure the quality of community life and identify aspects of 
the community that, if improved, would yield significant benefits. As 
an indicator set, the Quality of Life Progress Report provides 
information about the community by showing its history, its current 
status, and the areas requiring attention to reach the Jacksonville's 
goals. The Quality of Life indicator set provides a source of local, 
summary-level information about Jacksonville. Each annual update 
represents the community's report card, containing information used to 
inform the community, ensure public accountability, and guide decision 
makers to help promote and enhance the quality of life for all citizens 
in the community.

The Quality of Life project initially identified about 75 indicators to 
track. The latest report (see fig. 11) included 115 indicators focusing 
on nine areas: environment, economy, education, government, health, 
recreation, safety, social well-being, and transportation. Each of 
these areas contains between 8 and 19 individual indicators. The 
geographic scale of reporting includes Duval County, which encompasses 
Jacksonville's metropolitan area.

Figure 11: Jacksonville's 2003 Quality of Life Progress Report:

[See PDF for image] 

[End of figure] 

Development: The Quality of Life project began with the efforts of a 
citizen's task force, composed of about 100 individuals. The Chairman 
of the JCCI chose the head of a steering committee, which then selected 
committee members based on their volunteer experiences, leadership 
capabilities, and areas of expertise. The steering committee formed 
subcommittees/task forces for nine basic quality of life topic areas. 
For each topic area, the group selected various indicators based on the 
following criteria: validity; availability and timeliness; stability 
and reliability; understandability; responsiveness; policy relevance; 
and representativeness. The task forces periodically update the 
indicators and the associated targets. There was an update process 
carried out in 2000 that consumed almost 90 meetings over 6 months. The 
process included volunteers from various groups to assist in the 
review. Efforts have already been completed to revise the indicators, 
identify linkages, and set targets for 2005.

Use: The Quality of Life indicator set was developed to help track the 
progress that Jacksonville is making toward meeting established 
environmental and other goals. To this end, the City Council, Chamber 
of Commerce, citywide departments, and others all use this information. 
The biggest impact of the indicator set has been its ability to educate 
the public, highlight the environment, and increase community awareness 
of the environmental issues facing Jacksonville. In addition, the 
Quality of Life report has provided the essential information for 
decision makers to address various issues.

Next steps: JCCI will continually revise and update the indicators and 
associated targets to include in its annual progress reports. Recently, 
JCCI has begun developing indicators that address key issues in the 
community, such as illiteracy and racial disparity. In addition, JCCI 
has developed a Replication Kit for communities interested in 
establishing an indicator project, and provided direct consulting 
practices.

Environmental Indicators Project, West Oakland, California:

Overview: The Environmental Indicators Project (EIP) was created to 
assist policymakers and residents to use indicator information to 
initiate a dialogue among residents, policymakers, and the private 
sector to improve quality of life and create a healthy, safe 
environment in West Oakland, California. Community participation in the 
EIP development process was a critical part of achieving this goal.

The EIP began in 2000 with the partnership of the Pacific Institute (a 
nonprofit organization) and a West Oakland neighborhood organization. 
The EIP's 17 indicators represent a broad range of environmental 
concerns in the community, from issues of air quality and toxics to 
environmental health, land use, housing affordability, transportation, 
and civic engagement. The EIP includes "environmental indicators" that 
are broadly defined in an effort to integrate environmental measures 
with the community's social and economic well-being. The indicators are 
reported through indicator reports (see fig. 12) to the community and 
through brochures on groups of indicators relevant to specific 
community campaigns.

Figure 12: West Oakland's Neighborhood Knowledge for Change Report:

[See PDF for image] 

[End of figure] 

Development: The EIP established a task force of neighborhood residents 
to identify a core set of indicators that address issues of importance 
in the neighborhood. Participation in the indicator development process 
broadened community involvement beyond the staff of the community-based 
organizations to include residents who had previously not had access to 
such information. Task force members selected and developed the 
indicators by defining the term "environment" in the context of West 
Oakland; identifying environmental issues in the community; selecting 
the indicators community members would want to measure and track; and 
determining how such information could be incorporated into current 
advocacy, policy, and education. The Pacific Institute's team of 
researchers then collected and analyzed data from city, county, state, 
and national agencies to develop the indicators. An additional four 
indicators were selected by the community as important but were not 
reported on, either because (1) data were not available or (2) the 
available data were not reliable, consistent, or regularly updated. The 
EIP released its report in January 2002 and also designed brochures on 
groups of indicators relevant to the campaigns to make the information 
more accessible and understandable to community stakeholders, and to 
help educate residents on community advocacy efforts.

Use: Residents, policymakers, and agencies have used indicator 
information to begin to improving the quality of life for West Oakland 
residents. For example, the indicators provided evidence that a Red 
Star Yeast factory that was located in the community was releasing 
illegal amounts of toxic air pollutants and was subsequently closed. 
The EIP has also been valuable to the work of numerous community 
campaigns and in working with agencies because community testimonials 
can now be combined with the information presented through the 
indicators.

Next steps: The Pacific Institute will continue to work with community 
partners to develop a system that ensures that indicators remain 
accessible to, and are used by, the community. The Pacific Institute 
also plans to update the existing indicators and incorporate new ones 
as necessary.

[End of section]

Appendix IV: Selected Activities Identifying Need for More 
Comprehensive Environmental Information:

The following tables summarize major congressional attempts to address 
federal environmental data and indicator issues since 1970, as well as 
selected academic reports issued during the same period. None of the 
tables are exhaustive. Rather, the purpose of these lists is to 
illustrate significant legislative and academic milestones in federal 
environmental data and indicator management over the last 35 years. 
While there have been numerous such efforts, both Congress and the 
academic community had already identified and analyzed many of the 
fundamental issues confronting indicator development and data 
management by the close of the 1970s. Perhaps the most significant 
recent development is the focus since 1990 on the creation of an 
objective, nonpolitical environmental statistical agency within the 
federal government, an idea that has appeared in several recent 
legislative proposals to elevate the Environmental Protection Agency to 
Cabinet level. Two bills to elevate the EPA, one of which would 
establish a Bureau of Environmental Statistics, were introduced in the 
108th Congress.

Selected Legislation to Address Federal Environmental Data and 
Indicator Issues:

Table 7 presents selected Congressional bills introduced since 1970 
that deal with significant challenges involving federal environmental 
data management and indicator development. While Congress has been 
examining how best to address these challenges for some time, 
legislative consensus has yet to emerge on many key topics, including 
whether a Bureau of Environmental Statistics should be established--and 
if so, whether it should be done as part of legislation to elevate EPA 
to Cabinet status.

Table 7: Major Pieces of Legislation to Address Federal Environmental 
Data and Indicator Issues, 1970-2004:

Year introduced: 1970;
Bill: H.R. 17436;
Principal provisions: Would amend the National Environmental Policy Act 
of 1969 to create a National Environmental Data System to serve as the 
central national coordinating facility for the storage, analysis, and 
retrieval of environmental information to support environmental 
decisions in a timely manner. Would require each federal agency to make 
environmental data available to the Data System and would require data 
in the Data System to be available to Congress, federal agencies, 
states, and the public. The system would be operated by a director 
under the guidance of the Council on Environmental Quality. It would 
develop and publish environmental quality indicators for all of the 
regions in the United States;
Last action: Passed House and referred to the Senate Committee on 
Commerce.

Year introduced: 1970;
Bill: H.R. 18141;
Principal provisions: Similar to H.R. 17436. Would amend the National 
Environmental Policy Act of 1969 to provide for a National 
Environmental Data Bank for all data relating to the environment;
Last action: Hearing before the Subcommittee on Fisheries and Wildlife 
Conservation, House Committee on Merchant Marine and Fisheries.

Year introduced: 1971;
Bill: H.R. 56;
Principal provisions: Would create a National Environmental Data System 
that would provide for the development and utilization of information 
needed to support management of the environment. The Data System would 
serve as the central national facility for the selection, storage, 
analysis, retrieval, and dissemination of information, knowledge, and 
data specifically related to the environment. Would require data in the 
Data System to be available to Congress, federal agencies, states, and 
the public. The Data System would be operated by a director under the 
guidance of the Council on Environmental Quality and it would develop 
and publish environmental quality indicators for all of the regions in 
the United States;
Last action: Pocket Veto by President Richard Nixon.

Year introduced: 1984;
Bill: H.R. 5958;
Principal provisions: Would establish a National Commission on 
Environmental Monitoring to (1) investigate and study the nation's 
environmental monitoring programs and those international monitoring 
programs in which the United States participates; (2) recommend to 
Congress and the President a plan to improve environmental monitoring; 
and (3) advise and assist in the preparation of an environmental 
monitoring report;
Last action: Referred to subcommittees of House Committee on Merchant 
Marine and Fisheries and House Committee on Science and Technology.

Year introduced: 1990;
Bill: H.R. 3847;
Principal provisions: Would redesignate the Environmental Protection 
Agency as the Department of Environmental Protection and establish 
within it a Bureau of Environmental Statistics. Would require the 
Secretary of the department to establish an Advisory Committee on 
Environmental Statistics to (1) advise the director of the bureau and 
Congress on the collection and dissemination of statistical data, and 
(2) ensure that the statistics and analyses reported by the bureau are 
of high quality, publicly accessible, and not subject to political 
influence;
Last action: Passed House and referred to Senate Committee on 
Governmental Affairs.

Year introduced: 1990;
Bill: H.R. 3904;
Principal provisions: Would establish the National Environmental 
Institute Commission to (1) make recommendations to the President and 
Congress for the establishment of a National Environmental Institute, a 
Bureau of Environmental Information and Statistics, and an organization 
to examine public policies that affect the environment; and (2) 
identify areas of research that require long-term efforts to mitigate 
serious risk to the environment;
Last action: Referred to Subcommittee on Natural Resources, 
Agricultural Research, and Environment, House Committee on Science, 
Space, and Technology.

Year introduced: 1990;
Bill: S. 2006;
Principal provisions: Would elevate the Environmental Protection Agency 
to Cabinet-level status and rename the agency as the Department of the 
Environment. Would establish a Bureau of Environmental Statistics 
within the Department and create an Advisory Council on Environmental 
Statistics to advise the bureau on statistics and analyses, including 
whether the statistics and analyses disseminated by the bureau (1) were 
of high quality, and (2) were based upon the best available objective 
information. It also would authorize the Secretary of the Environment 
to make grants to, and enter into contracts with, state and local 
governments to assist in data collection;
Last action: Placed on Senate Calendar.

Year introduced: 1991;
Bill: S. 533;
Principal provisions: Would elevate the Environmental Protection Agency 
to Cabinet-level status and rename the agency as the Department of the 
Environment. Would establish a Bureau of Environmental Statistics 
within the Department and create an Advisory Council on Environmental 
Statistics to advise the bureau on statistics and analyses. It also 
would authorize the Secretary of the Environment to make grants to, and 
enter into contracts with, state and local governments to assist in 
statistic data collection. Would also direct the Secretary to enter 
into an agreement with the National Academy of Sciences for a study and 
report on the adequacy of the department's data collection procedures 
and capabilities;
Last action: Passed Senate and referred to the House Committee on 
Government Operations.

Year introduced: 1991;
Bill: S. 2132;
Principal provisions: Would require the Environmental Protection Agency 
to conduct a research program in environmental risk assessment in order 
to (1) ensure that the risk assessment process is based upon adequate 
environmental data and scientific understanding, and (2) provide for 
the most cost-effective use of environmental protection resources. 
Would direct the Administrator to conduct an environmental monitoring 
and assessment program to (1) design and evaluate methods and networks 
to collect monitoring data on the current and changing condition of the 
environment, (2) implement monitoring programs and manage data from 
such programs in formats readily accessible to the public, and (3) 
provide annual statistical reports of the results of such programs to 
Congress and the public;
Last action: Senate Committee on Environment and Public Works held 
hearing.

Year introduced: 1993;
Bill: H.R. 109;
Principal provisions: Would establish the Department of the Environment 
and create a Bureau of Environmental Statistics within the department 
to (1) compile, analyze, and publish a comprehensive set of 
environmental quality statistics, which should provide timely summary 
in the form of aggregates, multiyear averages, or totals and include 
information on the nature, source, and amount of pollutants in the 
environment and the effects of those pollutants on the public and the 
environment; (2) promulgate guidelines to ensure that information 
collected is accurate, reliable, relevant, and in a form that permits 
systematic analysis; (3) coordinate the collection of information by 
the department for developing statistics with related information-
gathering activities conducted by other federal agencies; (4) make the 
bureau's published statistics readily accessible; and (5) identify data 
gaps, review the gaps at least annually with the Science Advisory 
Board, and make recommendations to the department concerning research 
programs to provide information to fill the data gaps identified;
Last action: Referred to the Subcommittee on Legislation and National 
Security, House Committee on Government Operations.

Year introduced: 1993;
Bill: H.R. 3425;
Principal provisions: Would establish a Department of Environmental 
Protection and a Bureau of Environmental Statistics within the 
department to (1) collect, compile, analyze, and publish a 
comprehensive set of environmental quality and related public health, 
economic, and statistical data for determining environmental quality 
and related measures of public health, over both the short and long 
term, including assessing ambient conditions and trends and the 
distribution of environmental conditions and related public health 
conditions; (2) evaluate the adequacy of available statistical measures 
to determine the department's success in fulfilling statutory 
requirements; (3) ensure that data and measures referred to in this 
subsection are accurate, reliable, relevant, and in a form that permits 
systematic analysis; (4) collect and analyze such other data as may be 
required to fulfill the bureau's responsibilities and identify new 
environmental problems; (5) conduct specialized analyses and prepare 
special reports; and (6) make readily accessible all publicly available 
data collected;
Last action: Failed on House floor.

Year introduced: 1993;
Bill: S. 171;
Principal provisions: Would establish the Department of Environmental 
Protection and provide for a Bureau of Environmental Statistics within 
the department, as well as a presidential commission on improving 
environmental protection. Would require the bureau to issue an annual 
report on (1) statistics on environmental quality; (2) statistics on 
the effects of changes in environmental quality on human health and 
nonhuman species and ecosystems; (3) documentation of the method used 
to obtain and assure the quality of the statistics presented in the 
report; (4) economic information on the current and projected costs and 
benefits of environmental protection; and (5) recommendations on 
improving environmental statistical information. Would authorize the 
department to make grants to, and contracts with, state and local 
governments, Indian tribes, universities, and other organizations to 
assist in data collection. Would abolish the Council on Environmental 
Quality and transfer all of the council's functions to the Secretary of 
the new department;
Last action: Passed Senate;

not voted upon in House.

Year introduced: 2000;
Bill: H.R. 4757;
Principal provisions: Would require the Environmental Protection Agency 
to establish an integrated environmental reporting system, including a 
National Environmental Data Model that describes the major data types, 
significant attributes, and interrelationships common to activities 
carried out by the Administrator or state, tribal, and local agencies 
(including permitting, compliance, enforcement, budgeting, performance 
tracking, and collection and analysis of environmental samples and 
results). Would require EPA to use the model as the framework for 
databases on which the data reported to the Administrator through the 
integrated system would be kept;
Last action: Referred to Subcommittee on Health and Environment, House 
Committee on Transportation and Infrastructure.

Year introduced: 2000;
Bill: H.R. 5422;
Principal provisions: Similar to H.R. 4757, but with some 
modifications. For example, H.R. 5422 contained an authorization of 
appropriations;
Last action: Referred to Subcommittee on Health and Environment, House 
Committee on Transportation and Infrastructure.

Year introduced: 2001;
Bill: H.R. 2694;
Principal provisions: Would establish the Department of Environmental 
Protection and a Bureau of Environmental Statistics within the 
department to (1) collect, compile, analyze, and publish a 
comprehensive set of environmental quality and related public health, 
economic, and statistical data for determining environmental quality 
and related measures of public health, over both the short and long 
term, including assessing ambient conditions and trends and the 
distribution of environmental conditions and related public health 
conditions; (2) evaluate the adequacy of available statistical measures 
to determine the department's success in fulfilling statutory 
requirements; (3) ensure that data and measures referred to in this 
subsection are accurate, reliable, relevant, and in a form that permits 
systematic analysis; (4) collect and analyze such other data as may be 
required to fulfill the bureau's responsibilities and identify new 
environmental problems; (5) conduct specialized analyses and prepare 
special reports; and (6) make readily accessible all publicly available 
data collected;
Last action: Referred to the Subcommittee on Energy Policy, Natural 
Resources and Regulatory Affairs, House Committee on Government Reform.

Year introduced: 2003;
Bill: H.R. 2138;
Principal provisions: Similar to H.R. 2694. In addition, would require 
the bureau to (1) prepare and submit to Congress and the department an 
annual report on environmental conditions and public health conditions, 
using, to the maximum extent practicable, reliable statistical sampling 
techniques; and (2) make the annual report available to the public upon 
request, and publish a notice of such availability in the Federal 
Register. Would also require the statistical procedures and methodology 
of the Bureau of Environmental Statistics to periodically undergo peer 
review;
Last action: House Committee on Government Reform held hearing.

Source: GAO.

[End of table]

Selected Congressional Hearings Addressing Federal Environmental Data 
and Indicator Management Issues:

Table 8 highlights congressional hearings since 1970 that have 
addressed one or more salient aspects of the federal environmental 
information management challenge. As the table indicates, emphasis has 
shifted over time from creating a data bank centralizing all federal 
environmental information to the creation of a federal statistical 
agency that would be responsible for keeping environmental statistical 
information and establishing data quality standards. Hearings have also 
frequently examined the critical topic of environmental monitoring.

Table 8: Selected Congressional Hearings Addressing Federal 
Environmental Data and Indicator Management Issues, 1970-2004:

Hearing: Environmental Data Bank, 1970;
Committee: House Committee on Merchant Marine and Fisheries, 
Subcommittee on Fisheries and Wildlife Conservation;
Related bills: H.R. 17436 H.R. 17779 H.R. 18141;
Hearing purpose and description: The purpose of the hearing was to 
examine a proposed amendment to the National Environmental Policy Act 
of 1969, which would provide for the establishment of a National 
Environmental Data Bank. The Data Bank would serve as the central 
national depository of all information, knowledge, and data relating to 
the environment, including information, knowledge and data from the 
head of each department, agency, or instrumentality in the executive 
branch of the United States government as a result of its operations.

Hearing: Environmental Monitoring, 1977;
Committee: House Committee on Science and Technology, Subcommittee on 
the Environment and the Atmosphere;
Hearing purpose and description: The purpose of the hearing was to (1) 
examine the existing monitoring efforts of the federal agencies chiefly 
responsible for environmental monitoring; and (2) investigate the 
feasibility and practicality of developing and implementing a prototype 
monitoring system. The system could eventually be expanded into a 
comprehensive national or international monitoring program.

Hearing: Environmental Monitoring 2, 1978;
Committee: House Committee on Science and Technology, Subcommittee on 
the Environment and the Atmosphere;
Related bills: Draft bill;
Hearing purpose and description: The purpose of the hearing was to 
examine a draft bill developed after the 1977 hearings on Environmental 
Monitoring. The legislation would establish a coordinated, integrative, 
and cooperative prototype management system of selected, diverse 
environmental monitoring activities as a possible first step toward a 
national system to improve the effectiveness and efficiency of 
environmental monitoring activities. The President would establish and 
appoint a panel of 10 people, chaired by the Director of the Office of 
Science and Technology Policy, to develop a prototype monitoring system 
to demonstrate on a small scale how a national monitoring management 
system might work.

Hearing: National Environmental Monitoring, 1983;
Committee: House Committee on Science and Technology, Subcommittee on 
Natural Resources, Agricultural Research and Environment;
Hearing purpose and description: The purpose of the hearing was to 
explore the condition of the nation's environmental monitoring programs 
and (1) identify problems in monitoring efforts, and (2) provide 
recommendations that would lead to improvements in environmental 
monitoring.

Hearing: Environmental Monitoring Improvement Act, 1984;
Committee: House Committee on Science and Technology, Subcommittee on 
Natural Resources, Agricultural Research and Environment;
Related bills: Draft bill;
Hearing purpose and description: The purpose of the hearing was to 
examine a draft bill that would create a commission to act as the prime 
coordinating body for the nation's environmental monitoring efforts. 
The charge of the commission would be to clearly define the operational 
changes and the administrative coordination necessary to assure that 
cost-effective and statistically sound and reliable data are available 
to support U.S. environmental policy making.

Hearing: Establish a Department of Environmental Protection, 1989-1990;
Committee: House Committee on Government Operations, Subcommittee on 
Legislation and National Security;
Related bills: H.R. 3847;
Hearing purpose and description: The purpose of the hearing was to 
examine two bills that would elevate EPA to Cabinet status. One of the 
bills (H.R. 3847) would establish a Bureau of Environmental Statistics, 
which would make accessible a standardized set of environmental quality 
data to improve the effectiveness and objectivity of central 
environmental data collection and analyses so that the President, 
Congress, and the public can be adequately informed about conditions 
and trends in environmental quality and so that the department can 
better evaluate its programs.

Hearing: EPA Elevation, 2001-2002;
Committee: House Committee on Government Reform, Subcommittee on Energy 
Policy, Natural Resources and Regulatory Affairs;
Related bills: H.R. 2694;
Hearing purpose and description: The purpose of the hearing was to 
examine two bills that would elevate EPA to Cabinet status. One of the 
bills (H.R. 2694) would establish a Bureau of Environmental Statistics 
to provide environmental quality and related public health and economic 
information and analyses to meet the needs of the department and 
Congress.

Source: GAO.

[End of table]

Selected Academic Reports Addressing Federal Environmental Data and 
Indicator Management Issues:

Table 9 highlights a few of the most significant academic reports 
analyzing federal environmental information management since 1970. 
Collectively, these reports clearly indicate that most of the 
significant information challenges have long been recognized. Our 
report makes recommendations that, if implemented, would begin to 
address these long-standing challenges.

Table 9: Selected Academic Reports Addressing Federal Environmental 
Data and Indicator Management Issues:

Year: 1970;
Name of organization: National Academy of Sciences;
Description: Reported that the United States cannot effectively manage 
the environment without knowing what it is, what it was, and what it 
can be. Recommended giving the highest priority to developing a 
centralized comprehensive federal program for monitoring the 
environment, incorporating environmental quality indices;
Citation: National Academy of Sciences, Institutions for Effective 
Management of the Environment, report (part 1) of the Environmental 
Study Group to the Environmental Studies Board of the National Academy 
of Sciences, National Academy of Engineering, (Washington, D.C; January 
1970).

Year: 1982;
Name of organization: The Conservation Foundation;
Description: Reported that the nation had made progress in its attack 
on some conventional environmental problems; however, the information 
base on which sound environmental policy depends is inadequate and 
deteriorating. The nation has no monitoring data sufficient to describe 
accurately the extent or developing seriousness of any environmental 
problem;
Citation: The Conservation Foundation, State of the Environment 1982 
(Washington, D.C; 1982).

Year: 1988;
Name of organization: Paul Portney, Resources for the Future;
Description: Recommended the creation of a Bureau of Environmental 
Statistics because the U.S. does not adequately collect, analyze, and 
disseminate information about environmental conditions and trends. 
Environmental data are also not collected in a systematic way to make 
it useful to interested parties;
Citation: Paul Portney, "Reforming Environmental Regulation: Three 
Modest Proposals," Columbia Journal of Environmental Law, vol. 13 
(1988).

Year: 1997;
Name of organization: National Science and Technology Council;
Description: Proposed a conceptual framework for integrating the 
nation's environmental research and monitoring networks to deliver 
scientific data and information needed to produce integrated 
environmental assessments and enhance understanding, evaluation, and 
forecasting of natural resources;
Citation: National Science and Technology Council, Integrating the 
Nation's Environmental Monitoring and Research Networks and Programs: A 
Proposed Framework, a report by the Committee on Environment and 
Natural Resources (Washington, D.C; March 1997).

Year: 1998;
Name of organization: National Advisory Council for Environmental 
Policy and Technology;
Description: Reported that EPA information systems do not provide 
sufficient, appropriate, or accurate information to (1) inform decision 
making, (2) ensure accountability, or (3) document results and 
achievements. However, the systems have for the most part satisfied 
regulatory requirements for collecting environmental information;
Citation: National Advisory Council for Environmental Policy and 
Technology, EPA--Managing Information as a Strategic Resource: Final 
Report and Recommendations of the Information Impacts Committee, EPA 
100-R-98-002 (Washington, D.C; January 1998).

Year: 1999;
Name of organization: National Research Council;
Description: Addresses the question of whether the U.S. National Income 
and Product Accounts should be broadened to include activities 
involving natural resources and the environment. Concludes that the 
development of environmental and natural resource accounts is an 
essential investment for the nation;
Citation: National Research Council, Nature's Numbers: Expanding the 
National Economic Accounts to Include the Environment (Washington, D.C; 
National Academy Press, 1999).

Year: 2000;
Name of organization: National Council for Science and the Environment;
Description: Reported that the fragmented administrative jurisdictions 
among federal agencies charged with environmental stewardship compound 
difficulties in coordinating environmental research and in 
communicating scientific results to decision makers and the public. 
Changes in governmental institutions could significantly improve 
efficiency and communication among scientists and between scientists 
and decision makers;
Citation: National Council for Science and the Environment, 
Recommendations for Improving the Scientific Basis for Environmental 
Decisionmaking, report from the first National Conference on Science, 
Policy and the Environment (Washington, D.C; December 2000).

Year: 2002;
Name of organization: EPA Science Advisory Board;
Description: Reported that many scientists, most decision makers, and 
nearly all members of the public still have little understanding of the 
"health" or integrity of the nation's ecological systems. Recommended 
EPA would benefit from the development of a systematic framework for 
assessing and reporting on ecological condition by helping assure that 
required information is measured systematically and provide a template 
for assembling information across EPA and other agencies;
Citation: U.S. Environmental Protection Agency, Science Advisory Board, 
A Framework for Assessing and Reporting on Ecological Condition: An SAB 
Report (Washington, D.C; June 2002).

Source: GAO.

[End of table]

[End of section] 

Appendix V: Environmental Reporting by Private and Public 
Organizations:

Environmental reporting involves the disclosure of information on 
environmental performance and management practices that convey 
environmental impacts and the actions being taken to manage these 
impacts. Some private corporations and public institutions now conduct 
this type of environmental reporting. For example, some entities report 
environmental impacts, such as the amount of natural resources used, 
the amount of waste generated, and the amount of emissions released by 
a facility. Reports may also include information on the management 
efforts that are used to influence environmental impacts such as 
details on how a facility is implementing a pollution reduction 
program. The 1992 United Nations Conference on Environment and 
Development has recognized the importance of this type of information, 
encouraging private facilities to report "annually on their 
environmental records, as well as on their use of energy and natural 
resources" and "on the implementation of codes of conduct promoting 
best environmental practice."[Footnote 27]

Corporate reporting of environmental information is becoming 
increasingly prevalent in the United States and worldwide. A 2002 
survey of the Global Fortune Top 250 international companies (GFT 250) 
found that since 1999, there has been a 29 percent increase in the 
number of companies that publish separate reports on various aspects of 
corporate performance in addition to annual financial reports.[Footnote 
28] The majority of these separate reports contained environmental 
information. The United States had the largest number of reporting 
companies, with 32 of the 105 U.S. companies in the GFT 250 issuing a 
report--four more companies than reported in 1999. The survey also 
examined the top 100 companies in each of 19 different countries. The 
results show that Japan and the United Kingdom have the largest 
percentage of top 100 companies publishing reports--72 percent and 49 
percent, respectively. The United States was third with 36 percent of 
the top 100 U.S. companies reporting in 2002, an increase from 30 
percent in 1999. A separate survey conducted in 2001 found similar 
increases in reporting as results show that 50 percent of the GFT 100 
companies produced environmental reports, up from 44 percent in 
1999.[Footnote 29]

Corporate reporting of environmental information in the United States 
is sometimes a regulatory requirement. For example, certain facilities 
are required to submit information on the manufacture, process, and use 
of approximately 650 different types of toxic chemicals to the 
Environmental Protection Agency's Toxic Release Inventory (TRI) 
database. The Environmental Protection Agency reported that almost 
25,000 facilities submitted TRI information in 2001. Another form of 
mandatory reporting is the disclosure of information relating to 
environmental issues required in companies' financial filings with the 
Securities and Exchange Commission.

Companies may also voluntarily collect and report environmental 
information when it is not required because of the benefits that this 
information provides. The environmental information included in these 
reports can help corporations communicate the environmental impact of 
economic activities to a wide variety of stakeholders, such as local 
and planning authorities, community groups, the media, and the general 
public. Such communications can potentially benefit the corporation by 
enhancing its reputation and standing as environmentally responsible. 
This information also provides corporations benefits by identifying 
possible cost savings in both the resources used and operating costs 
and by identifying potential environmental risks, allowing corporations 
to better anticipate potential problems and avoid negative publicity on 
environmental issues. For example, this information can direct a 
corporation's attention to ways to change resource use that results in 
efficiency savings from lower energy, water, and material costs.

Reporting of standardized information is important in order to examine 
the progress of a facility over time and compare or aggregate 
information for many different facilities. Consequently, private and 
public facilities are adopting voluntary standards and guidelines for 
environmental reporting. A recent survey of multinational corporations 
identified some of the most influential voluntary standards now being 
used by corporations to standardize environmental 
information.[Footnote 30] Included on this list are the International 
Organization for Standardization 14000 standards, the Global Reporting 
Initiative, the World Business Council for Sustainable Development, the 
United Nation's Global Compact, and the Organization for Economic 
Cooperation and Development's Guidelines for Multinational 
Enterprises.

Verification of the quality of the information contained in voluntary 
reports is also important to assuage the inherent tension between a 
facility's desire to present its side of the story and a stakeholder's 
demand for greater transparency. Just as investors look to independent 
audits to certify the accuracy and completeness of financial reporting, 
stakeholders seek such assurances for the information contained in 
environmental reports. Even so, according to a 2002 study, only 3 
percent of those top 100 U.S. companies that reported information had 
their reports verified by third parties.[Footnote 31]

Environmental reporting is an important consideration for public 
governmental facilities as well. Executive Order 13148 calls for 
federal agencies to implement environmental management systems by 
December 31, 2005, at all appropriate agency facilities. The executive 
order states that these environmental management systems shall include 
measurable environmental goals, objectives, and targets to be reviewed 
and updated annually. According to the Office of the Federal 
Environmental Executive, more than 180 federal facilities have already 
developed and are implementing environmental management systems to 
ensure compliance with environmental requirements and integrate 
environmental accountability into decision making and planning. It also 
reports that, as of December 2002, hundreds of other facilities had 
initiated the education process needed to ensure commitment to the 
development of environmental management systems.

Whether the basis for environmental reporting is mandatory or 
voluntary, environmental reports contain information that can be used 
by a variety of stakeholders to monitor environmental impacts and 
inform decision making. For example, this information can inform 
community leaders and residents in local communities of environmental 
hazards, show how facilities are addressing specific environmental 
concerns, and provide an opportunity for the community to identify how 
a local facility is performing relative to other similar facilities. 
Employees can also use this information to understand a facility's 
existing occupational risks. In addition, information that identifies 
the environmental impacts associated with a product or service 
throughout its lifecycle can be of interest to customers and consumers 
and help inform the choices they make. Reporting can also yield 
information on a facility's environmental vision, environmental 
performance, future environmental plans, and environmental risks and 
liability. These issues may interest potential business partners, 
investors, insurers, and lenders. Finally, this information can further 
the understanding of government policy analysts regarding current 
environmental circumstances and inform government decisions on how best 
to achieve specific environmental objectives.

[End of section]

Appendix VI: Accounting for the Environment:

Environmental accounts can be used to develop indicators that examine 
the nexus between the environment and the economy. As a result, the 
development of environmental accounts is widely recognized as 
important. However, the United States currently has no federal effort 
to develop comprehensive environmental accounts.

Accounts Yield Indicators with Beneficial Uses:

Environmental accounts provide a framework that is used to link 
environmental information to the information that is contained in the 
national economic accounts. Combining this information allows 
environmental and economic issues to be examined jointly. For example, 
by linking information on the amount of pollution released during a 
manufacturing process with knowledge of the amount of economic output 
derived through that manufacturing, policymakers could better 
understand how a change in regulations, such as on pollution limits, 
might affect the ensuing economic performance of an industry.

Several federal agencies are responsible for managing and protecting 
the nation's environment and have developed strategic plans that 
highlight the importance of the interaction between the environment and 
the economy. For example, the strategic plan of the Environmental 
Protection Agency identifies procedures to ensure sound analysis of the 
economic effects of its environmental regulations, policies, and 
programs. The Department of the Interior's plan sets an objective of 
managing natural resources in a way that promotes responsible use while 
sustaining a dynamic economy. The Department of Agriculture's plan 
identifies the need to manage forests and rangelands that are resilient 
to natural and human disturbance while also managing for economic uses 
such as oil, natural gas, and timber. Finally, the National Oceanic and 
Atmospheric Administration's plan seeks to achieve a balance between 
the protection and use of the ocean's resources to ensure 
sustainability while also achieving an optimal contribution to the 
nation's economy.

Environmental accounts provide information that is useful in creating 
indicators to examine the interaction of the environment and the 
economy. The following are examples of these potential indicators.

* Policymakers could use efficiency indicators to determine the volume 
of waste created in production processes and allow for comparison with 
the resources used in production and the total economic output. 
Policymakers could use these indicators to measure and track the use of 
resources and to determine how best to improve the efficiency of 
resource use and minimize waste generation while considering the 
potential economic effects of such policies.

* Policymakers could use resource management indicators to determine 
the amounts of unharvested natural resources still available for future 
consumption. This information could provide policymakers with a better 
understanding of the rate of current resource use, allow for more 
effective long-term management of natural resources and help 
policymakers understand the potential economic effects resulting from 
changes in resources use.

* Policymakers could use environmental expenditure indicators to manage 
and track the amount of economic resources being devoted to abating 
pollution. Such indicators would allow policymakers to identify where 
resources are being spent to reduce pollution, evaluate the 
effectiveness of the nation's efforts, and determine the economic 
impacts on the economy resulting from the costs of abating pollution.

Importance of Environmental Accounts Recognized around the World:

There are several efforts under way to develop environmental accounts 
by governments and nongovernmental organizations. A recent report 
identifies 19 countries that are developing some type of environmental 
accounts in their statistical offices or other government 
ministries.[Footnote 32] Also, the United Nations, along with other 
international organizations, has developed guidelines to be used by 
both national and international agencies for compiling environmental 
accounts.[Footnote 33]

Canada and the Netherlands are currently developing environmental 
accounts alongside national economic accounts to inform policymakers in 
these countries. First published in 1997, the Canadian System of 
Environmental and Resource Accounts (CSERA) provides a comprehensive 
framework for understanding the environment and the economy by 
supplementing environmental information alongside information in the 
national economic accounts.[Footnote 34] According to Statistics 
Canada, while CSERA is a work in progress, information in the accounts 
has improved policymakers' knowledge of interactions between the 
environment and the economy in Canada. Statistics Netherlands has 
published a National Accounting Matrix including Environmental Accounts 
(NAMEA) for the years 1987 through 1992 and continues to further 
develop the accounts. NAMEA functions as an instrument for a variety of 
analyses, including the identification of the economic and 
environmental effects of consumption of certain products and the 
consequences of regulating energy use on environmental themes like 
greenhouse gases and economic issues, such as national income.

The World Bank and World Resources Institute have developed their own 
environmental accounts. The World Bank has developed a measure of net 
savings that calculates a nation's overall savings rate by including 
the value of a nation's natural resources along with traditional 
economic factors. The World Bank currently updates this measure 
annually for approximately 50 counties. This measure of adjusted net 
savings can be used to compare and contrast the traditional economic 
measures of savings in order to monitor the potential impacts of 
natural resource use. The World Resources Institute has created 
material flow accounts for several industrialized countries. These 
accounts track the physical flows of natural resources as they move 
through the economy, including extraction, production, fabrication, 
use, recycling, and final disposal. According to a World Resources 
Institute official, a goal of these accounts is to demonstrate to 
government agencies the value of this environmental and economic 
information for formulating public policy.

Finally, the National Research Council of the National Academy of 
Sciences has reported that environmental accounts can provide 
policymakers with information that would improve decision making 
resulting in substantial monetary benefit for the United States. The 
nation currently invests a substantial amount of money in pollution 
control to clean the air, water, and land, and environmental accounts 
could provide the information necessary to help identify how 
regulations may be refined, so that expenditures on pollution control 
would be allocated more efficiently. For example, the National Research 
Council estimates that improvements in regulations resulting in a 10 
percent reduction in pollution control expenditures would save the 
nation more than $10 billion per year.

The United States Has No Plans to Develop Federal Environmental 
Accounts:

In the United States, no federal effort to create comprehensive 
environmental accounts is either under way or planned. In 1992, the 
Bureau of Economic Analysis (BEA), within the Department of Commerce, 
began developing a set of comprehensive accounts called the Integrated 
Economic and Environmental Satellite Accounts (IEESA). BEA created 
prototype accounts for the mineral resources sector and planned to 
continue its IEESA work and develop accounts for other sectors, but in 
1995, a committee report accompanying the Department of Commerce's 
fiscal year 1995 appropriation directed BEA to suspend this effort and 
allow for an independent review of the IEESA. The National Academy of 
Sciences' National Research Council conducted this review and released 
its final report in 1999,[Footnote 35] recommending that Congress 
authorize and fund BEA to recommence its work developing the IEESA. 
However, Congressional appropriations committees up until fiscal year 
2002 directed BEA not to pursue the IEESA initiative. Although this 
restriction has now been lifted, to date no funding has been 
appropriated and BEA currently has no plans to continue with its work.

[End of section]

Appendix VII: The Uncertain Cost of Environmental Information:

The collection and provision of federal environmental data and 
statistics are costly, but it is uncertain how much the federal 
government spends each year on these activities. While there are no 
agreed-upon sources of the costs to the federal government of 
environmental information, there are two frequently cited sources that, 
despite known shortcomings, represent the best available federal 
estimates of such costs.

In July 1995, the National Science and Technology Council (NSTC) 
convened a team of federal scientists and program managers to develop a 
national framework for integrating and coordinating environmental 
monitoring and related research by amalgamating and building upon 
existing networks and programs. In 1997, the team's final report, 
Integrating the Nation's Environmental Monitoring and Research Networks 
and Programs, reported that the federal government spent about $650 
million on about 31 major federal environmental monitoring and research 
programs and networks in fiscal year 1995.[Footnote 36] The team 
arrived at this total by combining the amounts that agencies reported 
to the team on a project-by-project basis. The total was not 
disaggregated in NSTC's final report, and the effort has not been 
updated.

Additionally, the Office of Management and Budget (OMB) annually 
reports actual and estimated funding for major federal statistical 
programs[Footnote 37] in Statistical Programs of the United States 
Government, as required by the Paperwork Reduction Act.[Footnote 38] 
Major statistical programs differ in organizational structure and in 
the means through which they are funded. A particular agency may carry 
out some major statistical programs on its own. For example, according 
to OMB the sole mission of the Energy Information Administration, 
within the Department of Energy, is to develop energy statistics. Other 
agencies have statistical programs that are an outgrowth of their 
administrative responsibilities or that support their program planning 
and evaluation functions. In these cases, the budget for statistical 
activities comprises a portion of an agency's total appropriations, 
including an allocation of the salaries and operating expenses for the 
statistical program. Funding for statistical activities may increase or 
decrease as a result of the cyclical nature of surveys. Such increases 
or decreases should not be interpreted as changes in agency priorities, 
but rather as the normal and expected consequences of the nature of the 
programs. Agencies may also experience increases or decreases in their 
budgets when they conduct one-time surveys or studies in a particular 
fiscal year. Additionally, a statistical program may not always be 
executed by the agency that sponsors it. In these instances, the work 
is done on a reimbursable basis by another federal agency or by a state 
or local government or a private organization under contract. OMB's 
reported totals reflect statistical activities in support of the 
agency's mission, whether the activities are performed by the agency or 
by contract.

OMB divides federal statistical activities into four categories: Health 
and Safety; Social and Demographic; Natural Resources, Energy, and 
Environment; and Economic. Table 10 provides the direct funding levels 
that Congress appropriated for fiscal years 1998 through 2002 for 
statistical activities in the Natural Resources, Energy, and 
Environment category.

Table 10: Direct Funding for Major Environment, Energy, and Natural 
Resources Statistical Programs:

Millions of dollars; 

Agency: Forest Service; 
Fiscal year 1998: 19; 
Fiscal year 1999: 14; 
Fiscal year 2000: 23; 
Fiscal year 2001: 29; 
Fiscal year 2002: 29.

Agency: Natural Resources Conservation Service; 
Fiscal year 1998: 107; 
Fiscal year 1999: 107; 
Fiscal year 2000: 108; 
Fiscal year 2001: 113; 
Fiscal year 2002: 111.

Agency: NOAA; 
Fiscal year 1998: 49; 
Fiscal year 1999: 53; 
Fiscal year 2000: 54; 
Fiscal year 2001: 87; 
Fiscal year 2002: 87.

Agency: Office of Environment, Safety, and Health; 
Fiscal year 1998: 24; 
Fiscal year 1999: 24; 
Fiscal year 2000: 24; 
Fiscal year 2001: 33; 
Fiscal year 2002: 34.

Agency: Energy Information Administration; 
Fiscal year 1998: 66; 
Fiscal year 1999: 70; 
Fiscal year 2000: 72; 
Fiscal year 2001: 78; 
Fiscal year 2002: 78.

Agency: National Institute on Environmental Health Sciences; 
Fiscal year 1998: 26; 
Fiscal year 1999: 30; 
Fiscal year 2000: 39; 
Fiscal year 2001: 56; 
Fiscal year 2002: 65.

Agency: Fish and Wildlife Service; 
Fiscal year 1998: 6; 
Fiscal year 1999: 3; 
Fiscal year 2000: 4; 
Fiscal year 2001: 9; 
Fiscal year 2002: 9.

Agency: Minerals Management Service; 
Fiscal year 1998: 2; 
Fiscal year 1999: 2; 
Fiscal year 2000: 3; 
Fiscal year 2001: 3; 
Fiscal year 2002: 4.

Agency: National Park Service; 
Fiscal year 1998: 2; 
Fiscal year 1999: 2; 
Fiscal year 2000: 2; 
Fiscal year 2001: 2; 
Fiscal year 2002: 1.

Agency: Bureau of Reclamation; 
Fiscal year 1998: 2; 
Fiscal year 1999: 3; 
Fiscal year 2000: 3; 
Fiscal year 2001: 3; 
Fiscal year 2002: 4.

Agency: U.S. Geological Survey; 
Fiscal year 1998: 64; 
Fiscal year 1999: 60; 
Fiscal year 2000: 73; 
Fiscal year 2001: 83; 
Fiscal year 2002: 84.

Agency: Environmental Protection Agency; 
Fiscal year 1998: 144; 
Fiscal year 1999: 192; 
Fiscal year 2000: 202; 
Fiscal year 2001: 174; 
Fiscal year 2002: 148.

Agency: National Aeronautics and Space Administration; 
Fiscal year 1998: 17; 
Fiscal year 1999: 17; 
Fiscal year 2000: 17; 
Fiscal year 2001: 17; 
Fiscal year 2002: 17.

Total for major environment, energy, and natural resources statistical 
programs; 
Fiscal year 1998: 528; 
Fiscal year 1999: 577; 
Fiscal year 2000: 624; 
Fiscal year 2001: 687; 
Fiscal year 2002: 671.

Total for all federal statistical activities; 
Fiscal year 1998: 3,205; 
Fiscal year 1999: 4,167; 
Fiscal year 2000: 7,755; 
Fiscal year 2001: 4,179; 
Fiscal year 2002: 4,212.

Source: OMB.

Note: Totals reflect actual appropriations.

[End of table]

It is important to note that the totals produced through these efforts 
are not necessarily representative of the magnitude of federal 
investment in environmental information--the total produced by NSTC and 
the figures produced annually by OMB both likely have significant 
omissions. Moreover, the totals produced by these efforts do not 
necessarily cover similar activities, although there is likely 
significant overlap. OMB's classification includes issues (such as 
energy) and activities (such as statistical consulting or training) 
that were not necessarily included in the NSTC's calculations. GAO was 
not able to compare the various programs and subprogram activities that 
constitute the totals produced by these efforts. Reconciling the 
methodologies used by NSTC and OMB to produce these totals is beyond 
the scope of GAO's report.

In preparing this report, GAO used the estimate reported by NSTC to 
generally reflect the annual cost to the federal government of 
collecting environmental information--at least $600 million. However, 
this figure provides a limited snapshot of all spending related to 
collecting and maintaining information on the environment. Agency 
officials and other experts noted that the actual annual costs of 
environmental information to the federal government through monitoring, 
research, statistical, data management, and other activities will 
remain uncertain until a comprehensive assessment is performed that 
examines the completeness, overlap, gaps, and quality of the existing 
programs that produce environmental information.

[End of section]

Appendix VIII: Selected Options:

Experts who participated in the environmental indicator set meeting 
jointly convened by GAO and the National Academy of Sciences identified 
a number of short-term alternatives to assist environmental indicator 
set developers and users. These options were not independently 
evaluated by GAO and are presented in no specific order. Appearance in 
this appendix does not constitute an endorsement of the ideas.

* Congress should reinstate Section 201 of the National Environmental 
Policy Act (NEPA), requiring the Council for Environmental Quality 
(CEQ) to submit an annual report to the Congress on the environment.

* The Office of Management and Budget should hold a hearing to receive 
feedback from agencies on the Program Assessment Rating Tool.

* Congress should charge GAO or the Congressional Research Service with 
an annual review of environmental indicators, their adequacy, and 
utility.

* Federal agencies should pursue an executive order that would 
establish an interagency work group to deal with environmental 
information and data, specifically regarding the development of 
environmental indicators. One expert suggested using Executive Order 
13112 (National Invasive Species Council) as a model.

* Congress should commission a study by an independent expert 
organization, such as the National Academy for Public Administration or 
the National Academy of Sciences, to review appropriate institutional 
structures for housing an entity to coordinate the production of 
environmental information.

* Congress should charge an entity with starting the process of 
coordinating environmental information and developing and compiling 
existing and past environmental indicator efforts.

* Congress should consider acting upon the recommendations presented by 
the National Science and Technology Council's 1997 report Integrating 
the Nation's Environmental Monitoring and Research Networks and 
Programs: A Proposed Framework.

* Congress should task an agency with creating a fully searchable 
Internet clearinghouse to distribute information about developing and 
using environmental indicator sets, including links to related 
environmental data. Portal developers should ensure linked data are 
compliant with current Federal Geographic Data Committee standards.

* Congress should continue to support ongoing federal partnerships 
promoting integration of environmental data and interagency work on 
developing standards to ensure data interoperability.

[End of section]

Appendix IX: Comments from the Council on Environmental Quality:

EXECUTIVE OFFICE OF THE PRESIDENT: 
COUNCIL ON ENVIRONMENTAL QUALITY: 
WASHINGTON, D.C. 20503:

CHAIRMAN:

November 1, 2004:

Mr. John B. Stephenson:
Director, Natural Resources and Environment: 
U.S. General Accountability Office:
441 G Street, N.W.: 
Washington, D.C. 20548:

Dear Mr. Stephenson:

Thank you for the opportunity to review the proposed report entitled 
Environmental Indicators: Better Coordination Is Needed to Develop 
Environmental Indicator Sets that Inform Decisions (GAO-05-52). The 
proposed report is a very timely and comprehensive review of the many 
efforts underway to improve the reporting of indicators measuring 
environmental conditions. Your report properly documents the many 
advancements and challenges that experts in development of 
environmental indicators recognize.

As your proposed report describes, the Council on Environmental Quality 
(CEQ) has long been recognized by Congress and others as an appropriate 
institution to lead such efforts. Consistent with the sections of NEPA 
that are cited in the report on p. 36, Congress directed CEQ to, 
"review the adequacy of existing systems for monitoring and predicting 
environmental changes in order to achieve effective coverage and 
efficient use of research facilities and other resources", and to 
assist and advise the President by "collecting, collating, analyzing, 
and interpreting data and information on environmental quality, 
ecological research and evaluation." 42 U.S.C. 4372(d))(3)(7). Further, 
CEQ has been directed by Executive Order to "(e) Promote the 
development and use of indices and monitoring systems (1) to assess 
environmental conditions and trends, (2) to predict the environmental 
impact of proposed public and private actions, and (3) to determine the 
effectiveness of programs for protecting and enhancing environmental 
quality." Executive Order 11514, as amended by Executive Order 11991, 
May 24, 1977.

Several years ago, CEQ established an Interagency Working Group on 
Indicator Coordination. The Working Group developed a Framework for a 
National System of Indicators on Natural and Environmental Resources 
and has studied the current institutional arrangements for performing 
the various functions needed for development and operation of such a 
system. It has been drawing upon the work of a number of key "building 
block" projects that are selecting and identifying indicators for 
managing natural and environmental resources. These include the EPA 
Draft Report on the Environment, the Heinz Center's State of the 
Nation's Ecosystems Report, and the work of four roundtables that are 
using collaborative processes to develop criteria and indicators for 
sustainable management of natural resources, including forests, 
rangelands, minerals, and water resources. Further, four agencies with 
significant responsibility for monitoring and managing - the 
Departments of Agriculture, Commerce, Interior, and the Environmental 
Protection Agency - have begun studies of ways to improve their 
institutional arrangements for statistical reporting of the indicators 
that would be included in a comprehensive national system.

While your proposed report recognizes these steps, it suggests that a 
quasi-governmental agency also be studied. This approach may have 
merit, particularly if such an organization is created for reporting on 
a broader set of key national indicators. However, improvements in 
coordination and statistical reporting within the Federal government 
will be important irrespective of whether a quasi-governmental agency 
is formed. Gains can be achieved by moving ahead with improvements 
within Federal agencies, even while studying iterative organizational 
options.

One shortcoming in the proposed report is that its main focus is on 
the status of efforts in the Environmental Protection Agency (EPA). 
Other agencies have equally important missions for managing natural 
resources and the environment and collecting data on their conditions. 
For example, the Department of the Interior has responsibility for 
managing public lands, National Parks and National Wildlife Refuges, 
and for monitoring water resources; the Department of Agriculture has 
responsibility for monitoring forest land, managing the National 
Forests and implementing billions of dollars in conservation programs 
on agricultural land; and the National Oceans and Atmospheric 
Administration has responsibility for oceans, fisheries and marine 
sanctuaries.. Their contributions to the development and operation of a 
comprehensive system of indicators and their use of indicators in 
planning and management are essential to the effectiveness of the 
Federal government's environmental programs. Other agencies, such as 
NASA, National Science Foundation, the Center for Disease Control, 
OSHA, Department of Defense and Department of Energy also have a direct 
or indirect role in this issue. The GAO report should more clearly 
recognize that a comprehensive set of environmental indicators has the 
potential for benefiting environmental management across all Federal 
agencies.

On pages 7, 22 and 30, the proposed report identifies two closely 
related challenges to effective indicators: linking management actions 
to environmental conditions and integrating various indicators to 
better understand the environment. On page 37, the proposed report 
mentions the work of the Integration and Synthesis Group, an important 
effort to address these challenges. Over the past year, the group has 
made significant progress in the development of a systems-based 
conceptual framework designed to facilitate the selection, 
interpretation and reporting of indicators based on a better 
understanding of the linkages among management actions, social and 
economic processes and environmental processes. Participants in the 
Integration and Synthesis Group hope that the conceptual framework can 
be used as the basis for promoting greater integration among the 
various indicator projects and, as a basis for developing of a 
comprehensive national system of indicators that draws upon such 
projects.

I also recommend that the report take note of the Program Assessment 
Rating Tool (PART) recently developed by the Office of Management and 
Budget to improve program performance. Evaluation of federal agencies' 
indicator work through PART will enable both the executive and 
legislative branch of government to better understand program 
performance and identify opportunities for improvement.

Finally, any indicator system is only as strong as the relevance, scope 
and accuracy of the data that informs it. Your proposed report 
currently does not discuss the recent advancement, led by the federal 
government, of the Global Earth Observation System of Systems (GEOSS) 
which will include an Integrated Ocean Observing System. The 
Administration pressed for and obtained worldwide acknowledgement of 
the need for such a system in the Johannesburg Plan of Action that was 
produced at the World Summit on Sustainable Development. The 
Administration then obtained a concrete commitment to action from our 
G-8 partners. Two subsequent summits have lead to a formal partnership 
that continues. A detailed description can be found at http://
earthobservations.org/ I suggest that your report include a discussion 
of this effort.

In closing, I would like to emphasize the CEQ's commitment to improving 
our capacity to promote coordination and integration of information on 
environmental and natural resource conditions in the United States. The 
proposed GAO report documents many of the challenges that this effort 
involves and provides thoughtful recommendations for addressing them.	I 
look forward to the broader discussion that I know your report will 
stimulate.

Yours Sincerely,

Signed by: 

James L. Connaughton: 

[End of section]

Appendix X: Comments from the Department of the Interior:

United States Department of the Interior:

OFFICE OF THE ASSISTANT SECRETARY POLICY, MANAGEMENT AND BUDGET:  
Washington, D.C. 20240:

OCT 25 2004:

Mr. Ed Kratzer: 
Assistant Director: 
Natural Resources and Environment Division: 
U.S. Government Accountability Office: 
441 G Street, N. W.
Washington, D.C. 20548:

Dear Mr. Kratzer:

Thank you for providing the Department of the Interior the opportunity 
to review and comment on the Draft U. S. Government Accountability 
Office report entitled "Environmental Indicators: Better Coordination 
Is Needed to Develop Environmental Indicator Sets that Inform 
Decisions" (GAO-05-52). In general, we agree with the findings and 
recommendations in the report, including the recommendation that the 
Chair of the Council on Environmental Quality (CEQ) direct the 
Interagency Working Group on Indicator Coordination to study options 
for performing the functions necessary to guide the development, 
coordination, and integration of environmental indicator sets. This is 
essential to assure full consideration of the range of agency programs 
that are noted in the report, including their unique characteristics 
and complex interrelationships.

The enclosure provides specific comments from the Minerals Management 
Service and the National Park Service. We hope our comments will assist 
you in preparing the final report.

Sincerely,

Signed for: 

P. Lynn Scarlett: 
Assistant Secretary:
Policy, Management and Budget:

Enclosure: 

[End of section]

Appendix XI: GAO Contact and Staff Acknowledgments:

GAO Contact:

Ed Kratzer, (202) 512-6553:

Staff Acknowledgments:

Key contributions to this report were made by Nancy Bowser, Chase 
Huntley, Richard Johnson, Kerry Lipsitz, Jonathan McMurray, Mark 
Metcalfe, and Nathan Morris. Also contributing to this report were 
Jonathan Dent, Evan Gilman, Scott Heacock, R. Denton Herring, Kim 
Raheb, and Greg Wilmoth.

[End of section]

Bibliography:

[End of section]

Bakkes, J.A., G.J. van den Born, J.C. Helder, R.J. Swart, C.W. Hope, 
and J.D.E. Parker. An Overview of Environmental Indicators: State of 
the Art and Perspectives. Report commissioned by the United Nations 
Environment Programme. June 1994.

Bernard, James. State of the Science Ecological Indicators Report. 
Unpublished draft prepared for the U.S. Environmental Protection 
Agency. April 2002.

Bureau of Economic Analysis. "Integrated Economic and Environmental 
Satellite Accounts." Survey of Current Business. Washington, D.C; 
April 1994: 33-49.

Carter, J. "Environmental Priorities and Programs--a Message to 
Congress." Weekly Compilation of Presidential Documents, vol. 15, no. 
26 (July 2, 1979): 1353-1373.

Clarke, David P. "The State of the Nation's Ecosystems: Measuring the 
Lands, Waters, and Living Resources of the United States." Environment, 
vol. 45, no. 5 (2003): 40-43.

Department of Public Policy, University of North Carolina at Chapel 
Hill. Environmental Management Systems: Do They Improve Performance? 
Project final report: Executive summary prepared for the U.S. 
Environmental Protection Agency. January 2003.

Duncan, Joseph W., and Andrew C. Gross. Statistics for the 21st 
Century: Proposals for Improving Statistics for Better Decision Making. 
Chicago: Irwin, 1995.

Energy, Environment, and Resources Center, University of Tennessee. 
Expertise and the Policy Cycle, by Jack Barkenbus. Knoxville, 
Tennessee: September 1998.

Environment Canada. Environmental Indicators and State of the Reporting 
in Canada--Part 1: Current Trends, Status, and Perceptions. Draft 
background paper to a national environmental indicator and state of the 
environment reporting strategy; Washington, D.C; August 2003.

Environment Canada. Canadian Information System for the Environment: 
Sharing Environmental Decisions. Final report by the Task Force on a 
Canadian Information System for the Environment. Cat. No. EN21-207/
2001E. Hull, Quebec; October 2001.

Environment Canada. A National Strategy for Environmental Indicators 
and State of the Environment Reporting in Canada: Proposed Options. 
Draft report. Hull, Quebec; August 2003.

Environmental Council of States and U.S. Environmental Protection 
Agency Information Management Workgroup. National Environmental 
Information Exchange Network: Information Package. Washington, D.C; 
June 2001.

Flynn, Patrice, David Berry, and Theodore Heintz. "Sustainability and 
Quality of Life Indicators: Toward the Integration of Economic, Social 
and Environmental Measures." Indicators: The Journal of Social Health, 
vol. 1, no. 4 (fall 2002): 1-21.

Harwell, Mark A., Victoria Myers, Terry Young, Ann Bartuska, Nancy 
Gassman, John H. Gentile, Christine C. Harwell, Stuart Appelbaum, John 
Barko, Billy Causey, Christine Johnson, Agnes McLean, Ron Smola, Paul 
Templet, and Stephen Tosini. "A Framework for an Ecosystem Integrity 
Report Card: Examples from South Florida Show How an Ecosystem Report 
Card Links Societal Values and Scientific Information." BioScience, 
vol. 49, no. 7 (July 1999): 543-556.

Innes, Judith E. Knowledge and Public Policy: The Search for Meaningful 
Indicators, 2nd ed. New Brunswick, New Jersey: Transaction Publishers, 
2002.

International Council for Science. Making Science for Sustainable 
Development More Policy Relevant: New Tools for Analysis. ICSU Series 
on Science for Sustainable Development No. 8. Paris, 2002.

International Institute for Sustainable Development. "Review Paper on 
Selected Capital Based Sustainable Development Indicator Frameworks." 
Study for the Steering Committee of the National Roundtable on 
Environment and the Economy (Canada). Winnipeg, Manitoba; December 
2000.

National Academy of Public Administration. Evaluating Environmental 
Progress: How EPA and the States Can Improve the Quality of Enforcement 
and Compliance Information. Report for the U.S. Environmental 
Protection Agency. June 2001.

National Academy of Public Administration. Setting Priorities, Getting 
Results: A New Direction for EPA. Report to the U.S. Congress. Silver 
Spring, Maryland: Sherwood Fletcher Associates, 1995.

National Council for Science and the Environment. Recommendations for 
Improving the Scientific Basis for Environmental Decisionmaking. Report 
of the first National Conference on Science, Policy, and the 
Environment. Washington, D.C; December 2000.

National Research Council. Ecological Indicators for the Nation. 
Washington, D.C.: National Academy Press, 2000.

National Research Council. Our Common Journey: A Transition toward 
Sustainability. Washington, D.C.: National Academy Press, 1999.

National Research Council. Nature's Numbers: Expanding the National 
Economic Accounts to Include the Environment. Washington, D.C.: 
National Academy Press, 1999.

National Science and Technology Council. Integrating the Nation's 
Environmental Monitoring and Research Networks and Programs: A Proposed 
Framework. Report by the Committee on Environment and Natural 
Resources. Washington, D.C; March 1997.

O'Malley, Robin, Kent Cavender-Bares, and William C. Clark. "Providing 
'Better' Data: Not as Simple as It Might Seem." Environment, vol. 45, 
no. 4 (May 2003): 8-18.

Riche, Martha Farnsworth. The United States of America: Developing Key 
National Indicators. Paper presented at the Forum on Key National 
Indicators; Washington, D.C; February 2003.

Sustainability Institute. Indicators and Information Systems for 
Sustainable Development, by Donella Meadows. Report to the Balaton 
Group. Hartland, Vermont; September 1998.

Tonn, Bruce, and Robert Turner. Environmental Decision Making and 
Information Technology: Issues Assessment. NCEDR/99-01. National 
Center for Environmental Decision-Making Research. Knoxville, 
Tennessee; May 1999.

Train, Russell. "The Quest for Environmental Indices." Science 178, no. 
4057 (October 1972): 1.

United Nations, Integrated Environmental and Economic Accounting 2003: 
Handbook of National Accounting. ST/ESA/STAT/SER.F/61/Rev.1. (1992).

U.S. Environmental Protection Agency. A Framework for Assessing and 
Reporting on Ecological Condition: An SAB Report. EPA SAB-EPEC-02-009. 
Science Advisory Board. June 2002.

U.S. Environmental Protection Agency. Environmental Statistics and 
Information Division, Office of Policy, Planning, and Evaluation. A 
Conceptual Framework to Support Development and Use of Environmental 
Information in Decision-Making. EPA 239-R-95-012, Washington, D.C; 
April 1995.

U.S. Environmental Protection Agency. National Advisory Council for 
Environmental Policy and Technology. Managing Information as a 
Strategic Resource: Final Report and Recommendations of the Information 
Impacts Committee. EPA 100-R-98-002. Washington, D.C; January 1998.

World Bank. Policy Applications of Environmental Accounting, by G. 
Lange. Environmental Economic Series, Paper No. 88. New York, January 
2003.

World Conservation Union. Lessons Learned from Environmental 
Accounting: Findings from Nine Case Studies, by Joy Hecht. Washington, 
D.C; October 2000.

World Resources Institute. Environmental Indicators: A Systematic 
Approach to Measuring and Reporting on Environmental Policy Performance 
in the Context of Sustainable Development, by Allen Hammond, Albert 
Adriaanse, Eric Rodenburg, Dirk Bryant, and Richard Woodward. 
Washington, D.C; May 1995.

[End of section]

Related GAO Products:

Informing Our Nation: Improving How to Understand and Assess the USA's 
Position and Progress. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-05-01]
Washington, D.C.: November 10, 2004.

Watershed Management: Better Coordination of Data Collection Efforts 
Needed to Support Key Decisions. 
[Hyperlink, http://www.gao.gov/cgi- bin/getrpt?GAO-04-382]
Washington, D.C.: June 7, 2004.

Geographic Information Systems: Challenges to Effective Data Sharing.
GAO-03-874T. 
Washington, D.C.: June 10, 2003.

Forum on Key National Indicators: Assessing the Nation's Position and 
Progress. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03- 672SP]
Washington, D.C.: May 1, 2003.

Great Lakes: An Overall Strategy and Indicators for Measuring Progress 
Are Needed to Better Achieve Restoration Goals. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-515]
Washington, D.C.: April 30, 2003.

Environmental Protection: Observations on Elevating the Environmental 
Protection Agency to Cabinet Status. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-552T]
Washington, D.C.: March 21, 2002.

Environmental Information: EPA Needs Better Information to Manage Risks 
and Measure Results. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-01-97T]
Washington, D.C.: October 3, 2000.

Environmental Information: EPA Is Taking Steps to Improve Information 
Management, but Challenges Remain. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-99-261]
Washington, D.C.: September 17, 1999.

Environmental Protection: EPA's Problems with Collection and Management 
of Scientific Data and Its Efforts to Address Them. GAO/T-RCED-95-174. 
Washington, D.C.: May 12, 1995.

Environmental Protection: Meeting Public Expectations with Limited 
Resources. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED- 91-97]
Washington, D.C.: June 18, 1991.

Environmental Protection Agency: Protecting Human Health and the 
Environment through Improved Management. 
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-88-101]
Washington, D.C.: August 16, 1988. 

(360343):

FOOTNOTES

[1] It is difficult to determine exactly what the federal government 
spends each year on environmental information. Although the Office of 
Management and Budget annually publishes funding for natural resource, 
energy, and environmental statistics in Statistical Programs of the 
United States Government, we were not able to disaggregate the totals 
by program. Moreover, the National Science and Technology Council 
reported in 1997 that, in fiscal year 1995, the federal government 
spent about $650 million on environmental research and monitoring 
networks and programs, but that assessment has not been updated. See 
appendix VII for more information. 

[2] In addition, there are several governmentwide requirements that 
affect environmental data management. For example, the Information 
Quality Act requires the Office of Management and Budget to provide 
guidance to federal agencies for maximizing the quality of information 
they disseminate.

[3] Many organizations in the United States are developing 
comprehensive key indicator systems--organized, systematic efforts to 
produce selected economic, social, and environmental indicators--to 
assess position and progress toward specific goals. See GAO, Informing 
Our Nation: Improving How to Understand and Assess the USA's Position 
and Progress, GAO-05-01 (Washington, D.C.: Nov. 10, 2004). 

[4] National Science and Technology Council, Committee on Environment 
and Natural Resources, Integrating the Nation's Environmental 
Monitoring and Research Networks and Programs: A Proposed Framework 
(Washington, D.C; March 1997).

[5] National Council for Science and the Environment, Improving the 
Scientific Basis for Decisionmaking: A Report from the first National 
Conference on Science, Policy, and the Environment (Washington, D.C; 
December 2000).

[6] Effective May 15, 2000, the Federal Reports Elimination and Sunset 
Act (Pub. L. No. 104-66, § 3003) terminated the CEQ reporting 
requirement that had appeared in the National Environmental Policy Act.

[7] See appendix I for a more thorough description of our survey 
methodology and its limitations.

[8] P.L. 105-391 (1998).

[9] Note that GAO did not attempt to independently evaluate the costs, 
benefits, or risks of developing and using indicator sets that accrue 
from the positive impacts reported by indicator set users.

[10] The Federal Water Pollution Control Act Amendments of 1972--which, 
as amended, is commonly known as Clean Water Act--requires EPA to 
compile states' biennial reports on the quality of their waters into 
the National Water Quality Inventory. See 33 U.S.C.A. § 1315(b).

[11] For a recent example, see Jacksonville Community Council Inc., 
Making Jacksonville a Clean City (Jacksonville, Florida; spring 2002).

[12] GAO, Results-Oriented Government: GPRA Has Established a Solid 
Foundation for Achieving Greater Results, GAO-04-38 (Washington, D.C.: 
Mar. 10, 2004).

[13] GAO, Major Management Challenges and Program Risks: Environmental 
Protection Agency, GAO-01-257 (Washington, D.C.: Jan. 1, 2001).

[14] See GAO, Wetlands Overview: Problems with Acreage Data Persist, 
GAO/RCED-98-150 (Washington, D.C.: July 1998); and Results-Oriented 
Management: Agency Crosscutting Actions and Plans in Border Control, 
Flood Mitigation and Insurance, Wetlands, and Wildland Fire Management, 
GAO-03-321 (Washington, D.C.: Dec. 20, 2002).

[15] GAO, Watershed Management: Better Coordination of Data Collection 
Efforts Needed to Support Key Decisions, GAO-04-382 (Washington, D.C.: 
June 7, 2004).

[16] Council on Environmental Quality, Environmental Quality: The First 
Annual Report of the Council on Environmental Quality (Washington, 
D.C; 1970).

[17] William Clark, Thomas Jorling, and William Merrell, "Foreword," 
Designing a Report on the State of the Nation's Ecosystems (Washington, 
D.C.: H. John Heinz III Center for Economics and the Environment, 
1999).

[18] An official from the Heinz Center reported that efforts have been 
made to enhance the likelihood that future reports will be able to 
quantify this metric.

[19] 42 U.S.C. 4321.

[20] These indictor sets include those developed by the Roundtable on 
Sustainable Forests, Sustainable Rangelands Roundtable, Sustainable 
Water Resources Roundtable, Sustainable Minerals Roundtable, EPA's 
Environmental Indicators Initiative, and the Heinz Center's State of 
the Nation's Ecosystems project.

[21] GAO, The Government Performance and Results Act: 1997 
Governmentwide Implementation Will Be Uneven, GAO/GGD-97-109 
(Washington, D.C.: June 2, 1997).

[22] GAO, Results-Oriented Government: GPRA Has Established a Solid 
Foundation for Achieving Greater Results, GAO-04-38 (Washington, D.C.: 
Mar. 10, 2004).

[23] GAO, Major Management Challenges and Program Risks: Environmental 
Protection Agency, GAO-01-257 (Washington, D.C.: Jan. 1, 2001).

[24] EPA is not the only agency to struggle with this issue. See GAO, 
Major Management Challenges and Program Risks: Department of the 
Interior, GAO-03-104 (Washington, D.C.: Jan. 1, 2003); and Department 
of Agriculture: Status of Achieving Key Outcomes and Addressing Major 
Management Challenges, GAO-01-761 (Washington, D.C.: Aug. 23, 2001).

[25] As we noted in our January 2003 report on EPA's major management 
challenges and program risks, the indicators initiative has the 
potential to make a substantial contribution to measuring EPA's 
progress within an overall framework of ecological and human health, 
assisting EPA's strategic planning efforts, and facilitating a 
transition to performance-based management driven by environmental 
goals. See GAO, Major Management Challenges and Program Risks: 
Environmental Protection Agency, GAO-03-112 (Washington, D.C.: Jan. 1, 
2003).

[26] GAO-03-112, 4.

[27] United Nations, United Nations Conference on Environment and 
Development, Agenda 21 (1992).

[28] KPMG, KPMG Global Sustainability Services, KPMG International 
Survey of Corporate Sustainability Reporting 2002 (De Meern, The 
Netherlands; June 2002).

[29] Corporate Social Responsibility Network, The State of Global 
Environmental and Social Reporting, (Shrewsbury, United Kingdom; 2001).

[30] World Bank, International Finance Corporation, Race to the Top: 
Attracting and Enabling Global Sustainable Business, Business Survey 
Report, by J. Berman and T. Webb (Washington, D.C; October 2003).

[31] KPMG, KPMG International Survey of Corporate Sustainability 
Reporting 2002.

[32] G. Lange, Policy Applications of Environmental Accounting, Paper 
88, World Bank Environmental Economic Series (Washington, D.C; January 
2003).

[33] United Nations, Integrated Environmental and Economic Accounting 
2003: Handbook of National Accounting (1992). 

[34] CSERA was first published in 1997, then again in 2000, and updates 
are planned for 2004.

[35] National Research Council, Nature's Numbers: Expanding the 
National Economic Accounts to Include the Environment (Washington, 
D.C.: National Academy Press, 1999). 

[36] National Science and Technology Council, Integrating the Nation's 
Environmental Monitoring and Research Networks and Programs: A Proposed 
Framework, a report by the Committee on Environment and Natural 
Resources (Washington, D.C; March 1997).

[37] OMB reports on programs that receive direct funding of at least 
$500,000 on statistical activities, which include: (1) the planning of 
surveys and other techniques of data collection; (2) personnel; (3) 
collection, processing, or tabulation of statistical data for 
publication, dissemination, research, analysis, or program management 
and evaluation; (4) publication of data and studies; (5) methodological 
research; (6) data analysis; (7) forecasts or projections made 
available for governmentwide or public use; (8) publication of data 
collected by others; (9) secondary data series or models that are an 
integral part of generating statistical series or forecasts; (10) 
management or coordination of statistical operations; and (11) 
statistical consulting or training.

[38] Paperwork Reduction Act of 1995, 44 U.S.C. §3504(e)(2).

GAO's Mission:

The Government Accountability Office, the investigative arm of 
Congress, exists to support Congress in meeting its constitutional 
responsibilities and to help improve the performance and accountability 
of the federal government for the American people. GAO examines the use 
of public funds; evaluates federal programs and policies; and provides 
analyses, recommendations, and other assistance to help Congress make 
informed oversight, policy, and funding decisions. GAO's commitment to 
good government is reflected in its core values of accountability, 
integrity, and reliability.

Obtaining Copies of GAO Reports and Testimony:

The fastest and easiest way to obtain copies of GAO documents at no 
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains 
abstracts and full-text files of current reports and testimony and an 
expanding archive of older products. The Web site features a search 
engine to help you locate documents using key words and phrases. You 
can print these documents in their entirety, including charts and other 
graphics.

Each day, GAO issues a list of newly released reports, testimony, and 
correspondence. GAO posts this list, known as "Today's Reports," on its 
Web site daily. The list contains links to the full-text document 
files. To have GAO e-mail this list to you every afternoon, go to 
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order 
GAO Products" heading.

Order by Mail or Phone:

The first copy of each printed report is free. Additional copies are $2 
each. A check or money order should be made out to the Superintendent 
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or 
more copies mailed to a single address are discounted 25 percent. 
Orders should be sent to:

U.S. Government Accountability Office

441 G Street NW, Room LM

Washington, D.C. 20548:

To order by Phone:

 

Voice: (202) 512-6000:

TDD: (202) 512-2537:

Fax: (202) 512-6061:

To Report Fraud, Waste, and Abuse in Federal Programs:

Contact:

Web site: www.gao.gov/fraudnet/fraudnet.htm

E-mail: fraudnet@gao.gov

Automated answering system: (800) 424-5454 or (202) 512-7470:

Public Affairs:

Jeff Nelligan, managing director,

NelliganJ@gao.gov

(202) 512-4800

U.S. Government Accountability Office,

441 G Street NW, Room 7149

Washington, D.C. 20548: