Defense Management: Tools for Measuring and Managing Defense	 
Agency Performance Could Be Strengthened (13-SEP-04, GAO-04-919).
                                                                 
GAO was mandated to assess the effectiveness of defense agency	 
performance contracts as management tools. As agreed, GAO also	 
reviewed other tools (performance plans and balanced scorecards) 
and focused on three defense agencies--the Defense Logistics	 
Agency (DLA), the Defense Information Systems Agency (DISA), and 
the Department of Defense Education Activity (DODEA). GAO	 
addressed (1) the extent that the defense agencies initially used
performance contracts, including whether this tool addressed	 
attributes associated with results-oriented management; (2)	 
defense agencies' efforts to implement performance plans using	 
lessons learned from the initial contracts; and (3) the extent	 
DOD established mechanisms to share lessons learned. GAO reviewed
the content of these tools, but not the actual or reported	 
performance. DISA has not yet finalized its scorecard, thus this 
report discusses only DISA's plans for its scorecard.		 
-------------------------Indexing Terms------------------------- 
REPORTNUM:   GAO-04-919 					        
    ACCNO:   A12273						        
  TITLE:     Defense Management: Tools for Measuring and Managing     
Defense Agency Performance Could Be Strengthened		 
     DATE:   09/13/2004 
  SUBJECT:   Agency missions					 
	     Internal controls					 
	     Performance measures				 
	     Strategic planning 				 
	     Performance plans					 

******************************************************************
** This file contains an ASCII representation of the text of a  **
** GAO Product.                                                 **
**                                                              **
** No attempt has been made to display graphic images, although **
** figure captions are reproduced.  Tables are included, but    **
** may not resemble those in the printed version.               **
**                                                              **
** Please see the PDF (Portable Document Format) file, when     **
** available, for a complete electronic file of the printed     **
** document's contents.                                         **
**                                                              **
******************************************************************
GAO-04-919

                 United States Government Accountability Office

           GAO	Report to the Committee on Armed Services, U.S. Senate

September 2004

DEFENSE MANAGEMENT

      Tools for Measuring and Managing Defense Agency Performance Could Be
                                  Strengthened

                                       a

GAO-04-919

[IMG]

September 2004

DEFENSE MANAGEMENT

Tools for Measuring and Managing Defense Agency Performance Could Be
Strengthened

What GAO Found

Since fiscal year 1998, the Department of Defense (DOD) has implemented
various tools to help manage and oversee the performance of defense
agencies. Between fiscal year 1999 and 2003, DLA, DISA, and DODEA
initially used "performance contracts"-internal management agreements- to
bring specific problems to the attention of senior DOD and agency
leadership. While the contracts produced some useful information for
decision makers, this tool would have been more effective for assessing
performance, making resource allocation decisions, and taking corrective
actions if DOD had required the agencies to include certain attributes
associated with results-oriented management. Such attributes include
aligning agency performance goals and measures with agency strategic plans
and departmentwide goals; identifying individuals accountable for
achieving results; providing a comprehensive view of organizational
performance; linking resource needs to performance; discussing data
quality; and providing contextual information, including external factors
that affect reported performance.

Beginning in fiscal year 2003, DOD renamed the performance contracts as
"performance plans" and gave the defense agencies the option to use a
"balanced scorecard" approach, a tool used in the public and private
sectors to assess organizational performance. Based on experiences using
the initial contracts, DOD took steps to strengthen performance plans and
scorecards by revising the oversight and review process, requiring
performance measures to align with agency and departmentwide goals, and
requiring measures to provide a more comprehensive view of agency
performance. DLA's scorecard, DODEA's performance plan, and DISA's plans
for the agency's scorecard incorporated these changes and other attributes
to varying degrees. While these tools have the potential to provide
information useful to decision makers, they would be strengthened if DOD
had required the agencies to include additional attributes such as
designating specific individuals responsible for achieving results;
identifying the relationship between resource needs and performance;
reporting on data quality; and providing contextual information to allow
top leaders to understand the extent of progress made, take corrective
actions to achieve goals, and establish realistic performance goals for
future years. With these attributes, decision makers would potentially
gain additional insights into agency performance and areas needing greater
management attention.

DOD has developed mechanisms, such as a performance management Web site
and roundtables, to help agencies share lessons learned from implementing
performance plans and scorecards. In response to GAO's suggestions during
this review, DOD recognized the need to continue to hold roundtables more
frequently. DLA and DISA have also proactively shared their experiences
with each other.

                 United States Government Accountability Office

Contents

  Letter 1

Results in Brief 4 Background 7 While Used to Varying Degrees, Performance
Contracts That We

Reviewed Could Have Been More Effective Had They Included

Certain Attributes Associated with Results-Oriented

Management 10 The Plan and Scorecard We Reviewed Show Improvement, but

Could Be Strengthened by Including Certain Key Attributes 19 DOD Captures
and Shares Lessons Learned 35 Conclusions 35 Recommendations for Executive
Action 36 Agency Comments and Our Evaluation 37

Appendix I Scope and Methodology

Appendix II	Department of Defense's Risk Management Framework

Appendix III	Fiscal Year When Defense Agencies Implemented
Their First Performance Contract, Plan, or Scorecard 42

Appendix IV	Certain Key Attributes Associated with Results-Oriented
Management

Appendix V Comments from the Department of Defense 48

Appendix VI Key Contact and Staff Acknowledgments 51

                            Related GAO Products 52

Tables

Table 1: Time Frames for DLA's, DISA's, and DODEA's Use ofPerformance
Contracts, Performance Plans, or Balanced Scorecards

Table 2: Extent To Which DLA's, DISA's, and DODEA'sPerformance Contracts
Incorporated Certain Key Attributes Associated with Results-Oriented
Management

Table 3: Extent To Which DLA's Scorecard and DODEA's Plan Incorporated
Certain Key Attributes Associated withResults-Oriented Management

3 13 21

Figures

Figure 1: Depiction of How DLA's May 2004 Scorecard Supports
Departmentwide Goals and Cascades to Lower Organizational Levels

Figure 2: Example of the Type of Information in DODEA's Performance Plan
That OSD Officials Monitor, as of May 2004

Figure 3: Detailed Data on DODEA's Pupil to Teacher Ratio as of May 2004
Figure 4: Format of a Sample Measure from DISA's Draft September 2003
Balanced Scorecard

Figure 5: DOD's Risk Management Framework

Figure 6: Defense Agencies

Figure 7: Defense Field Activities

23

25

26

34 41 43 44

Abbreviations

DISA Defense Information Systems Agency DLA Defense Logistics AgencyDOD
Department of Defense DODEA Department of Defense Education ActivityOSD
Office of the Secretary of Defense PA&E Program Analysis and Evaluation
USD/P&R Under Secretary of Defense for Personnel and Readiness

This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed in
its entirety without further permission from GAO. However, because this
work may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this material
separately.

United States Government Accountability Office Washington, DC 20548

September 13, 2004

The Honorable John W. WarnerChairman The Honorable Carl Levin Ranking
Minority MemberCommittee on Armed Services United States Senate

Department of Defense (DOD) agencies and field activities, such as
theDefense Logistics Agency, are a key part of DOD's business operations
and provide numerous support services, ranging from information technology
and education of servicemembers' dependents to logistics support for the
department.1 The agencies' customers include the military services, other
agencies, and military personnel and their families. During the past2
decades, service officials have expressed concern about inadequate
oversight and performance of the defense agencies; specifically, the lack
of customer responsiveness and inefficient business processes in light
ofgrowing defense agency budgets.

As part of a departmentwide effort, known as the Defense ReformInitiative,
aimed at improving DOD's business operations, in 1998, the Deputy
Secretary of Defense directed the defense agencies to developagreements
between the defense agencies and the Under and Assistant Secretaries to
which they report. Known as "performance contracts" at the time, these
agreements were intended to help improve the Office of the Secretary of
Defense's (OSD) oversight of these agencies as well as their performance.
According to a DOD official, these performance contractswere not intended
to be contracts in the legal sense, but rather were one of a number of
management tools-such as strategic plans-used by the defense agencies to
improve performance by setting performance goals, focusing the attention
of OSD and agency leaders on customer concerns, and bringing management
visibility over key areas of the agencies'

1 Defense agencies and DOD field activities perform similar support
functions. However, the field activities are smaller and serve a more
limited portion of DOD than do defense agencies. For purposes of this
report, we use the term "defense agencies" to refer collectively to
defense agencies and field activities.

performance.2 In fiscal year 2003, DOD refined its approach to performance
management and renamed the contracts as "performance plans" in order to
more accurately reflect the intent of the agreements. According to DOD,
these plans are intended to provide a more comprehensive view of agency
goals and performance and align with agency strategic plans and
departmentwide goals. DOD gave the defense agencies the option to use a
"balanced scorecard"-a form of performance plan.3 Balanced scorecards
organize performance measures by organizational drivers-including
financial, customer, and internal business processes as well as workforce
learning and growth-to help measure performance, make improvements, and
assess how well organizations are positioned to perform in the future.
Although the plans and scorecards themselves are not intended to resolve
or prevent problems, OSD and defense agency officials intend to use these
tools to monitor agency performance, identify developing problems and
corrective actions to improve business practices, meet customer needs, and
set funding priorities.

We have conducted an extensive body of work over the last decade
identifying leading practices in results-oriented management, including
performance measurement and reporting. In our prior work on government
performance, we identified attributes associated with results	oriented
management that help inform management decision making.4 For this report,
we identified seven key attributes that we feel are the most important for
accurately assessing the strengths and weaknesses of programs and making
improvements. These attributes include aligning performance measures with
strategic goals, assigning accountability for achieving results,
developing measures that demonstrate results, developing measures that
provide a comprehensive view of agency

2 For purposes of this report, we use the term "performance contracts" to
include both performance contracts as well as accompanying annual
performance contract reports that discuss agencies' progress toward
meeting performance contract goals.

3 The balanced scorecard is a private-sector concept introduced by Robert
Kaplan and David Norton in 1992 to assess organizational performance and
is used by several government agencies.

4 Some of these products include GAO, Results-Oriented Government: GPRA
Has Established a Solid Foundation for Achieving Greater Results,
GAO-04-38 (Washington, D.C.: Mar. 10, 2004); The Results Act: An
Evaluator's Guide to Assessing Agency Annual Performance Plans,
GAO/GGD-10.1.20 (Washington, D.C.: Apr. 1, 1998); and Executive Guide:
Effectively Implementing the Government Performance and Results Act,

GAO/GGD-96-118 (Washington, D.C.: June 1, 1996).

performance, linking resource needs to performance, discussing data
quality, and including contextual information.

The Senate Committee on Armed Services report accompanying the National
Defense Authorization Act for Fiscal Year 2004 directed us to assess the
effectiveness of defense agency performance contracts as management
tools.5 Because of the evolution from performance contracts to plans or
balanced scorecards, this report evaluates all of these tools. Of the 14
defense agencies that presently use these tools, we focused our work on
3-the Defense Logistics Agency (DLA), the Defense Information Systems
Agency (DISA), and the Department of Defense Education Activity
(DODEA)-based on their levels of experience using performance contracts,
different types of missions, and types of services and customers.

Table 1 shows the time frames during which these agencies used performance
contracts, performance plans, or balanced scorecards.

Table 1: Time Frames for DLA's, DISA's, and DODEA's Use of Performance
Contracts, Performance Plans, or Balanced Scorecards

Fiscal year

           Agency Performance contract  Performance plan   Balanced scorecard 
              DLA             1999-2001    Not applicable           2003-2004 
             DISA             2000-2002              2003                2004 
            DODEA             2000-2003              2004      Not applicable 

Sources: DLA, DISA, and DODEA.

Note: Because DLA was transitioning to a balanced scorecard, with DOD's
approval, the agency did not develop a performance contract or plan during
fiscal year 2002. DODEA officials decided to use a performance plan rather
than exercise their option to develop a balanced scorecard for fiscal year
2004. DISA submitted a sample draft balanced scorecard to the Office of
the Secretary of Defense, Program Analysis and Evaluation in September
2003 in order to meet reporting requirements.

Although we obtained DISA's draft balanced scorecard dated September 2003,
we did not evaluate it against certain key attributes associated with
results-oriented management that we discuss in this reportbecause the
scorecard had not been finalized by the time we completed our review.
However, we do discuss DISA's plans for using a scorecard

5 S. Rep. No. 108-46, at 347 (2003).

and addressing the attributes, as applicable. Our objectives were to
evaluate

(1) the extent to which the defense agencies initially used performance
contracts to manage and measure agency performance, including whether the
contracts addressed certain key attributes associated with
results	oriented management, such as performance measurement and
reporting;

(2) the status of efforts by the defense agencies to implement performance
plans, including the extent to which these plans reflect any lessons
learned from DOD's experience with the initial performance contracts and
address certain key attributes associated with results-oriented
management; and

(3) the extent to which DOD has established mechanisms to capture and
share lessons learned from the implementation of performance plans.

To address these objectives, we collected data from and interviewed
knowledgeable OSD, defense agency, and service officials about the use of
performance contracts, plans, and balanced scorecards, as applicable, at
DLA, DISA, and DODEA. We obtained these agencies' performance contracts,
plans, and scorecards, as applicable, and analyzed them against certain
key attributes associated with results-oriented management that we
identified in our prior reports. We did not review actual or reported
agency performance, only the content of the tools and how they were used.
We also did not validate procedures that the defense agencies have in
place to ascertain the reliability of data used in their performance plans
and scorecards. We conducted our review between July 2003 and May 2004 in
accordance with generally accepted government auditing standards. Details
about our scope and methodology appear in appendix I.

Between fiscal year 1999 and 2003, DLA, DISA, and DODEA-the three agencies
we reviewed-initially used performance contracts to varying degrees to
identify performance issues, elevate management challenges and customer
concerns to the attention of senior OSD and defense agency leaders, and
take corrective action. The contracts were an important step toward more
results-oriented management practices for the defense agencies. However,
the contracts would have been more effective tools for informing decision
making and managing and measuring agency performance had they included
additional attributes associated with results-oriented management. The
contracts showed progress made and encouraged discussion among top agency
and OSD leaders about actions needed to achieve targeted results,
particularly in terms of quality,

Results in Brief

quantity, cost, and timeliness of services. For example, DODEA officials
told us that the agency's fiscal year 2003 performance contract included a
measure to achieve optimum pupil to teacher ratios that influenced DOD to
fund the hiring of 200 full-time teachers. However, DOD guidance did not
require performance contracts to include certain key attributes that we
have identified in prior reports that are associated with results-oriented
management. While other management documents may reflect some of these
attributes, it is important that they also be reflected in the contracts
to provide a comprehensive picture of goals and performance. Consequently,
the performance contracts could have been strengthened by including the
following attributes:

o 
Aligns goals and measures with agency and departmentwide goals.
Performance goals and measures for some of the agencies we reviewed only
partially aligned with agency and departmentwide goals and none of the
contracts cascaded goals and measures to lower organizational levels.

o 
Assigns accountability for achieving results. Individuals at lower
organizational levels accountable for accomplishing specific goals and
taking corrective actions were not identified.

o 
Demonstrates results. Contracts were inconsistent in using a combination
of clearly defined output and outcome goals accompanied by trend data.

o 
Provides a comprehensive view of agency performance. With the exception of
DODEA, the contracts we reviewed lacked some measures that could have
provided officials with a comprehensive view of factors that drive
organizational performance, such as financial, customer, and internal
business processes. The contracts, to some extent, did reflect the
different concerns of customers and stakeholders.

o 
Links resource needs to performance. Contracts did not discuss the
relationship between resource needs and performance outcomes.

o 
Discusses data quality. Contracts did not discuss whether data were
reliable, valid, or verifiable, or identify data sources.

o 
Provides contextual information. Contracts lacked some contextual
information to help officials assess the significance of factors that
affected reported performance, such as how performance measures

help achieve mission and departmentwide goals or the impact of contingency
operations on achieving performance goals.

Based on experiences using the initial performance contracts, DOD took
steps to strengthen performance plans and scorecards by revising the
oversight and review process as well as guidance to require that,
beginning with fiscal year 2004, performance measures align with agency
and departmentwide goals and provide a more comprehensive view of agency
performance. However, these tools could have been strengthened had DOD
required the defense agencies to include certain key attributes associated
with results-oriented management. For example, while DLA's scorecard
included a combination of output and outcome measures accompanied by trend
data, it did not include clearly defined measures. DODEA's performance
plan measures also were not consistently clearly defined and not always
accompanied by trend data. The scorecard and plan we reviewed also lacked
other attributes-not required by DOD-that could be useful to officials for
discussing and understanding competing priorities and resource allocation
constraints, assessing and understanding the extent of progress made
against existing performance targets, making improvements, and
establishing realistic performance goals for future years. Such attributes
include

o 
identification of specific individuals who are responsible for achieving
results;

o  linkage of resource needs to performance;

o 
discussion of data quality, including data sources, as well as whether
data were reliable, valid, and verifiable; and

o 
inclusion of contextual information such as external factors that could
affect performance and relevant information about the agencies' missions,
business lines, and customers.

DOD developed some mechanisms-such as a performance management Web site
and roundtable discussions-to help the defense agencies capture and share
experiences and lessons learned from developing performance plans and
balanced scorecards, and some agencies are proactively sharing such
information with each other. Based on our suggestions during this review,
DOD recognized the need to continue to hold roundtables to discuss
balanced scorecards. A DOD official also told us that DOD plans to revise
and update its performance management Web site to include defense
agencies' performance plans and balanced scorecards as well as guidance.
Some defense agencies, such as DLA and DISA, are sharing their experiences
with each other to leverage lessons learned.

This report contains recommendations to DOD aimed at making performance
plans and scorecards more informative for decision making and further
strengthening the potential of these tools for measuring and managing
agency performance. In commenting on a draft of this report, DOD generally
concurred with our recommendations.

                                   Background

Evolution of Performance Contracts

In 1997, the Secretary of Defense issued the Defense Reform Initiative
report outlining a plan for reforming the department's business
operations. DOD established a management oversight structure to help
sustain the direction and emphasis of these efforts. As part of its
efforts to improve its business operations, DOD began requiring selected
defense agencies to develop annual performance contracts for fiscal year
1999. These contracts were internal management agreements between the
principal staff assistants-also known as Under and Assistant
Secretaries-in the Office of the Secretary of Defense, the Deputy
Secretary, and the defense agencies. DOD intended the performance
contracts to improve oversight of the defense agencies by (1) identifying
programming and budgeting issues, (2) providing agency directors with
clear objectives and performance targets, (3) establishing open and direct
communications with customers to, among other things, demonstrate how
their actions affect agency performance, and (4) changing the way DOD does
business. The contracts included improvement goals for the agencies in
terms of cost, productivity, quality, and responsiveness to customers. DOD
also required the defense agencies to assess their progress toward
achieving performance contract goals in annual performance reports. A
Defense Management Council-chaired by the Deputy Secretary and consisting
of key civilian and military leaders-was created to oversee Defense Reform
Initiative efforts, and a Defense Agency Task Force-chaired by the Deputy
Director of Resource Analysis in the Office of the Secretary of Defense,
Program Analysis and Evaluation (OSD/PA&E) and consisting of senior-level
representatives from the Under Secretaries of Defense, the service
secretaries, and the Joint Staff-was formed to oversee the development and
review of performance contracts.

DOD initially selected 4 defense agencies to implement performance
contracts during fiscal year 1999, and gradually expanded the requirement
to a total of 10 agencies by fiscal year 2002. At that time, DOD issued
guidelines that established a standard format for performance contracts
and recommended that certain attributes associated with results-oriented

management be included. DOD initially excluded defense agencies with
intelligence-or research-related missions from the requirement because
officials believed it would be more appropriate to focus primarily on
certain agencies with businesslike missions and develop lessons learned
based on their experiences. The 4 agencies DOD initially selected were

o  the Defense Logistics Agency, a revolving fund activity;

o  the Defense Finance and Accounting Service, a businesslike agency;

o 
the Defense Contract Audit Agency, a small and appropriated fund agency;
and

o  the Defense Health Program.

As part of the current administration's focus on business transformation,
DOD issued guidance in February and August 2003 that refined its approach
to performance management and renamed the contracts as performance plans.
These plans, first implemented in September 2003, build on the performance
goals and measures established in the performance contracts and are
intended to include a more comprehensive set of performance measures that
align with agency strategic plans and departmentwide goals set forth in
DOD's risk management framework. DOD uses the risk management framework,
established in the 2001 Quadrennial Defense Review Report, to consider
trade-offs among defense objectives and resource constraints. The
framework consists of four dimensions of risk:

o 
Force management-the ability to recruit, retain, train, and equip
sufficient numbers of quality personnel and sustain the readiness of the
force while accomplishing its many operational tasks;

o 
Operational-the ability to achieve military objectives in a near-term
conflict or other contingency;

o 
Future challenges-the ability to invest in new capabilities and develop
new operational concepts to dissuade or defeat mid-to long-term military
challenges; and

o 
Institutional-the ability to develop management practices and controls
that use resources efficiently and promote the effective operation of the
defense establishment.

These risk areas form the basis for DOD's annual performance goals and for
tracking associated performance results. Appendix II depicts DOD's risk
management framework.

DOD gave the defense agencies the option to use a balanced scorecard to
develop performance goals and report on progress. Currently, according to

an official in OSD/PA&E, DOD requires 10 of 27 defense agencies with
businesslike missions to have either performance plans or balanced
scorecards. Five defense agencies use performance plans, and the remainder
have exercised the option to develop balanced scorecards. In addition,
this official told us that 3 intelligence agencies have developed
performance plans in order to demonstrate how they are coordinating with
DOD on strategic and budgetary planning in the post-September 2001
environment. Appendix III depicts the first fiscal year when each defense
agency began using performance contracts, plans, or scorecards.

DOD Organizations Responsible for Development of Performance Contracts and
Plans

The Director, PA&E, has responsibility for overseeing the development of
defense agency performance plans, including reviewing the progress of each
defense agency. Specifically, this official oversees the development of
defense agency performance plans and balanced scorecards, provides
guidance for interpreting and applying DOD's risk management framework,
and, in conjunction with the relevant Under or Assistant Secretary of
Defense, reviews the progress of each agency in building its performance
plan or scorecard and reporting on its overall performance. The Under and
Assistant Secretaries of Defense are responsible for overseeing the
defense agencies and implementing performance plans and balanced
scorecards.

Our Prior Work Identifies Attributes Associated with Results-Oriented
Management

Measuring performance allows organizations to track the progress they are
making toward their goals and gives managers critical information on which
to base decisions for improving their programs. Our prior work on
results-oriented management in the federal government indicates that
agencies that are successful in measuring performance and achieving
business transformation goals strive to establish goals and measures at
all levels of the agency that

o 
align goals and measures with the agency's strategic plan and
departmentwide goals;

o  assign accountability for achieving results;

o  demonstrate results;

o  provide a comprehensive view of agency performance;

o  link resource needs to performance;

o  discuss data quality, including reliability and data sources; and

o 
provide contextual information to help officials evaluate the significance
of underlying factors that may affect reported performance.

While these may not cover all attributes associated with results-oriented
management, we feel they are some of the most important for
accuratelyassessing the strengths and weaknesses of programs and making
improvements. Without these attributes, decision makers may not have
ascomplete information for measuring and managing an agency's performance
as needed. Each of these attributes is discussed in appendix IV.

Between fiscal year 1999 and 2003, DLA, DISA, and DODEA-the three agencies
we reviewed-used performance contracts to varying degrees to identify
performance issues, demonstrate progress made, and encourage discussions
to improve agency performance. The contracts were an important step toward
more results-oriented management practices for thedefense agencies.
However, the contracts could have been more effective had DOD guidance
required the defense agencies to include certainattributes associated with
results-oriented management. For example, the contracts addressed key
business lines, but did not consistently include information that could
have helped demonstrate results or provide somecontextual information that
could help officials assess progress and understand the significance of
underlying factors that may affect reported performance.

While Used to Varying Degrees, Performance Contracts That We Reviewed
Could Have Been More Effective Had They Included Certain Attributes
Associated with Results-Oriented Management

Defense Agencies Used Performance Contracts to Varying Degrees

Officials at the three defense agencies we reviewed used performance
contracts to varying degrees as part of their strategic planning systems
to demonstrate progress made and increase visibility over the quality,
quantity, cost, and timeliness of products and services for their key
business lines. In some instances, the contracts helped OSD and defense
agency officials to identify developing problems and assess the
effectiveness of corrective actions, such as justifying additional funding
to achieve performance goals and implementing or revising internal
processes to meet customer expectations. We did not verify the actual or
reported performance data included in the performance contracts or in
these examples. The agencies we reviewed provided the following examples
showing how they used performance contracts.

DLA officials told us that they used their performance contracts to focus
management attention and monitor efforts to improve the agency's response
time for providing logistics commodities, such as clothing and textile
items, medical and pharmaceutical supplies, and weapon system spare parts.
In its fiscal year 2001 performance contract, DLA established a
performance measure for its supply centers to process a greater percentage
of requisitions within shorter time frames. DLA reported that by
increasing emphasis on logistics response time goals to officials in its
Ohio-based Defense Supply Center and dedicating more trucks to its Defense
Distribution Center, the agency had processed 96 percent of requisitions
of commodities in 5.8 days during fiscal year 2001 rather than the
targeted 9.3 days.

A DISA official told us that the performance contracts helped DISA reduce
the cost of providing services in response to customer concerns. For
example, DISA developed cost-related performance measures for delivering
computing services, such as "Provide mainframe information processing
services while incurring no more than the unit costs listed." As a result
of increased management attention, DISA reported in its fiscal year 2002
annual performance report, which assessed progress against contract goals,
that it had reduced the cost per unit for providing OS/390 mainframe
processing services by 24 percent, from $38.26 in fiscal year 2001 to
$29.04 in fiscal year 2002. DISA officials told us that the high-level
attention made possible by the performance contracts helped ensure that
cost reduction goals were established and pursued.

DODEA officials told us that in order to meet OSD requirements, they
developed and submitted performance contracts that included measures taken
directly from their community strategic plan. However, they primarily used
their community strategic plan-containing strategic goals, objectives, and
performance measures-rather than the performance contract, which mirrored
the strategic plan, to manage and measure systemwide school performance.
We determined that DODEA's contract did not show the relationship between
resource needs and performance, but performance data helped bring
attention to initiatives and programs needing additional funding and
provided support for funding requests. For instance, DODEA officials noted
that a measure in DODEA's fiscal year 2003 performance contract influenced
DOD's decision to provide $114 million to hire 200 full-time teachers so
DODEA could achieve lower pupil to teacher ratios and, thereby, further
improve student performance.

Contracts Did Not Fully Incorporate Certain Key Attributes

Although useful in some respects, the performance contracts for the
agencies we reviewed would have been stronger tools if they had fully
incorporated certain key attributes that are associated with
results	oriented management and that contribute to effective government
performance. In previous reports, we identified seven such attributes.
However, the contracts we reviewed incorporated aspects of these
attributes to varying degrees. We found that the contracts did not
identify individuals accountable for achieving results or clearly
demonstrate relationships between resource needs and performance. We also
determined that the contracts, with the exception of DODEA, did not
provide officials with a comprehensive view of agency performance,
including factors that drive organizational performance, such as
financial, customer, and internal business processes. Furthermore, defense
agency contracts were inconsistent in demonstrating results by providing
trend data to show agency progress in meeting established goals over time,
and in discussing data quality issues and providing contextual information
to help officials understand the measures used. We based our analysis on
the most recent performance contracts developed by DLA, DISA, and DODEA
because DOD officials told us that these would reflect lessons learned
from prior iterations. Our analysis did not include other internal
management documents, such as strategic plans, that may have addressed
these attributes. We also did not evaluate the usefulness or
appropriateness of the measures themselves. Table 2 summarizes the extent
to which the performance contracts we reviewed included the key attributes
we have identified.

Table 2: Extent To Which DLA's, DISA's, and DODEA's Performance Contracts
Incorporated Certain Key Attributes Associated with Results-Oriented
Management

                                 DLA DODEADISA

Attribute

                            FY 2001 FY 2002 FY 2003

Aligns goals and measures with agency and departmentwide goals

 Each of the agency's strategic plan goals is supported by performance measures
                    Measures align with departmentwide goals

measures

            Identifies individuals accountable for achieving results

                              Demonstrates results

  Includes a combination of output- and outcome-oriented measures Measures are
  clearly defined Provides trend data to demonstrate progress against targeted
                                  performance

Provides comprehensive view of agency performance

Measures address factors that drive organizational performance (financial,
customer, and internal business processes and workforce learning and
growth)

            Measures address cost, quality, quantity, and timeliness

Measures address key business lines

                Incorporates customer and stakeholder priorities

Links resource needs to performance

Measures identify resources (e.g., human capital and information
technology) needed to achieve performance goals

                             Discusses data quality

Discusses data verification and validation procedures, including
reliability Identifies sources of data for each measure

Provides contextual information

Provides context or explanation for understanding measures and reported
performance

Legend: Fully incorporated; Partially incorporated; Not incorporated

Source: GAO analysis of DLA's, DISA's, and DODEA's performance contracts.

Notes: "Fully incorporated" indicates that all measures exhibited that
particular attribute. "Partially incorporated" indicates that some of the
measures, but not all, exhibited that attribute, whereas "not
incorporated" means that none of the measures did.

a Performance contract measures did not cascade to lower organizational
levels. However, agency officials told us that they shared strategic goals
and objectives from their strategic plans with all organizational levels
and made these available to customers and stakeholders.

Performance Contracts Did Not Always Align with Agency and Departmentwide
Goals

Contracts Did Not Assign Accountability for Achieving Results to
Individuals at Lower Organizational Levels

b According to a DLA official, performance appraisals for top agency
leadership reflected performance contract measures. However, DLA officials
were not able to provide a standardized performance appraisal template to
demonstrate this.

Additional information about the extent to which the performance contracts
for the agencies we reviewed included these attributes is discussed below.

DOD did not require performance contract measures to align with agencies'
strategic plans or departmentwide goals until fiscal year 2004. As a
result, some of the agencies' performance contract measures did align,
while other measures did not. Each agency's performance contract included
statements conceptually linking performance goals with strategic plans and
departmentwide goals. For example, DISA's fiscal year 2002 performance
contract stated that "The performance measures used in this contract
directly support the goals and objectives in the strategic plan," and
DISA's strategic plan directly related to DOD strategic goals and
objectives. We found that all of DODEA's performance measures aligned with
the agency's strategic plan goals and could be linked to departmentwide
goals. With the exception of the following example, all of DLA's fiscal
year 2001 performance contract measures aligned with the agency's
strategic plan goals, but not all aligned with departmentwide goals. DLA's
contract lacked performance measures that addressed its strategic plan
goal to ensure that the agency's workforce is able to deliver and sustain
world-class performance. Consequently, the performance contract did not
provide DLA and DOD officials with information to help assess recruiting
needs or workforce skills and training requirements that contribute to
meeting customer needs.

Furthermore, performance goals and measures used in the contracts we
reviewed did not cascade from the highest to the lowest levels of the
agencies to reinforce accountability throughout the organization.

While defense agency performance contracts identified individuals
accountable for achieving performance results at the most senior levels of
DOD and the defense agencies, they did not identify individuals
responsible for tracking performance or making improvements at lower
organizational levels.

Defense Agency Task Force officials were directed to review the defense
agencies' performance contracts and make recommendations to the Deputy
Secretary of Defense on whether the performance contracts satisfied DOD
guidance. The Defense Management Council was directed to

Contracts Did Not Consistently Provide Information to Help Assess Progress
and Demonstrate Results

recommend ways to reduce costs incurred by the defense agencies while
improving customer service. DOD officials told us that after the events of
September 11, 2001, OSD's attention on the performance contracts
temporarily diminished, less strict oversight occurred, and PA&E officials
did not hold agencies to meeting reporting requirements. However,
beginning in April 2002, the Deputy Secretary of Defense began requiring
the Under and Assistant Secretaries of Defense to provide quarterly
updates concerning the defense agencies' performance on key performance
contract measures and the corrective actions to be taken if performance
goals were not expected to be met. DOD officials were not able to provide
historical information concerning whether these reviews occurred as
frequently as required or whether corrective actions were taken.

Although top leaders had oversight of the performance contracts, none of
the performance contracts we reviewed identified individuals, such as
business managers, responsible for achieving results at lower
organizational levels as a way to reinforce accountability for achieving
results and maintain focus on organizational priorities. However, in a
related internal management report called the Chief Financial Executive
4th Quarter Fiscal Year 2002 Performance Contract Report-which DISA used
internally to report progress against performance contract goals- DISA
identified business managers responsible for monitoring each performance
measure.

Defense agency performance contracts did not consistently provide
information to help assess progress and demonstrate results. Although the
agencies' performance contracts included milestones or time frames for
completing specific initiatives, they differed in the degree to which
measures were clearly defined and outcome-oriented and the extent to which
they included trend data. For instance, we found that some measures in
DODEA's performance contract were not clearly defined. One example is:

"improve student achievement and education quality consistent with
Presidential and national educational goals. Per pupil costs will not
increase more than 7% over the previous year."

This measure is actually two measures. Furthermore, the measure does not
define the specific presidential and national education goals-such as

Contracts Did Not Provide a Comprehensive View of Organizational
Performance

the No Child Left Behind Act of 2001-against which the agency measured its
progress.6

In addition, DODEA and DISA did not consistently report trend data for
each measure in their performance contracts to show progress made over
time, and DLA limited trend data to the prior year. For example, in its
fiscal year 2003 performance contract results, DODEA reported that it had
met goals for teacher certification and school accreditation, but the
agency did not identify the baseline or percentage of teachers certified
or schools accredited.

Each agency's performance contract included measures that addressed key
business lines as well as cost, quality, quantity, and timeliness of
services, but, with the exception of DODEA, did not provide as
comprehensive a view of agency performance as possible because the
measures did not address all drivers of organizational performance,
including customer, financial, and internal business processes, and
workforce learning and growth. Doing so could have helped officials assess
how well the agencies were positioned to perform in the future.

We found that each agency's most recent performance contract included
measures assessing aspects of each key business line. For example, DLA's
fiscal year 2001 contract included 26 measures that supported its 5
business lines, including Defense Automated Printing Service, Defense
Distribution, Defense National Stockpile Center, Defense Reutilization and
Marketing Service, and Supply (Energy and Non-Energy). DISA's fiscal year
2002 contract included 29 measures that addressed 4 business lines,
including Computing Services, Enterprise Acquisition Services, Joint
Warfighting and DOD-wide Enterprise Capabilities, and Telecommunication
Services. DODEA's fiscal year 2003 performance contract included 15
measures that addressed its 2 business lines- domestic and overseas
education.

While measures addressed the agencies' key business lines as well as cost,
quality, and timeliness of services provided, they did not all address all
drivers of organizational performance. For example, the contracts did not
always reflect the different concerns of customers and stakeholders.
According to defense agency officials, PA&E officials prescribed certain

6 The No Child Left Behind Act of 2001, among other things, focused
attention on closing the achievement gaps among various groups of
students.

Linkage between Resource Needs and Performance Outcomes Was Not Clearly
Established

Quality of Data Was Not Discussed

measures, although agency officials believed these were not value added
and did not align with agency strategic goals. For instance, PA&E
officials required DODEA officials to identify Management Headquarters
Costs- and specifically the "cost per school"-in DODEA's performance
contract. Agency officials told us that this measure had limited value
because it provided aggregated data rather than specific information for
individual school's costs. Furthermore, according to a former senior
official in the Office of the Secretary of Defense for Acquisition,
Technology, and Logistics, the performance contracts did not reflect the
needs of the chiefs of staff of the services and the Secretary of Defense,
who were key customers and stakeholders. This official believed that the
measures were too technical and "in the weeds" to help inform strategic
management decisions and were more appropriate for lower-level managers to
use when monitoring program performance.

None of the contracts we reviewed identified the resources needed to meet
performance goals or explained how additional resources could have
contributed to achieving expected levels of performance. Such information
could have helped officials to better understand how trade-offs among
different investments-such as human capital, information technology, and
workforce training-affected performance outcomes and drove costs. For
instance, DLA established a fiscal year 2001 goal of improving inventory
accuracy by 4.2 percent over the previous year's performance, but it did
not provide additional data in its performance contract that would have
helped officials to determine the resources needed to achieve this goal,
if any. DOD officials stated that they began to establish links between
resource needs and performance outcomes by requiring the defense agencies
to submit their performance contracts at the same time as their Program
Objectives Memorandum.7

None of the performance contracts we reviewed discussed the quality of
performance data, including reliability, data sources, or data
verification and validation procedures. Consequently, when reviewing the
performance contracts, top officials in DLA, DISA, and DODEA may not have
had information needed to determine whether reported data and performance
results were accurate, reliable, or current enough to use in decision
making and that corrective actions would produce the desired

7 Program Objectives Memorandum is the final product of the programming
process within DOD. The Program Objectives Memorandum displays the
resource allocation decisions of the DOD components in response to, and in
accordance with, Defense Planning Guidance.

Contextual Information Could Have Better Informed Decision Making

outcomes. For example, an October 2000 internal audit initiated by the
DISA Inspector General concluded that data reported in DISA's

1st Quarter Fiscal Year 2000 Chief Financial Executive Report contained
errors or bias and lacked sufficient supporting documentation.8 The
Inspector General reported that these conditions occurred because DISA had
not established a systematic approach for collecting and reporting
performance data. The Inspector General concluded that, as a result,
reported data provided limited assurance that DISA had valid and reliable
information to manage agencywide performance. The Inspector General
concluded that DISA should establish a data collection and reporting
process and definitive procedures to ensure performance information was
sufficiently complete, accurate, and consistent to document actual
performance and support decision making. Senior DISA officials told us
that the agency did not initially implement related recommendations
because DISA discontinued use of performance contracts. However, according
to these officials, DISA is currently taking steps to address these
recommendations by setting the expectation that senior officials will be
personally accountable for the quality of data included in DISA's balanced
scorecard and establishing internal focal points for data collection and
reporting. In addition, DISA officials are reviewing commercial products
to track data electronically.

The performance contracts provided some contextual information about some
external factors outside of the agencies' control that may have impacted
the achievement of performance goals and provided general information
about the agencies' mission, business lines, and customers. However, the
performance contracts did not always provide some contextual information
to help officials understand each agency's accomplishments for that fiscal
year; determine how measures related to agency and departmentwide goals;
assess agency progress; understand the impact that external factors, such
as contingency operations, could have on the agency's ability to meet its
performance targets; and determine how resource needs should be
prioritized. For example, DISA reported in its fiscal year 2002
performance contract under the Joint Warfighting and DOD-wide Enterprise
Capabilities business line that it planned to, "Successfully complete
fielding of a DOD standard spectrum management information system
(Spectrum XXI) software version 3.0 by December 2001." However, it is
unclear how this linked specifically to DISA's

8Defense Information Systems Agency Inspector General. Audit of DISA's
Performance Contract. Final Report. Project No. 2000-H-601. (Washington,
D.C.: October 24, 2000).

strategic plan or DOD's goals or what external factors could have impacted
DISA's intended performance.

The Plan and Scorecard We Reviewed Show Improvement, but Could Be
Strengthened by Including Certain Key Attributes

In fiscal year 2003, DOD renamed the performance contracts as performance
plans and, based on the agencies' experiences using performance contracts,
revised the oversight and review process and guidance. DOD required
performance measures, beginning in fiscal year 2004, to align with agency
and departmentwide goals and provide a more comprehensive view of agency
performance. However, the plan and scorecard we reviewed could be stronger
tools for informing officials about agency performance if they
incorporated some additional key attributes associated with
results-oriented management such as linking resource needs to performance,
discussing data quality, and providing contextual information for
understanding the measures and reported performance.

DOD Strengthens Oversight Process and Requires Inclusion of Additional Key
Attributes, but Some Are Still Lacking

Six years after implementation of the performance contracts, the original
intent-to assist senior leaders at the defense agencies and within OSD to
oversee agency performance and to encourage the agencies to become more
responsive to customer needs-remains unchanged. However, beginning in
fiscal year 2004, DOD renamed the performance contracts as performance
plans and gave the defense agencies the option to use either performance
plans or balanced scorecards. These tools were to build on the performance
contracts, and to include additional attributes associated with
results-oriented management that were previously lacking under the
contracts. These changes centered on (1) revising the oversight and review
process to strengthen accountability for achieving results, (2) requiring
performance measures to align with agency strategic goals and DOD's risk
management framework, and (3) requiring the measures to be more
comprehensive in scope to better demonstrate results.

The defense agencies we reviewed are at different stages of developing and
implementing performance plans and balanced scorecards, and DOD has
required these tools to be in place for less than a year. Consequently, it
is too soon to determine how useful they will be for managing and
measuring agency performance. However, DLA's balanced scorecard and
DODEA's performance plan showed progress in incorporating many of the
attributes associated with results-oriented management that we have
identified, but because DOD did not require the defense agencies to
include all of the attributes discussed in this report, the plan and
scorecard we reviewed continue to lack certain information-such as the

relationship between resource needs and performance, discussions of data
quality, and some contextual information-that could strengthen these tools
for managing defense agency performance. Although we obtained DISA's draft
balanced scorecard dated September 2003, we did not evaluate it against
the key attributes discussed in this report because it had not been
finalized by the time we completed our review. However, we do discuss
DISA's plans for addressing several of the key attributes in the balanced
scorecard that the agency is developing.

Table 3 shows the extent to which DLA's balanced scorecard and DODEA's
performance plan incorporated certain key attributes associated with
results-oriented management. We based our analysis specifically on the
performance plan and balanced scorecard, although other agency internal
management documents may have addressed these attributes. We also did not
evaluate the usefulness or appropriateness of the measures themselves.

Table 3: Extent To Which DLA's Scorecard and DODEA's Plan Incorporated
Certain Key Attributes Associated with Results-Oriented Management

                      DLA DODEA Attribute FY 2004 FY 2004

Aligns goals and measures with agency and departmentwide goals

 Each of the agency's strategic plan goals is supported by performance measures
                    Measures align with departmentwide goals

a

Goals and measures cascade from the corporate level to the lowest level of
the agency

                  Assigns accountability for achieving results

Establishes a foundation to hold top leadership accountable for achieving
                                    results

Performance appraisals for top agency leadership reflect performance goals
and measures

Identifies individuals accountable for achieving results

                              Demonstrates results

  Includes a combination of output- and outcome-oriented measures Measures are
  clearly defined Provides trend data to demonstrate progress against targeted
                                  performance

Provides comprehensive view of agency performance

Measures address factors that drive organizational performance (financial,
customer, and internal business processes and workforce learning and
growth)

            Measures address cost, quality, quantity, and timeliness

Measures address key business lines

Incorporates customer and stakeholder priorities

Links resources needs to performance

Measures identify resources (e.g., human capital and information
technology) needed to achieve performance goals

Discusses data quality

Discusses data verification and validation procedures, including
reliability

Legend: Fully incorporated; Partially incorporated; Not incorporated

Source: GAO analysis of DLA's balanced scorecard and DODEA's performance
plan.

Notes: "Fully incorporated" indicates that all measures exhibited that
particular attribute. "Partially incorporated" indicates that some of the
measures, but not all, exhibited that attribute, whereas "not
incorporated" means that none of the measures did.

a Performance measures did not cascade to lower organizational levels.
However, DODEA officials told us that they shared strategic goals and
objectives from their strategic plans with all organizational levels and
made these available to customers and stakeholders.

Varying Progress Made in Aligning Performance Goals and Measures with
Agency and Departmentwide Strategic Goals and Cascading These to Lower
Organizational Levels

Additional information about the extent to which DLA's balanced scorecard
and DODEA's performance plan include these attributes is discussed on the
following pages.

The balanced scorecard and performance plan we reviewed showed varying
progress in meeting DOD's requirement to align performance goals and
measures with agency strategic plan goals and DOD's risk management
framework beginning in fiscal year 2004. However, DLA and DODEA fulfilled
DOD's requirement that every organization's performance goals and measures
be consistent with those of its next higher and lower organizations. DLA
and DODEA have adopted different strategies for cascading their scorecard
and plan, respectively. DLA's corporate-level balanced scorecard aligns
with its strategic plan goals to transform DLA's operations over the next
2 to 5 years to better support its customers and departmentwide
performance measures outlined in DOD's risk management framework. DLA has
also cascaded its corporate scorecard to each of its Defense Supply
Centers located in Richmond, Virginia, Columbus, Ohio, and Philadelphia,
Pennsylvania, as well as other field activities including the Defense
Energy Support Center, Defense Distribution Center, and Defense
Reutilization and Marketing Service. Each center's or field activity's
scorecard identifies "how and to what extent" their strategies support
DLA's corporate-level balanced scorecard and is expected to reflect DLA's
overall vision of "Right Item, Right Time, Right Place, Right Price, Every
Time...Best Value Solutions for America's Warfighters." According to DLA
officials, in this way, DLA provides a consistent vision to its customers,
stakeholders, and employees. They told us that this process also allows
customers, stakeholders, and employees to understand how corporate-level
strategies are supported by functional, day-to-day measures.

Figure 1 depicts how DLA's corporate-level scorecard aligns with
departmentwide goals and measures and how the Defense Supply Center
Richmond's internal scorecard supports these goals. DLA officials told us
that the goals within the institutional risk area of DOD's risk management
framework cascade down to the Under Secretary of Defense for Acquisition,
Technology, and Logistics' balanced scorecard. This scorecard has four
strategies that cascade down and inform DLA's scorecard. In turn, the four
strategies in DLA's corporate-level scorecard cascade down to its defense
supply centers. Figure 1 depicts specifically how the three strategies
within the customer portion of DLA's scorecard cascade down to the Defense
Supply Center Richmond. The supply center has developed four measures that
it uses to assess its performance in terms of customer service on a
monthly basis.

Figure 1: Depiction of How DLA's May 2004 Scorecard Supports
Departmentwide Goals and Cascades to Lower Organizational Levels

              Source: GAO presentation of DOD and DLA information.

DODEA's fiscal year 2004 performance plan does not include goals and
measures that support all of its community strategic plan goals and
objectives such as (1) developing, promoting, and maintaining a network

Oversight and Review Process Strengthened to Reinforce Accountability

of meaningful partnerships and alliances to enhance social, emotional, and
academic growth and (2) maximizing resources or developing safe,
well-managed, and disciplined environments conducive to learning. However,
DODEA has linked each performance measure in its performance plan with the
risk management framework area that it supports. For example, DODEA linked
the measure "In fiscal years 20042009, no less than 95 percent of the
Department of Defense Domestic Dependent Elementary and Secondary Schools
teachers will be professionally certified in the subject area and grade
level to which they are assigned within three years of employment" with
force management risk, and specifically the departmentwide goal to
maintain a quality workforce. DODEA officials told us that they continue
to communicate these goals and performance measures to employees and
customers systemwide through their community strategic plan, and school
improvement plans reflect these goals, as relevant.

DOD has strengthened the accountability of OSD and top agency leadership
for achieving performance results through the development of OSD-level
scorecards and semiannual reviews of defense agency performance, and the
agencies we reviewed are taking steps to strengthen accountability at
lower organizational levels. However, DODEA's performance plan does not
identify specific individuals who are responsible for making improvements.
The Under and Assistant Secretaries of Defense continue to have
responsibility for reviewing the defense agencies' performance plans or
scorecards and providing direct feedback. However, some have developed
balanced scorecards-such as the Under Secretary of Defense for Personnel
and Readiness (USD/P&R)-and have taken steps to incorporate data from
quarterly reviews of agency performance measures in their balanced
scorecards while continuing to provide feedback to agency officials on a
regular basis. For example, prior to approving DODEA's fiscal year 2004
performance plan, the Deputy Under Secretary of Defense for Military
Community and Family Policy within USD/P&R asked the agency to incorporate
updated information on its customer satisfaction survey and to explain
annual variations in the cost-per-pupil index. OSD and agency officials
also review performance data during periodic internal meetings. For
example, figure 2 shows the high-level DODEA performance measures that
officials within USD/P&R monitor on a quarterly basis in their balanced
scorecard called Monitoring the Status of the Force. Part A: Balanced
Scorecard. USD/P&R officials are responsible for monitoring the force
management risk area, and within this area, performance strategies related
to "maintain workforce satisfaction." Specifically, these officials
monitor DODEA's progress in meeting three performance contract measures.

Figure 2: Example of the Type of Information in DODEA's Performance Plan
That OSD Officials Monitor, as of May 2004

Source: GAO presentation of DOD data.

Additional details about each performance measure are also available to
USD/P&R officials, as depicted in figure 3.

    Figure 3: Detailed Data on DODEA's Pupil to Teacher Ratio as of May 2004

Number of pupils 18

17 FY 2000 FY 2001 FY 2002 FY 2003

Actual performance

Department of Defense Elementary and Secondary Schools

Department of Defense Dependent Schools

Source: DOD.

DOD officials told us that the Office of the Under Secretary of Defense
for Acquisition, Technology, and Logistics that oversees DLA and the
Assistant Secretary of Defense for Network Information and Integration
that oversees DISA are developing balanced scorecards that are expected to
incorporate certain performance measures from DLA's and DISA's balanced
scorecards, respectively, in order to help maintain oversight and monitor
the agencies' progress toward achieving specific performance goals.

DOD has also implemented semiannual performance reviews at which time the
defense agencies brief PA&E officials, respective Under and Assistant
Secretaries of Defense, and agency officials on progress to date against
performance goals. DOD began holding midyear reviews during May 2004, and
officials expect to complete these reviews by August. A DOD official told
us that in the future, the first cycle reviewing the prior year's
performance will occur between September and December and the second
cycle, the midyear review, will occur between May and August. DOD
officials expect results from these meetings to be reflected in the
agencies' fiscal year 2005 performance plans and scorecards. During a

Agencies' Progress in Demonstrating Results Varies

May 2004 midyear review of DLA's balanced scorecard, the Director, PA&E,
recommended that the agency include more specific information on its
customer surveys, including the population surveyed and the way customer
feedback is obtained.

By assigning responsibility for each performance measure to specific
individuals, some defense agencies are reinforcing accountability for
achieving performance goals at all levels of the organization. For
example, in addition to implementing performance appraisals for Senior
Executive Service staff that are linked to strategic goals and objectives
against which ratings are assessed, DLA identified individuals in its
balanced scorecard and charged them with monitoring, reporting on, and
implementing corrective actions for each scorecard measure. A senior DISA
official told us that the agency recently approved linking balanced
scorecard outcomes with Senior Executive Service members' and supervisory
employees' performance appraisals beginning in fiscal year 2005 in order
to institutionalize its balanced scorecard and improve accountability for
results. DISA also is reinforcing accountability by requiring its senior
leaders (scorecard owners), who are accountable for achieving results, to
provide status updates during corporate-board meetings of senior agency
officials. In contrast, DODEA's performance plan does not identify
specific individuals accountable for monitoring, reporting on, and
implementing corrective actions for each performance measure. However,
DODEA officials told us that the agency includes performance plan
objectives in each manager's performance appraisal, and both monetary and
nonmonetary awards are used as incentives to meet or exceed goals and
measures.

DLA's balanced scorecard and DODEA's performance plan showed varying
progress in the extent to which they included outcome-oriented measures
and trend data to help demonstrate results, and the clarity of performance
measures continues to be an issue. We determined that DLA's 2004 scorecard
includes outcome-oriented measures and trend data that help to demonstrate
progress made. For example, one of DLA's outcome-oriented financial
measures-the achievement of accuracy in non-energy sales
forecasting-includes baseline and trend data for the 3 preceding fiscal
years. However, DLA's 2004 scorecard measures include terms that are
generally not clearly defined, as demonstrated by the measure "Percentage
of reengineered functionality implemented in BSM." In contrast, DODEA's
performance plan includes almost all of the same measures as did its
fiscal year 2003 performance contract, and these measures are not always
clearly defined or accompanied by trend data. As we previously mentioned,
DODEA does not clearly define the specific

Agencies' Scorecard and Plan Show Progress in Providing a Comprehensive
View of Agency Performance

presidential and national education goals against which it measures
progress. Furthermore, DODEA's performance plan does not consistently
include baseline information or trend data to help officials gauge
progress. DODEA did not indicate the percentage of teachers who are
certified in the subject area and grade level to which they are assigned
within 3 years of employment, although the agency set a goal of no less
than 95 percent. The plan also does not indicate the percentage of schools
that have full high-speed connectivity to the Internet, although it
establishes a goal of 100 percent.

Our analysis showed that DLA's scorecard and DODEA's performance plan
included measures for key business lines that provide officials with
information about agency performance and, to varying degrees, address the
four drivers of organizational performance as well as cost, quality, and
timeliness of services provided, and that all of the agencies we reviewed
are taking steps to better reflect customers' and stakeholders'
priorities.

DLA organizes its corporate-level scorecard by drivers of performance-
including learning and growth and customer, financial, and internal
business processes-and includes strategies and supporting measures that
are cascaded to its defense supply centers and field activities. As would
be expected, lower organizational level scorecards used by the defense
supply centers include performance measures that address the key business
lines for which they are responsible and provide more specific assessments
of cost, quality, timeliness, and customer satisfaction. For example, the
Defense Supply Center Richmond's scorecard includes performance measures
focusing on aviation weapons systems and environmental logistics support,
whereas the Defense Supply Center Philadelphia's scorecard focuses on the
provision of clothing, textiles, and medical equipment, among other items.
The learning and growth quadrant of DLA's corporate scorecard contains a
measure to identify gaps between workforce competencies and skills
required to meet mission requirements for 100 percent of DLA positions by
the end of fiscal year 2007. The Defense Supply Center Richmond's
scorecard includes several measures that address employee certification,
including acquisition training certification to ensure that the supply
center's workforce is fully proficient in the acquisition process. This
measure, in turn, aligns with the departmentwide policy of establishing
education and training for each defense acquisition position. Similarly,
DODEA's plan includes measures that address its two key business
lines-domestic and overseas education-and are cascaded to schools
worldwide through its community strategic plan.

Unlike DLA's balanced scorecard, DODEA's performance plan does not
specifically categorize performance measures by the factors that drive
organizational performance, but we identified at least one measure or
initiative that can be linked with each organizational driver. For
example, DODEA's performance plan includes a measure to limit increases in
per pupil cost to no more than 7 percent over the prior fiscal year. This
measure addresses both financial and internal business processes that
affect organizational performance. DODEA addresses workforce learning and
growth by establishing a goal to certify no less than 95 percent of
teachers in the subject and grade levels to which they are assigned within
3 years of employment. Finally, DODEA includes measures that address
customer satisfaction such as students' performance and parents'
satisfaction with local schools.

Our analysis showed that the defense agencies we reviewed are also taking
steps to better address specific customer and stakeholder concerns about
performance. Both DODEA and DLA conduct periodic reviews of agency
performance and make adjustments to meet customers' and stakeholders'
priorities. For instance, DLA's balanced scorecard Executive Board,
comprised of DLA's Vice Director, designated leaders for each area of
organizational performance (quadrant leaders), the Chief Information
Officer, and a Lead Center representative, conducts monthly reviews of
enterprisewide operational measures. Each quarter, DLA's Corporate Board,
comprised of DLA's Director, Vice Director, and the heads of DLA's major
organizational components, meets with commanders from the field to review
the agency's performance from the corporate level down to the supply
centers. The defense supply centers conduct weekly reviews to monitor
progress, identify developing problems, and take corrective actions.
According to DLA officials, to make information more timely, accessible,
and relevant for measuring and managing the agency, DLA posts the agency's
balanced scorecard on its internal Web site to allow all employees,
including agency managers and leaders, to share performance data. This Web
site is also available to all of DLA's military customers and
stakeholders. DLA also continues to obtain customer and stakeholder
feedback through a variety of mechanisms, such as customer surveys,
reports from customer service liaisons embedded with deployed military
service units, and meetings between the agency's Director and leaders of
the military services. Moreover, DLA is taking steps to address the unique
concerns of each military service through the development of
performance-based agreements that include specific performance measures
that are based on collaborative, direct negotiations between DLA and each
military service. These agreements are intended to create and sustain
end-to-end customer support by aligning DLA's resources with

Funding Not Yet Linked with Performance

the logistics solutions that will meet the services' needs and maximize
their effectiveness.

In addition, we determined that the defense combat support agencies- such
as DLA and DISA-are taking steps to incorporate additional strategies and
measures in their balanced scorecards that respond to customers' changing
needs during contingencies and war. For example, DLA's scorecard assesses
readiness by measuring the percentage of items that the military services
identify as requiring surge and sustainment support (e.g., hardware items,
clothing and textiles, and medical supplies) to ensure appropriate
inventory levels to accommodate short production lead times. DLA officials
told us that to reduce procurement lead time and decrease delivery delays
to the agency's primary customer, the warfighter, the agency has developed
22 long-term contracts, called strategic supplier alliances, with its
major contractors. According to these officials, having these agreements
in place enhances customer support through the inclusion of better
performance metrics, such as reduced administrative lead times and
increased stock availability. DISA officials told us that as a combat
support agency, DISA's mission is to support the warfighter and that the
strategies and performance measures DISA is developing focus on wartime
requirements but also are applicable to peacetime environments. For
example, DISA is developing measures that will support its customer
strategy to eliminate bandwidth constraints in response to customer
complaints that limited bandwidth is a problem, especially during
contingency operations. According to DISA officials, removing current
bandwidth limitations can provide DISA's customers with better and
timelier information during combat and peacetime.

A major goal of DOD is to strategically link major decisions to program
and budget development, and the defense agencies we reviewed are in the
initial stages of attempting to do this. DOD has directed that the Under
Secretary of Defense, Comptroller, and the Director, PA&E, review program
performance and assess the degree to which budget estimates sustain and
improve programmatic results and, as appropriate, recommend alternate
solutions or funding adjustments to correct resource needs if performance
goals are not met. Senior DOD officials told us that this continues to be
a challenging undertaking. The defense agencies we reviewed are using
their performance plans and scorecards to bring attention to priority
management projects. Although these tools still do not discuss resource
implications, DODEA included goals and measures from its fiscal year 2004
performance plan in its fiscal year 2004-2005 operation and maintenance,
defensewide, biennial budget estimates. We believe that

Quality of Data Is Not Discussed

Some Contextual Information Is Lacking

inclusion of the agency's priorities can help officials better consider
appropriate funding levels and trade-offs.

Although DISA is still developing its balanced scorecard, senior agency
officials told us that the agency plans to use the scorecard to bring
visibility and management attention to priority projects in order to fund
them appropriately and to improve collaboration internally. These
officials told us that DISA is in the process of aligning its investment
and scorecard strategies and expects initiatives in its scorecard to be
assigned funding priority. To this end, DISA is developing a Program
Objectives Memorandum that aligns investments with its scorecard. During
corporate-board meetings when senior agency officials review the agency's
performance, DISA requires each measure owner to discuss the context and
status of initiatives, including the need for additional resources.
According to DISA officials, the agency has already experienced some
benefits when testing this approach. DISA officials identified a
disconnect between customer expectations, performance levels, and funding
for an initiative to develop a Web-based portal to share information about
DISA's business activities internally. The goal of the portal is to help
employees perform critical operations faster and more effectively.
Officials told us that although internal customers identified this
initiative as a priority, the portal did not receive dedicated funding
and, as a result, was neither as robust as planned nor on target to meet
development time frames. DISA officials told us that inclusion of the
project in its balanced scorecard is expected to raise these challenges to
the attention of senior leaders to encourage corrective action and
appropriate funding levels.

The balanced scorecard and performance plan we reviewed do not provide
decision makers with confidence that performance data are credible because
they do not include specific information on the procedures to verify and
validate performance information. In addition, no information on the
sources of data is included. Thus, decision makers may not have complete
information needed to determine whether reported data and performance
results are accurate.

The scorecard and plan we reviewed did not include some contextual
information that could have helped to inform decision makers because DOD
guidance did not require the defense agencies to do so. However, a sample
measure from DISA's draft September 2003 balanced scorecard depicts the
type and amount of contextual information that agency officials have
considered incorporating.

DLA's balanced scorecard lacks information about the agency's mission and
customers that had been included in its performance contract. In contrast,
DODEA's performance plan provides general information about the agency's
mission, business lines, and customers. However, neither agency clearly
articulates how its performance strategies and measures help achieve
mission or departmentwide goals or improve customer responsiveness. Also,
information about external factors, such as recent contingency operations,
that can affect achievement of performance goals is limited or absent. For
example, DODEA's performance plan includes a measure to maintain pupil to
teacher ratios for kindergarten through grade 12 between 18:1 and 24:1
during fiscal years 2004-2009. However, DODEA's performance plan does not
explain that studies have demonstrated that lower pupil to teacher ratios
contribute to improved student performance or how this measure supports
presidential and national goals against which the agency evaluates itself.
Providing this contextual information could assist officials with
realizing how human capital-related strategies, such as staffing, affect
the agency's goals to continue to improve student performance in reading,
math, and languages, and help establish links between resource needs and
performance outcomes.

In September 2003, DISA developed a draft balanced scorecard with a sample
measure that depicts the type and amount of contextual information that
agency officials have considered incorporating. The sample measure, as
currently structured, incorporates most of the attributes associated with
results-oriented management that we have discussed in this report,
including contextual information. However, the sample measure does not
include some aspects of data quality such as reliability, validity, and
verification, and it is unclear whether DISA plans to discuss external
factors that could affect the achievement of performance goals or the link
between resource needs and performance beyond funding problems. For
example, as currently structured, the draft scorecard would

(1) align measures with agency and departmentwide goals. The scorecard is
expected to identify the quadrant of DISA's scorecard that the measure
supports (e.g., customer) and the Assistant Secretary of Defense's
balanced scorecard goal and risk management framework area with which it
aligns.

(2) assign accountability for achieving results. The scorecard is
anticipated to identify individuals responsible for monitoring and
reporting on performance and making corrections, as needed for

each measure.

(3) demonstrate results. The scorecard is expected to define the measure
and present trend data showing performance improvements over 4 fiscal
years.

(4) discuss data quality. The scorecard is expected to identify data
sources but does not include information about the way data are collected
or the verification and validation procedures.

(5) provide contextual information. The scorecard is expected to include
historical data, and an assessment of the measure's status compared to
prior years, and identify any challenges that need attention (e.g., hot
issues) such as funding problems.

Figure 4 shows the content that DISA officials told us that they expect to
include and the way this information addresses the attributes we have
identified in this report.

 Figure 4: Format of a Sample Measure from DISA's Draft September 2003 Balanced
                                   Scorecard

Source: GAO analysis of DISA data.

Notes: PSA=Principal Staff Assistant. TBD=To Be Determined.

DOD Captures and Shares Lessons Learned

OSD/PA&E officials have developed some mechanisms to help the defense
agencies capture and share experiences and lessons learned from developing
performance plans and balanced scorecards. Some defense agencies, such as
DLA and DISA, are sharing their experiences with each other to leverage
lessons learned.

OSD/PA&E has taken some steps to help the defense agencies share their
experiences and lessons learned with developing performance plans and
balanced scorecards, and DOD plans additional initiatives. In response to
our suggestions during this review, an OSD/PA&E official decided to
continue to hold semiannual roundtables during which defense agency
officials could voluntarily share their experiences developing and using
performance plans or balanced scorecards. OSD/PA&E officials also told us
that they plan to revise and update a departmentwide Web site originally
used to disseminate agencies' performance contracts and related guidance
and to incorporate this information, as well as defense agencies'
performance plans and scorecards, into DOD's performance management Web
site. The original site was not updated as frequently as it could have
been, and firewalls used to protect information from unauthorized access
precluded defense agencies with domain names that did not end in ".mil"
from gaining access. For example, DODEA officials were not able to access
the original Web site because the agency's World Wide Web domain address
ends in ".edu." Consequently, DODEA could not leverage other agencies'
contracts and performance measures when developing its own. DODEA
officials told us that they were not even aware that the Web site existed,
but that it would have been helpful for identifying some performance
measures.

In addition, some defense agencies that we reviewed are proactively
sharing lessons learned from developing and implementing balanced
scorecards. For example, DISA officials told us that they contacted DLA to
learn how it had developed and implemented its balanced scorecard. In
addition, officials at both agencies told us that they have or are
developing internal Web sites to facilitate performance measurement.
DISA's internal Web site serves as a collaborative forum and repository
for balanced scorecard material, and provides related guidance on
developing balanced scorecards.

Conclusions 	DOD has improved its process for measuring and managing
defense agency performance by taking steps to include certain attributes
associated with results-oriented management in defense agencies'
performance plans and scorecards that were not included in the

performance contracts. Specifically, changes made to DOD's process for
overseeing and reviewing defense agencies' balanced scorecards and
performance plans have the potential to further strengthen the
accountability of top leadership for overseeing the defense agencies and
achieving business transformation goals. While it is too early to
determine the extent to which these plans and scorecards will be fully
useful, these tools still do not incorporate some specific attributes that
would strengthen them. Until all performance plans and scorecards identify
individuals responsible for monitoring progress and taking corrective
actions, DOD's ability to ensure accountability and continuous assessment
at lower organizational levels is hindered. Furthermore, without
performance measures that are clearly defined and accompanied by trend
data and absent a discussion of data quality and contextual information-
such as information about the agency's missions, business lines, and
customers as well as external limitations that could affect achievement of
performance goals-senior OSD and agency leaders will not have the
necessary information for fully assessing the extent to which progress has
been made against existing performance targets and determining realistic
performance goals for future years. Finally, without demonstrating the
relationship between resource needs and performance outcomes, DOD
officials do not have all of the information needed to make improvements
and consider trade-offs among competing priorities.

To make performance plans and scorecards more informative and useful for
decision making, and further strengthen these tools' potential for
measuring and managing defense agency performance, we recommend that the
Secretary of Defense direct the Director, Program Analysis and Evaluation,
to revise guidance to require all defense agencies to address the
following five recommendations in their plans and scorecards:

o  	identify individuals accountable for achieving results at lower
organizational levels;

o  	include measures that are clearly defined and include trend data for
at least the past fiscal year's performance to help assess progress;

o  	identify resources needed to achieve performance goals and inform
budget decisions;

o  	discuss data quality, including the reliability, validity, and
limitations of performance measures as well as data sources; and

o  	provide contextual information to better understand how performance
measures support the agency's mission.

Recommendations for Executive Action

Agency Comments
and Our Evaluation

In written comments on a draft of this report, DOD concurred or partially
concurred with all of our recommendations. DOD's comments are reprinted in
their entirety in appendix V. DOD also provided technical comments that we
incorporated as appropriate.

Specifically, DOD concurred with our recommendations to revise guidance to
require defense agencies to include information on individuals accountable
for achieving results at lower organizational levels; include measures
that are clearly defined and include trend data for at least the past
fiscal year's performance; discuss data quality, including the
reliability, validity, and limitations of performance measures as well as
data sources; and provide contextual information to better understand how
performance measures support the agency's mission. DOD stated that it
would revise guidance to either implement or clarify these
recommendations.

DOD partially concurred with our recommendation to revise guidance to
require all defense agencies to identify resources needed to achieve
performance goals and inform budget decisions in their performance plans
or balanced scorecards. DOD stated that it may not be possible to fully
cost performance goals given the complex nature of defense operations and
that the appropriate process for proposing resource alternatives is the
Planning, Programming, Budgeting, and Execution System. However, DOD did
acknowledge that the defense agencies' performance plans should be
synchronized with this process by demonstrating how Planning, Programming,
Budgeting, and Execution System decisions are being implemented and by
providing performance information and metrics that can be used to inform
the department's annual budget and execution review and the allocation of
resources. We believe DOD is moving in the right direction by trying to
synchronize defense agency performance management with the budget process.
However, we continue to believe that the defense agencies' performance
plans and balanced scorecards would be an appropriate vehicle to discuss
resource implications. Full cost information may not be available, but the
performance plans and balanced scorecards should discuss resource
implications and include available cost information. Inclusion of this
information would help decision makers understand the connection between
performance goals and funding levels and evaluate competing needs at all
phases of the planning and budgeting process.

We are sending copies of this report to other interested congressional
committees; the Secretary of Defense; the Secretaries of the Army, the
Navy, and the Air Force; the Commandant of the Marine Corps; and the
Director, Office of Management and Budget. We will make copies available
to others upon request. In addition, the report will be available at no
charge on the GAO Web site at http://www.gao.gov.

If you or your staff have any questions regarding this report, please call
Sharon Pickup at (202) 512-4300 or Patricia Dalton at (202) 512-6806.
Additional contacts and staff acknowledgments are provided in appendix

VI.

Sharon L. Pickup
Director, Defense Capabilities and Management

Patricia A. Dalton
Director, Strategic Issues

                       Appendix I: Scope and Methodology

To meet our first two reporting objectives, we interviewed knowledgeable
Department of Defense (DOD), defense agency, and service officials.
Specifically, we interviewed officials in the offices of the Under and
Assistant Secretaries of Defense who act as principal staff assistants for
the three defense agencies reviewed. We also collected data from and
interviewed officials in the Office of the Secretary of Defense, Program
Analysis and Evaluation and the Office of the Comptroller. We also
reviewed our extensive prior work on results-oriented management and
identified seven attributes that contribute to effective government
performance. While these do not cover all attributes associated with
results-oriented management, we believe they are the most important for
accurately assessing the strengths and weaknesses of programs and making
improvements. These include (1) aligning goals and measures with agency
and departmentwide goals; (2) assigning accountability for achieving
results; (3) demonstrating results; (4) providing a comprehensive view of
agency performance by addressing factors that drive organizational
performance; (5) linking resource needs to performance; (6) discussing
data quality; and (7) providing contextual information. We systematically
analyzed the measures contained in selected performance contracts,
performance plans, and balanced scorecards against these attributes. We
focused our analysis on the Defense Logistics Agency (DLA), the Defense
Information Systems Agency (DISA), and the Department of Defense Education
Activity (DODEA). We judgmentally selected these three defense agencies to
review based on (1) their experience developing and implementing
performance contracts, plans, and balanced scorecards; (2) different types
of missions; and (3) types of services and types of customers. We analyzed
the most recent performance contracts used by DLA, DISA, and DODEA, and we
limited our analysis of balanced scorecards and performance plans to DLA's
fiscal year 2004 balanced scorecard and DODEA's fiscal year 2004
performance plan. Although we obtained DISA's draft balanced scorecard
dated September 2003, we did not evaluate it against the attributes we
discuss in this report because DISA had not yet finalized its scorecard by
the time we completed our review. However, we did discuss DISA officials'
plans for developing a balanced scorecard. We did not evaluate the
usefulness or appropriateness of the measures contained in the contracts,
plans, and balanced scorecards or verify actual or reported performance.
In addition to analyzing the performance contracts, performance plans, and
balanced scorecards, we also reviewed the defense agencies' strategic
plans and annual performance contract reports. Further, we reviewed the
cascaded-balanced scorecards that are used at lower organizational levels,
specifically, the fiscal year 2004 scorecards for DLA's supply centers
(Richmond, Virginia, Philadelphia, Pennsylvania, and Columbus, Ohio) and

Appendix I: Scope and Methodology

for its Defense Energy Support Center, Defense Distribution Center, and
Defense Reutilization and Marketing Service. We also reviewed school
improvement plans for three elementary and middle schools located at Fort
Bragg, North Carolina, to determine whether they aligned with DODEA's
community strategic plan and performance contract goals and measures.

To assess the extent to which DOD has established mechanisms to capture
and share lessons learned from the implementation of performance plans and
balanced scorecards, we interviewed knowledgeable DOD, defense agency, and
service officials and reviewed pertinent documents.

We performed our work between July 2003 and May 2004 in accordance with
generally accepted government auditing standards. We obtained comments on
a draft of this report from DOD and incorporated comments where
appropriate.

Appendix II: Department of Defense's Risk Management Framework

In 2001, the Department of Defense (DOD) established a risk management
framework to use in considering trade-offs among defense objectives and
resource constraints. The framework consists of four dimensions of risk
with supporting performance measures that are being cascaded to the
defense agencies and activities as well as departmentwide. Although not
depicted, financial considerations underlay each of DOD's risk areas.

                   Figure 5: DOD's Risk Management Framework

                                  Source: DOD.

Appendix III: Fiscal Year When Defense Agencies Implemented Their First
Performance Contract, Plan, or Scorecard

This appendix shows the fiscal year during which each defense agency and
activity implemented its first performance contract or performance plan.
The defense agencies and activities that implemented performance contracts
later converted these to performance plans or scorecards. We use the
generic term "performance plan" in figures 6 and 7 to include both
performance plans and balanced scorecards.

    Appendix III: Fiscal Year When Defense Agencies Implemented Their First
                    Performance Contract, Plan, or Scorecard

                           Figure 6: Defense Agencies

First Performance Contract/Plan Implemented

Fiscal Year 1999

Fiscal Year 2000

Fiscal Year 2001

Fiscal Year 2002

Fiscal Year 2004

DOD agencies without performance plans as of fiscal year 2004

Source: GAO analysis and presentation of DOD data.

ASD ATSD DUSD PDUSD USD

Assistant Secretary of Defense
Assistant to the Secretary of Defense
Deputy Under Secretary of Defense
Principal Deputy Under Secretary of Defense

Under Secretary of Defense

    Appendix III: Fiscal Year When Defense Agencies Implemented Their First
                    Performance Contract, Plan, or Scorecard

                       Figure 7: Defense Field Activities

First Performance Contract/Plan Implemented ASD Assistant Secretary of
Defense

ATSD Assistant to the Secretary of DefenseFiscal Year 1999

DUSD Deputy Under Secretary of Defense

Fiscal Year 2000 PDUSD Principal Deputy Under Secretary of Defense

DOD field activities without performance USD Under Secretary of
Defenseplans as of fiscal year 2004

Source: GAO analysis and presentation of DOD data.

Appendix IV: Certain Key Attributes Associated with Results-Oriented
Management

Establishing goals and measuring performance are essential to successful
results-oriented management practices. Measuring performance allows
organizations to track the progress they are making toward their goals and
gives managers critical information on which to base decisions for
improving their programs. Our body of work on results-oriented management
practices has identified key attributes of success. This work indicates
that agencies that are successful in achieving business transformation
goals, such as those being focused on by the Department of Defense in its
performance contracts and plans, strive to establish practices and
performance systems at all levels of the agency that include the seven key
attributes described in this appendix.

o  	Aligns goals and measures with agency and departmentwide goals.
Performance goals and measures should align with an agency's long-term
strategic goals and mission as well as with higher-level departmentwide
priorities, with the relationship clearly articulated. Such linkage is
important in ensuring that agency efforts are properly aligned with goals
(and thus contribute to their accomplishment) and in assessing progress
toward achieving these goals. Goals and measures also should cascade from
the corporate level of the agency to the operational level to provide
managers and staff with a road map that shows how their day-to-day
activities contribute to achieving agency and departmentwide performance
goals. In addition, measures used at the lowest levels of the agency to
manage specific programs should directly relate to unit results and
upwards to the corporate level of the agency.

o  	Assigns accountability for achieving results. We have previously
reported that the single most important element of successful management
improvement initiatives is the demonstrated commitment of top leaders in
developing and directing reform efforts. Top leadership must play a
critical role in setting results-oriented goals and quantifiable measures
that are cascaded to lower organizational levels and used to develop and
reinforce accountability for achieving results, maintain focus on the most
pressing issues confronting the organization, and sustain improvement
programs and performance, especially during times of leadership
transition. One way to reinforce accountability is through the use of
employee performance appraisals that reflect an organization's goals.

o  	Demonstrates results. A key element of any major program is its
ability to establish clearly defined goals and performance measures to
monitor and report progress to management, identify corrective actions,
set priorities, and allocate resources. Performance measures

Appendix IV: Certain Key Attributes Associated with Results-Oriented
Management

should show an organization's progress toward achieving an intended level
of performance or results and include quantifiable targets or other
measurable values to help assess progress and trend data toward
accomplishing goals. Using a combination of output measures that assess
the actual level of work accomplished or services provided and outcome
measures that assess the actual results, effects, or impacts of a program
or service compared to the intended purpose can also help determine
progress made.

o  	Provides a comprehensive view of agency performance. For each key
business line, performance goals and measures should provide a
comprehensive view of performance, including customers' and stakeholders'
priorities. Goals and measures should address key performance dimensions
such as (1) factors that drive organizational performance-including
financial, customer, and internal business processes, and workforce
learning and growth-and (2) aspects of customer satisfaction, including
timeliness, quality, quantity, and cost of services provided. Doing so can
allow managers and other stakeholders to assess accomplishments, make
decisions, realign processes, and assign accountability without having an
excess of data that could obscure rather than clarify performance issues.

o  	Links resource needs to performance. One of the ways that performance
management can be promoted is if this information becomes relevant for (1)
identifying resources (e.g., human capital, information technology, and
funding) needed to achieve performance goals; (2) measuring cost; and (3)
informing budget decisions. When resource allocation decisions are linked
to performance, decision makers can gain a better understanding of the
potential effect of budget increases and decreases on results.

o  	Discusses data quality. A key factor that contributes to the
usefulness of performance data is the degree to which officials are
confident that the performance information is credible. Useful practices
for helping decision makers to assess the quality and value of performance
data include discussion of data sources and standards and methods-
including data collection and reliability, verification and validation
procedures, and proposals to review these procedures-to assess the quality
of performance data. Reliability refers to whether standard procedures for
collecting and calculating results can be applied to measures so that they
produce the same results if applied repeatedly to the same situation.
Validity is the extent to which data adequately represent actual
performance. Verification is the assessment of data

Appendix IV: Certain Key Attributes Associated with Results-Oriented
Management

completeness, accuracy, consistency, timeliness, and related quality
control practices.

o  	Provides contextual information. Performance reporting systems should
include information to help clarify aspects of performance that are
difficult to quantify or to provide explanatory information such as
factors that were within or outside the control of the agency. This
information is critical to identifying and understanding the factors that
contributed to a particular result and can help officials measure, assess,
and evaluate the significance of underlying factors that may affect
reported performance. In addition, this information can provide context
for decision makers to establish funding priorities and adjust performance
targets and assess means and strategies for accomplishing an
organization's goals and objectives.

Appendix V: Comments from the Department of Defense

                                 Now on p. 36.

                                 Now on p. 36.

                                 Now on p. 36.

Appendix V: Comments from the Department of Defense

                                 Now on p. 36.

                                 Now on p. 36.

Appendix VI: Key Contact and Staff Acknowledgments

GAO Contact Marilyn K. Wasleski (202) 512-8436

Acknowledgments 	Gabrielle M. Anderson, Krislin M. Bolling, Elizabeth H.
Curda, Benjamin T. Licht, William A. McKelligott, Elizabeth G. Mead,
Marjorie L. Pratt, and Robert K. Wild also made significant contributions
to the report.

Related GAO Products

Future Years Defense Program: Actions Needed to Improve
Transparency of DOD's Projected Resource Needs. GAO-04-514.
Washington, D.C.: May 7, 2004.

Department of Defense: Further Actions Needed to Establish and
Implement a Framework for Successful Business Transformation.
GAO-04-626T. Washington, D.C.: March 31, 2004.

Results-Oriented Government: GPRA Has Established a Solid
Foundation for Achieving Greater Results. GAO-04-38. Washington, D.C.:
March 10, 2004.

Satellite Communications: Strategic Approach Needed for DOD's
Procurement of Commercial Satellite Bandwidth. GAO-04-206.
Washington, D.C.: December 10, 2003.

Defense Inventory: Several Actions Needed to Further DLA's Efforts to
Mitigate Shortages of Critical Spare Parts. GAO-03-709. Washington,
D.C.: August 1, 2003.

Results-Oriented Cultures: Creating a Clear Linkage between Individual
Performance and Organizational Success. GAO-03-488. Washington, D.C.:
March 14, 2003.

Defense Management: New Management Reform Program Still Evolving.
GAO-03-58. Washington, D.C.: December 12, 2002.

Tax Administration: IRS Needs to Further Refine Its Tax Filing Season
Performance Measures. GAO-03-143. Washington, D.C.:
November 22, 2002.

Performance Budgeting: Opportunities and Challenges. GAO-02-1106T.
Washington, D.C.: September 19, 2002.

Information Technology: Defense Information Systems Agency Can
Improve Investment Planning and Management Controls. GAO-02-50.
Washington, D.C.: March 15, 2002.

Defense Management: Actions Needed to Sustain Reform Initiatives and
Achieve Greater Results. GAO/NSIAD-00-72. Washington, D.C.:
July 25, 2000.

Related GAO Products

Management Reform: Elements of Successful Improvement Initiatives

GAO/T-GGD-00-26. Washington, D.C.: October 15, 1999.

Performance Plans: Selected Approaches for Verification and Validation of
Agency Performance Information. GAO/GGD-99-139. Washington, D.C.: July 30,
1999.

The Results Act: An Evaluator's Guide to Assessing Agency Annual
Performance Plans. GAO/GGD-10.1.20. Washington, D.C.: April 1, 1998.

Executive Guide: Effectively Implementing the Government Performance and
Results Act. GAO/GGD-96-118. Washington, D.C.: June 1, 1996.

GAO's Mission

Obtaining Copies of GAO Reports and Testimony

The Government Accountability Office, the audit, evaluation and
investigative arm of Congress, exists to support Congress in meeting its
constitutional responsibilities and to help improve the performance and
accountability of the federal government for the American people. GAO
examines the use of public funds; evaluates federal programs and policies;
and provides analyses, recommendations, and other assistance to help
Congress make informed oversight, policy, and funding decisions. GAO's
commitment to good government is reflected in its core values of
accountability, integrity, and reliability.

The fastest and easiest way to obtain copies of GAO documents at no cost
is through GAO's Web site (www.gao.gov). Each weekday, GAO posts newly
released reports, testimony, and correspondence on its Web site. To have
GAO e-mail you a list of newly posted products every afternoon, go to
www.gao.gov and select "Subscribe to Updates."

Order by Mail or Phone 	The first copy of each printed report is free.
Additional copies are $2 each. A check or money order should be made out
to the Superintendent of Documents. GAO also accepts VISA and Mastercard.
Orders for 100 or more copies mailed to a single address are discounted 25
percent. Orders should be sent to:

U.S. Government Accountability Office 441 G Street NW, Room LM Washington,
D.C. 20548

To order by Phone: 	Voice: (202) 512-6000 TDD: (202) 512-2537 Fax: (202)
512-6061

To Report Fraud, Contact:

Waste, and Abuse in Web site: www.gao.gov/fraudnet/fraudnet.htm

E-mail: [email protected] Programs Automated answering system: (800)
424-5454 or (202) 512-7470

Gloria Jarmon, Managing Director, [email protected](202)
512-4400Congressional U.S. Government Accountability Office, 441 G Street
NW, Room 7125 Relations Washington, D.C. 20548

Public Affairs 	Jeff Nelligan, Managing Director, [email protected] (202)
512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149
Washington, D.C. 20548

                           PRINTED ON RECYCLED PAPER

           Presorted Standard Postage & Fees Paid GAO Permit No. GI00

United States
Government Accountability Office
Washington, D.C. 20548-0001

Official Business
Penalty for Private Use $300

Address Service Requested
*** End of document. ***