This is the accessible text file for GAO report number GAO-02-738 
entitled 'Defense Logistics: Improving Customer Feedback Program Could 
Enhance DLA's Delivery of Services' which was released on September 23, 
2002. 

This text file was formatted by the U.S. General Accounting Office 
(GAO) to be accessible to users with visual impairments, as part of a 
longer term project to improve GAO products' accessibility. Every 
attempt has been made to maintain the structural and data integrity of 
the original printed product. Accessibility features, such as text 
descriptions of tables, consecutively numbered footnotes placed at the 
end of the file, and the text of agency comment letters, are provided 
but may not exactly duplicate the presentation or format of the printed 
version. The portable document format (PDF) file is an exact electronic 
replica of the printed version. We welcome your feedback. Please E-mail 
your comments regarding the contents or accessibility features of this 
document to Webmaster@gao.gov. 

This is a work of the U.S. government and is not subject to copyright 
protection in the United States. It may be reproduced and distributed 
in its entirety without further permission from GAO. Because this work 
may contain copyrighted images or other material, permission from the 
copyright holder may be necessary if you wish to reproduce this 
material separately. 

United States General Accounting Office: 
GAO: 

Report to Congressional Committees: 

September 2002: 

Defense Logistics: 

Improving Customer Feedback Program Could Enhance DLA's Delivery of 
Services: 

GAO-02-776: 

GAO Highlights: 

Highlights of GAO-02-776, a report to the Committee on Armed Services, 
U.S. Senate, and the Committee on Armed Services, House of 
Representatives. 

Why GAO Did This Study: 

The Defense Logistics Agency supports America’s military forces 
worldwide by supplying almost all consumable items—from food to jet 
fuel—that the military services need. The Floyd D. Spence Defense 
Authorization Act for Fiscal Year 2001 mandated that GAO conduct 
reviews of the agency, including its relationship with its military 
service customers. For this report, GAO determined (1) how customers 
perceive the quality of the agency’s service, (2) how useful its 
approaches are for obtaining customer feedback, and (3) whether 
opportunities exist to enhance its initiatives for improving customer 
service. 

What GAO found: 

Military service customers at eight judgmentally selected locations GAO
visited had mixed views of the Defense Logistics Agency’s services—
satisfied with aspects of routine service, such as the delivery time for
routine parts, but dissatisfied with other areas, such as the 
detrimental impact that the agency’s service has had on their 
operations. Customers cited difficulties, for example, in getting 
critical weapons systems parts by the needed time. 

The agency’s approach for obtaining systematic customer service
feedback is limited. It: 

* lacks an integrated method to obtain adequate data on problems; 

* does not effectively use surveys or local representatives to obtain 
feedback to identify the importance or depth of customers’ issues; 

* has not adequately defined or identified its customers; and; 

* does not provide a “single face” to its customers, thus fragmenting
accountability for customer satisfaction. 

Agency management acknowledged that the agency has not been customer 
focused and has been slow to respond to customer support concerns. The 
agency is acting to improve its customer relationships and provide a 
single face to its customers. But these initiatives do not fully 
address the limitations in its current approaches to obtain feedback and
do not incorporate other soliciting and analytical approaches, such as
those used in the private sector. Research of best practices for 
customer satisfaction suggests that multiple approaches and the 
integration of feedback data are needed to effectively listen to and 
understand customers’ perceptions and needs and to take appropriate 
actions to meet those needs. 

Figure: Defense Logistics Agency’s Process for Providing Customers with 
Needed Materiel: 

[See PDF for image] 

This figure is an illustration of the various Defense Logistics 
Agency’s processes for providing customers with needed materiel, with 
the following data depicted: 

Customers: 
* Requisitions to Agency supply centers; 
* Orders from Agency supply centers to Agency distribution depots; 
* Material is delivered. 

Customers: 
* Requisitions to Agency supply centers; 
* Orders from Agency supply centers to Manufacturers/Vendors; 
* Material is delivered. 

Customers: 
* Orders to vendors; 
* Material is delivered. 

[End of figure] 

What GAO Recommends: 

GAO recommends that the Secretary of Defense direct the Defense 
Logistics Agency, along with the military services, as appropriate, to: 

* develop a comprehensive customer-feedback plan to better determine 
customer needs and solutions to the needs; 

* determine who its customers are and their needs, and; 

* clarify guidance for customer representatives to help create a 
“single face” for customers. 

DOD generally concurred with GAO’s recommendations and agreed that DLA 
needs to increase its focus on customer satisfaction. 

This is a test for developing Highlights for a GAO report. The full 
report, including GAO's objectives, scope, methodology, and analysis is 
available at [hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-776]. 
For additional information about the report, contact Charles I. Patton, 
Jr. (202-512-4412). To provide comments on this test Highlights, 
contact Keith Fultz (202-512-3200) or e-mail HighlightsTest@gao.gov. 

[End of section] 

Contents: 

Results in Brief: 

Background: 

Customer Satisfaction with DLA Services Is Mixed: 

Usefulness of Customer Feedback Approaches Has Been Limited: 

Initiatives for Achieving a Better Customer Focus Could Be Enhanced 
Through Improved Customer Feedback Approaches: 

Conclusions: 

Recommendations for Executive Action: 

Agency Comments and Our Evaluation: 

Appendix I: Scope and Methodology: 

Appendix II: Comments from the Department of Defense: 

Appendix III: GAO Contact and Staff Acknowledgements: 

Table: 

Table 1: DLA Customer Segments and Illustrative Military Commands: 

Figures: 

Figure 1: DLA’s Supply-Chain Management Process: 

Figure 2: Example of Relationship between DODAACs and Army Customer 
Activities: 

Figure 3: AT&T Customer Feedback and Listening Strategies: 

Figure 4: DLA Customer Locations Visited by GAO: 

Abbreviations: 

DLA: Defense Logistics Agency: 

DOD: Department of Defense: 

DODAACs: DOD Activity Address Codes: 

GAO: General Accounting Office: 

[End of section] 

United States General Accounting Office: 
Washington, DC 20548: 

September 9, 2002: 

The Honorable Carl Levin: 
Chairman: 
The Honorable John W. Warner: 
Ranking Minority Member: 
Committee on Armed Services: 
United States Senate: 

The Honorable Bob Stump: 
Chairman: 
The Honorable Ike Skelton: 
Ranking Minority Member: 
Committee on Armed Services: 
House of Representatives: 

The Defense Logistics Agency (DLA) performs a critical role in 
supporting America’s military forces worldwide by supplying almost every
consumable item—from food to jet fuel—that the military services need to
operate. To fulfill this role, the agency oversees a staff of more than 
28,000 civilian and military employees who work in all 50 states and 27 
foreign countries. It manages approximately 4 million supply items and 
processes over 23 million requisitions annually. DLA reported that, in 
fiscal year 2001, these operations resulted in sales to the military 
services of about $15 billion, of which $12 billion was for supplies. 

This report is one in a series mandated under the Floyd D. Spence 
Defense Authorization Act for Fiscal Year 2001. [Footnote 1] The act 
directed that we review DLA’s efficiency and effectiveness in meeting 
customer requirements, the application of best business practices, and 
opportunities for improving the agency’s operations. As agreed with 
your offices, this report focuses on the relationship between DLA and 
its military service customers. More specifically, we determined (1) 
how customers perceive the quality of service they receive, (2) how 
useful the agency’s approaches are for obtaining customer service 
feedback, and (3) whether there are opportunities to enhance the 
agency’s initiatives for improving customer service. To address these 
objectives, we used a case study approach to obtain customers’ views. 
Our scope was limited to a judgmentally selected number of materiel 
management customers. We visited eight military service customer 
locations within the continental United States. The results of our work 
at these locations are not projectable to the agency as a whole. 
However, studies conducted by the Joint Chiefs of Staff, DLA surveys, 
and comments from agency headquarters officials suggest that many of 
the issues we raise in this report are systemic in nature. The details 
on our objectives, scope, and methodology are in appendix I. 

Results in Brief: 

Customers at the eight locations we visited expressed both satisfaction
and dissatisfaction with the services the agency provides. While they 
were satisfied with some aspects of routine service, such as the 
delivery time for routine parts and certain contractor service 
arrangements, customers also raised a number of points of 
dissatisfaction, particularly with regard to the detrimental impact of 
DLA’s service on their operations. For example, many customers cited 
difficulties in getting critical weapons systems parts in time to meet 
their needs, resulting in equipment readiness deficiencies as well as 
the cannibalization of other equipment to obtain needed parts. Not 
getting accurate and timely information on the status and/or 
availability of critical items frustrated other customers. Some of the 
difficulties that customers encountered in trying to get parts from DLA
included inaccurate dates from automated systems on the status of 
deliveries, difficulty in obtaining additional information on the 
availability of parts, and a lack of support from DLA in identifying 
alternate vendors or other means to obtain critical items that were 
unavailable through DLA. 

The agency’s approach for obtaining customer service feedback has been 
of limited usefulness because it lacks a systematic integrated approach 
for obtaining adequate information on customer service problems. For
example, DLA has not adequately defined or identified all of its 
customers, leaving it without a sufficient means to initiate and 
maintain contact with its many thousands of customers to solicit 
meaningful feedback. In addition, although DLA reaches out to selected 
customers through satisfaction surveys and the use of local customer 
support representatives at various locations, these mechanisms do not 
provide the customer feedback that DLA needs to identify the 
significance or depth of issues that particularly trouble its 
customers. Furthermore, the satisfaction survey response rates are too 
low to provide meaningful statistical analyses of customer 
satisfaction. Lastly, DLA’s current customer support system does not 
provide a “single face” to its customers, leaving accountability for 
ensuring high customer satisfaction fragmented throughout the agency. 

While DLA has initiatives under way to improve its customer service, 
there are opportunities to enhance these initiatives to provide for an 
improved customer feedback program. DLA management at the highest 
levels has acknowledged that the agency has not been as customer 
focused as it should be, has been slow to respond to customer-support 
concerns, and is taking actions to improve its customer relationships. 
However, the agency’s initiatives do not completely address the 
limitations we identified in its current approaches for obtaining 
customer service feedback. For example, while DLA’s new strategy lays 
out a means to provide a single face to its customers, it does not 
incorporate other approaches, such as those used in the private sector, 
to solicit and analyze feedback from those customers. Research on best 
practices in the area of customer satisfaction suggests that multiple 
approaches are needed to effectively listen to customers about their 
perceptions of quality service and needs. Such approaches include 
customer service surveys, telephone interviews, and customer complaint 
programs. Best practices research also highlights the need to integrate 
all data obtained through various customer feedback approaches so that 
service providers can completely understand customer perceptions and 
take appropriate actions to meet customer needs. 

This report includes recommendations for executive action to help DLA
better identify customers’ needs and solutions for meeting them through
an integrated customer feedback framework. The Department of Defense
(DOD) generally concurred with our recommendations and agreed that DLA 
needs to increase its focus on customer satisfaction. The department’s 
comments on our report are reprinted in their entirety in appendix II. 

Background: 

DLA is a DOD Combat Support Agency under the supervision, direction,
authority, and control of the Under Secretary of Defense for 
Acquisition, Technology, and Logistics. DLA’s mission is to provide its 
customers—the military services and federal civilian agencies—with 
effective and efficient worldwide logistics support as required. 
[Footnote 2] DLA buys and manages a vast number and variety of items 
for its customers, including commodities such as energy, food, 
clothing, and medical supplies. DLA also buys and distributes hardware 
and electronics items used in the maintenance and repair of equipment 
and weapons systems. 

Customers determine their requirements for materiel and supplies and
submit requisitions to any of four DLA supply centers. [Footnote 3] The 
centers then consolidate the requirements and procure the supplies for 
their customers. DLA provides its customers with requested supplies in 
two ways: some items are delivered directly from a commercial vendor 
while other items are stored and distributed through a complex of 
worldwide distribution depots that are owned and managed by both DLA 
and the military services. DLA refers to this ordering and delivery 
process as materiel management or supply-chain management. [Footnote 4] 
Figure 1 provides a snapshot of this process. 

Figure 1: DLA’s Supply-Chain Management Process: 

[See PDF for image] 

This figure is an illustration of the various Defense Logistics 
Agency’s processes for providing customers with needed materiel, with 
the following data depicted: 

Customers: 
* Requisitions to Agency supply centers; 
* Orders from Agency supply centers to Agency distribution depots; 
* Material is delivered. 

Customers: 
* Requisitions to Agency supply centers; 
* Orders from Agency supply centers to Manufacturers/Vendors; 
* Material is delivered. 

Customers: 
* Orders to vendors; 
* Material is delivered. 

Source: GAO’s analysis of DLA’s process. 

[End of figure] 

Because DLA is the sole supplier for many critical items that can 
affect the readiness of the military services, the agency strives to 
provide its customers with the most efficient and effective logistics 
support. Thus, DLA has adopted a policy to provide customers with “the 
right item, at the right time, right place, and for the right price, 
every time.” In an effort to institutionalize this customer support 
concept, DLA has adopted the Balanced Scorecard approach [Footnote 5] 
to measure the performance of its logistics operations. The scorecard, 
a best business practice used by many private and public organizations, 
is intended to measure DLA’s performance by integrating financial 
measures with other key performance indicators around customers’ 
perspectives; internal business processes; and organization growth, 
learning, and innovation. 

Customer Satisfaction with DLA Services Is Mixed: 

Our work showed that customers at the eight locations we visited 
expressed satisfaction and dissatisfaction with the services the agency
provides. On the one hand, customers are generally satisfied with DLA’s
ability to quickly respond to and deliver requests for routine, high-
demand, in-stock items; provide customers with an easy-to-use ordering 
system; and manage an efficient prime vendor program. On the other hand,
customers at some locations were dissatisfied that, among other things,
DLA is unable to obtain less frequently needed, but critical, items and
parts and provide accurate and timely delivery status information. Some
customers did not express an opinion on the overall quality of customer
service. 

Customers Generally Satisfied with Routine Services: 

One aspect of DLA customer support is to provide customers with 
supplies when they need them. Common supplies include vehicle parts 
such as pumps, hoses, filters, and tubing. Timeliness, which sometimes
requires deliveries to be made in a day or less, can vary with 
customers, depending on the particular item. However, customers at all 
locations we visited commented that they were generally satisfied with 
DLA’s ability to provide most supply items in a time frame that meets 
their needs. Customers stated that the majority of the routine, 
frequently demanded supplies they order through DLA are delivered 
quickly—a view that is also supported by a February 2002 DLA 
performance review. The review concluded that the majority of 
requisitions (over 85 percent) was filled from existing inventories 
within DLA’s inventory supply system. Similarly, a 2001 Joint Staff 
Combat Support Agency Review Team assessment of DLA’s support to the 
unified commands indicated that overall, DLA received outstanding 
comments regarding its ability to provide its customers with timely 
supplies and services. [Footnote 6] 

Customers were also satisfied with the ease in ordering supplies such as
the pumps, hoses, and filters mentioned above. Customers stated that 
even though they conduct large amounts of business through DLA, they 
had few problems with the ordering process. This occurs because, 
according to some customers, ordering is facilitated by effective on-
line systems that work well and have readily available information. 

Another method that DLA uses to ensure customer satisfaction is its 
prime vendor program, which DLA instituted to simplify the procurement 
and delivery of such items as subsistence and medical or pharmaceutical
supplies that commonly have a short shelf life. The program enables
customers to directly interact with vendors, thereby reducing the 
delivery time for these supplies. Two customers of these DLA-managed 
prime vendor programs told us the programs effectively reduced delivery 
time. For example, at one location, prime vendors reduced the delivery 
time of food items from 7 days—the time it took to deliver the items 
when purchased from DLA—to 2 days for items purchased directly from 
prime vendors. [Footnote 7] The customers we spoke with at a medical 
supply unit told us they were so pleased with the prime vendor’s quick 
delivery time that they intend to obtain even more medical supplies 
from the prime vendor. They also told us that the prime vendor provides 
an additional service in the form of monthly visits to assess customer 
satisfaction with its services. The unit pointed out that DLA’s 
customer support representatives [Footnote 8] are less likely to make 
such frequent visits. 

Customers Also Expressed Dissatisfaction with Some DLA Services: 

Although customers seemed pleased with the way DLA handles routinely
available items, some raised concerns over the agency’s ability to 
provide critical items such as weapon system parts, timely and accurate
information on the status of ordered items, and proactive management for
high-priority requisitions. A Combat Support Agency Review Team 
assessment in 1998 also surfaced similar issues. Additionally, customers
we talked to criticized how DLA manages customer-owned assets in DLA
warehouses. 

Difficulties in Obtaining Critical Parts: 

As previously noted, DLA strives to provide the timely delivery of all
supplies and parts, including common consumable supply items like food;
clothing and hardware; and critical parts for weapons systems such as 
tanks, helicopters, and missiles. Customers at four locations we visited
told us that DLA was not able to timely deliver some critical items, 
such as weapons systems parts, which significantly affected their 
equipment readiness. A number of customers told us that the items they 
have difficulty obtaining from DLA are those that are more costly or
infrequently required. At two locations, customers used parts from
existing equipment (known as “parts cannibalization”) because they were
unable to obtain the parts they needed. At two other locations, 
customers said they grounded aircraft and/or deployed units without 
sufficient supplies. Customers at one location experienced an over-6-
month delay in obtaining helicopter parts. As a result, customers at 
this location told us that some of the unit’s helicopters were unable 
to fly their missions. We reported in November 2001 that equipment 
cannibalizations adversely affect the military services, resulting in 
increased maintenance costs, and lowered morale and retention rates 
because of the increased workload placed on mechanics. [Footnote 9] 

One customer also told us that DLA does not provide adequate 
information about items requiring long procurement lead times. The 
customer stated that having this information more readily available 
would aid customers in making decisions about the types and quantities 
of items they should retain to minimize the impacts of long DLA lead 
times. 

The 1998 Combat Support Agency Review Team’s assessment conducted
at military service field activities found that even though DLA met its
overall supply availability goal of 85 percent, the remaining 15 
percent of items that were not available “almost certainly includes a 
number of items that are critical to the operation of essential weapon 
systems.” The assessment attributed this shortfall to flaws in DLA’s 
requirements determination models, which are used to estimate 
customers’ demands so that DLA can maintain sufficient inventory 
quantities. 

The study further stated that customers are not satisfied with the 
delivery time for items that are not in stock. In fact, in April 2002, 
the overall logistics response time was almost 100 days for nonstocked 
items—a problem that appears to have persisted for the last several 
years, in spite of efforts to reduce this time. Customers at four 
locations provided us with examples of back-ordered items having lead 
times in excess of 1 year, such as navigational instruments and 
airframe parts. In discussing this issue further with DLA headquarters 
officials, they acknowledged that this is a problem and are working on 
a number of initiatives to address customers’ concerns. 

Inaccurate and Untimely Status Information: 

Customers need accurate and timely information on the status of their
orders so they can plan equipment maintenance schedules to optimize the
readiness of existing equipment. However, customers at six locations 
were frustrated with obtaining accurate and timely information from DLA 
item managers and the automated systems that are intended to provide 
status information on requisitions. Customers at three locations said 
that when they tried to directly contact item managers by telephone, 
the managers often could not be reached and voice-mail messages were 
seldom returned. 

Furthermore, military service customers told us that DLA’s automated
requisition systems often do not contain accurate status data. Of 
particular concern to customers are the expected shipping or delivery 
dates posted on the automated systems. These dates show when parts will 
be available and allow units to coordinate maintenance schedules. If 
the dates are incorrect, units cannot effectively plan to have 
equipment available to be repaired. We discussed this concern with DLA 
headquarters officials, who told us they are investigating the problem. 

Lack of Proactive Management for High-Priority Requisitions: 

Another significant concern raised by customers at three locations was
that DLA is not proactive in seeking alternate ways to obtain critical 
items that are not immediately available within DLA’s supply system. DLA
typically places such items on back order, which, to meet mission needs,
places a burden on customers to find their own means to obtain the
necessary items right away. A number of customers at these three
locations said they felt that DLA, in an effort to be more customer 
focused, should do more to seek out alternate sources of supply to 
alleviate these high-priority back orders. Some customers also remarked 
that the required efforts for them to call vendors and solicit bids is 
a problem for their unit because of limited staffing levels and lack of 
contracting capabilities. 

In one instance, an aviation supply unit requisitioned a critical part 
from DLA that was needed to repair a helicopter unable to fly its 
mission. This requisition was placed on back order by DLA, and delivery 
was not expected to occur until 8 months later. Because of the critical 
nature of the needed part, the unit had to search for other means to 
obtain the part sooner. In fact, the unit directly contacted the same 
vendor that DLA was working with to fill the back orders and learned 
that the vendor had stock on hand and would be able to ship the item 
immediately. The unit subsequently purchased the part from that vendor 
instead of waiting for it to be available from DLA. 

In another instance, a DLA item manager informed an aircraft 
maintenance depot customer that $2 million worth of critical parts for a
helicopter engine overhaul program would be placed on back order 
because the parts were not available from the DLA vendor. In researching
listings for property to be disposed of, [Footnote 10] the customer 
found the required parts—still new and unopened in the manufacturers’ 
container—available for redistribution or sale within DLA’s disposal 
system. As a result, the customer initiated a shipping request to 
procure the $2 million in helicopter parts for only the cost to ship 
the items. 

Ineffective Management of DLA Warehouses: 

DLA manages all warehousing functions at locations where a DLA 
distribution depot [Footnote 11] is collocated with a military 
activity. Management functions include, among other things, logging in 
and storing equipment. During the course of our interviews, customers 
raised concerns over DLA’s handling of these functions. At three of the 
sites we visited, the customers perceived that their assets were not 
being serviced and maintained as required. Their concerns centered on 
DLA’s process for recording the ownership of equipment and the 
commingling of different customers’ inventories. 

To assign asset ownership, DLA “codes” items in its automated inventory
system. That is, DLA assigns unique codes to differentiate between Army,
Navy, Marine Corps, Air Force, and DLA-owned assets. However, customers 
at three locations we visited stated that in numerous instances, DLA 
assigned inventory items to the wrong management account, thus creating 
the possibility that an item ordered and paid for by one unit or 
service could be issued to another. One location we visited had 
documented over $1 million worth of items coded into the wrong 
management account. Another location identified $621,000 worth of 
incorrectly coded items. Before the errors were corrected, neither 
activity could access the materials they needed. As a result, both 
locations invested unnecessary amounts of time and money in correcting 
DLA’s error. During our review, we brought this issue to the attention 
of DLA officials, who indicated that they would investigate the 
problem. 

Customers also expressed concerns about the commingling of service-owned
assets with DLA-owned assets in DLA-managed warehouses. Like inaccurate 
coding, commingling creates a significant risk that items will be 
issued by the warehouse to someone other than the purchasing unit. As a
result, the items would not be available to the true owner when needed.
Also, for equipment items that need periodic inspection and repair, 
there is a risk the owner will expend resources to perform maintenance 
or repairs but not be able to retrieve the item because DLA mistakenly 
issued that item to a different requisitioning entity or military 
service. As a result, the “true owner” could have needlessly spent 
resources on items given to somebody else and also be left with items 
still needing repair. In discussions with DLA headquarters officials, 
they acknowledged the problem and told us that DLA is taking steps to 
address it with a National Inventory Management Strategy, which is part 
of DLA’s goal to better manage its supply chain effectiveness. 

Usefulness of Customer Feedback Approaches Has Been Limited: 

DLA’s approach for obtaining customer service feedback has been of 
limited usefulness because it lacks a systematic integrated approach for
obtaining adequate information on customer service problems. As a 
result, the agency does not have the information necessary to identify 
its customers’ concerns, and more importantly, to initiate actions for
improving customer service, thereby placing at risk DLA’s ability to 
meet its overall goal of providing quality service to the war fighter. 
In particular, DLA has not (1) adequately identified all of its 
customers, (2) effectively solicited customer feedback, and (3) clearly 
identified those accountable for ensuring customer satisfaction. 

DLA Has Not Adequately Identified All of Its Customers: 

Obtaining good meaningful feedback from customers means knowing who 
those customers are. DLA broadly defines a “customer” as someone who
purchases items or directly causes products to be bought, but DLA has 
not identified who those individuals are from the multitude of 
organizations it deals with. DLA’s current portfolio of customers is 
identified by approximately 49,000 address codes, known as DOD Activity 
Address Codes (DODAACs). [Footnote 12] The military services assign 
DODAACs to various organizations and activities for ordering supplies. 
However, these address codes, a legacy of a system built in the 1960s, 
contain little information about the customer’s organization beyond a 
physical address. No meaningful customer contact point is associated 
with the codes or, in many cases, a specific organization that DLA can 
use as a basis for interaction with the customers using their services. 
As a result, DLA has no effective process to initiate and maintain 
contact with its customers for soliciting feedback. Without such a 
customer interface process, DLA has no routine means to understand 
customers’ needs and to take appropriate corrective actions to address 
those needs. 

Our efforts to identify and interview DLA customers were hindered 
because a single DODAAC does not necessarily equate to a single 
customer. In many cases we found that one organization interacts with
DLA using a number of DODAACs. For example, DLA’s customer database
shows over 580 DODAACs for Fort Bragg. However, according to DLA and
Army officials, the number of Fort Bragg customer organizations 
interacting with DLA for these same DODAACs is smaller. The reason for
this is that, in part, central order points at Fort Bragg are 
responsible for submitting and tracking orders for a number of smaller 
organizations, thereby covering multiple DODAACs. In addition, each of 
these organizations also uses multiple DODAACs to differentiate between
various types of supply items, such as repair parts and construction
materials. For example, one DODAAC is used for ordering numerous
repair parts while another is used for ordering construction materials. 
One of these customer organizations at Fort Bragg is the Division 
Support Command of the 82nd Airborne Division, which interacts with DLA 
for supplies ordered using 159 different DODAACs. Thus, many DODAACs
could represent only one customer. Figure 2 illustrates the relationship
between the DODAACs used by DLA to define customers and the Division
Support Command. 

Figure 2: Example of Relationship between DODAACs and Army Customer
Activities: 

[See PDF for image] 

This figure is an illustration of the relationship between DODAACs and 
Army Customer activities. The following data is depicted: 

DODAACs: 

1. W36B2U: 
Name: PR LT MNT CO MSB A; 
2. W36BY3: 
Name: XR E CO 782D DX; 
3. W36BYM: 
Name: PR 505 INF 1 BN HHC; 
...
159. W36NOT: 
Name: SR 82 AVN D CO ASL. 

This is equivalent to: 

Army Customer: 
Division Support Command, 82nd Airborne Division. 

Source: GAO’s analysis of DLA- and Army-provided data. 

[End of figure] 

DLA Does Not Adequately Solicit Customer Feedback: 

A principal aspect of DLA’s strategic plan is for managers to focus on
customers’ needs and improve customer satisfaction by listening to
customers about the quality of service they receive—both good and bad—
and making changes necessary to enhance that service. DLA uses
customer surveys, customer support representatives, and focus groups to
obtain feedback from its customers on their level of satisfaction with 
the services DLA provides. For example, DLA conducts quarterly mail-out
surveys to measure overall customer satisfaction levels. It also places
customer support representatives at selected customer organizations to
assist customers in planning, implementing new supply initiatives, and
solving problems. However, we noted several weaknesses in these
methods. Specifically, (1) the satisfaction survey response rates are 
too low to provide meaningful statistical analyses of customer 
satisfaction, (2) the survey instrument does not provide a sufficient 
means to understand why customers may be less than satisfied, and (3) 
customer support representatives are more reactive than proactive in 
soliciting customer feedback. 

Quarterly Mail-out Surveys Have Low Response Rates: 

The quarterly mail-out surveys that DLA uses to measure customer
satisfaction elicit a relatively low number of responses from DLA
customers, significantly limiting its usefulness in soliciting customer 
feedback. The survey response rates were too low to provide meaningful
statistical analyses of customer satisfaction. The response rate for the
33,000 surveys that DLA mailed out in fiscal year 2001 averaged around 
23 percent, and only about 20 percent for the August 2001 cycle (the 
latest cycle where results have been made available). As such, less 
than one quarter of DLA’s customers are providing input on how they 
perceive DLA support and what problems they are experiencing that may 
need to be addressed. 

Large survey organizations like Gallup attempt to get response rates of
between 60 and 70 percent for their mail surveys. Experts on customer
satisfaction measurement have stated that although survey response rates
are never 100 percent, an organization should strive to get its rate as 
close as possible to that number. [Footnote 13] They suggest that 
ideally, organizations can obtain response rates of over 70 percent. 
The experts also noted that organizations conducting surveys commonly 
make the mistake of assuming that if a final sample size is large, the 
response rate is unimportant. This leads organizations to accept 
response rates well under 25 percent. However, such low rates can lead 
to serious biases in the data. 

Having an inadequate understanding of who its customers are likely
contributes to DLA’s problem with low response rates. The surveys are
mailed to addresses associated with the DODAACs and include with each
survey a message asking that the survey be provided to a person most
familiar with requisitioning and ordering supplies. However, during the
fiscal year 2001 survey period, over 2,200 of the 33,000 surveys mailed
(about 7 percent) were returned to DLA as “undeliverable” or were
delivered to people who were no longer customers. Furthermore, another
128 respondents noted in their survey returns that they do not consider
themselves to be customers. DLA officials stated that the undeliverable
rate increases when there are many units that move to other locations or
when service officials do not update DODAACs for changed addresses. 

Surveys Are Insufficient for Identifying Causes of Customer 
Dissatisfaction: 

The quarterly mail-out survey asks customers to rate their overall
satisfaction with DLA products and services, along with specific 
aspects of support, such as providing products in time to meet needs 
and effectively keeping customers informed. While these surveys provide 
general aggregate information on the levels of customer satisfaction, 
they do not provide the means to understand why customers may be less 
than satisfied. For example, a number of customers we interviewed voiced
concern over the fact that status dates for back-ordered items were 
either sometimes wrong or varied between different inventory systems. 
The survey might indicate only an overall low level of satisfaction in 
the area of keeping customers informed but would not provide a reason. 
If this problem were systemic throughout DLA, there would be less of an
opportunity to take immediate corrective action. Most recently, in June
1999, DLA supplemented a quarterly survey with two focus groups
targeted at soliciting specific customer feedback on DLA’s communication
efforts. While DLA determined the focus groups to be an excellent
feedback mechanism, the sample size was too small for DLA to run a
statistical analysis of the data obtained; and the topics for 
discussion were limited to customer communication. 

DLA officials stated that they use a number of methods to obtain 
customer feedback. These include analyses of survey results, focus 
groups, and structured interviews. However, they acknowledged that the 
usefulness of these methods is somewhat limited owing either to low 
response rates; limited discussion topics; small sample sizes; or, in 
the case of structured interviews, the fact that the most recent ones 
were conducted in 1997. 

DLA’s own survey results also indicate the flaws with its survey
techniques. For example, DLA’s fiscal year 2000 survey results show that
customers rated as “low satisfaction” their ability to reach the right 
DLA person to meet their needs. However, the survey noted that “due to 
its high importance to customers and the myriad of interpretations of 
‘less than satisfied’ responses to this attribute, more information 
will need to be gathered” to determine what issues are preventing 
customers from reaching the right person. This indicates that DLA’s 
survey was not adequate to get behind the underlying causes of customer 
dissatisfaction. In fact, with respect to low satisfaction ratings, the 
survey reports for fiscal years 2000 and 2001 recommended that DLA 
conduct one-on-one interviews to identify why customers were not 
satisfied with DLA services. 

Another difficulty that DLA encounters in using mail-out satisfaction
surveys to identify customer problems is that the surveys are designed 
to protect the confidentiality of the respondents, which limits DLA’s 
ability to follow up with customers for adequate feedback. As a result, 
there is no means to follow-up with customers expressing low 
satisfaction levels to identify specific problems or to determine what, 
if any, corrective actions are needed. During our meetings with DLA 
customers, we were able to identify specific problems only by engaging 
in a dialogue with them about their experiences. In conducting these in-
depth discussions on aspects of the supply process such as placing 
orders, obtaining the status of outstanding requisitions, receiving 
supply items, and obtaining customer service, we were able to ask 
follow-up questions to determine exactly what problems they were 
experiencing in some of these areas. 

Customer Support Representatives Not Proactive in Soliciting Feedback: 

Another method DLA uses to facilitate customer service is the placement
of customer support representatives at key customer locations. The use 
of these on-site representatives has the potential to provide DLA with 
a good link to its customers. In fact, some customers at three 
locations we visited specifically noted their satisfaction with the 
assistance the representatives provided. However, according to DLA 
headquarters officials, customer support representatives have been more 
reactive in that they help customers resolve only specific problems or 
assist in implementing new initiatives as requested. DLA headquarters 
officials told us that the representatives neither proactively solicit 
feedback on a regular basis from the multitude of customers in their 
geographical area nor reach out to identify the types of problems 
customers are experiencing. 

Furthermore, not all representatives are in contact with all DLA 
customers at their assigned locations. For example, at one location we 
visited, the representative was working closely with a specific 
customer organization. According to officials at this location, the 
representative has been very helpful to them in resolving supply 
problems and implementing new initiatives. However, a number of other 
customers at this location said they do not use the customer support 
representative at all because they use other options, such as call 
centers. Some customers noted that they were not even aware that there 
was such a representative in the area. The Combat Support Agency Review 
Team’s assessment in 1998 also found that some customers were unaware 
that customer support representatives even existed. The study 
identified a need for DLA to improve its interaction with customers and 
suggested that DLA “get out more and visit the customers” to identify 
and correct problems. Headquarters officials told us they assign 
customer support representatives to DLA’s larger customers, which 
account for about 5 percent of the overall customer population and 80 
percent of the agency’s business. Officials also stated they recognize 
that the customer support representative program is not as effective as 
it should be. As a result, the agency currently has initiatives under 
way to (1) provide more customer support representatives and training, 
(2) standardize the representatives’ roles, and (3) make the 
representatives more proactive in serving customers. 

Current Customer Feedback Framework Is Too Fragmented and Lacks 
Accountability: 

An important part of providing effective customer service is simplifying
customers’ access to the organization, such as through centralized 
contact points. In addition, best practices research emphasizes the 
need for a single, centralized management framework for receiving 
customer feedback so that all information about the customers can be 
linked together to facilitate a more complete knowledge of the customer.
However, DLA does not provide a “single face” to its customers for
addressing their issues. To obtain assistance, customers sometimes need
to navigate through a number of different channels, none of which are
interconnected. This process causes confusion with customers and
fragmented accountability throughout DLA for customer satisfaction. 

When customers order multiple types of supply items, they must use many
channels, depending on the type of item, to obtain assistance from DLA.
However, as DLA has noted, there is no single DLA contact point
responsible for resolving customers’ problems for all the items they
requisition. For example, the supply centers are responsible for 
managing specific weapons system parts or types of commodities. As 
such, problem resolution is performed through each supply center, 
depending on the type of item the customer is ordering. To obtain 
assistance with requisitions, customers must contact the appropriate 
supply center, generally through its customer “call center,” which is 
an activity dedicated to provide customer assistance for the particular 
items. In addition, Emergency Supply Operation Centers are available at 
each supply center for high priority items. Also, customers can contact 
individual item managers at the supply centers to resolve problems with 
their orders. At three locations, some customers told us they are 
sometimes confused over whom to call and reported difficulties with 
getting in touch with the right person to resolve their problems. 
Customers at four locations were also frustrated with the quality of 
assistance provided by DLA, noting that while some of the DLA 
representatives were helpful, others were not able to give them the 
assistance they needed. 

To illustrate further, one aviation supply unit we visited had high-
priority, back-ordered requisitions from each of the three DLA supply 
centers in Richmond, Virginia; Columbus, Ohio; and Philadelphia, 
Pennsylvania. As a result of these back orders, some of the unit’s 
aircraft were unable to operate because of maintenance needs. In order 
to get assistance with these requisitions, either to request help in 
expediting the order or to obtain better status information, unit 
supply personnel needed to contact the call centers or the Emergency 
Supply Operation Centers at each of the supply centers, depending on 
the item. If there were a single DLA point of contact, the unit could 
go to that contact for assistance with all the items on its list of 
priority requisitions. 

Another problem with DLA’s having many separate lines of communication 
with its customers is that meaningful information about those customers 
is not collected centrally for analysis. For example, each of the 
supply centers accumulates vital information about customer 
satisfaction through its contacts with customers. For instance, 
customers express specific problems they are having when getting help 
through the call centers. They might also convey information on 
problems they are having to various supply center teams conducting on-
site visits for purposes of training or other liaison activities. 
However, this information is neither shared between the supply centers 
nor provided to the DLA corporate level for a global review. As a 
result, no analysis of this information can be made to identify 
systemic problems or any accountability at one place for a given 
customer to ensure that its concerns are being addressed. 

Initiatives for Achieving a Better Customer Focus Could Be Enhanced
Through Improved Customer Feedback Approaches: 

While DLA has initiatives under way to improve its customer service, 
there are opportunities to enhance these initiatives to provide for an 
improved customer feedback program. DLA has recognized that it is not 
as customer focused as it should be and is developing a new strategy to 
improve its relationship with its customers. This new strategy, 
referred to as the Customer Relationship Management initiative, lays 
out an improved approach to customer service that creates a single DLA 
face to customers and focuses on customer segments to develop a better 
understanding of the customer. However, DLA’s initiatives do not 
completely address the limitations we identified in its current 
approaches for obtaining customer service feedback, such as by 
improving the way that it solicits feedback from individual customers. 
Research on best practices for customer service shows that successful 
organizations utilize multiple approaches to listen to their customers. 
These approaches include transaction surveys, customer interviews, and 
complaint programs that provide qualitative and quantitative data. The 
research also points to a need for centrally integrating all customer 
feedback so that managers can achieve a better understanding of 
customers’ perceptions and needs. 

DLA Is Developing a Strategy to Improve the Relationship with Its 
Customers: 

In February 2002, DLA’s Deputy Director stated that DLA “has been
internally focused rather than customer focused” and that its culture 
has been to talk to customers only “when problems arose.” To address 
this problem, DLA has begun a multimillion-dollar initiative aimed at 
focusing its business operations to better deliver important customer 
outcomes and actively managing relationships with its customers. This 
effort, known as Customer Relationship Management, is being developed 
in conjunction with DLA’s broader strategic planning initiatives such 
as Business Systems Modernization and implementation of the Balanced 
Scorecard approach to performance measurement. To implement Customer 
Relationship Management, DLA expects to spend about $73 million during 
fiscal years 2002-2008. According to DLA officials, when this effort is 
complete, DLA expects its customer service program to be on the same 
level as those in place at leading organizations in the private sector. 

The concept of the Customer Relationship Management initiative is a step
in the right direction toward significantly improving DLA’s relationship
with its customers. For example, part of the management initiative is a
plan to radically change the focus of its business practices and 
improve its interactions with customers. To do this, DLA is grouping 
customers by business segment, collaborating with these segments to 
achieve a better understanding of their needs, and tailoring logistics 
programs to the unique needs of the segments. Examples of business 
segments include deployable combat forces, industrial facilities, and 
training activities. Table 1 illustrates the proposed customer 
segments, which will include major military service commands. 

Table 1: DLA Customer Segments and Illustrative Military Commands: 

Segment: Deployed;
Army: Commanders-in-chief by geographic area of responsibility (e.g., 
U.S. European Command, U.S. Pacific Command, U.S. Central Command); 
Navy: [Empty]; 
Marine Corps: [Empty]; 
Air Force: [Empty]. 

Segment: Deployable; 
Army: Forces Command; 
Navy:Commander-in-Chief, U.S. Atlantic Fleet; 
Marine Corps: II Marine Expeditionary Force; 
Air Force: Air Combat Command. 

Segment: Training; 
Army: Training and Doctrine Command; 
Navy: Chief of Naval Education and Training; 
Marine Corps: Marine Corps Combat Development Command; 
Air Force: Air Education and Training Command. 

Segment: Industrial; 
Army: Army Materiel Command; 
Navy: Naval Air Systems Command; 
Marine Corps: Marine Corps Materiel Command; 
Air Force: Air Force Materiel Command. 

Segment: Other; 
Army: [A]; 
Navy: Naval Supply Systems Command; 
Marine Corps: [A]; 
Air Force: [A]. 

[A] No Army, Marine Corps, or Air Force commands designated by DLA for 
this segment. 

Source: DLA. 

[End of table] 

In an effort to streamline the numerous customer-reporting channels
currently in place, DLA plans to establish a multilevel-focused account
manager structure and increase accountability. DLA hopes that this 
effort will reduce the number of channels a customer must navigate to 
obtain assistance and focus accountability for customer satisfaction on 
account managers rather than on item managers. DLA plans to establish 
account managers at three levels: 

* National Account Managers are to collaborate with military services 
at the departmental level, for demand planning and problem resolution. 

* Customer Account Managers are to be the “single DLA face” to each
customer segment. These managers are to collaborate with executives at
the segment level to develop service-level agreements that outline
customer segment needs and to resolve issues at the segment level. 

* Customer Support Representatives are working-level DLA personnel who,
on a day-to-day basis, work with specific customers within a segment,
providing on-site assistance as appropriate. 

In addition, DLA plans to place its existing customer contact points, 
such as call centers and Emergency Supply Operation Centers, under the
control of account managers instead of the supply centers. 

DLA Improvement Efforts Do Not Include New Approaches to Obtain 
Customer Feedback: 

Although the Customer Relationship Management initiative is conceptually
sound, the program’s implementation actions do not completely address
the limitations we identified in its current practices. For example, 
the new strategy does not lay out milestones for implementing the 
program or specific improvements on how DLA solicits detailed feedback 
from its individual customers on their perceptions of service and the 
specific problems they are experiencing. The strategy also does not 
include a process for developing actions in response to issues that 
customers have identified and involving customers in that process. 
Furthermore, even though the plans include making account managers 
responsible for collecting customer feedback and exploring the idea of 
using Web-based tools to obtain customer feedback, they do not lay out 
specific tools or processes to accomplish this. 

To further illustrate, under the new Customer Relationship Management
plan, an account manager would be created with responsibility for all
customers within the U.S. Army Forces Command, which represents the
Army’s deployable forces segment. (See table 1.) This manager would
work with the Army’s customer representatives to identify customers’
needs at the Forces Command level and reach formal agreements on
service. However, there is no revised set of tools in the plan for 
collecting detailed feedback on an ongoing basis from the individual 
customer organizations representing the more than 6,600 DODAACs 
(address codes that represent mailboxes, locations, or people) in the 
Forces Command. 

Furthermore, the improvement initiatives do not provide for actions to
link military service customer DODAACs to specific accountable 
organizations. Under the Customer Relationship Management program,
DLA has developed a customer profile database that links DODAACs to
major military commands, such as the U.S. Army Forces Command. It also
plans to link each DODAAC to a business segment through this database
sometime in the future. However, as noted previously, the major command
and business segment levels comprise numerous DODAACs. Interaction
with customers to get detailed feedback on their level of satisfaction
requires better identification of customer organizations beyond the data
currently associated with a DODAAC. 

Best Practice Organizations Use Multiple Approaches: 

Studies examining best practices in the area of customer service have
found that leading organizations use multiple approaches to listen to 
their customers’ concerns. [Footnote 14] In particular, a 2001 Mid-
American Journal of Business study pointed out that best practice 
companies [Footnote 15] use multiple tools to gather these data rather 
than relying on a single method such as a customer survey, which might 
be too narrow in scope and limited in its application to fully capture 
customers’ concerns. [Footnote 16] The 2001 Mid-American Journal study 
and others concluded that the best approach for obtaining customer 
feedback is to use a broad measurement system with more than one 
listening tool to capture customers’ input from many different 
perspectives. 

Using different tools alone is not enough to effectively obtain customer
feedback. Centrally linking the feedback obtained is also important. 
Best practices research shows that information obtained through various
methods needs to be integrated in order to gain a more complete
understanding of customers. Thus, by linking all the various feedback
tools in a standard and consistent manner, the organization would have
better diagnostic information to guide improvement efforts. 

On the basis of our discussions with private sector experts and our
reviews of literature on customer service best practices, leading
organizations such as AT&T WorldNet Services, U.S. West, and Eastman
Chemical combine quantitative and qualitative listening tools to obtain
customer feedback and then centrally integrate the data in one location.
Quantitative tools include such methods as customer satisfaction surveys
and customer complaints, which can provide measurable data for use in
performance scorecards. Qualitative tools include focus groups, 
personal interviews, and observation and are used by organizations to 
provide a more in-depth understanding of their customers. According to 
the research, not all tools are appropriate for all organizations, and 
the research points out that careful selection is therefore important. 
Examples of “listening” tools being used by the best practices 
organizations we identified through our reviews of best practice 
studies follow: 

* Customer satisfaction surveys. Research shows that most major
organizations use listening tools such as relational and critical 
incident surveys [Footnote 17] to periodically capture customers’ 
overall perceptions about their organization and to measure 
satisfaction with specific transactions soon after they occur. These 
surveys can be administered through the mail, such as with DLA’s 
quarterly satisfaction survey; by telephone; in person; or 
electronically via the Internet. However, feedback from mail and 
electronic-based surveys can be more limited than that obtained through 
other methods because there is no opportunity to probe the respondent 
for better, more-detailed information. AT&T WorldNet Services, U.S. 
West, Eastman Chemical, and Hewlett-Packard are among the leading 
organizations that are turning to critical incident surveys in 
conjunction with other tools to learn more about customers’ perceptions.
Critical incident surveys are becoming more popular in the private 
sector because they provide information related to specific processes, 
which can be used to make specific improvements. 

* Customer complaints. Gathering complaint data is a standard practice 
for most companies. All aspects of the customer complaint process are
measured and tracked through this mechanism. Information collected and
analyzed from this approach includes the nature of the complaint, speed 
of resolution, and customer satisfaction with the resolution. Eastman
Chemical, for example, uses customer complaint data in conjunction with
a survey tool to obtain customer feedback. It organizes the complaint 
data along the same attributes as the survey data. 

* Benchmark surveys. Benchmark surveys gather perceptions of 
performance from the entire market. These surveys usually gather 
customer perceptions of performance about top competitors in an 
industry. This allows the company to examine its customer-perceived
strengths and weaknesses in the overall marketplace. Best practices 
companies, such as Sun Microsystems, use this information primarily in
the strategic planning process to identify their competitive advantage 
in the marketplace and to identify opportunities and shortfalls in the
industry. While continuous improvement may be a result of this listening
tool, the real value, according to the research in this area, comes from
breakthrough thinking to gain a sustainable advantage. 

* Won-lost and why surveys. “Lost” customers—those who do not replace
orders with a company—can be an excellent source of valuable 
information. Some companies, such as Eastman Chemical, employ “won-lost
and why” surveys to measure actual customer behavior and the rationale 
behind the behavior. This survey is utilized on a current basis, being 
administered to customers soon after they are “won” or “lost” (i.e., 
decide to drop a company). For example, if a customer is won or lost, 
the company then probes the customer as to why its business was won or 
lost. For companies with a large number of customers, this tool may be
implemented in a survey. 

* Focus groups. Organizations use focus groups to get better information
from customers than survey results provide. In these groups, customers
are probed about why they answered survey questions the way they did.
DLA has used focus groups to get detailed feedback on a single topic, 
but as noted previously, the number of individuals making up the focus 
groups was too small to draw agency-wide conclusions. AT&T Universal 
Card Services (now part of Citigroup) conducts multiple focus groups 
per year to discuss a wide range of topics. In these forums, both 
satisfied and dissatisfied customers discuss the company’s service, 
products, and processes. 

* Customer interviews. Conducting interviews with customers can provide 
a way to get very detailed information about their specific needs and
problems. Like focus groups, this tool is used by leading customer 
service organizations to probe survey respondents as to why they 
answered survey questions a certain way. U.S. West identifies 
dissatisfied customers from its surveys and follows up with them to 
determine what problems they are having and how they can be fixed. 

* Customer observation. In performing observations, organizations send
teams to visit customers where they observe how those customers interact
on a daily basis with the organization. This tool complements verbal
information obtained through customer interviews and focus groups in
that it provides confirmation to and a deeper understanding of that
information. 

* Management listening. Using this tool, managers listen in on actual
customer calls to the organization to learn first-hand about what
customers are experiencing. In an example of this technique, one best 
practice company encourages all of its managers, including the chief
executive officer, to listen to customer calls. 

* Customer service representatives. Collecting information from those
employees who are in continuous direct contact with customers provides
valuable information to best practice organizations. Often, these 
representatives are among the first to recognize customer problems. As
mentioned previously, DLA uses customer support representatives to 
obtain feedback. However, according to DLA officials, it does not 
currently have enough representatives assigned to its customers, and the
representatives generally are not proactive in obtaining customer 
feedback. Furthermore, while DLA’s representatives provide headquarters
with monthly written reports on customer support, best practice 
organizations have taken this a step further by using electronic 
feedback mechanisms. Research shows that best practice organizations 
have their customer service representatives gather ideas, perceptions, 
and opinions from customers and report them electronically through a 
corporate intranet system. These data are then coded and distributed 
throughout the organization, thereby centrally integrating the feedback 
information. 

Figure 3 shows an example of how multiple approaches can be linked, as
illustrated by AT&T Universal Card Services’ use of a “Customer 
Listening Post” team. 

Figure 3: AT&T Customer Feedback and Listening Strategies: 

[See PDF for image] 

This figure is an illustration of AT&T Customer Feedback and Listening 
Strategies as follows: 

Scenario One: 

Customer Feedback and Listening Strategies: 

Level one: Customer Impacting Business Processes: 
Level two: Solicit Feedback - Strategies; 
Level three: 
* Customer Expectation Research; 
* Performance Research; 
Level four: Organizational Improvement Activities: Modification of 
Existing Policies/Procedures and Standards; 
Level five: Customer Listening Post Team; 
Level six: Short term/Long term. 

Level four feeds information back to level one to allow the process to 
cycle. 

Scenario Two: 

Customer Feedback and Listening Strategies: 

Level one: Customer Impacting Business Processes: 
Level two: Operations Feedback - Strategies; 
Level three: 
* Direct Customer Feedback; 
* Process Management; 
Level four: Organizational Improvement Activities: Modification of 
Existing Policies/Procedures and Standards; 
Level five: Customer Listening Post Team; 
Level six: Short term/Long term. 

Level four feeds information back to level one to allow the process to 
cycle. 

Scenario one and two occur simultaneously. 

Note: AT&T Universal Card Services integrates methods such as monthly 
satisfaction surveys, telephone surveys/interviews, data mining (from 
customer commendations, letters, and phone calls), annual focus groups, 
and monthly management’s listening to customer calls. 

Source: Best Practices, LLC. 

[End of figure] 

Conclusions: 

While high-quality service to its customers is an overall goal, DLA 
lacks the information necessary to systematically assess the quality of 
service it is providing its customers with. Indications are that 
customers, while satisfied in some areas, are dissatisfied in others. 
The failure to address areas of dissatisfaction means opportunities to 
improve supply readiness are being missed. DLA is in the process of 
developing a program to improve its customer service relationships, but 
it currently does not have in place an effective mechanism that 
systematically gathers and integrates information on customer service 
views so that solutions can be identified to better meet their needs. 
The agency’s current practices do not always surface these concerns, or 
more importantly, provide information on why they exist or how they can 
be corrected. To its credit, DLA is undertaking a number of initiatives 
to improve the effectiveness of its customer relationship improvement 
efforts. However, these initiatives do not completely address the 
limitations of its current approaches for obtaining customer feedback 
because DLA (1) has not yet fully determined who its customers are or 
how best to serve their needs; (2) has not established the means to 
determine the underlying causes for customer dissatisfaction in order 
to fully achieve its strategic goal of providing customers with the 
most efficient and effective worldwide logistics support; and (3) lacks 
a centralized, customer-driven integrated framework in which to solicit
feedback from its customers. Also, customer mail-out surveys are 
insufficient for identifying the causes of customer dissatisfaction. 
Finally, DLA is not yet making full use of best practice techniques, as 
discussed in this report, to identify and address customers’ concerns. 

Recommendations for Executive Action: 

To improve DLA’s ability to determine its customers’ needs, identify
solutions for better meeting those needs, improve the supply readiness 
of military units, and improve the efficiency and effectiveness of depot
maintenance repair activities, we recommend that the Secretary of 
Defense direct the Under Secretary of Defense for Acquisition, 
Technology, and Logistics to require the Director of DLA, as part of the
agency’s customer relationship improvement efforts, to take the 
following actions: 

* Develop a comprehensive plan for obtaining customer feedback that
includes but is not limited to the following actions: 
- Work with the military services to arrive at a mutually agreed
determination of the military organizations that function as DLA
“customers.” In doing so, both DLA and the services should identify
officials accountable for providing and receiving customer feedback. 
- Develop a customer feedback program that uses a variety of approaches 
such as those depicted in the best practices research discussed in this 
report. In developing this program, pilot tests could be used to 
determine which approaches meet agency and customer needs. 
- Establish milestones for implementing the customer feedback program 
and for identifying the office accountable for its implementation. 
- Integrate all customer feedback into an overall assessment to provide
managers with a better understanding of customers’ perceptions and 
concerns. 
- Establish a process for developing actions in response to issues that
are identified from the customer feedback program and involve customers 
in that process. 
- Establish processes for providing customers with information on 
actions that are being taken to address customer feedback issues. 

* Improve the usefulness of its customer survey instruments by 
identifying ways to improve customer response rates, such as the use of 
effective follow-up procedures. 

* Clarify guidance for customer support representatives to ensure that 
they are responsible for routinely contacting customers to obtain 
customer feedback. 

We also recommend that the Secretary of Defense direct the Secretaries 
of the Army, Navy, and Air Force to identify specific organizations 
that will be responsible for working with DLA in establishing a 
mutually agreed determination of those activities, organizations, and 
individuals that function as DLA “customers” and for working with DLA 
as it implements its customer feedback program. 

Agency Comments and Our Evaluation: 

The Department of Defense provided written comments on a draft of this
report, which are reprinted in their entirety in appendix II. DOD 
generally concurred with our recommendations and agreed that DLA needs 
to increase its focus on customer satisfaction. The department also 
noted that DLA is taking or is planning to take a number of actions to 
respond to our recommendations. For example, under DLA’s Customer 
Relationship Management program, DLA National Account Managers are to 
identify customer organizations in concert with their military service 
negotiating partners. In addition, DOD intends to use its Defense 
Logistics Executive Board as a forum to obtain input from each of the 
services on the specific organizations that will be responsible for 
working with DLA on customer feedback issues. Furthermore, DLA intends 
to better integrate customer feedback into an overall assessment and to 
improve its processes for providing customers with information on 
actions that are being taken to address customers’ issues. 

DOD did not agree with our recommended action that DLA develop a 
customer feedback program that uses a variety of approaches, such as
those depicted in the best practices research discussed in this report. 
DOD stated that DLA’s use of feedback mechanisms should not be dictated 
by the best practices research we discussed. It further stated that DLA 
should continue to have the latitude to use its customer satisfaction 
measurement resources in the most efficient manner. Our discussion of 
best practice approaches was only intended to illustrate various 
techniques that some best practices organizations use to improve the 
ways they collect and analyze customer feedback. It was not our intent 
to prescribe specific approaches that DLA should use. Rather, we 
included examples of some of the approaches to best illustrate the 
concept of using multiple and integrated customer feedback approaches 
to better listen to customers’ opinions and concerns. We continue to 
believe that DLA’s customer feedback program could benefit from 
studying best practice organizations, such as those discussed in this 
report as well as others, to identify additional feedback approaches 
that could be pilot-tested and implemented to help strengthen its 
current customer feedback efforts. 

We are sending copies of this report to the Secretary of Defense; the
Secretary of the Army; the Secretary of the Navy; the Secretary of the 
Air Force; the Commandant of the Marine Corps; the Director, Defense
Logistics Agency; the Director, Office of Management and Budget; and
other interested congressional committees and parties. We will also make
copies available to others upon request. In addition, the report will be
available at no charge on the GAO Web site at [hyperlink, 
http://www.gao.gov]. 

Please contact me on (202) 512-4412 if you or your staff have any 
questions concerning this report. Major contributors to this report are
included in appendix III. 

Signed by: 

Charles I. Patton, Jr. 
Director, Defense Capabilities and Management: 

[End of section] 

Appendix I: Scope and Methodology: 

To determine how customers perceived the quality of service they 
received, we examined customer satisfaction studies and surveys such as
the Defense Logistics Agency’s (DLA) fiscal year 2000 and fiscal year 
2001 quarterly satisfaction surveys and the Joint Staff Combat Support 
Agency Review Team’s 1998 and 2001 assessments. In addition, we 
performed a case study analysis using a judgmentally selected sample of 
DLA customers that included the use of structured interviews to identify
customers’ perceptions and levels of satisfaction with DLA service. The
details of our customer selection process, interview techniques, and
sampling methodology follow: 

* We initially selected customers using DLA-provided databases of its 
“top” military customers, which DLA primarily based on sales volume. DLA
identified customers by Department of Defense Activity Address Codes
(DODAACs) or military installation. We compiled the DLA information
into a single database that included over 800 customer records 
accounting for about $5.6 billion of DLA’s total $7.8 billion nonfuel 
supply sales (about 72 percent) to the military services for fiscal 
year 1999, the most recent available data at the time of our review. 

* We judgmentally selected customers from the database to maximize our
coverage of the following significant variables: dollar sales, 
geographic location, DLA-defined customer type (i.e., deployed and 
deployable forces, industrial organizations, training activities, and 
the “other” segment), commodity type, and military service branch. We 
did not validate the accuracy of the DLA sales data, since the data’s 
purpose was to provide us with general customer sales activity. 

* Because the DLA-provided customer DODAAC and installation data did
not provide us with sufficient information about specific customer
organizations and related points of contact, we held discussions with 
DLA and military service officials to further define customers and 
subsequently visited those customer organizations and activities. 

* We conducted over 50 structured interviews with customers at more than
20 selected activities. We designed the interview questions on the 
basis of aspects of DLA’s supply process: submitting requisitions, 
following up on the status of open requisitions, contacting DLA for 
customer service, and receiving supplies. We also discussed other 
factors related to DLA support, such as the availability, price, and 
quality of DLA-provided supply items. Some customers did not express an 
opinion on the overall quality of customer service. 

* Our initial sample of DLA customers included customers from more than
20 locations throughout the continental United States and overseas,
covering multiple customer types within each military service. However,
because of the September 11, 2001, terrorist attacks on the World Trade 
Center in New York, and the Pentagon in Washington, D.C., we did not 
complete our planned visits. As a result, we limited our visits to eight
military service customer locations within the continental United 
States, as shown in figure 4. Our selection of customers included all 
four military services and each of the DLA customer types except for 
deployed forces. 

* Because we did not draw a statistical sample and we limited our 
selection of customers, the results of our work cannot be projected to 
DLA as a whole. However, DLA surveys, Combat Support Agency Review Team
assessments, and comments from DLA officials suggest that many of the
issues we raise are systemic problems. 

Figure 4: DLA Customer Locations Visited by GAO: 

[See PDF for image] 

This figure is a map of the continental United States, depicting the 
following sites visited by GAO: 

Marine Corps Base, Quantico, VA: 
Langley Air Force Base, VA: 
U.S. Atlantic Fleet/Fleet and Industrial Supply Center, Norfolk, VA: 
Fort Bragg, NC: 
Camp Lejeune/Marine Corps Air Station, New River, NC: 
Marine Corps Logistics Base, Albany; GA: 
Corpus Christi Army Depot, TX: 
Tinker Air Force Base/Oklahoma City Air Logistics Center, OK. 

[End of figure] 

To determine how useful the agency’s approaches are for obtaining
customer service feedback, we met with DLA headquarters officials to
discuss current processes and planned initiatives for measuring customer
service and obtaining feedback. We also discussed with DLA customers,
feedback mechanisms such as the use of DLA customer support 
representatives and quarterly surveys. We reviewed relevant reports,
briefing documents, and other key information related to the agency’s
processes and mechanisms for soliciting customer feedback. Additionally,
we examined the agency’s customer feedback survey techniques and
methods, such as the use of quarterly mail-out surveys and focus 
groups. 

Furthermore, we conducted an extensive literature search of best 
practice organizations to determine popular techniques for collecting 
customer feedback, and their advantages and disadvantages. 

To determine whether there are opportunities to enhance DLA’s 
initiatives to improve customer service, we performed a comparative 
analysis between DLA’s current practices and planned initiatives, and 
best practices that we identified through extensive literature 
searches. We reviewed related DLA planning documents and met with 
agency officials to discuss the agency’s plans. Through our literature 
search, we identified relevant research performed in the area of best 
practices in customer satisfaction. We reviewed a number of pertinent 
studies and held discussions with customer satisfaction experts from 
industry and academia to identify methods and techniques used in 
leading organizations to obtain meaningful feedback from their 
customers. 

We performed our work from March 2001 to June 2002 in accordance with
generally accepted government auditing standards. 

[End of section] 

Appendix II: Comments from the Department of Defense: 

Deputy Under Secretary Of Defense For Logistics And Materiel Readiness: 
3500 Defense Pentagon: 
Washington, DC 20301-3500: 

August 12, 2002: 

Mr. Charles I. Patton, Jr. 
Director, Defense Capabilities and Management: 
U.S. General Accounting Office: 
Washington, D.C. 20548: 

Dear Mr. Patton: 

This is the Department of Defense (DoD) response to the General 
Accounting Office (GAO) draft report, GAO-02-776, "Defense Logistics: 
Improving Customer Feedback Program Could Enhance DLA's Delivery of 
Services," dated July 8, 2002 (GAO Code 350159). 

The DoD agrees that DLA needs to increase its focus on customer 
satisfaction, and generally concurs with the draft report's specific 
recommendations. Detailed comments on the GAO recommendations are 
provided in the attachment. The DoD appreciates the opportunity to 
comment on the draft report. 

Sincerely, 

Signed by: 

Allen W. Beckett: 
Principal Assistant: 

Attachment: As stated: 

GAO Draft Report Dated July 8, 2002: 
(GAO CODE 350159): 
"Defense Logistics: Improving Customer Feedback Program Could Enhance 
Dla's Delivery Of Services" 

Department Of Defense Comments To The GAO Recommendations: 

Recommendation 1: To improve DLA's ability to determine its customers' 
needs, identify solutions for better meeting those needs, improve the 
supply readiness of military units, and improve the efficiency and 
effectiveness of depot maintenance repair activities, the GAO 
recommended that the Secretary of Defense direct the Under Secretary of 
Defense for Acquisition, Technology, and Logistics to require the 
Director of DLA, as part of the agency's customer relationship 
improvement efforts, to take the following actions: 

* Develop a comprehensive plan for obtaining customer feedback that 
includes but is not limited to the following actions; 
- Work with the military services to arrive at a mutually agreed 
determination of the military organizations that function as DLA 
"customers." In doing so, both DLA and the services should identify 
officials accountable for providing and receiving customer feedback. 
- Develop a customer feedback program that uses a variety of approaches 
such as those depicted in the best practices research discussed in this 
report. In developing this program, pilot tests could be used to 
determine which approaches meet agency and customer needs. 
- Establish milestones for implementing the customer feedback program 
and for identifying the office accountable for its implementation. 
- Integrate all customer feedback into an overall assessment to provide 
managers with a better understanding of customer perceptions and 
concerns. 
- Establish a process for developing actions in response to issues that 
are identified from the customer feedback program and involve customers 
in that process. 
- Establish process for providing customers with information on actions 
that are being taken to address customer feedback issues. 

* Improve the usefulness of its customer survey instruments identifying 
ways to improve customer response rates, such as the use of effective 
follow-up procedures. 

* Clarify guidance for customer support representatives to ensure that 
they are responsible for routinely contacting customers to obtain 
customer feedback. (pp. 32-33/GAO Draft Report) 

DOD Response: Generally concur. DoD agrees that DLA should work with 
the Military Services to identify DLA customers. The National Account 
Managers established under DLA's Customer Relationship Management 
program will identify the customer organizations in concert with their 
negotiating partners. DoD does not concur with the GAO recommendation 
that DLA's use of feedback mechanisms should be dictated by the best 
practices research discussed in the report. DLA should continue to have 
the latitude to uses its customer satisfaction measurement resources in 
the most efficient manner. DoD concurs that DLA should establish 
milestones for implementing the customer feedback program and for 
identifying the office accountable for its implementation. DLA has 
established such milestones in its Balanced Scorecard Customer Quadrant 
and has established the DLA Readiness and Customer Support office as 
the accountable office. DoD concurs that DLA should strive to integrate 
customer feedback into an overall assessment. DLA does this to the 
extent possible in the DLA Readiness and Customer Support office. Once 
fully implemented, the Customer Relationship Management program will 
provide a comprehensive integration capability. DoD concurs that DLA 
should establish a process for responding to issues that are identified 
through the customer feedback program. DLA currently reacts to customer 
feedback at both the field activity level and the Agency level. As DLA 
proceeds with the Customer Relationship Management program, these 
processes will be reengineered. DoD concurs that DLA should establish 
processes for providing customers with information on actions that are 
being taken to address issues identified by customers. DLA currently 
involves and informs customers, but will seek ways to improve the 
process as Balanced Scorecard implementation progresses. DoD agrees 
that DLA should clarify guidance for customer support representatives 
to ensure that they are responsible for routinely contacting customers 
to obtain customer feedback. 

Recommendation 2: The GAO recommended that the Secretary of Defense 
direct the Secretaries of the Army, Navy, and Air Force to identify 
specific organizations that will be responsible for working with DLA in 
establishing a mutually agreed determination of those activities, 
organizations, and individuals that function as DLA "customers" and for 
working with DLA as it implements its customer feedback program. (p. 
33/GAO Draft Report) 

DOD Response: Concur. The Defense Logistics Executive Board will be 
used as the forum to obtain Military Department input on the specific 
organizations that will be responsible for working with DLA on customer 
feedback issues. 

[End of section] 

Appendix III: GAO Contact and Staff Acknowledgements: 

GAO Contact: 

Kenneth R. Knouse, Jr. (202) 512-9280: 

Acknowledgements: 

Elizabeth G. Mead, Cary B. Russell, David R. Warren, Jeffrey A. Kans, 
Jack Kriethe, David Schmitt, Patricia Albritton, Brian G. Hackett, 
Latrealle Lee, and Stanley J. Kostyla also made significant 
contributions to this report. 

[End of section] 

Footnotes: 

[1] P.L. 106-398, sec. 917. 

[2] Since the early 1990s, DLA has been striving to better define and 
refine its understanding of “customer.” Currently, the agency defines 
its military customers, or war fighters, as those who purchase items, 
and directly cause products to be bought or not bought, and the 
commanders-in-chief of the military services. For this report, we did 
not include DLA’s interaction with its federal civilian customers. 

[3] DLA’s four supply centers are (1) Defense Supply Center, Columbus, 
Ohio, which is responsible for land, maritime and missile support; (2) 
Defense Energy Support Center, Fort Belvoir, Va., the lead center for 
comprehensive energy solutions, such as contract support and the 
management of petroleum-based fuels; (3) Defense Supply Center, 
Richmond, Va., which is responsible for air, aviation, and space 
support; and (4) Defense Supply Center, Philadelphia, Pa., the lead 
center for troop support items, such as food, clothing, and medical 
supplies. 

[4] DLA performs five major business functions: distributing materiel 
ordered from its inventory; purchasing fuels for DOD and the U.S. 
government; storing strategic materiel; marketing surplus DOD materiel 
for reuse, reutilization, or disposal; and providing numerous 
information services, such as item cataloging, for DOD and the U.S. and 
selected foreign governments. 

[5] The Balanced Scorecard, introduced by Professor Robert Kaplan and 
Dr. David Norton in 1992, is a strategic management system for 
describing, implementing, and managing strategy at all levels of an 
organization by linking objectives, initiatives, and measures to an
organization’s strategic plan. 

[6] Under 10 U.S.C. 193, the Joint Staff conducts a biennial Combat 
Support Agency Review, including a review of DLA. The January 2001 
review of DLA surveyed the unified commands and Joint Staff directors 
with responsibility to the Commander, Joint Chiefs of Staff. The review 
focused on services that DLA provides the unified commands with. 

[7] Although customers were satisfied with DLA’s prime vendor program 
in these instances, in recent years, the DOD Office of Inspector 
General reported that the program has failed to demonstrate an 
effective shift to commercial, industrial-base resources as an 
integrated logistics solution or provide the best value for DLA 
customers. As a result, the prime vendor program did not reduce total 
logistics costs, improve financial accountability, streamline defense 
infrastructure, or add value to the defense supply system. 8 DLA places 
customer support representatives at selected locations such as those 
with high business volume or readiness needs to monitor the agency’s 
overall success of its relations with its customers. The 
representatives are to provide a corporate face to particular customer 
sites. 

[8] DLA places customer support representatives at selected locations 
such as those with high business volume or readiness needs to monitor 
the agency’s overall success of its relations with its customers. The 
representatives are to provide a corporate face to particular customer 
sites. 

[9] See Military Aircraft: Services Need Strategies to Reduce 
Cannibalizations, GAO-02-86 (Washington, D.C.: Nov. 21, 2001). 

[10] Often, when items are not immediately available, customers can 
check excess property listings provided by DLA’s Defense Reutilization 
and Marketing Service to see if the needed parts are available 
elsewhere. 

[11] In an effort to reduce warehousing costs, DOD decided in 1989 to 
consolidate military service and DLA warehousing functions. This 
resulted in the collocation of both militaryservice-owned and DLA-owned 
parts in the same warehouse, referred to as a Distribution Depot. 

[12] A DODAAC is a six-position numeric code that uniquely identifies a 
unit, activity, or organization that has the authority to requisition 
and/or receive materiel. 

[13] See J. Anton and D. Perkins, Listening to the Voice of the 
Customer, 16 Steps to a Successful Customer Satisfaction Measurement 
Program, The Customer Service Group (New York City: 1997). 

[14] See M.S. Garver and R.L. Cook, “Best Practice Customer Value and 
Satisfaction Cultures,” Mid-American Journal of Business, vol. 16, no. 
1 (2001); M.S. Garver, “Modeling Best Practices for Government 
Agencies: Implementing Customer Satisfaction Programs” (Jan. 28, 2002); 
Best Practices, LLC, “Achieving World-Class Customer Service: An 
Integrated Approach” (copyright 1998-2001); Federal Benchmarking 
Consortium, Serving the American Public: Best Practices in Customer-
Driven Strategic Planning (Feb. 1997). 

[15] Best practice companies used in the research met at least three of 
the following criteria: “has won a quality award such as the Malcom 
Baldridge award; has been discussed as a best practices company in a 
respected publication; has presented a best practice at a customer 
value and satisfaction practitioner conference; is respected as a 
leading edge company; met the researchers’ best practices 
characteristics such as innovation, uniqueness, and high data 
utilization rates.” 

[16] See M.S. Garver, “Listening to Customers,” Mid-American Journal of 
Business, vol. 16, no. 2 (2001). 

[17] Relationship surveys are viewed as traditional customer 
satisfaction surveys and are administered to customers typically on a 
quarterly basis; transaction surveys are typically short in length and 
are used to measure the performance of a specific process. They are 
administered immediately following a certain type of service encounter, 
event, or interaction with the customer. 

[End of section] 

GAO’s Mission: 

The General Accounting Office, the investigative arm of Congress, 
exists to support Congress in meeting its constitutional 
responsibilities and to help improve the performance and accountability 
of the federal government for the American people. GAO examines the use 
of public funds; evaluates federal programs and policies; and provides 
analyses, recommendations, and other assistance to help Congress make 
informed oversight, policy, and funding decisions. GAO’s commitment to 
good government is reflected in its core values of accountability, 
integrity, and reliability. 

Obtaining Copies of GAO Reports and Testimony: 

The fastest and easiest way to obtain copies of GAO documents at no 
cost is through the Internet. GAO’s Web site [hyperlink, 
http://www.gao.gov] contains abstracts and fulltext files of current 
reports and testimony and an expanding archive of older products. The 
Web site features a search engine to help you locate documents using 
key words and phrases. You can print these documents in their entirety, 
including charts and other graphics. 

Each day, GAO issues a list of newly released reports, testimony, and 
correspondence. GAO posts this list, known as “Today’s Reports,” on its 
Web site daily. The list contains links to the full-text document 
files. To have GAO e-mail this list to you every afternoon, go to 
[hyperlink, http://www.gao.gov] and select “Subscribe to daily E-mail 
alert for newly released products” under the GAO Reports heading. 

Order by Mail or Phone: 

The first copy of each printed report is free. Additional copies are $2 
each. A check or money order should be made out to the Superintendent 
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or 
more copies mailed to a single address are discounted 25 percent. 

Orders should be sent to: 

U.S. General Accounting Office: 
441 G Street NW, Room LM: 
Washington, D.C. 20548: 

To order by Phone: 
Voice: (202) 512-6000: 
TDD: (202) 512-2537: 
Fax: (202) 512-6061: 

To Report Fraud, Waste, and Abuse in Federal Programs Contact: 

Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: 

E-mail: fraudnet@gao.gov: 

Automated answering system: (800) 424-5454 or (202) 512-7470: 

Public Affairs: 
Jeff Nelligan, managing director, NelliganJ@gao.gov: 
(202) 512-4800: 
U.S. General Accounting Office: 
441 G Street NW, Room 7149:
Washington, D.C. 20548: