Achieving World Class Maintenance Status
Measuring and assessing maintenance performance
By Paul D. Tomlingson
“World class” performance is the ultimate objective
of many maintenance organizations. It marks the organization as a leader in its
industry and sets it apart as the ultimate achiever. But, what is world class,
and how can it be achieved? World class maintenance could result when an organization
consistently produces reliable equipment by conducting an effective program using
accurate, timely, complete information. The challenge is to determine what constitutes
a world class maintenance organization and then derive a specific set of performance
standards that, if honestly met, universally identify the qualifying organization
as world class within its industry. Most maintenance organizations admit a need
to improve. With world class performance as a target, they should take steps to
assess their “as-is” performance status and determine what they must
improve to meet the target. Evaluation is the first step of improvement. An evaluation
establishes the current performance level by identifying those activities needing
improvement as well as those being performed well. The most important byproducts
of an effective evaluation will be the education of personnel about specific improvements
and their purpose, and obtaining their genuine commitment to helping achieve the
improvements. An effective evaluation must compare the demonstrated performance
of the subject organization against a set of standards that are consistent with
the type of industrial maintenance organization being assessed. The evaluation
procedure should be an established management practice that initially establishes
the organization’s “as is” performance level; then, at regular
intervals measures progress toward meeting the standards. Successful maintenance
is not a stand-alone activity. Maintenance planning, for example, cannot be evaluated
in isolation. It must be examined in the light of how well purchasing and warehousing
support material needs or how effectively the information system allows the planner
to manage planned jobs from inception to completion. If operations were to consistently
elect to try to meet elusive production targets ahead of making equipment available
for scheduled maintenance, poor maintenance performance (by operations default)
would be the unfortunate result. Maintenance would be equally unsuccessful if
their maintenance program were undocumented “folklore,” inadequately
communicated to other departments that want to help but can only guess how. Maintenance
would surely fail if the mine manager thought of maintenance as a necessary evil—a
cost reduction challenge to be quickly solved by an ill informed workforce reduction.
Who Evaluates?
We often think of consultants as evaluators. They can be neutral third parties
with experience in various types of operations. But, their evaluation may come
with a huge price tag and might cause prolonged disruption of mine operations.
Mine personnel might be spectators rather than participants in the evaluation
process, and the employees’ unique, pertinent and factual knowledge of actual
mine circumstances could be overlooked. Therefore, a mine should choose wisely
if they want consultants to evaluate them. And, there are evaluation techniques,
equally as effective as those provided by consultants, that are less disruptive
and costly and produce reliable results. For example: A cross-section of mine
personnel, rating maintenance against a series of performance standards that touch
on everything from mine manager- ship to production cooperation to staff department
support to preventive maintenance, planning, scheduling and the effective use
of information. If that cross-section were to consist of managers who watch all
departments interact, to staff departments that support maintenance, plus production
people who use maintenance services and maintenance themselves, it is possible
that a good picture of actual performance could result. Moreover, if each group
were represented by a vertical slice of their personnel, the results might be
even better. Suppose, for instance, that the maintenance group included the maintenance
manager, several supervisors, planners, a maintenance engineer and various craftsmen.
Consider what might result if the maintenance manager learned that a new procedure
was considered impractical by his supervisors, and the craftsmen for whom it was
intended never heard of it. The outcome of the combined ratings would confirm
that the maintenance manager should have conferred with his key personnel as future
procedures are developed. He would have learned a valuable lesson first hand and
be inclined to correct it—particularly if the mine manager has read the
same report.
The Value of Benchmarking and KPI’s
Is benchmarking a suitable evaluation technique for achieving maintenance improvement?
The idea of learning from other's good ideas or benefiting from their successful
experiences as well as avoiding their mistakes is a philosophy of long standing.
But make certain that what is benchmarked has value to your improvement needs.
Benchmarking is only a comparison of pertinent practices, not a contrast of performance.
Therefore, while helpful in the overall aspect of improvement, benchmarking alone
will not bring about the improvement necessary to achieve world class performance.
Rather, it might help to identify those organizations that have achieved it and
encourage the innovative adaptation of their best practices to the aspiring maintenance
organization. Furthermore, serious flaws exist in the prevailing notion that key
performance indices (KPI’s) are an effective performance measurement tool
and that actions based on their results can help a maintenance organization achieve
top level performance status. Key performance indices are useful in revealing
a trend toward realizing a visualized performance target. However, the field data
on which the indices are often based is usually developed and submitted by the
same people who are to be evaluated. Unless the monitoring organization can verify
the validity of the ingredient data, the resulting indices may be suspect. Agreement
on which performance indices best reveal actual circumstances can be difficult
and the logic for their inclusion questioned. For example, one popular index is
the amount of work that a maintenance organization plans. Inconsistencies abound;
preventive maintenance, for instance, is often considered to be planned and scheduled.
In reality, PM services were planned when the PM program was initially developed.
Thus, PM services are repetitively scheduled not planned on a week to week basis
as is other work embraced by this index such as overhauls or major component replacements.
For reasons such as this, the intended interpretation of indices must be established
with care. While a typical mine manager who looks at an array of performance indices
can observe relative scores, he often cannot direct a specific corrective action
as a result. By contrast, an evaluation technique built on a cross-section of
mine personnel rating a series of appropriate, pertinent performance standards
will identify exactly what is in need of improvement and establish the proper
corrective actions, by urgency and priority. To illustrate: Too little planning
often means that the PM inspection and testing program fails to find deficiencies
with sufficient lead-time to be able to plan the work rather than react to it.
If specific standards assess planning as inadequate and identify poorly executed
PM inspection and testing along with poor PM schedule compliance by operations,
the mine manager now has the right improvement target and a specific improvement
action. If this knowledge causes him to look in on a typical weekly operations
and maintenance scheduling meeting only, to find an indifferent planner talking
to empty operations chairs, he can focus on fixing the problem of too little planning,
permanently. By contrast, a single KPI simply telling him there is too little
planning only frustrates him. Performance indices are often limited to examining
only direct maintenance activities such as planning without assessing underlying
factors that influence the success of planning—quality of material support
provided by warehousing or purchasing, for example. From the mine manager’s
view, the contrast between looking at indices and reviewing the details of a well-conceived
self-evaluation is the difference between looking out the window to guess how
you are doing versus “management by walking around.”
What’s the Frequency?
Overly frequent evaluation of weekly indices soon loses its appeal—similar
to the work order procedure that requires reporting of delay causes for each job.
Such procedures often get off to an enthusiastic start but are ultimately abandoned.
Problems revealed in this week’s index are seldom able to be acted on in
the short interval. By contrast, self-evaluation applied at longer intervals (~6
months) is less intrusive and more anticipated. Random sampling, one of the principles
underlying a self-evaluation, is more realistic than KPI’s. It is welcomed
as an opportunity to see how the organization stacks up against a pertinent series
of performance standards. Thereafter, a periodic repeat of the evaluation acts
as a report card on improvement progress. This report card is also welcomed because
it represents the progress made by the same personnel who identified necessary
improvements needs at the start of the self-evaluation process. Most importantly,
self-evaluation allows dedicated, skilled craftsmen and others to participate
in the assessment process. They seize the opportunity to offer frank and objective
assessments knowing that otherwise they might never have been asked. When combined
with the responses of the balance of the cross-section of the mine population,
the result is a more reliable assessment with specific insights into the right
corrective actions. If an evaluation were to cover hundreds of performance standards,
all rated by a crosssection of knowledgeable, caring mine personnel, the results
could help propel maintenance toward positive corrective actions leading to the
real attainment of world-class status. Repetitions of the evaluation at thoughtful
intervals would act to measure interim improvement progress toward that goal.
A self-evaluation—assuming it contains the right standards—has considerable,
direct improvement potential. Personnel know the mine well. They are familiar
with people in other departments and how they must interact successfully. As they
rate the standards, they are likely to be frank and objective in identifying and
prioritizing actions or procedures that should be changed or improved. They know
that they are going to be directly affected by the potentially beneficial outcome
they visualize. They concurrently make a genuine commitment to help implement
changes they see as practical and necessary. Unlike spectators to third party
evaluations by consultants or employees who observe the irate mine manager’s
demand for more planning, self-evaluation participants can impact their own futures
directly and they know it.
Defining Evaluation Standards
An inevitable question is: What standards should apply, and who says they are
the right ones? Developing the right standards is a task that must precede any
evaluation effort that compares current performance against them. The standards
must be based on a well-conceived, fully documented, well-understood and effectively
executed maintenance program. A well–defined and effective maintenance program
spells out the interaction of all departments as they request or identify work,
classify it to determine the best reaction, plan selected work to ensure it is
accomplished efficiently, schedule the work to ensure it is performed at the best
time with the most effective use of resources. In addition, the maintenance program
specifies how work is assigned to personnel in a way that assures each person
has a full shift of bona–fide work. Then, as work is performed, the program
establishes work control procedures to ensure quality work, completed on time.
In addition, the program specifies how completed work is measured to ensure timely
completion, under budget with quality results. The maintenance program should
also prescribe a means of periodic evaluations to identify and prioritize improvement
needs. The maintenance program is the soul of the overall maintenance effort.
It defines the basics of what maintenance does, who does what, how they do it
and why. When the program is well-conceived, fully-documented and well-understood
across the operation it will be effectively executed and produce quality results.
By sharp contrast, a maintenance department uncertain of these basics cannot determine
how they will organize or even select the right information systemmuch less use
it effectively. The end result is a totally reactionary maintenance effort. A
quality maintenance program is fundamental to the development of performance standards
and to the ultimate success of maintenance. Few maintenance organizations have
such programs, fewer yet have documented them and very few have bothered to explain
how they do their work to even their own people. Customers in operations and supporting
departments guess at what is expected of them, fail to be of proper assistance
to maintenance and, in exasperation, usually ask, “what program?”
Mine managers often express the same frustration. For maintenance departments
without such programs, life is difficult. Developing standards is out of the question
and attempts to adopt advanced strategies like reliability centered maintenance
(RCM) or total productive maintenance (TPM) fail most of the time. Similarly,
implementation of modern techniques for improving equipment reliability may fail
as well. More, importantly, the fundamentals of maintenance management are never
mastered. A totally reactionary maintenance organization can never hope to achieve
top level performance. But, the organization that has a well-defined program can
establish the standards it must meet and, with the aid of an established self-evaluation
procedure, move toward and achieve world class status.
Paul D. Tomlingson is principal of management consulting firm Paul D.
Tomlingson Associates (www.tomlingson. com). He can be contacted by e-mail at
pdtmtc@sprynet.com