You are on page 1of 223

BottomLine05.

book Page iii Friday, August 19, 2005 9:54 AM

BOTTOM-LINE
AUTOMATION

2nd Edition

Peter G. Martin
BottomLine05.book Page iv Friday, August 19, 2005 9:54 AM

Notice
The information presented in this publication is for the general education of the reader.
Because neither the author nor the publisher have any control over the use of the information
by the reader, both the author and the publisher disclaim any and all liability of any kind aris-
ing out of such use. The reader is expected to exercise sound professional judgment in using
any of the information presented in a particular application.
Additionally, neither the author nor the publisher have investigated or considered the
affect of any patents on the ability of the reader to use any of the information in a particular
application. The reader is responsible for reviewing any possible patents that may affect any
particular use of the information presented.
Any references to commercial products in the work are cited as examples only. Neither the
author nor the publisher endorse any referenced commercial product. Any trademarks or
tradenames referenced belong to the respective owner of the mark or name. Neither the author
nor the publisher make any representation regarding the availability of any referenced com-
mercial product at any time. The manufacturers instructions on use of any commercial prod-
uct must be followed at all times, even if in conflict with the information in this publication.

Copyright 2006 ISAThe Instrumentation, Systems, and Automation Society

All rights reserved.

Printed in the United States of America.


10 9 8 7 6 5 4 3

ISBN 1-55617-962-6

No part of this work may be reproduced, stored in a retrieval system, or transmitted in any
form or by any means, electronic, mechanical, photocopying, recording or otherwise, without
the prior written permission of the publisher.

ISA
67 Alexander Drive
P.O. Box 12277
Research Triangle Park, NC 27709

Library of Congress Cataloging-in-Publication Data in process


BottomLine05.book Page v Friday, August 19, 2005 9:54 AM

Dedication
To Liz, Derek and Erin
BottomLine05.book Page ix Friday, August 19, 2005 9:54 AM

List of Figures
Figure 1.1 Trends for Surviving in a Global Economy .......................................4
Figure 1.2 Traditional Product Costing ............................................................11
Figure 1.3 Vollmann Triangle .......................................................................13
Figure 1.4 The Convergence: Survival in a Global Economy...........................15
Figure 2.1 The Technology Trend ....................................................................17
Figure 2.2 Technology Focus ............................................................................26
Figure 2.3 Computer Integrated Manufacturing ..............................................30
Figure 2.4 Operations Craftsmanship ..............................................................37
Figure 2.5 Transitions in Operations................................................................42
Figure 2.6 Lifecycle Cost Profile .......................................................................46
Figure 2.7 Lifecycle Economic Profile ..............................................................47
Figure 2.8 Lifecycle Economic Model...............................................................48
Figure 2.9 Lifecycle Economic Profile Results ..................................................49
Figure 2.10 Lifecycle Cost Breakdown................................................................49
Figure 3.1 The Quality Trend ...........................................................................53
Figure 3.2 The Evolution of Quality Improvement .........................................56
Figure 3.3 PDCA................................................................................................60
Figure 3.4 Juran Trilogy ....................................................................................62
Figure 3.5 Quality Indicators or Performance Measures ..................................72
Figure 3.6 Quality Indicators and Performance Measures Converge...............72
Figure 3.7 Lean Enterprise ................................................................................74
Figure 4.1 The Cost Accounting Trend ............................................................82
Figure 4.2 Cost Accounting and Manufacturing Performance ........................85
Figure 4.3 Traditional Product Costing ............................................................87
Figure 4.4 Cost Management Systems Activity-based Costing....................92
Figure 4.5 Vollmann Triangle .......................................................................95
Figure 4.6 Vollmann Decomposition ...........................................................97
Figure 4.7 Strategy ............................................................................................98
Figure 4.8 Action Planning...............................................................................98
Figure 4.9 The Automation Economic Gap....................................................101
Figure 4.10 Process Plant Decomposition ........................................................104
Figure 4.11 Plant Level Vollmann Decomposition ..........................................105
Figure 4.12 Dynamic Performance Measure Hierarchy....................................105
Figure 4.13 Performance Dashboard ................................................................106
Figure 4.14 The Performance Measure Hierarchy ............................................107
Figure 4.15 Current Plant Accounting Model ..................................................111
Figure 4.16 Production Model-based Accounting System ...............................112
Figure 4.17 Performance-based Plant Accounting Model ................................113
Figure 4.18 Model for Cost and Performance Measurement (Stage IV)...........114
Figure 5.1 The Convergence: Survival in a Global Economy.........................121
Figure 5.2 Improving Economic Performance ...............................................123
Figure 5.3 The Convergence ...........................................................................129
Figure 5.4 Lean Enterprise ..............................................................................129
Figure 6.1 Asset Availability and Asset Utilization Are Inverse
Functions as They Approach Their Maximum Ranges .................139
Figure 6.2 An Inverse Relationship Puts Maintenance Personnel and
Operators at Odds ..........................................................................140

B O T T O M - L I N E A U T O M A T I O N ix
BottomLine05.book Page x Friday, August 19, 2005 9:54 AM

L I S T O F F I G U R E S

Figure 6.3 Supporting Assets Improve the Efficiency and


Effectiveness of the Primary Assets............................................... 143
Figure 6.4 Overall Plant Breakdown .............................................................. 148
Figure 6.5 Operational Measures are Seldom Aligned with
Accounting Measures.................................................................... 156
Figure 6.6 Describes the Instantaneous Relationship of Business Value
from a Single Asset with Respect to EU and EA ........................... 159
Figure 6.7 Expanded Equation to Show Asset Value Over Longer Periods ... 160
Figure 6.8 Instantaneous Asset Set Value ...................................................... 164
Figure 6.9 A Simplified, Two-Dimensional View of the Asset Value Sector.. 165
Figure 6.10 To Calculate an Assets Total Value Over a Time Period,
Use the Integral Over That Period ................................................ 165
Figure 6.11 Aligning Metrics from the Plant Floor to the Business
Measures Is Critical ....................................................................... 167
Figure 6.12 Developing Initial Conditions from History ................................ 169
Figure 7.1 Solvent Recovery Overview .......................................................... 181
Figure 7.2 Incinerator Overview .................................................................... 182
Figure 7.3 Solvent Recovery Operator Dashboard......................................... 186
Figure 7.4 Incineration Operator Dashboard ................................................ 187

x B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page xi Friday, August 19, 2005 9:54 AM

List of Abbreviations
ABC activity-based costing
CIM computer integrated manufacturing
CMS cost management systems
CRT cathode ray tube
DCS distributed control system
DEC Digital Equipment Corporation
DPM dynamic performance measures
HP Hewlett Packard
IBM International Business Machines
I/O input/output
JIT just-in-time manufacturing
LCC lifecycle cost
LCE lifecycle economics
MIS management information system
NPV net present value
PDCA plant-do-check-act
PDP programmable data processor
PID proportional, integral, derivative
PLC programmable logic controller
QA quality assurance
RFP request for proposal
ROI return on investment
SPC statistical process control
SQC statistical quality control
TQM Total Quality Management
YEL years of expected life

B O T T O M - L I N E A U T O M A T I O N xi
BottomLine05.book Page xiii Friday, August 19, 2005 9:54 AM

ACKNOWLEDGEMENTS
I would like to express my sincere gratitude to a number of people
who significantly helped to make this book and the background research
a reality.
A number of Invensys Foxboro professionals contributed signifi-
cantly to this work, including, but not limited to: Janice Able, Dan
Antonellis, Russ Barr, Joe Barricella, Ron Beene, Barry Boyle, John Brkich,
Larry Brown, Daniel Cervantes, Bob Cook, Mike Cook, Debbie Con-
neally, Ricardo Cortejoso, Marty Culverhouse, Garry Cusick, Trevor Cus-
worth, Mark Davidson, Nancy Delhommer, Donna Distante, Rik Evans,
Stan Devries, John Eva, Carlos Fernandez, Joe Fillion, Kevin Fitzgerald,
Jay Galasso, Dan Gilmore, Steve Gollemme, Harpreet Gulati, Bruce Hend-
erson, Jim Hetzer, Neil Holden, Richard Howells, Liwei Huang, Fayyaz
Hussain, Graham Jester, Ken Johnson, Rob Hasselbaum, Gary Hodgson,
Jim Holt, Larry Hueni, Hesh Kagan, Randy Karg, Bill Ketelhut, Jeff
Kissling, Mike Kletti, John Krause, John Kuenzler, Krishnakumer Kumer,
Don Kylin, Larry Lablanc, Peter Lovelace, Bruce MacLeod, Leo McEvoy,
Rebecca Marshall-Howarth, Bob Meadower, Janice Miller, Paul Miller, Jim
Monahan, Ron Pariseau, Melanie Russell, George Sarney, Neil Schmidt,
Tracey Sledge, Jim Sommers, Jack Spencer, Dick Staun, Marilyn Tarallo,
Laurens van der Tang, Rick Whitmyre, Dave Williams, Lew Wright, Gene
Yon and Barry Young. I am indebted to this group for their support.
A number of people from various organizations throughout the pro-
cess industries also helped significantly in this effort including: Dave
Adler, Wilfrido Arroyo, Rich Baker, Jeffery Baumer, Malcolm Beaverstock,
George Bozant, Gene Bylinsky, Andy Chatha, Colin Christie, Dr. Robin
Cooper, Bob Cox, Tony DeHerrara, Tom Fisher, Tom Fiske, Steve Fur-
bacher, Jack Garrity, Jim Getchell, Jack Hickey, Dick Hill, Bill Huff, Mark
Hughes, Wendy Johnson, Mark Kluesmer, Lowell Koppel, Terry Landano,
David Lauer, Steve Leus, Rick Lissa, Owen Martin, Jose Louis Martinez,
Dick McAllister, Gerard McCarry, Dan Miklovic, Jim Montague, Philippe
Moro, Charles Morris, Joge Nader, John Nero, Aundra Nix, Clay Noble,
Larry Obrien, Doug Palmer, Sam Parino, Andy Peters, Ronald Root, Steve
Sarnecki, Don Schuette, Ron Simmons, John Snodgrass, Greg Stedronsky,
John Stolle, Len Sugarman, Donna Takeda, Tom Vollmann, Ray Walker,
John Wason, Jim Wolf, and Dave Woll.
I would also like to thank those at ISA who helped with the editing
and publishing of the book especially: Chip Lee, Susan Colwell, Matt
Lamoreaux, Shandra Botts, Greg Hale and Bob Burns.
I would also like to express my appreciation for the ongoing encour-
agement and support of my family. I thank my parents, Jack and Jean
Martin have been a constant source of encouragement, my brothers and
sisters and their families and Atwell and Nan Collins.
I especially thank my wife Liz and my children Derek and Erin for
their ongoing love and support.

B O T T O M - L I N E A U T O M A T I O N xiii
BottomLine05.book Page xv Friday, August 19, 2005 9:54 AM

FOREWORD

A Solution
Without a
Problem

BACKGROUND ECONOMIC
FORCES LEAD TO DYNAMIC
PERFORMANCE MEASURES
Automation systems have been used to help enable and opti-
mize manufacturing processes for decades. When the digital com-
puter was introduced as a viable tool for automation, process
engineers often seemed to focus more attention on the tool than
on the actual automation objectives. Suppliers of automation
technology were typically selected on the basis of the state-of-
the-art technological features they included in their systems
rather than the improvements their automation technology
could offer the manufacturing operation. Issues such as band-
width, pixel resolution, and color palette seemed to be given
greater weight than performance-enhancing software capabilities,
such as automatic continuous loop tuning. Returns on automa-
tion investments were rarely, if ever, even calculated.

It was in this environment that, in the late 1980s, The Foxboro


Company introduced its I/A Series system, which was designed to
provide as many state-of-the-art features to the marketplace as
was then possible. The Foxboro Company had invested signifi-
cant amounts of money in the design and development of the I/A
Series system and fully expected that it would help revolutionize

B O T T O M - L I N E A U T O M A T I O N xv
BottomLine05.book Page xvi Friday, August 19, 2005 9:54 AM

F O R E W O R D : A S O L U T I O N W I T H O U T A P R O B L E M

process manufacturing. Unfortunately, after a number of I/A


Series systems were installed and operating, an assessment
showed that these systems were being implemented in the exact
same way as traditional proprietary distributed control systems
(DCS). These early assessments showed that most of the advanced
features of the I/A Series system that had been critical during the
selection process were not even being utilized.

At this point, the company commissioned an applied research


program to determine how automation systems should be used to
effectively improve plant performance. The premise of this exer-
cise was that the engineers and technologists in process manufac-
turing operations had become too enamored of technology and
were driving automation suppliers like The Foxboro Company to
continually upgrade to new technological features with no con-
sideration for what manufacturing or operational problem they
would solve. Automation orders were being won or lost on fea-
tures that were seldom used. Clearly, this technology-focused
approach was good for neither manufacturers nor automation
suppliers.

The applied research program was structured as an iterative


interview process targeted at the management levels above the
technologists in process manufacturing operations. In particular,
plant managers, operations managers, production managers, as
well as vice presidents of manufacturing, operations, and engi-
neering were the focus of the interviews. Initially, four researchers
interviewed hundreds of operations managers, plant managers,
and corporate executives in over 700 different manufacturing
operations in North America, Latin America, Europe and Asia over
a four- to five-year period. The interviews focused on process
manufacturing operations in the chemical, refining, paper, min-
ing, food, pharmaceutical, water and wastewater, cement, clay
and glass industry segments. In essence, the Foxboro program was
structured to find real problems in manufacturing operations that
the new I/A Series technology could address.

The iterative interview approach involved selecting a small


number (approximately twenty) of managers for interviews. Once
these initial interviews were conducted, the research team

xvi B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page xvii Friday, August 19, 2005 9:54 AM

F O R E W O R D : A S O L U T I O N W I T H O U T A P R O B L E M

developed a presentation that summarized their results and


conclusions. This presentation was then used to communicate the
findings back to as many of the participants of the initial group as
possible as well as to an expanded group. Their feedback enabled
the researchers to fine-tune the results and conclusions and to
gather additional information. This process was continued
throughout the program with an ever-expanding group of
managers.

The initial interviews did not bear much fruit. The research
team was not really sure what it was searching for, and conse-
quently the managers, although very cooperative, tended to talk
in generalities. The primary information gathered in these early
interviews focused on the business and market forces driving the
process manufacturing industry segments. Interestingly, this first
round of interviews revealed that the driving forces the managers
identified did not vary much from segment to segment. All man-
agers expressed concerns over the globalization of markets and
resulting tough competitive economic environment that was just
starting to emerge. Shareholder value seemed to be becoming
more important than, or at least as important as, customer satis-
faction. The picture they painted was one in which survival
depended on significantly outperforming the competition eco-
nomically. Though, interestingly, many of the managers pre-
dicted that corporate consolidation would accelerate because of
globalization, it is doubtful any of them expected the level of con-
solidation through buy-outs and mergers that has occurred in the
past few years.

The subsequent rounds of interviews proved to be much more


valuable, and they have actually continued for over a decade now.
The turning point in the interview process was the research
teams acceptance of these driving forces of globalization, compe-
tition, and shareholder value. The team was then able to turn the
process around and ask the managers what they were doing to
address these forces within their companies to ensure that their
firms would be among the survivors. The results were fascinating.
Each of the managers described the programs they were investing

B O T T O M - L I N E A U T O M A T I O N xvii
BottomLine05.book Page xviii Friday, August 19, 2005 9:54 AM

F O R E W O R D : A S O L U T I O N W I T H O U T A P R O B L E M

in that they believed would help them survive, and, once again,
there was an amazing consistency in the managers answers.

The research team categorized the programs the managers


were investing in into four areas. The first was technology. In
spite of most of the managers negative opinions about the effi-
cacy of automation technology, all believed that if they fell
behind in technology they would be at a competitive disadvan-
tage. Many of them stated that they were uncomfortable about
continuing to invest in such technology because of the lack of
measurable benefits but that they were afraid not to.

The second investment category the managers identified


involved quality. During the late 1980s, the quality movement hit
North America and Europe. Japan had been utilizing statistical
control and quality team approaches since the early 1950s, when
W. Edwards Deming and Joseph Juran brought the quality move-
ment to Japan. By the 1980s, it was clear that Japan had become a
world manufacturing power, and many Japanese companies
appeared to be outperforming traditional manufacturers both
financially and in terms of quality. Western manufacturers were
scrambling for ways to catch up quickly.

The third investment category was empowerment of personnel


through automation technology. Several managers alluded to
this, although until the past few years evidence of true employee
empowerment has been scarce. Since this empowerment was
enabled by the technology, this category was coupled with the
first category as the program evolved.

The final investment area focused on major changes that


appeared to be taking place in the financial reporting structures of
manufacturing operations. These changes were initially identified
under the banners of cost management systems (CMS) or
activity-based costing (ABC). At first, the impact of this cate-
gory on the research team was fairly low since the team came
from The Foxboro Company, a technology-based automation
supplier. But its importance became more and more evident as
the interviews progressed.

xviii B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page xix Friday, August 19, 2005 9:54 AM

F O R E W O R D : A S O L U T I O N W I T H O U T A P R O B L E M

The managers also expressed considerable concern over the


disconnect they were detecting between the corporate strategies
they were trying to execute and the behaviors and actions they
witnessed in day-to-day manufacturing operations. Several man-
agers were openly frustrated that they had invested millions of
dollars in automation only to find that their manufacturing
plants were not operating any better than they had decades ear-
lier with much lower levels of automation. Interestingly, later
interviews with the plant engineering and operations personnel
of many of these plants revealed that they thought their plants
were running much better as a result of automation technology
and their own efforts.

As the interview process progressed it became apparent that


better was defined very differently by management and the plant
engineering and operations personnel. The interview team
returned to a number of these operations merely to try to discern
what the differences were. Management typically defined better in
terms of reduced cost, increased yields, increased production and
the like, while the plant operations and engineering personnel
used phrases such as ease of use, openness, and ease of engi-
neering. Clearly, there was a major gap in perspective between
the two groups.

Once the gap was identified, the interview process focused on


how the more effective use of automation might eliminate it.
During the 1980s, much work had been done in the areas of per-
formance measurement systems, cost management systems, and
activity-based costing. All of this work shared the goal of getting
at the issues behind the gap between management and engineer-
ing/operations, but none of it had found a solution.

The Foxboro research team began working with Dr. Thomas


Vollmann of Boston University to try to gain a better understand-
ing of the knowledge gap issues. Dr. Vollmann had a solid under-
standing of these knowledge gap issues as they related to discrete
manufacturing operations, and he helped identify a top-down
analysis process that would start with a corporate strategy, the
associated action plans, and ways of measuring the performance
of the action plan and then systematically de-compose the per-

B O T T O M - L I N E A U T O M A T I O N xix
BottomLine05.book Page xx Friday, August 19, 2005 9:54 AM

F O R E W O R D : A S O L U T I O N W I T H O U T A P R O B L E M

formance measures throughout the operation. This Vollmann


Decomposition, as we refer to it, enables the economic perfor-
mance measures for each function, process, and activity in the
organization evaluated according to the corporate strategic per-
formance measures. If plant economic performance measures
(money saved or extra money earned) are derived and determined
according to this process, every function and person will be mea-
sured according to its contribution to the corporate plan. In this
way, the gap in perspective between the plant-level personnel and
the managers could be overcome.

The Vollmann Decomposition analysis proved to be a very


powerful tool. The problem was that for performance measures to
be effective in process manufacturing operations they had to be
calculated in terms of the same time constant as the manufactur-
ing process itself. If an operator is trying to manage and operate a
process in real time and the performance measures are only pro-
vided on a daily basis, it will be very difficult for the measures to
positively impact manufacturing performance.

Two members of the Foxboro applied research team, Dr. Mal-


colm Beaverstock and Peter Martin, developed and patented an
approach for modeling real-time economic performance measures
in process manufacturing operations. Their models are con-
structed from the data from the process instrumentation installed
to monitor and control manufacturing processes. The resulting
metrics are referred to as dynamic performance measures (DPMs)
and can be used by operation, engineering, and maintenance per-
sonnel as decision-support tools during the real-time operation of
the plant.

Over the past decade, DPMs have been installed in approxi-


mately thirty different process plants in industries such as chemi-
cal, gas, pharmaceutical, paper, mining, and food. In every case,
the plant engineering and operations personnel have learned how
to make decisions that result in significant ongoing gains in eco-
nomic value to the manufacturing operation. DPMs, if properly
derived, can thus be used as the primary metrics for continuous
improvement programs in process plants. When they are, organi-

xx B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page xxi Friday, August 19, 2005 9:54 AM

F O R E W O R D : A S O L U T I O N W I T H O U T A P R O B L E M

zations work toward common objectives and enjoy significant


improvements in economic impact, quality, and morale.

The remainder of this book will present the results and conclu-
sions of the Foxboro applied research project without going
through all of the iterations and changes that took place through-
out the program. The material will be presented in its final form,
with some supporting background information where necessary.
This form reflects the inputs from all of the teams interviews as
well as information from related books and professionals the
team consulted. Placing the material in this larger context ensures
that the reader gains a better understanding of what the managers
being interviewed were trying to convey. I have chosen not to
continually point out that what is being conveyed is a summary
of inputs from multiple interviews. For the purposes of readability
and flow, just the results are presented. Specific individuals and
works consulted throughout the research process are identified
either by quoting and footnoting where appropriate or by just cit-
ing the name of the source in the text.

The material presented is the state of the program at the time


this book was written. It became clear to the applied research
team that manufacturing performance is a very dynamic subject
that will continuously evolve. As a result, the interview and data-
gathering process associated with this applied research program
will be ongoing.

B O T T O M - L I N E A U T O M A T I O N xxi
BottomLine05.book Page vii Friday, August 19, 2005 9:54 AM

Table of Contents
List of Figures ......................................................................................................ix

List of Abbreviations..........................................................................................xi

Foreword: A Solution Without a Problem ................................................... xv

Chapter 1: The Drive to Economic Performance ........................................... 1


Globalization What It Takes to Survive ........................................1
Pathway to Performance....................................................................4
The Technology Trend ......................................................................5
The Quality Trend .............................................................................8
The Accounting Trend.....................................................................10
The Convergence .............................................................................15
Notes ................................................................................................15

Chapter 2: Technology and the Bottom Line .............................................. 17


The Technology Trend ....................................................................17
Technology for Manufacturing .......................................................18
Technology for Technology ............................................................21
Computer Integrated Manufacturing (CIM) A Noun or a Verb?28
CIM in the Process Industries..........................................................29
Lights Out Manufacturing ...............................................................33
Labor and Education .......................................................................36
Technology for the Bottom Line .....................................................43
Notes ................................................................................................52

Chapter 3: Quality and the Bottom Line ..................................................... 53


The Quality Trend ...........................................................................53
Quality for Manufacturing ..............................................................55
SQC in Manufacturing.....................................................................59
Quality for Technology ...................................................................64
SQC Fever in the Process Industries ................................................65
Quality for the Bottom Line ............................................................68
The Convergence of Lean Manufacturing and
Quality Improvement ..............................................................73
Lean Manufacturing Concepts ........................................................74
Lean Process Plants ..........................................................................76
Notes ................................................................................................79

Chapter 4: Cost Accounting and the Bottom Line ...................................... 81


The Cost Accounting Trend ............................................................81
Cost Accounting and Manufacturing..............................................83
Cost Management Systems..............................................................91
Accounting for the Bottom Line .....................................................94
Dynamic Performance Measures ...................................................102
Dynamic Performance Measures and Accounting ........................108
Notes ..............................................................................................117

Chapter 5: The Convergence....................................................................... 121


The Technology Trend and Dynamic Performance Measures ......122
The Quality Trend and Dynamic Performance Measures .............124
The Accounting Trend and Dynamic Performance Measures ......126
The Convergence ...........................................................................128
A New Management Paradigm ......................................................129
Notes ..............................................................................................132

B O T T O M - L I N E A U T O M A T I O N vii
BottomLine05.book Page viii Friday, August 19, 2005 9:54 AM

T A B L E O F C O N T E N T S

Chapter 6: The Economic Optimization of Industrial Assets ................... 135


Introduction.................................................................................. 135
Primary Plant Floor Functions in Industrial Operations.............. 137
Performance Measures for Primary Plant Floor Functions........... 139
Plant Asset Analysis ...................................................................... 141
Human Assets ............................................................................... 146
Traditional Approaches to Optimizing Plant Assets .................... 147
Fundamentals of Mathematical Optimization ............................. 150
Traditional Approaches to Industrial Optimization..................... 152
Optimizing Plant Assets for Business Performance ...................... 157
Asset Set Optimization.................................................................. 160
Requirement for Performance Vectors for Asset Sets ................... 162
Business Value Analysis ................................................................ 165
A Pragmatic Approach for Optimizing Business Value
from Plant Assets................................................................... 167
Barriers to Deployment of Business Performance Optimization . 170

Chapter 7: Case Studies ............................................................................... 173


Introduction.................................................................................. 173
Case Study 1: Dynamic Performance Measure Program at a
Major Pharmaceutical Company .......................................... 174
Case Study 2: Dynamic Performance Measures and
Culture at Dynegy Midstream Services................................. 188
Lessons Learned ............................................................................ 193
Notes ............................................................................................. 196

Bibliography ................................................................................................... 197

Index ................................................................................................... 209

viii B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 1 Friday, August 19, 2005 9:54 AM

CHAPTER 1

The Drive to
Economic
Performance

G L O B A L I Z A T I O N W H A T I T TA K E S
TO SURVIVE
Throughout the 1970s and 1980s business management
forums were dominated by discussions about the increasing glo-
balization of industry. But it was not until the 1990s that the full
impact of globalization was actually realized: a very tough busi-
ness environment in which only the best-prepared and best-man-
aged companies survive. In many cases, the surviving companies
have thrived to an unprecedented degree by growing both organi-
cally and through acquisitions.

In some respects, the decade of the 1990s was a buy-or-be-


bought decade. The fiscally strong companies have consumed the
weaker ones and often become quite large. But size alone has not
been enough to prevent acquisition. Some of the larger compa-
nies in the world, even in traditionally untouchable industry seg-
ments, have been acquired in the past decade or so, sometimes
even by smaller firms with fiscal strength and tough manage-
ment.

Globalization has created a difficult environment marked by


an incredible degree of organizational downsizing or perhaps, to
be more politically correct, rightsizing. Some organizations are

B O T T O M - L I N E A U T O M A T I O N 1
BottomLine05.book Page 2 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

now doing an increased level of business with half of their origi-


nal workforce. Some of the downsizing has been the result of
technological advances, but most of it is motivated by a desire to
reduce labor costs as much as possible without hampering pro-
ductivity. The impact on the people working in these environ-
ments has been harsh. Longevity and loyalty seem to have lost
their value in many organizations, and many employees feel they
are regarded as mere piece-parts and numbers rather than valued
contributors.

Globalization has led to several other trends that businesses


must respond effectively to if they want to survive. Inconsistency
and instability in the global resource base for raw materials and
human resources have put new pressures on manufacturers.
Those that have traditionally focussed successfully on one geo-
graphic market segment now face new competitors from different
parts of the world entering their traditional market. Often for
these competitors raw materials are easier and less costly to access
and of higher quality, giving them favorable competitive advan-
tage. Human resources often follow a similar pattern. Twenty
years ago, developing areas of the world tended to have an abun-
dance of human resources who were low cost but typically also
lower skilled and less educated. Manufacturers in the industrial-
ized regions had no choice but to pay higher wages for the higher
skills and better education of their local workforce. But this edu-
cation and skill gap is rapidly closing, presenting manufacturers
in developed regions with a very challenging competitive envi-
ronment. They cannot survive unless they change their tradi-
tional practices.

Another key driver created by globalization is the shift in the


criteria used to select automation solutions from technical con-
siderations to economic considerations. A decade ago almost all
decisions about automation purchases depended on the prefer-
ence by the technical community within the manufacturing oper-
ation of a state-of-the-art technology or feature. This resulted in a
significant level of capital spending for automation technology
that we now see provided little or no economic return.

2 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 3 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

Ten years ago, the phrase shareholder value was seldom, if ever,
heard. Today, this phrase can be heard in almost every business
discussion, even among the technical communities of manufac-
turing operations. The concern with shareholder value is intro-
ducing a new and extremely beneficial perspective into the
process of selecting and using automation assets in manufactur-
ing operations. This concern has been intensified by reductions in
capital spending and capital budgets. In the 1980s, companies
expected returns on capital in the range of 15 to 18 percent.
Today, the focus on shareholder value has often resulted in expec-
tations for returns on capital in the 50 percent or greater range.
This drive has presented manufacturers with significant chal-
lenges in obtaining the capital they need to enhance and upgrade
their technology base.

The driving economic forces that have resulted from globaliza-


tion are all a great cause for concern among manufacturers. Busi-
ness managers of these firms realize that to survive and thrive
within this tough competitive environment they have to be bet-
ter than their competitors. In the past, they could surpass com-
petitors through superior sales and marketing strategies. In the
future, they know they will have to win through more effective
manufacturing strategies and better execution of those strategies.

Manufacturers have experimented with dozens of different


programs, such as Total Quality Management, to set themselves
apart, but most have enjoyed little success. Unfortunately, the
path to survival is neither clearly defined nor without obstacles.

But the path does exist. It leads to the realization of new levels
of plant performance through more effective use of all plant
resources. The objective of this book is to take you down this
path. Manufacturers who start along it today will likely be the
world-class manufacturers of the future, the ones who survive to
tell of their success.

B O T T O M - L I N E A U T O M A T I O N 3
BottomLine05.book Page 4 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

PATHWAY TO PERFORMANCE
Manufacturers have been struggling to use technology to opti-
mize manufacturing strategies ever since the computer was intro-
duced as a tool for automation. From the beginning, it seemed
obvious that this new digital technology offered a tremendous
potential to improve on the manufacturing process. However, no
matter how much effort was applied, the results never seemed to
meet the expectations.

The growth in technology, especially computer technology,


has resulted in some peculiar human behaviors. Technologies are
typically invented to address sets of problems that humans have
encountered. But the newer and more complex the technology,
the more people seem to focus on the technology itself rather
than the problems to be solved.

This is exactly what has happened in the manufacturing


industries. Computers are ideal tools for addressing many manu-
facturing issues, but after they were introduced they became the
object of focus rather than the problems facing manufacturers. As
a result, computer technology has never reached its full problem-
solving potential in manufacturing.

Today, much of the hype surrounding com-


puter-based technology is behind us. Users
Technology
are increasingly focused not on the tool, but
on its use. This is an exciting and encourag-

Quality
ing change, one that is vitally necessary if
manufacturers are to move to a new para-
digm for the management of plant or mill
Accounting performance. This new paradigm is bottom-
line automation, a discipline that has the
Figure 1.1 Trends for potential to significantly improve the performance of plant opera-
Surviving in a Global
Economy tions by emphasizing a complementary relationship between peo-
ple and technology. If that potential is realized the result will be
that ever-elusive idealworld-class manufacturing.

One of the most startling aspects of bottom-line automation is


that there is nothing terribly revolutionary about it. As Peter
Drucker said in his book Innovation and Entrepreneurship, the

4 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 5 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

greatest praise an innovation can receive is for people to say: This


is obvious, why didnt I think of it?1 Bottom-line automation is
in many respects a natural evolutionary step in manufacturing
systems. It is based on three seemingly unrelated or independent
trends that can actually be made to converge into a single com-
prehensive pathway leading to enhanced manufacturing perfor-
mance.

The concept of the convergence of seemingly independent


trends into a knowledge-based innovation is not new. As Drucker
also pointed out, a characteristic of knowledge-based innova-
tions is that they are almost never based on one factor, but on
the convergence of several different kinds of knowledge, not all of
them scientific or technological.2 In the case of bottom-line
automation there are three trends (see figure 1.1). One is purely
technological but the other two, quality improvement and
performance measurement, are not. All of three have been
independently viewed by manufacturing management as part of
the solution to the difficult economic driving forces we discussed
earlier. Lets briefly discuss each of these trends.

T H E TE C H N O L O G Y TR E N D
For about the past thirty years, industrial manufacturers have
looked longingly to computer-based automation to solve their
manufacturing problems. They often focused on the computer
system replacing automation functions that had traditionally
been accomplished with non-computer-based automation equip-
ment such as electronic relays and/or analog controllers. In prac-
tice, however, the payback for the functional replacement of older
technologies by computer-based technology never fully occurred.
Doing the exact same job with a newer technology seldom pro-
vides significant returns on the investment.

Computers were introduced into the actual automation of pro-


cess manufacturing plants in the 1960s. When Digital Equipment
Corporation (DEC) proved that a computer could be manufac-
tured for a reasonably low price, computer technology was seen as

B O T T O M - L I N E A U T O M A T I O N 5
BottomLine05.book Page 6 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

a viable candidate for many applications that it had traditionally


not been considered for. One of the most intriguing areas of
application was the manufacturing process itself.

Once installed, a primary problem faced by industry was that


to be used effectively this new machine required specialists in
computer technology, and these specialists were few and far
between. Also, most of the people who had a reasonable knowl-
edge of computers knew little or nothing about manufacturing.
The result was that the same companies that were vendors of the
traditional automation equipment performed the initial applica-
tions of computers to manufacturing plants, typically on a cus-
tomized basis. Computers showed promise, but the custom
application costs were very high, and just getting the computer-
based systems to the same level as the previous automation sys-
tems required a great expenditure of resources.

The emphasis on computer technology in manufacturing


started to evolve significantly in the 1970s. The availability of
standard applications software made the computer as a tool for
manufacturing automation much more viable. It allowed the
computers to be configured to solve many of the repetitive applica-
tions found in manufacturing rather than programmed from
scratch each time. Although the computer was becoming easier
and more economical to apply, it was used, for the most part,
only to replace automation functions that had previously been
accomplished with older technology. Many of the manufacturing
computers included software packages such as linear and nonlin-
ear programming, but these advanced techniques were imple-
mented sparingly. The result was that the computer was
becoming an increasingly common tool for automating manufac-
turing facilities, but using this new technology led to few func-
tional automation enhancements. It was simply a case of a new
technology replacing an older technology.

Toward the end of the 1970s a new trend in manufacturing


came into vogue, a movement to tie all of the plants computers
together. It was assumed that if all the various computers in a
plant were linked together the plant would operate as an inte-
grated facility. In manufacturing, this trend became known as

6 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 7 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

computer integrated manufacturing (CIM). Computer inte-


grated manufacturing became a goal unto itself in that every
manufacturer, or at least every technologist in a manufacturing
facility, tried to achieve some degree of interconnectivity, and
each was absolutely convinced that good things would result
from interconnecting computers. Nobody seemed quite sure what
these good things would be, but there was general agreement
that, whatever they were, interconnectivity would make them
happen.

Predictably, the CIM movement did not meet the high expec-
tations of industry. In 1988, I talked to a member of a mill-wide
automation team for a large pulp and paper company. He related
to me that over the previous five years his company had spent
over $50 million on CIM, with not one additional dollar of profit
to show for the investment.

This is a prime example of the focus on the technology rather


than the true problems in the plant. Connecting computers
together into a plantwide network overcomes some of the barriers
to implementing solutions, but it does not solve manufacturing
problems. Because of its high expense, however, the CIM trend
was exactly what the manufacturing world needed to drive some
reality back into automation strategies and planning. Computer
integrated manufacturing should have been helping plants
accomplish automation, and automation should be implemented
to solve plant problems. If a new development solves an automa-
tion technology problem, however, it is only of value if it results
in true functional enhancements in the end application of the
technology. In other words, connecting the various computers
together, in and of itself, did not make the manufacturing opera-
tion perform any better.

However, it is incorrect to believe that the CIM movement had


no value; it resulted in some significant advancement. Communi-
cations, operating system, and database management standards
were developed during this period, and these tools helped to
break down some traditional technological barriers. But the most
significant positive accomplishment of this period was that it

B O T T O M - L I N E A U T O M A T I O N 7
BottomLine05.book Page 8 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

refocused manufacturers back on the functional issues of manu-


facturing and away from the automation technology issues.

This decade-by-decade technology drive created a tremen-


dously fertile environment for manufacturing improvement into
the 1990s. The focus on technology for its own sake is, for the
most part, behind us. Technological tool sets, especially in the
area of application software, are now abundant. The reduced cost
of the technology now enables creative new forms of physical and
functional distribution. And the emerging new automation sys-
tems have begun to take on many of the characteristics of higher-
level information systems. These new tool sets provide a technol-
ogy framework that is ideally suited to addressing the current
opportunities facing manufacturing. Todays challenge is to take
advantage of these powerful new tool sets. The time has come to
apply them directly to the problems manufacturers face.

THE QUALITY TREND


The preoccupation with technology for its own sake doesn't
only take place in the computer arena. It often happens when any
new development emerges and seems to offer a magic cure for
pressing problems. One such movement is statistical quality control
(SQC).

Back in the early 1930s, Walter A. Shewhart, a researcher with


Bell Laboratories, published a groundbreaking book on the appli-
cation of statistics to the control of manufacturing processes. His
book, Economic Control of Quality of Manufactured Product (1931),
provided the foundation for a movement that would take shape
some twenty to thirty years later.

Shewhart, who W. Edwards Deming referred to as the master


of SQC, recognized that no matter how precise a manufacturing
process was, in the real world it could never produce exactly the
same output each time a product was made. There are variations
that occur in nature that man-made manufacturing processes
cannot overcome. Shewhart recognized that statistics provided an
excellent platform for dealing with this natural variability. His

8 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 9 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

book addressed how to apply statistical methods to the control of


manufacturing processes. A good part of the fundamental con-
cept behind the application of statistical analysis to manufactured
product was that the process is in statistical control as long as
the variability of measurement is random and within reasonable
standard deviation limits. If a process is in statistical control, it is
performing as well as can be expected within its natural limita-
tions. The essence of Shewharts theory is that productivity
improves as variability is reduced.

W. Edwards Deming, a statistician who worked with Shewart at


what is now AT&T Bell Laboratories before World War II, was very
familiar with Shewarts work and is now credited with starting the
quality movement in the 1950s. In 1950, Deming went to Japan
and conducted a number of sessions on SQC and how this meth-
odology could significantly improve manufacturing productivity.
Because Japan was trying to recover from the impact of World
War II, his message created a significant amount of interest. Obvi-
ously, many Japanese manufacturers followed Deming's advice,
and his personal guidance and its results, combined with other
business and manufacturing practices, have been staggering.

As Deming worked on applying statistical techniques in manu-


facturing plants, he developed a valuable insight: the statistics by
themselves were not enough. For the statistical methods to be
effective, workers had to be actively involved in the application of
the SQC information. Deming became a strong advocate of
worker participation and the use of structures, such as quality cir-
cles, to promote worker participation. In this way, he not only
showed the value of SQC techniques; he also helped to point out
that this technology applied to manufacturing processes was not
enough to ensure success. There are human issues that are just as
important, if not more so.

Joseph M. Juran was another of the leaders of the quality


movement who had also worked with Shewart, but his focus was
more on managing for quality improvement than on statistical
techniques. Juran also went to Japan in the early 1950s, and he
did much work building an entire framework for managing for
quality improvement. Jurans framework is the basis for the man-

B O T T O M - L I N E A U T O M A T I O N 9
BottomLine05.book Page 10 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

agement structures in a number of world-class corporations that


have instituted successful quality improvement programs and are
today realizing substantial results.

The point behind extending the quality movement from SQC


to an overall management structure is not that the statistics are
inappropriate, but rather that the SQC technology by itself is not
enough. Today, the quality movement is vitally strong, moving
ahead rapidly, and continuing to show great promise. Meanwhile,
the followers of Shewart, Deming, and Juran have come to realize
that the quality movement is a major management and technol-
ogy movement that must be implemented as a comprehensive
program to improve the economic performance of manufacturing
operations. Jurans laid-back, team problem-solving approach has
since been supplanted in some of the leading global companies
by more rigorous performance-based approaches. One example of
this is the Six Sigma program that is gaining momentum in man-
ufacturing operations. Such performance-based quality programs
have refocused many of the useful tools and concepts of statistical
control into the performance-based global environment manufac-
tures must deal with today. Though these programs are starting to
show great promise, they are limited by the lack of appropriate
performance-based quality indicators.

Unfortunately, some of todays manufacturing technologists


continue to approach SQC techniques with a belief in technology
for its own sake. As has been demonstrated many times, this
approach is almost never successful. Technological advancements
must have clear objectives to achieve success.

THE ACCOUNTING TREND


One of the key issues in manufacturing that has come to the
forefront in the past few years has involved the way manufactur-
ing performance has been measured. In the past, manufacturing
operations often had to develop profiles of their performance to
report to the outside world, generally to secure some financial
backing or to report to financial backers. These systems evolved

10 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 11 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

into computer-based cost accounting systems. In many respects,


the first computer-based cost accounting systems were designed
to report information to the financial community in the same
way as earlier manual reporting systems. For a manufacturing
operation, one of the traditional measures developed by this kind
of system has been cost-per-unit-product-made (see figure 1.2).
This statistic is calculated by counting the amount of each prod-
uct or product type manufactured over a given period of time, cal-
culating the cost of the operation, and assigning the costs to each
product produced by the plant. Back in the early 1900s, this sta-
tistic had significance since most plants produced a very small
number of different products, and a large majority of the costs in
a plant were directly assignable to a product or product line. So,
in a macro sense, this statistic had some historical validity and
was certainly reasonable enough for the purposes of the financial
community.

Since the cost-per-unit-product-made


statistic was readily available from the cost Direct Costs + Allocated Costs
Cost per Unit Product =
accounting system and seemed to have Quantity of Product Produced
statistical merit, it became the primary
measure of manufacturing performance for almost all manufac- Figure 1.2 Traditional
Product Costing
turing facilities. Manufacturing managers were extremely cogni-
zant of what their costs and volumes were because their pay
increases, incentives, and promotions were directly driven by
their operations performance with respect to this statistic.

As manufacturing facilities became more complex and the


number of products and product lines produced by a single facil-
ity increased, the cost-per-unit-product-made became more diffi-
cult to compute directly. Today, many production people will tell
you that often less than 10 percent of their total plant costs can
be directly assigned to a product or product line. This means that
over 90 percent of the plants costs have to be assigned using an
allocation algorithm of some sort. This algorithm might be based
on sales volume, capital expenditures, or any number of other
factors associated with the products. The point is this: the statistic
cost-per-unit-product-made is now rarely a valid indicator of the
cost performance of manufacturing. Moreover, since manufactur-

B O T T O M - L I N E A U T O M A T I O N 11
BottomLine05.book Page 12 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

ing managers have little control over the numerator of that statis-
tic, their logical goal is to make as much of their products as they
possibly can in order to enlarge the denominator and thereby
reduce the resulting statistic.

A new cost accounting approach referred to as cost manage-


ment systems (CMS) has recently emerged to replace cost-per-
unit-product-made. One of the primary early factors behind the
development of cost management systems was to gain a better
sense of the assignability of all the costs in an organization. This
was accomplished through a technique known as activity-based
costing (ABC), where all activities associated with a product or
product line are accounted for and assigned to their cost profile.
Functions that have been traditionally considered overhead cost
functions, such as product development and sales, are now
assigned to products directly. For example, if a salesperson goes to
a customers site to sell a particular product, then the cost of the
salesperson's call is assigned to the product being sold.

Activity-based costing has evolved significantly since its incep-


tion. Many organizations have benefited greatly from de-com-
posing traditional organizational activities that have been
grouped and hidden in traditional cost accounting systems and
developing cost information on an activity basis in their stead.
This method has helped organizations identify activities within
their operation that have little or no economic value-added. In
some operations, even some manufacturing operations, these
non-value-added activities could be eliminated, creating signifi-
cant economic savings to the organization in areas that would
probably have continued to go undetected with traditional cost-
ing approaches. However, much controversy surrounds the ques-
tion of how activity-based costing may be applied in a process
manufacturing environment in which the manufacturing activi-
ties are highly interrelated, connected by piping, and difficult to
change. The general feeling seems to be that these techniques
should apply and help add value, but few advocates have been
able to determine how to accomplish this.

12 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 13 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

Early cost management systems provided a


better profile of the actual cost of a facility and a Strategy
reasonably accurate lifecycle cost statistic for a
product or product line, but they didnt really
help measure manufacturing performance. First,
many of the activities associated with and
assigned to a product are beyond the control of
the manufacturing operation. Second, even if Action Measure
the statistic was perfectly accurate, the
assumption would be that cost-per-unit-product-made is the cor- Figure 1.3 Vollmann
Triangle
rect and only statistic for measuring the performance of manufac-
turing, which is not necessarily the case. Third, this statistic is
typically calculated after the fact and reported to manufacturing
biweekly or monthly. By the time manufacturing personnel know
how they are doing, either good or bad, the product that the sta-
tistics were calculated on is long gone, and there is nothing they
can do about it.

To determine whether or not cost-per-unit-product-made is


the correct way to measure manufacturing performance, a simple
model developed by Dr. Thomas Vollmann is quite useful. Figure
1.3 shows a triangle representing the appropriate process for
implementing and evaluating manufacturing strategies. The com-
panys manufacturing strategy is at the top of the triangle. Once
the strategy has been determined, an action plan for achieving
the strategy can be developed and implemented. Once the action
plan is implemented, it should be measured to determine if the
strategy is being met. This so-called Vollmann Triangle offers a
true measure of manufacturing performance. If the strategy is
strictly cost-based, then cost-per-unit-product-made may in fact
be an appropriate measure. But if the strategy has any other basis,
different performance measures must be developed.

Throughout the interview process for Foxboros research pro-


gram interviewees confirmed that most of todays manufacturing
strategies are implemented and evaluated in two stages. First, the
manufacturing strategy is developed and used to develop an
action plan. Second, implementation of the action plan is initi-
ated. But, unfortunately, the measurement of manufacturing per-

B O T T O M - L I N E A U T O M A T I O N 13
BottomLine05.book Page 14 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

formance is cost-per-unit-product-made, regardless of whether


this statistic measures the objectives of the strategy. Just-in-time
(JIT) manufacturing programs are a good recent example of man-
ufacturing action plans that are diametrically opposed to the cost-
per-unit-product-made measure. Just-in-time is an action plan
structure based on a strategy of meeting customer demands while
lowering inventory costs through the minimization of an opera-
tions storage capacity. The goal is to make only as much of the
product as the customer needs at any point in time, and no more.
Note how opposed this is to the cost-per-unit-product-made sta-
tistic, which promotes making as much product in a given period
as is physically possible.

The effectiveness of a JIT program should be measured in


terms of customer satisfaction and the reduction of process stor-
age. The measure must match the strategy! But, the traditional
cost-per-unit-product-made measure seldom does. It is no wonder
that so many new manufacturing programs fail. Organizations tend
to perform to their measures, whether or not the measures are the cor-
rect ones.

One final point on measuring manufacturing performance is


that the correct measures, once determined, must be presented to
the appropriate people in the operation within a time frame that
will allow for true continuous improvement. If the measures are
made and presented to the operations people in charge of manag-
ing the process on an end-of-week or end-of-month basis, there
will be little they can do to effect improvement. There is just too
much dead time between the making of the product and the mea-
surement of performance.

The key to success for manufacturing managers is to select the


right performance indicators, institute realistic measurements,
and present them to the right people on a timely basis (as the
product is being made) so they can continually improve
performance.

14 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 15 Friday, August 19, 2005 9:54 AM

C H A P T E R 1 : T H E D R I V E T O E C O N O M I C P E R F O R M A N C E

THE CONVERGENCE
The three trends we have discussed thus far
the changes in technology, quality control, and Technology
accountinghave evolved significantly in the past
Bottom
few years, and each of them, on its own merits, Line
Quality
seems to offer manufacturers a ray of hope for Improvement

improved performance and new paths to manu-


Accounting
facturing excellence. In fact, lets now consider the
Dynamic Performance Management (DPM)
method that will demonstrate that these three trends are not Figure 1.4 The
Convergence: Survival in a
unrelated but actually converge into a single comprehensive path- Global Economy
way to excellence: bottom-line automation (see figure 1.4).

That method, DPM, combines the essence of three seemingly


independent trends into one complementary relationship. Inter-
estingly, most manufacturers have already invested in one or
more aspects of this approach, and thus the transition to a com-
prehensive bottom-line automation program has already begun
and is now well within reach.

This convergence, by itself, is not what is important. It is the


plant performance and organizational effectiveness that results
from the convergence. In todays very difficult competitive manufac-
turing marketplace, those manufacturers who significantly improve
upon their plant performance through approaches such as bottom-line
automation will become the world-class manufacturers.

The significance of this new paradigm is that it is in fact a path


toward survival! The only manufacturers who survive to see the
future will be world-class manufacturers. To become world-class
manufacturers, todays managers must find a way to achieve new
levels of manufacturing performance. The gauntlet has been
thrown at their feet.

NOTES
1. Drucker, Peter F., Innovation and Entrepreneurship: Practices and
Principles. New York: Harper and Row, 1985, page 135.

2. Drucker, Innovation and Entrepreneurship, page 111.

B O T T O M - L I N E A U T O M A T I O N 15
BottomLine05.book Page 17 Friday, August 19, 2005 9:54 AM

CHAPTER 2

Technology
and the
Bottom Line

THE TECHNOLOGY TREND


Over the past century process
instrumentation and automation Technology
for
Manufacturing Technology
technology have evolved through Technology for Technology
Technology for the
three major phases (see figure 2.1). Bottom Line
Bottom
Each of these phases has had a pro- Quality Line
Improvement
found impact on manufacturing. To
understand the current move toward Accounting
bottom-line improvement, we must
understand the major issues behind
the transition from phase to phase. The first phase is referred to as Figure 2.1 The Technology
Trend
the Technology for Manufacturing phase, and it coincided with
the early process manufacturing operations. The second phase,
Technology for Technology, began around 1970 and continues
right up to the present in the majority of process plants. The third
phase is just starting to take shape and is destined to have a pro-
found impact on the way in which automation is used. This phase
is called the Technology for the Bottom Line and promises to
lead to unprecedented economic performance improvements in
the process operations that take advantage of it.

B O T T O M - L I N E A U T O M A T I O N 17
BottomLine05.book Page 18 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

TECHNOLOGY FOR
MANUFACTURING
Though digital technology is currently driving manufacturing
toward a unified approach to instrumentation and automation,
instrumentation and automation in the process industries and
discrete manufacturing industries evolved very differently. One of
the primary reasons for this divergence is that automation is not a
basic requirement for discrete manufacturing. Discrete manufac-
turing operations involve the assembly of piece parts, which is a
very visible operation that can be effectively accomplished by
people. Many discrete product assembly lines began as human
processes without any automation, and worked quite well. Auto-
mation was essentially introduced in discrete manufacturing to
replace humans in assembly line processes with machines, such as
programmable logic controllers, which were faster, never tired,
and worked in a highly repeatable and reliable manner.

In process manufacturing, on the other hand, a basic level of


instrumentation is required just to enable the manufacturing to
take place. Many process plant parameters are totally invisible to
humans because the process is hidden within piping or tanks.
Instrumentation is required to make these critical parameters visi-
ble. Photographs of early process plants clearly show the gauges
that were needed so plant personnel could see the flows, tem-
peratures, pressures, and levels hidden within pipes, vessels, and
tanks. This basic physical fact made the economic value proposi-
tion for introducing instrumentation into process plants very
compelling and put the companies providing that instrumenta-
tion in a strong economic and technical position vis--vis plant
personnel. These companies did not need to focus their energies
on selling their economic value proposition because they pro-
vided manufacturing-enabling technology. Essentially, process
manufacturing could not be accomplished without it.

Early process plants were primarily manually operated and


employed a very basic level of monitoring instrumentation. Per-
sonnel were responsible for monitoring and when necessary
adjusting valves or other actuators to manually control the pro-
cessesnecessitating a large number of operators. The amount of

18 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 19 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

instrumentation that a single plant employee could manage was


often constrained by physical distance more than individual
human capacity. Many process plants were spread out over large
areas, and the instrumentation and valves could only be posi-
tioned at the point where the process was being measured or
manipulated. Furthermore, many manufacturing processes
involved hazardous materials. Stationing personnel close enough
to use the instrumentation in hazardous environments often
meant exposing them to very dangerous conditions. Plant engi-
neers challenged the instrumentation companies to help solve
these problems.

The companies responded by producing mechanical controls.


The earliest controllers were typically very simple mechanical link
devices. For example, level controllers similar to those still found
in toilets today were sometimes used to ensure that levels in tanks
did not exceed certain predefined limits. These devices did not
typically improve the economic efficiency of the plant. Rather,
they overcame barriers in the operation so plants could manufac-
ture new products that had been difficult or impossible to manu-
facture before. The relationship between the instrumentation
companies and process manufacturers was very symbiotic during
this phase of the evolution of instrumentation and automation.
They worked together to enable new forms of process manufac-
turing.

Mechanical controllers were typically limited by distance.


They needed to be located and operated right at the process. This
was often a difficult and dangerous proposition. Initially, the
instrumentation companies addressed this challenge by introduc-
ing pneumatic controllers, which used air signals to transmit
measurement information to a gauge or controller, which
returned an air signal to drive the valves. These controllers could
be located at a distance from the process vessels. This allowed the
operators to be located in a much safer environment and oversee
a much larger section of the plant by bringing multiple control-
lers and display stations together at a central point. Pneumatics
also enabled better levels of control, making automatic control of
the process more feasible. With the introduction of automatic

B O T T O M - L I N E A U T O M A T I O N 19
BottomLine05.book Page 20 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

pneumatic controllers, operators were still responsible for setting


the desired operating points (set points) into the controllers, but
the controllers did the second-by-second manipulation of the
valves. This further freed operators to control even larger plant
domains. The net result was that process manufacturers could
control their plants much better with fewer people, and the plant
personnel could enjoy improved workplace safety. Because of
pneumatic control technology the first full-scale control rooms
were implemented in process plants.

As good as pneumatic technology was for mechanical control


systems, it had its own limitations. If transmission lines were very
long, the nature of air signaling system meant delays in the trans-
mission of the signals from the instrument to the controller and
from the controller to the valve. These delays often led to control
problems in the plants because they introduced dead time into
the control loop that reduced control below desirable levels.
Pneumatic systems were also expensive to maintain. Once again,
process manufacturers challenged the instrumentation compa-
nies, which were now known as instrument and control compa-
nies, to help resolve these issues. In response, the instrument
companies developed and introduced electronic analog control
systems, which were designed to emulate pneumatic systems, but
used electronic current rather than air for signaling. They were
very effective and provided excellent control over much longer
distances.

Whether control was effectuated manually through basic


instrumentation, mechanically, pneumatically, or through elec-
tronic analog control systems, most process manufacturing man-
agers viewed automation technology as absolutely essential to
enabling the operation of their plant. The economic value propo-
sition of all these systems was simple and huge. If you had them
you could make product; if you didnt you couldnt make any-
thing. This led to a very close, technology-based relationship
between the instrument and control companies and process plant
personnel.

During this first phase, most of the field salespeople for the
instrument and control companies were professional engineers

20 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 21 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

who understood the automation technology and could help cus-


tomers design automation solutions using the technology. Many
field salespeople spent over half their time helping customers
engineer, install, and operate the automation systems. The rela-
tionship between suppliers and users was typically a very positive,
consultative relationship focused on making the plant operate.
The automation and instrumentation technology was an absolute
necessity to make most process plants work, and thus this phase
in the evolution of automation systems is referred to as the Tech-
nology for Manufacturing phase. Many of the managers inter-
viewed by Foxboros research team expressed very positive
memories of the collaborative working relationship they enjoyed
with the instrument and control companies during this phase. In
fact, they indicated that their working relationship with these
companies had changed significantly since the pneumatic and
electronic analog system days, and they longed to have the old
relationships back.

TECHNOLOGY FOR TECHNOLOGY


General public awareness of computer technology began to
surge in the 1950s. This awareness initiated one of the most fasci-
nating technological periods mankind has ever experienced.
Computer technology seemed to become more of a cult than a
science, and the manufacturing world became one of the cults
followers.

It is hard to convey the profound impact that the develop-


ment of the smart machine had on the world to people who did
not actually live through this period. Almost overnight a mytho-
logical structure developed based on the premise that computers
would take over and run the industrial world. Manufacturing
personnel gravitated into two camps: either they began to wor-
ship computers as pseudo-saviors, or they cowered in fear of the
computers awesome and mysterious capability. Meanwhile, most
of the people who had originally worked in the computer field
had already begun to focus much more on the computer as a
technology than on the problems that technology could address.

B O T T O M - L I N E A U T O M A T I O N 21
BottomLine05.book Page 22 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

Several key discoveries by extraordinary scientists and mathe-


maticians led to the invention of the computer. The computer
was first commercialized by Remington Rand Corporation in the
early 1950s with the introduction of the UNIVAC I. IBM and Bur-
roughs also commercialized computer systems in this period.
These computers were designed for either business or scientific
applications, were very expensive, and were extremely difficult to
use. As a result, at first only the largest organizations in business
and government represented a viable market for this new tech-
nology.

The technology behind the digital computer was a mystery to


most people. The limitations of these machines were not at all
clear, and many perceived the computer to be a tool of virtually
unlimited power. During this period, decisions were often made
to buy computers with no consideration for what the machine
would actually do for the organization installing it. Many com-
puters were acquired because everyone else has one, and com-
panies didnt want to be left out. With such important decisions
and expenditures being made on this basis, its clear that the
focus of these organizations had obviously shifted from the busi-
ness mission to the technology mission: technology for technol-
ogys sake.

In the early 1960s, Digital Equipment Corporation (DEC) over-


came one of the primary barriers to the application of computers
in manufacturing operations when it introduced the minicom-
puter. These computers offered less computing power, but were
much lower in price than the computers of the traditional ven-
dors. Once computers became affordable, the rush was on to
determine where and how they could be used. One of the obvious
areas was manufacturing, and DECs programmable data proces-
sor (PDP) product lines eventually developed an extremely large
installed base in manufacturing segments.

When digital electronic technology became viable for use in


process control during the late 1960s the Technology for Manu-
facturing phase came to an abrupt end. With all the advantages
of the electronic analog systems, but much more flexibility, the
digital process control computer was seen as the technology that

22 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 23 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

would overcome all of the limitations and barriers of existing pro-


cess automation.

Digital computer technology was very immature when it was


introduced to manufacturing, and most people did not under-
stand it, and to a degree feared it. These systems required new and
unique skill sets that were not readily available. It is important to
keep in mind that at the time the instrumentation and control
technology used in manufacturing plantsrelay, pneumatic, and
electronic analogwas considered to be very high technology.
Most instrument and control companies and most process manu-
facturing companies had large engineering, maintenance, and
operations staffs who were familiar with the traditional technolo-
gies and were the leading technologists in their companies. But
traditional control technologies required a very different set of
talents than digital technology both to develop and to apply.
Most instrument and control companies were forced to hire a
whole new generation of employees who understood digital com-
puter technology well, but had little experience with instrumen-
tation or control.

These digital gurus carried with them a high-tech aura


even greater than that of the existing staff, and an organizational
schism often developed between the traditional instrument and
control people and the new computer people. Some instrument
and control companies went so far as to have two separate divi-
sions, one for the instrument and control peoplethe analog
divisionand one for the computer peoplethe digital divi-
sion. Though the two groups clearly needed each other to apply
the digital computer technology to process automation they had
two very different vocabularies. The control people would talk
about PID control and P&IDs while the computer people
would refer to bits, bytes, pixels, and bandwidth, terms
that seemed to have little to do with controlling process plants.
The result of this computerization push and the schisms it caused
was the most technology-focused period in the history of industry,
one that seemed to live by the principle, technology for its own
sake.

B O T T O M - L I N E A U T O M A T I O N 23
BottomLine05.book Page 24 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

Many plant and process engineers and operators understood


the traditional pneumatic and electronic analog automation tech-
nology quite well. They knew how to design, implement, and/or
operate an analog control system. Typically, however, they knew
nothing at all about how to program and operate a digital com-
puter. For computer technology to be successful, a structure had
to be developed to accommodate this lack of computer familiar-
ity. Otherwise, traditional analog control was destined to remain
the furthest limit of process automation technology.

The solution was the block concept, in which software struc-


tures could be written to run in a real-time environment and to
imitate the action of the traditional analog modules. Software
control packages were now designed in terms of blocks or mod-
ules of functionality that directly matched the traditional pieces
of hardware available in an analog control system. These software
blocks were comprised of programs, subprograms, and files
arrayed in an architecture that would support the real-time opera-
tional characteristics and functionality of process plants. Rather
than the wiring that connected the analog hardware, these soft-
ware blocks were linked together by programming code. To
engineers or operators, a software block could essentially per-
form the same functions as the corresponding physical analog
components, such as a PID controller, signal selector, or alarm
unit, making the application of computers in process environ-
ments much more acceptable.

Now a plants engineering staff could set up control strategies


via a fill-in-the-blank configuration process, wherein each compo-
nent required in the control system was activated or identified by
simply filling in its specification data. Multiple blocks could be
chained together to form control loops (input block + control
block + output block). This new method of defining process con-
trol was pioneered by the Foxboro Company, and has been used
ever since to construct continuous control strategies.1

To satisfy the old-school process operators, these software


blocks were linked to displays that imitated the traditional analog
panel boards. That is, the user interfaces the operators saw on

24 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 25 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

their CRTs mimicked traditional analog control faceplates,


allowing them to more comfortably visualize their tasks.

Because the block structure allowed plant engineers and opera-


tors to use their existing knowledge when utilizing the new tech-
nology, it finally made viable the use of computers in process
control. At the same time, assuming the guise of the traditional
analog systems to make plant engineers comfortable with the
brave new world of automation placed an artificial limit on the
flexibility and potential for higher levels of automation that com-
puters actually offered. As a result, most of the early digital con-
trol systems were implemented as exact functional replacements
of either the pneumatic or electronic analog systems that pre-
ceded them. History has shown that the exact functional replace-
ment of one technology with a newer one seldom leads to
performance breakthroughs. Much of the early promise of digital
computing in process manufacturing was lost.

In the 1970s, the mainstream computer industry next began to


interconnect multiple computers through evolving digital com-
munication networks. The industrial sector followed the com-
puter industrys lead in a trend that came to be known as
distributed control. The industry tended to try to avoid the word
computer in defining these new systems because of the negative
stigma computers still had for many in the industrial market-
place. This was accomplished by either giving them functional
names, such as digital process controller, or by just alluding to
the fact that they were microprocessor-based components and
omitting the word computer altogether.

Some early distributed control systems (DCS) were developed


merely by networking existing process computers. As digital com-
munication technology became accepted within industry, the
next step was to move to a more functionally distributed structure
in which each major control system function would have a sepa-
rate computer, or set of computers, dedicated to it. A communica-
tion network interconnected the different functions. In this case,
automation vendors continued to manufacture the large multi-
purpose process computers and make them part of the DCS struc-
ture, but they also made computer modules that were dedicated

B O T T O M - L I N E A U T O M A T I O N 25
BottomLine05.book Page 26 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

to functions such as control, process I/O, and human interfacing.


The vendors also designed their own proprietary communication
networks to tie all these computers together. Distributed control
systems became the standard for automating the process areas of
industrial plants from the mid-1970s through the early 1980s.

The unfortunate aspect of the


Supplier A Supplier B Supplier C Supplier D evolution from early digital pro-
Chip Set cess computers to DCSs was that
Bandwidth most of the innovations were on
Protocol the side of computing technology.
LAN Length
As new processors, color moni-
LAN Media
tors, or communication
Pixel Resolution
approaches became available the
Color Palette
automation companies would
MIPs
invest tremendous sums of
Figure 2.2 Technology money to incorporate them into their systems. As manufacturers
Focus
tried to decide which DCS would be most suitable for their opera-
tions, the decision criteria more often than not became which
supplier offered the latest state-of-the-art computer features.
Many industry trade journals would run annual DCS comparison
issues in which they would conduct detailed analyses to deter-
mine which DCS was the best. Figure 2.2 shows a blanked out ver-
sion of an actual comparison chart right out of one of the
industry trade journals of the late 1980s. It clearly shows the DCS
selection criteria that journal, and many of its customers, felt
were important. This figure does not attempt to show all of the
criteria presented in the source article, merely a representative
sample of the first high-priority criteria. Note how difficult it is to
relate the criteria in the chart to whether or not the system can
help operate a process plant any better. That issues such as the
color palette and pixel resolution of the CRT screens were consid-
ered more important than the availability of advanced functions
that might help the plant perform better is nothing short of
amazing. A study of a number of requests for proposals (RFP) for
automation systems made to automation companies during this
same period revealed that the primary issues of concern were
right in line with those presented in figure 2.2. This is a sure indi-
cator that the people selecting the automation solutions for their

26 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 27 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

plants were enamored of the computing technology and so


focused on the technology itself rather than its potential to solve
new manufacturing problems.

This abnormal focus on the technology of automation rather


than the objectives of automation led to a long period in which
significant capital investments were made with minimal, if any,
returns. Process automation systems were being implemented as
strict functional replacements of their analog predecessors, and
the resulting plant performance improvements were very disap-
pointing. The traditional technology-enables-solutions mindset
that had been so successful with early instrument and control sys-
tems had evolved into an aberrant preoccupation with technol-
ogy after the introduction of digital control automation systems.
Delegating the automation decisions in process plants to the
technologists, who were typically not measured in terms of plant
or process economic performance improvement, helped to justify
and reinforce this technology focus. This gave the technologists
much power within their organizations and ensured that the
selection, implementation, and operation of automation technol-
ogy was going to be driven by technology objectives instead of
business or financial objectives. This is not to say that the result-
ing decisions were always at odds with financial and business
objectives, only that the technological objectives greatly out-
weighed them.

As more and more computer-based automation systems were


installed, manufacturing management began to feel growing dis-
appointment over the lack of tangible results. Manufacturers were
investing millions of dollars in these systems, yet in many cases
their plants did not run any better. For all the catchphrases and
slogans bandied about to celebrate these technology-driven initia-
tives the promise of the smart machine was not being realized.
A review of two of these initiativesCIM and lights out manufac-
turingmay be instructive.

B O T T O M - L I N E A U T O M A T I O N 27
BottomLine05.book Page 28 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

COMPUTER INTEGRATED
MANUFACTURING (CIM) A
NOUN OR A VERB?
By the middle of the 1970s, there were many personnel in
industry whose primary job was to deal with the new digital com-
puter technology. In truth, many of them were in a unique posi-
tion of influence in their organizations because not many other
employees understood computers. As manufacturing manage-
ment began to openly express disappointment over the new tech-
nologys lack of performance, it became clear that the jobs of
these technocrats might be in serious jeopardy.

The automation technologists responded in the same way that


had worked so many times before. They proposed throwing still
more technology at the problem. One reason this tactic had suc-
ceeded so often in the past was that they were the only ones who
understood the technology! If an operations manager spoke out
against them and his lack of knowledge was made evident, his
career could be severely damaged. To decide whether to adopt
new technologies that they didnt understand, management
turned to their technologists, even though they were the ones
who stood to gain the most from these technology-focused solu-
tions in the first place.

The new technology solution proposed in the mid 1970s was


based on the fact that there were already many, many computers
in factories, plants, and mills, each performing a subset of the
total automation function and operating somewhat indepen-
dently. If all of these computers could somehow be tied together,
something good would surely happen. Nobody seemed to know
quite what good was, but everybody agreed that, whatever it
was, it could be made to happen. This concept became known as
computer integrated manufacturing or, more commonly, CIM,
and soon everybody was CIM-ing (to use the verb form) in one
way or another.

Almost overnight, CIM experts appeared on the scene. They


were in very high demand and, of course, highly paid. Though no
two of them could quite agree on what CIM was, each of them

28 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 29 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

presented equally convincing arguments supporting their points


of view.

The tactic worked; very few managers were in a position to


argue. Industry magazines were festooned with CIM articles.
Computer integrated manufacturing was portrayed as the savior
of industry and manufacturing. Everybody wanted it. Everybody
could do it. But still nobody quite knew what it was.

CIM IN THE PROCESS INDUSTRIES


During the CIM period little distinction was made between
discrete and process manufacturing operations. The perception
seemed to be that if CIM, or any major initiative, was good for
one it had to be good for the other. That automation in discrete
manufacturing had evolved very differently from automation in
process manufacturing was conveniently ignored.

Much of the fervor associated with the CIM movement was


initially directed at the discrete manufacturing industries. The
automation in these segments had evolved around a work cell
concept and had led to islands of automation. The mission of
CIM in these operations was to promise to interconnect the vari-
ous islands and by doing so provide better logistic control over
the facility.

In the process industries, which had not evolved from a work


cell idea at all, CIM was nevertheless approached in the same
basic way. In many process plants, however, the DCSs that had
been installed encompassed much more of the plant than was
encompassed by a single work cell in the discrete industries. The
focus of the CIM movement in the process industries, therefore,
was on connecting the DCSs to the plants business systems. This
was almost universally accomplished by having the DCS vendors
develop gateways that would allow their DCSs to connect to the
business systems in the plant (figure 2.3). Since there were many
different types and vintages of business systems, even in the same
organizations, an intermediate computer, often referred to as a
host computer, was often employed to enable the information

B O T T O M - L I N E A U T O M A T I O N 29
BottomLine05.book Page 30 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

connectivity. DEC PDP11s, DEC


Business System VAXs, DEC MicroVAXs, HP1000s,
and HP9000s probably accounted
for over 85 percent of the interme-
Host Computer
diate computers in use.

Although both the business sys-


Gateway
tems and the DCSs were based on
digital computer technology, two
DCS
very different departments were
2 3 n
1
typically responsible for each in
process plants. The plant engineer-
ing department was usually respon-
sible for the DCS and the
Figure 2.3 Computer Management Information Systems (MIS) department for the busi-
Integrated Manufacturing
ness systems. In most plants, these two different departments did
not work very closely together and really did not understand
what the other did. In many plants visited by Foxboros research
team, before we could start a meeting involving both the plant
engineering and MIS departments, we had to wait for them to
introduce themselves to each other! Some general perceptions
worked to create conflict and insecurity between the two groups.
For example, at the time there was a general feeling that the MIS
professionals should be responsible for all the computer technol-
ogy in a plant. This caused the plant engineering professionals
considerable concern because the MIS group typically had no
notion about plant operations. The MIS group was equally con-
cerned: they knew computers, but they had no clue about the
hardware and software of DCSs. Further, MIS functions are typi-
cally transactional while process functions need to operate in
real time. Trying to get two groups that were so disjointed to
work together around a common goal proved very difficult.

The net result was that the two groups often agreed to a parti-
tioning of technical responsibilities. The boundary was typically
set up as the connection between the gateway and the host com-
puter. So it became the engineering groups responsibility to get
the information across the gateway and into the host computer,
and the MIS group would take it from there. Since the engineers

30 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 31 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

typically had no idea what the business systems would do with


the data they were providing, they tended to try to provide to the
host all the data they could pull together. This was accomplished
by scanning the databases within the DCS at a regular interval,
typically every few minutes, and providing a snapshot of as many
of the various flows, temperatures, levels, and pressures around
the plant as was possible. Often a single snapshot involved thou-
sands of points of data. This time-based record was passed
through the gateway to the host computer at this time interval.
The host computer would insert the record into its database. The
database in the host was typically set up as a circular file with a
large number of records. Each time a new snapshot was provided,
the software in the host would insert it into the next available
record. When the database in the host filled up, the software
would insert the new snapshot record right over the oldest record
in the file. In this way, at any point in time, the hosts database
would contain all of the values of the plants process measure-
ments in time slices from the present back to the number of time
slices defined by the size of the databaseoften about a week or a
month. This meant that the host had an enormous amount of
data about the plant for that period.

The thinking behind this approach was that the MIS team
should be able to discern anything it might need to know about
the plant operations from this data. The reality was that to the
MIS personnel, who really were not very familiar with plant oper-
ations, this huge data set was essentially useless information. We
conducted a simple study in the late 1980s to try to determine
how much of the data that had been sucked up into the host
computer was ever used and found that over 95 percent of it was
overwritten before it was ever accessed. If we had undertaken a
more rigorous study the percentage probably would have gone up
significantly. As a matter of fact, a number of plants reported that
none of the data was ever accessed by the MIS team.

One other interesting anomaly about this kitchen-sink


approach was the fundamental direction of the data flow. Our
interviews with plant CIM teams revealed that one of the primary
objectives of most CIM initiatives was to improve the plants

B O T T O M - L I N E A U T O M A T I O N 31
BottomLine05.book Page 32 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

operational performance. In other words, the objective of the pro-


gram was improving the performance of the plant in manufactur-
ing its products. But all of the data flow was up to the host
computer, away from the point of objective. In most of the plant
sites we observed no information was being sent to the plant to
make it run better. It is amazing that this fundamental design flaw
was overlooked. This clearly reflects a technology-focused
approach rather than a business-focused one.

One thing became abundantly clear, however, as well as uni-


versally accepted: the need to be able to connect equipment from
different manufacturers together. The move was now on to
develop technological solutions that could tie together different
systems from different vendors. New businesses sprang up exclu-
sively to accomplish this interconnectivity. They went by a num-
ber of different names, but were mostly referred to as systems
integrators.

Connecting disparate systems was a very expensive proposi-


tion, and automation users soon expressed their displeasure at the
expense. Systems integrators and automation vendors responded
by insisting that the reason for such high costs was the lack of
communication standards that would make a general connectiv-
ity solution possible. The automation users accepted this argu-
ment, and in fact they responded in a much more positive way
than many of the technology vendors really expected them to. A
major effort was thus initiated to develop communication stan-
dards. Manufacturers and vendors alike believed communication
standards would make it more cost effective to tie these comput-
ers together, and since connectivity was the essence of CIM, these
connectivity standards would finally allow them to accomplish
their ultimate goal at much lower cost.

The technological focus throughout this entire period was on


the interconnectivity of all the computers and related equipment
within a plant. The unfortunate assumption was that if all the
computers were connected together, the plant would run better.
Industry was obviously in for another major disappointment. The
lessons of the past should have made it clear that this automation
technology drive would not succeed if it did not support more

32 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 33 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

automated functionality in the process of manufacturing the


products themselves. Connecting equipment together does not
add functionality; it merely overcomes a technology barrier to
existing functionality. The evolution of these communication
standards merely allowed process manufactures and suppliers to
accomplish nothing at a much lower cost.

The entire period of automation system evolution, dating from


the time the digital computer was introduced as a feasible plat-
form for manufacturing automation in the early 1960s, has been
one of technology focus rather than solution focus. The majority
of CIM activities represented the culmination of this evolution,
and they were, without doubt, the most extreme example of tech-
nology for its own sake ever experienced in the manufacturing
sector.

As the CIM movement expanded and the fervor enveloped


more and more manufacturers, experts preached the virtues of
CIM through seminars, articles, and books. But to those who lis-
tened and tried to understand, it became apparent that no two
experts agreed on what CIM was, and it became even less clear
whether CIM was a noun or a verb.

LIGHTS OUT MANUFACTURING


The challenges posed by the global market environment have
been the subject of countless conversations among manufactur-
ing professionals. For years, most manufacturers had their market
sectors largely to themselves. Not any more. New competitors
from emerging industrial regions are not only entering these mar-
ket segments; they are taking them by storm. They are offering
similar products, often at much lower prices, and sometimes even
with higher quality.

These new manufacturers are located in regions where labor


costs are considerably lower than in the industrialized world.
Often raw materials are readily accessible to these firms and thus
much less expensive. And these new manufacturers dont even
have the expense of inventing their manufacturing processes;

B O T T O M - L I N E A U T O M A T I O N 33
BottomLine05.book Page 34 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

they can copy the same processes that the established manufac-
turers have already invented and developed. In some cases, they
can even improve on these processes because they are starting
with the very latest processing equipment.

How can traditional manufacturers compete? With high labor


and raw materials costs and aging manufacturing plants, it could
seem as though the battle is lost before its even begun. Lets take
a closer look at the problem; a solution may be staring us right in
the face, if only we take the time to look. There is no question
that if traditional manufacturers behave as they have for years, if
they do not reevaluate their critical resources, the new global
competitive environment will drive them out of business.

Until now, the resource that has been the brunt of the greatest
criticism is the human elementin other words, labor. Ever since
labor organized years ago, labor costs to manufacturers have
steadily increased at a much faster rate than productivity, and in
some cases productivity has actually decreased. As a result, manu-
facturers have been expending significant effort determining how
to reduce the size and cost of the labor force. One example of this
is reflected in the catchphrase of the early 1980s, lights out man-
ufacturing, in which technologists envisioned computer tech-
nology totally replacing thereby eliminating the requirement for
any lights in the plant. The theory was that initially such technol-
ogy would increase capital expenditures, but eventually it would
significantly reduce the ongoing operating cost of a plant.

This approach was very consistent with the technology focus


in manufacturing at the time, and just as unsuccessful. Lights out
manufacturing strategies have not been realized, especially in pro-
cess plants. The use of technology has led to some reductions in
the labor force, but not to the expected levels, and certainly not
to the extent of total replacement. No one has yet discovered how
computer technology in process manufacturing can be used suc-
cessfully without people. In fact, its possible that our key to suc-
cess will be to change our perspective with respect to the human
element in manufacturing operations. Those who have emphasized
treating people as a resource, not a burden, have begun reaping great
benefits.

34 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 35 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

Like many of the other slogans in manufacturing of the past


thirty years, lights out manufacturing got its start in discrete
manufacturing operations. These operations began as assembly
lines. As automation technology evolved, it was found that tech-
nology, such as programmable logic controllers (PLCs), could
replace people on the assembly line. The PLCs were faster, cost
less to operate, and were more efficient because, unlike people,
they did not get tired. Replacing all the personnel on an assembly
line with automation technology began to seem feasible, and the
lights out movement gained momentum.

Process manufacturing plants did not evolve from assembly


lines. Unlike assembly lines, where every action and response is
well defined and visible, there is considerable variability in pro-
cess plantsthe range of things that could happen in response to
a given action are manifold. For example, if the chemical compo-
sition of a catalyst or one of the reagents in a chemical reaction
varies slightly from what is expected, the result of the process
may be unknown. Responding to this type of variability correctly
requires an intimate understanding of the operation or a level of
craftsmanship that can only be provided by the experienced and
knowledgeable people operating the plant. Eliminating the
human element from the process operation is not as simple and
straightforward as in discrete operations.

To gain a better understanding of the subtleties of applying


lights out manufacturing to the process industries, we focused on
the evolution of the operations personnel in process manufactur-
ing environments. One of the best and most concise sources for
this information, In the Age of the Smart Machine by Shoshana
Zuboff, focuses on the changing level of craftsmanship of the
labor force from the beginning of the Industrial Revolution to
today. Though the scope of this excellent study is broader than
just the labor force and craftsmanship, it explained the increasing
quality of the labor forces output in a very clear and direct way.

In our initial interviews, the industrial managers we spoke to


seldom talked about the labor force and craftsmanship or, if they
did, they focused more on the high cost of the labor force than on
its increased capability. Whenever we raised the issue of crafts-

B O T T O M - L I N E A U T O M A T I O N 35
BottomLine05.book Page 36 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

manship from a more positive point of view, the managers often


seemed perplexed by our questions. This could have meant that
they thought the fact that operations people were becoming
increasingly valuable was obvious and thus did not merit discus-
sion, or that they had never considered operations people in
terms of value or quality before. Our follow-up discussions indi-
cated that the latter was more often the case. Moreover, when the
managers came to this realization, it was fascinating to see how
excited many of them became.

The transformation in craftsmanship of the operators in pro-


cess plants is well documented, especially in Shoshana Zuboffs
book. So, we will make no attempt here to repeat it. Most readers
who are familiar with the current industrial climate and condi-
tions will easily recognize this issue.

The labor force of the industrialized world is no longer the


uneducated, unskilled mass it was fifty years ago. In fact, the edu-
cational level of the populations of most industrialized countries
is quite high today. If we continue to view this resource as we did
half a century ago, we will certainly not be taking advantage of an
indispensable resource that stands ready to help. As Thomas
Peters and Robert Waterman conclude in In Search of Excellence,
Treating people not money, machines or minds as the nat-
ural resource may be the key to it all.2

LABOR AND EDUCATION


Before the Industrial Revolution, many of the goods on the
market were produced by single-family operations. In the early
stages of the revolution, factories frequently came into being
when one of these family-run operations bought out others so as
to consolidate resources and manufacture goods more efficiently.
The need to intensify production was the driving force behind
the establishment of the early factories and workshops.3 Typi-
cally, the family that took the initiative to consolidate became the
factorys management, and the families that joined with them
became the workforce. One noteworthy aspect of this arrange-

36 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 37 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

ment was that everybody in the factory management and labor


knew how to make the product. They were all craftsmen, and
thus a high degree of craftsmanship and pride was demonstrated
throughout the manufacturing operation.

This type of factory must have been a pleasure to run. The


managers knew that if they or any of the workers were not avail-
able, the others in the factory could easily cover for them. Indica-
tions are that relationships between the owners and the labor
force in these early mills and plants were often very close. In a
way, the early success of this model was very misleading, in that it
caused some people to believe that the transition to an industrial
economy was going to go much easier than it actually was.

The bathtub curve shown in


Management
figure 2.4 provides a geometric Labor
model of the history of labor C High
r Trade Information
from the start of the Industrial a
Craftsman Craftsman
f
Revolution to today. It may help t
s
m Scientific Informate
to clarify the discussions to follow a Management
n Automate
on the evolution of the labor s
Organization
of Labor
h
movement and to identify where i Educate
p
that movement stands today and Low
1850 Time 1995
how we can make the most of it.

As the first generation of labor began to be replaced, the indi- Figure 2.4 Operations
Craftsmanship
vidual craftsmanship of the labor force of the early family-run fac-
tories was lost. The children of these laborers had not been
brought up in the factory the way their parents had and so hadnt
acquired the intimate knowledge of the craft. Also, and perhaps
more significantly, a major population shift was occurring from
the rural agricultural areas to the industrial cities. It took only 30
years, from 1830 to 1860, to transform both Western Europe and
the Eastern United States from rural and farm-based societies into
industry-dominated big-city civilizations.4

Almost overnight there was a huge mass of uneducated and


unskilled people available at a very low cost. It was also clear that
there were not enough highly skilled craftsmen available to effec-
tively expand manufacturing operations. Manufacturers realized

B O T T O M - L I N E A U T O M A T I O N 37
BottomLine05.book Page 38 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

that to survive they had to take advantage of this vast unskilled,


uneducated human resource.

It was a major challenge. Typically, the few skilled craftsmen


who were available were recruited to join the management force
in plants and mills. Thus, they would help to organize and direct
the unskilled laborers. But the work habits of the new labor force
were not entirely appropriate for manufacturing operations. They
were accustomed to working on farms, setting their own sched-
ules, and working somewhat independently. This mode of opera-
tion did not fit the new manufacturing environments. Workers
often were late to work and took unscheduled breaks. Manage-
ment tried all manner of tactics to gain control over the situation,
but with limited success. Elaborate systems of fines were devel-
oped, minutely tailored to extinguish particular expressions of
the impulsive body.5 The primary method of coercion employed
by the foremen was to combine authoritarian combativeness
with physical intimidation in order to extract maximum effort
from the worker.6 But these tactics failed.

It was clear that something significant had to be done to turn


this impasse into a workable manufacturing system. A more ratio-
nal approach was required, one that focused on the resources
available and getting the most out of them. As Zuboff notes in
The Age of the Smart Machine, The man who emerged as the chief
symbol of the rational approach to management was Frederick
Taylor.7 The approach that he championed became known as
Scientific Management, or sometimes as Taylorism.

The agenda for Scientific Management was to increase pro-


ductivity by streamlining and rationalizing factory operations.8
In theory, this could be accomplished by applying the methods of
science to the study of the manufacturing processes. This required
the acquisition of new and detailed knowledge about the way in
which labor worked. As Zuboff observes, The principal method
of acquiring such knowledge was the time study, later with the
influence of Frank Gilbreth, the time and motion study. Here
'expert' observations of worker performance made it possible to
translate actions into units of time and reconstruct them more
efficiently.9 Once this data was gathered, it, in combination

38 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 39 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

with other systematic information regarding tools and materials,


laid the foundation for a new division of labor within the fac-
tory.10

The by-product of these studies was that management found


ways to minimize the amount of skill and training time associ-
ated with efficient operations.11 The basic concept was to divide
an entire manufacturing operation into well-defined work units
and then train each laborer to do only the functions within his or
her work unit. The workers did not learn how to make the entire
product, or how to be craftsmen, nor did they have to as long as
they could accomplish the operations in their area. Often the
workers didnt even understand how their actions contributed to
the overall fabrication of the product. With this limited perspec-
tive, even with years of experience the workers had little added
knowledge and were no more valuable to the operation than
when they had first joined.

The implementation of Scientific Management as the only rea-


sonable approach to managing manufacturing operations marked
the beginning of a strong and well-defined schism between man-
agement and labor. As Zuboff puts it, Scientific Management
legitimized a new conception of managerial responsibility in
coordinating and controlling the complexities of the factory as it
entered the era of mass production.12 The laborers did the physi-
cal work with no reasonable hope of improving their lot in life.
Management controlled the laborers, made much more money,
and was part of a different social class. Taylor had believed that
the transcendent logic of science, together with easier work and
better, more fairly determined wages, could integrate the worker
into the organization, and inspire a zest for production. Instead,
the forms of work organization that emerged with Scientific Man-
agement tended to amplify the divergence of interest between
management and workers.13

The result of this situation was very predictable. The labor


force had to organize to protect its interests and increase its
opportunities. This organization effort took two basic forms
depending on where in the world they took place: labor unions
and communism. In either case, the labor force became a major

B O T T O M - L I N E A U T O M A T I O N 39
BottomLine05.book Page 40 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

political force, and as such, labor had power. One of the primary
objectives of the application of this power was to make their chil-
drens opportunities greater than their own.

Many social programs evolved that focused on broadening the


possibilities of the children of the labor force. One such program
was public education. Although public education had been
around for some time in some form, the labor movement gave it
new emphasis. Workers believed that if they could get their chil-
dren an education, the chances that their children would get bet-
ter jobs and have better lives would significantly increase. It
worked. Initially, the laborers children became educated enough
to read, write, and perform basic arithmetic. By World War II,
much of the population in the industrialized world had received a
secondary school education. At that time, secondary school edu-
cation by itself opened up opportunities for the children of the
working class to move into management positions. After World
War II, it was not unusual for members of this generation to
attain college degrees. This educational movement was not iso-
lated to a few of the laborers, but represented a widespread trend.
The workforce continued to become better educated, and at the
same time workers became much better qualified and motivated
to make significant contributions to the manufacturing environ-
ment.

As we noted earlier in this chapter, the computer was intro-


duced into the manufacturing process in the late 1960s. Certainly
in the process industries, at least, the computer has since become
a primary interface between the operators and the manufacturing
operations. Operators sit for hours each week watching the pro-
cess through a window provided by CRTs linked to the comput-
ers. This window presents a wealth of information, on a real-time
basis, concerning how the plant is running now and how it has
been running over the past few hours. This technology has
enabled operators to monitor and take responsibility for ever-wid-
ening areas of the plant.

Because todays operators are more highly educated, as they sit


in front of their CRTs they learn about the plant in ways nobody
in the operation ever has before. They watch the plant day after

40 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 41 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

day, and they come to understand process relationships that


nobody else understands. In essence, they are developing a new
craftsmanship. They are becoming the information craftsmen of
their age.

While working on an automation project in a process plant a


few years ago, a project team was reviewing the plant diagrams to
scope out the work to be done. As they reviewed the diagrams,
one of the team members asked if there was any relationship
between two of the process variables. The plant engineer
responded that he had studied the piping and instrumentation
diagrams and found that no relationship existed. One of the pro-
cess operators that worked in that plant area countered that he
had observed a relationship between the variables, and a small
disagreement ensued. The operator contended that even though
he had not studied the diagrams, he had watched the plant oper-
ate for several years and could see that the two variables reacted
to each other. A more detailed study revealed that the operator
was indeed correct. This is the type of nuanced, experience-based
knowledge that these laborers bring to their jobs. It represents
an entirely new form of craftsmanship.

Unfortunately, because of tradition, the information that is


being presented to the majority of these craftsmen is essentially
the same as was presented to their predecessors. In the meantime,
the level of control exercised by automation systems has
increased significantly, giving the process operators more avail-
able time. Unfortunately, the mindset of management has been
to avoid giving them any more responsibility because that would
mean giving up authority.

Yes, this new workforce is skilled and educated. They have


built an information base on the operation of the plant that
nobody else possesses. Yes, they are an expensive resource com-
pared to labor in the developing areas of the world. But the labor
force in the newly industrialized regions does not possess this
new form of craftsmanship. If the manufacturers in the industrial-
ized world want to compete with the manufacturers in the devel-
oping regions, they must use this craftsman-based labor force in
an entirely different way than the unskilled and uneducated labor

B O T T O M - L I N E A U T O M A T I O N 41
BottomLine05.book Page 42 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

force is used. These operators and laborers can accomplish more


than basic process and machine control; they can now be used to
manage the performance of the entire plant. But while they have
the skills and experience to manage plant performance, they lack
two important criteria: information and management commitment.

Taylorism has outlived its usefulness. Today, we need a new


paradigm for plant performance. We need a paradigm that makes
full use of the new and valuable resources available in our plants:
the information craftsmen that some still think of as mere labor-
ers. Ideally, this new paradigm will result in the closing of the
management/labor schism by empowering the labor force with
information. Clearly, this would represent a very uncomfortable
trend for traditional management since, as Shoshana Zuboff
points out, the prospect of shared information threatens the dis-
tinction between managers and the managed. For many manag-
ers, sharing information and maximizing opportunities for all
members to become more knowledgeable is felt to be akin to trea-
son.14

Despite resistance, the new operations


Past Future paradigm is emerging through a variety
Lights Out Automation Automation & Information of subtle and not-so-subtle transitions
Management vs Labor Empowered Teams now arising in many of the leading
Hierarchy Lean and Flat process manufacturing plants (see fig-
High Overhead No Overhead ure 2.5). The focus on replacing opera-
High Paper Usage Electronic Communications tions personnel with automation
Operation by Exception Management technology is giving way to a focus on
providing appropriate levels of
Figure 2.5 Transitions in automation and information so that process operators can effec-
Operations
tively utilize their new information craftsmanship to help drive
the plants performance. The management/labor schism is begin-
ning to be replaced by empowered work teams or quality teams
chartered with driving continuous process improvement through-
out their operations. The layered hierarchical organizational
structures that buried plant operations under piles of bureaucracy
are being superseded by much leaner and flatter organizational
structures. The overbearing overhead associated with hierarchical
organizations is giving way to much more action-oriented opera-

42 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 43 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

tions. Digital computer technology is making possible the replace-


ment of the traditional stacks of paper and reports with electronic
communications that provide needed information much faster
and free up operators time to focus on performance. The net
result is that the basic operators job is changing from the tradi-
tional operation-by-exception model to a model of performance
managers who are responsible for driving the performance of the
plant.

There is one major element missing in this transition, how-


ever. If the operators are to become performance managers and
drive plant performance in the right direction, they must have
continual knowledge of what the plants performance is. This
type of economic performance data has traditionally been kept
from operators. For this transition to the new operations para-
digm to reap its true potential benefits this data must not only be
made available to the operators but also in a time frame that is
consistent with the job they doand that is real time.

Given the advanced state of computer-based technologies


today, real-time performance information can now be made avail-
able to everyone in an operation, right up to the front lines. Once
this information is made available, once the organizations atti-
tude is adjusted to consider the labor force as craftsmen and
performance managers, and once authority is delegated to the
workforce, plants will perform as never before. Concepts like
lights out manufacturing in process plants will be relegated to
the history books where they belong. And a new era of manufac-
turing performance will begin.

TECHNOLOGY FOR THE BOTTOM


LINE
When the digital computer was first introduced to process
manufacturing in the late 1960s, its promise was unbounded.
Many manufacturing managers saw computer technology as the
key to driving the performance of their plants to new levels and
gaining real competitive advantages in their manufacturing lines.
For the most part, however, after over thirty years of applying this

B O T T O M - L I N E A U T O M A T I O N 43
BottomLine05.book Page 44 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

technology in the process industries this vision has still not been
realized.

As our interviews with senior manufacturing managers pro-


gressed over the past decade it became very clear that there was a
major disconnect between the executives running the operations
and the technologists responsible for the automation systems.
When asked what their primary motivators were when they pur-
chased automation systems, the managers stated that they were
the desire to:

Improve plant quality


Improve safety
Increase manufacturing flexibility
Improve operations reliability
Improve decision-making
Improve regulatory compliance
Increase product yields
Increase productivity
Increase production
Reduce manufacturing costs

Although few would argue with this list, these criteria are sel-
dom taken into consideration during the actual purchase of an
automation system or during the systems life cycle.

Most of the listed criteria have a direct impact on the ongoing


economic performance of the manufacturing operation. In spite
of this, in most cases the only economic factor associated with
the use of computers in process plants has been price and price
alone. Even at first glance, however, it is clear that price is but one
of the economic variables involved in the assessment of the eco-
nomic returns of an automation system. When the decision-mak-
ing and major capital investments for ongoing improvement
processes are left to technologists, price typically becomes the
only economic variable in the equation. This price focus has
turned industrial automation systems into commodities in the
eyes of the manufacturing operations. As commodities, the value
of the technology is only assessed when obsolescence makes the
ongoing cost of keeping the system running unacceptable. This
commodity focus is a major contributing factor to the disappoint-

44 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 45 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

ing economic returns that have generally resulted from automa-


tion.

In recent years, there has been a growing trend among manu-


facturers to try to develop ongoing relationships with automation
suppliers. Although volume price discounts have been one key
aspect of these relationships, they have also often caused the
manufacturers and automation suppliers to take into account a
larger range of economic factors than just price. The first new fac-
tor to be scrutinized was lifecycle costs. That is, instead of just
analyzing the current price of a system, manufacturers would esti-
mate the entire cost of the system over its useful life span (see fig-
ure 2.6). The basic equation for analyzing lifecycle cost is as
follows:

LCC = Price + Project Engineering + Installation +


NPV (Ongoing Annual Costs)

where
LCC = Lifecycle Costs
Price = Automation system price
Project Engineering = Total cost to engineer the project
Installation = Total cost to install the system
(including start-up)
NPV = The net present value function of the ongoing costs
Ongoing Annual Costs = The annual engineering,
operations, and maintenance cost of the system

The net present value function calculates the time value of


money over extended periods.

This lifecycle cost perspective challenged automation suppliers


to provide features in their automation systems that would reduce
the initial engineering and installation costs as well as the ongo-
ing operation and maintenance costs. Generally, lifecycle costs
tend to be high at the beginning of the life cycle due to the
expenses of initial engineering, installation, and the products
purchase price. After start-up, they level off for several years, but
then start to rise toward the end of the lifecycle because of the
costs associated with obsolescence: repairs, retraining, and other

B O T T O M - L I N E A U T O M A T I O N 45
BottomLine05.book Page 46 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

maturity costs. Note that the


system price tends to be a
reasonably small compo-
nent of the overall automa-
$ tion system cost. In fact,
studies have placed the aver-
P age product price at less than
r
i 35 percent of a typical
c
e project's cost, without even
taking into consideration the
Time ongoing costs. This
Figure 2.6 Lifecycle Cost demonstrates one of the deficiencies of the price-only approach.
Profile
The expansion of economic perspective from price to lifecycle
costs was a major step forward for industry, but that perspective
was still very limiting. A cost-only focus tends to relegate the
automation system to the category of a necessary evil. From
this perspective, the only economic justification for upgrading or
replacing a system is that the cost of maintaining the current sys-
tem is too high because it has become obsolete. This is the per-
spective of most industrial automation users today; most of them
still use a cost-only economic evaluation standard. If there is no
perceived economic benefit to the manufacturing operation from
installing an automation system then that system is certainly not
meeting managements objectives, and probably nothing but the
most rudimentary system should ever be deployed.

Many automation system users have talked about determining


the returns on their automation investments, but to calculate
returns you need more than the lifecycle costs: you need to have
an accurate measure of the lifecycle benefits the manufacturing
operation derives from using the automation system. In the inter-
views we conducted with manufacturing executives, we found
none who were actually measuring the benefits created by auto-
mation. Most admitted that they did not know how to get at
these metrics in any reasonable way and that their current cost
accounting systems did not provide detailed enough economic
data to be able to infer the benefit value.

46 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 47 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

The economic benefits automation systems create for manu-


facturing operations essentially occur in two major areas. The first
is the manufacturing cost savings that result from reduced power
consumption, reduced raw material costs, and reduced manpower
requirements. The second is the increase in production that can
be gained through better asset utilization. Of these, the only one
that manufacturers regularly monitored was the reduced man-
power automation made possible, because it is relatively easy to
measure. The other elements of the benefit calculation are vari-
ables that constantly change as the products are being produced
and are, therefore, very difficult to measure.

If all these economic benefit variables were measured, the


return on the automation investment could be determined
through the following equation:

Economic Return on Automation = Lifecycle Benefits Lifecycle Costs

This equation represents the classic investment profile for any


capital investment offering a return (see figure 2.7). In theory, the
benefit of most capital investments should be capable of being
managed so that they provide either constant returns over the life
cycle or, as in the case of automation systems, continuously
increasing economic benefits over the life cycle.

As the concept of lifecycle economic profiles for automation


systems began to gain recognition during the late 1990s,
the author was asked to lead a
session on the subject at the Life-cycle Benefit
1996 ISA (The Instrumenta-
tion, Systems, and Automa-
tion Society) Technical $
Conference. At that conference
and at subsequent meetings, a P Life-cycle Cost
r
number of professionals from i
c
companies such as E. I. e
DuPont, General Foods, Eli
Time
Lilly, and Dow Chemical
contributed much data from several automation projects that Figure 2.7 Lifecycle
Economic Profile
helped us build a lifecycle economic profile. A high-level view of

B O T T O M - L I N E A U T O M A T I O N 47
BottomLine05.book Page 48 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

the profile we used is shown in figure 2.8. This profile defines the
benefit of an automation system as the net present value (NPV) of
the annual manufacturing cost savings and the annual produc-
tion increases that result from the automation system. The net
present value is a function that calculates the current value of
money paid on a regular basis over a number of years. It is appro-
priate when trying to make an up-front decision about the best
economic value offered by a number of possible choices that will
either be generating economic costs or benefits for a set period.
An automation system does both. YEL (years of expected life) is
an indicator that the NPV
Economic Profile = Lifecycle Benefits- Lifecycle Costs calculation should be done
over the number of years
(
Annual Annual
NPV Mfg. Cost + Production
YEL Savings Increases ) that the automation system
is expected to operate in the
plant. This helps to deter-
Annual Annual Annual
System + Initial Eng. + Installation +
Price Costs Costs
NPV
YEL
( Eng. + Ops. + Maint.
Costs Costs Costs
) mine the value of automa-
tion systems that may have
different expected life cycles
Figure 2.8 Lifecycle once installed. The lifecycle cost calculation has two basic compo-
Economic Model
nents: the project costs and the ongoing costs. The project costs
can be captured by the three general categories of price, initial
engineering costs, and installation costs. The ongoing costs can
also be calculated using the net present value function; they
include annual engineering costs, annual operations costs, and
annual maintenance costs.

As part of my data-gathering project I analyzed several auto-


mation projects and from the data collected an actual economic
profile emerged. Considerable data was available for the cost side
of the profile, but only limited data was available to analyze the
benefit side. This is not surprising given the lack of focus indus-
try-wide on the benefits component of the equation. The analysis
of this data produced a number of interesting results. First, as fig-
ure 2.9 depicts, most of the projects for which the benefit side was
measured at all, showed a continuous decline in the benefit value
over the systems life cycle. This result was shared with several of
the professionals who contributed the data; they were not overly
surprised by it. Many of them attribute the continuous benefit

48 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 49 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

decline to the gradual obsoles-


cence of the system. This con-
clusion was later demon-
Life-cycle Benefit
strated to be false since many
projects do realize continuous
$
improvement in the eco-
nomic benefits derived from
P Life-cycle Cost
the automation system. In r
i
fact, the decline seems to be c
e
associated more with the lack
of measurements of the bene- Time
fit side than an actual decline
caused by obsolescence. One of the basic principles of process Figure 2.9 Lifecycle
Economic Profile Results
control is that if you cant measure the controlled variable, you
cant control it. The same is true for financial variables such as
lifecycle benefits. It also appears that if the variable is not mea-
sured, it will most probably move in the wrong direction, which is
exactly what the data showed.

Another interesting aspect of


System Initial Eng. Installation Annual Annual Annual
the data we collected was that the Price Costs Costs Eng. Ops. Maint.
Costs Costs Costs
lifecycle cost data was actually
distributed across the set of data, Average first 5
23.2% 28.5% 16.1% 9.3% 7.6% 15.3%
year % of cost
as shown in figure 2.10. For
example, the price, which had
traditionally been the primary economic variable in the automa- Figure 2.10 Lifecycle Cost
Breakdown
tion system decision process, represented less than a quarter of
the cost of the first five years of system use. Between the initial
engineering costs and the five-year lifecycle engineering costs, the
engineering of the automation system accounted for 37.8 percent
of the five-year costsconsiderably greater than the purchase
price of the system. Perhaps the most interesting result of the
analysis came on the benefit side of the model. The benefit data
collected, which was not statistically valid due to the small sam-
ple size of projects for which benefit data was available, showed
that the benefit-to-cost ratio over the first five years was 3.4:1.
This means that the average return on automation technology for
projects for which the benefit data was measured was very strong.
The analysis team did not believe that the 3.4:1 ratio was repre-

B O T T O M - L I N E A U T O M A T I O N 49
BottomLine05.book Page 50 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

sentative of the average return on automation realized in indus-


try. In fact, the automation users who were measuring the benefit
that automation provided were considered to be among the best
performers. This means that the 3.4:1 ratio is a best-practices
result rather than an average result.

When this data was shared with a larger number of automa-


tion users, the consensus seemed to be that these results were not
surprising. In fact, most of those interviewed readily accepted the
results. This caused us to ask why more users werent focused on
the economic benefit created by automation. Most of those we
asked did not have any specific response, but when we followed
up with a more detailed analysis we uncovered three industry
practices that are significant barriers to an effective total lifecycle
economic approach to automation.

The first barrier was the replacement automation approach to


automation projects, which is almost universally used in industry.
This approach begins at most industrial plants when their capital
budget for an automation system upgrade is approved. At that
point, a project team is typically established to determine the
specifications for the new automation system to be installed. In
most of these cases, the specification is developed by looking at
what the currently installed system does and then building the
new system specification around it. This leads to requests for pro-
posal (RFP) seeking competitive bids from automation suppliers,
requests whose specifications exactly describe the old system
already in place. The suppliers know that the one who will win
the order is the one who meets the specification at the lowest
price. Thus, even though they may have added many perfor-
mance-enhancing capabilities to their computer-based automa-
tion systems since the old system was installed, they propose the
lowest cost system they can. Typically, this means that all of the
advanced capability the suppliers have invested in is left out of
the proposal. The net result is that the new system being installed
is an exact functional replacement for the system being retired.
Replacing old technology with new technology that does the
exact same thing seldom leads to breakthrough improvements.

50 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 51 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

When this phenomenon is pointed out to the users project


team their response is typically, dont worry, once the new sys-
tem is up and running we will take full advantage of all of the
advanced capabilities. This is when the second barrier to a total
lifecycle economic approach to automation starts to take effect:
the project team approach. Project teams of highly qualified engi-
neers are established when capital budgets for automation
projects are approved. The project team works on the project
throughout the project cycle. Once the project is initiated, the
project team goes away. Some of the members may go on to other
projects. Some may remain in the plant to manage ongoing engi-
neering activities. But the human resources required to take
advantage of the advanced features of the automation system are
now no longer available. As a result, the later on that everyone
was hoping to take advantage of never happens. The initial bene-
fit, if any, provided by an automation system is typically the best
benefit that will be realized from the system over its life cycle.
From that point there is a continuous degradation of benefit.

The good news in all of this is that while this continuous deg-
radation of economic performance goes on, nobody knows it.
This is because in most cases the benefit side of the economic profile
is not measured. This is the third barrier to success. Measurement
systems are typically only valued if they are providing good news.
The news that might be provided by a measurement system that
continuously communicates the economic benefit due to auto-
mation would probably not be very positive. But since these mea-
surements are not being made, most manufacturers do not feel
the pain of poor performance. In this case, no news is good news.
Or is it? This approach was fine when process manufacturers
could not help but make significant profits no matter how good
or bad their operations were. But in todays tough global busi-
ness environment, this approach is just not acceptable.

One promising result of all of this analysis was that when the
professionals in industrial operations paid attention to the eco-
nomic benefit measureseven if poorly calculated on an infre-
quent basisthey were able to realize phenomenal results. What
if these econometrics could be made and provided to the opera-

B O T T O M - L I N E A U T O M A T I O N 51
BottomLine05.book Page 52 Friday, August 19, 2005 9:54 AM

C H A P T E R 2 : T E C H N O L O G Y A N D T H E B O T T O M L I N E

tions personnel in real time? How much better could the opera-
tion perform? Research has shown that the 3.4:1 benefit-to-cost-
ratio is a very conservative indicator of what could actually be
accomplished through better decision-making based on better
and more timely information. The real-time econometrics of
plant performance are called Dynamic Performance Measures
(DPM). We will address the implementation and use of DPMs
later in this book.

NOTES
1. Shaw, William T., Computer Control of Batch Processes. Cock-
eysville, MD: EMC Controls, 1982, page 100.

2. Peters, Thomas, and Robert H. Waterman Jr., In Search of


Excellence: Lessons from Americas Best-Run Companies. New
York: Harper & Row, 1979, page 39.
3. Zuboff, Shoshana, In the Age of the Smart Machine. New York:
Basic Books, 1988, page 31.
4. Drucker, Peter F., Innovation and Entrepreneurship: Practices and
Principles. New York: Harper & Row, 1985, page 90.
5. Zuboff, In the Age of the Smart Machine, page 33.
6. Zuboff, In the Age of the Smart Machine, page 35.
7. Zuboff, In the Age of the Smart Machine, page 41.
8. Zuboff, In the Age of the Smart Machine, page 42.
9. Zuboff, In the Age of the Smart Machine, page 42.
10. Zuboff, In the Age of the Smart Machine, page 42.
11. Zuboff, In the Age of the Smart Machine, page 42.
12. Zuboff, In the Age of the Smart Machine, page 44.
13. Zuboff, In the Age of the Smart Machine, page 45.
14. Zuboff, In the Age of the Smart Machine, page 238.

52 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 53 Friday, August 19, 2005 9:54 AM

CHAPTER 3

Quality and
the Bottom
Line

THE QUALITY TREND


The idea of quality in manufacturing has probably undergone
more change over the past fifty years than any other in industry.
Perhaps the primary reason for this is that quality is more of a
concept than a well-defined attribute and one that can apply to
many different things including products, organizations, compa-
nies, and individuals. In fact, the word quality has been used
interchangeably to describe all of the above.

The interviews Foxboros


research team conducted with
Technology
senior manufacturing executives
Quality Quality Quality
revealed that the quality trend in for for for the Bottom
Manufacturing Technology Bottom Line
manufacturing has undergone three Quality Line
Improvement
major transitions in the past half
century (see figure 3.1). Most of this Accounting

emphasis on quality has focused on


maintaining and improving specific
measurable attributes of products being manufactured. This Figure 3.1 The Quality
Trend
period had been referred to as the Quality for Manufacturing
period of the quality revolution. During the late 1980s the focus
seemed to turn, at least in the process industries, to a much stron-
ger technology emphasis that was based on statistical process con-

B O T T O M - L I N E A U T O M A T I O N 53
BottomLine05.book Page 54 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

trol (SPC) tools. This period has been referred to as the Quality
for Technology phase. Currently, there is a shift underway to a
more performance-focused view of quality that is significantly
changing the way quality is being addressed in manufacturing
operations. It is referred to as the Quality for the Bottom Line
phase.

The concept of quality has been important for as long as peo-


ple have made products. But it is a very difficult concept to quan-
tify. If a product exhibits the characteristics that the market needs
or wants, it is considered to be of reasonable quality. The more
desired characteristics a product exhibits, the higher its quality is
deemed to be. Because almost all markets are dynamic, however,
with needs and wants that change over time and across segments,
this relative definition of quality means manufacturers may be
uncertain at any given time that their products have the neces-
sary quality. Thus, one of the first steps in measuring the quality
of a product is to develop a comprehensive understanding of the
characteristics the market wants.

The quality issue has only grown in importance as globaliza-


tion has impacted many segments of industry in the past ten
years. This globalization has manifested itself in two major ways.
First, manufacturers have begun to encounter new competition in
market segments where there have traditionally been only estab-
lished, well-known competitors. New competitors frequently alter
the markets perception of what constitutes quality because they
introduce the consumer to a wider range of products, influencing
their understanding of product characteristics, including quality.
Staying on top of dynamic new definitions of product quality
requires great effort on the part of manufacturers and, in turn, a
corresponding degree of flexibility in their manufacturing pro-
cesses. Even more challenging, these new product modifications
require manufacturers to continuously monitor the market to
determine whether their products' characteristics are those con-
sumers require at any given time for any market segment.

The second aspect of the globalization of the marketplace is


that manufacturers are being forced to market their products in
geographical markets they had not previously participated in.

54 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 55 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

When they move into these new markets manufacturers often


find that their new customers value different product characteris-
tics than customers in their traditional segments. This puts addi-
tional pressure on the manufacturer to accommodate the new
requirements. Thus, todays globalized market makes it more diffi-
cult than ever to manufacture high-quality products on a consis-
tent basis.

QUALITY FOR MANUFACTURING


The first reasonably vigorous approach manufacturers took to
ensure product quality has traditionally been referred to as qual-
ity assurance (QA). Very simply, QA means checking that the
required characteristics of a product were properly built in during
manufacture. If the required characteristics are present, the prod-
uct is released to the marketplace and sold. If they are not, then
the product must be reworked, sold at a different grade (and usu-
ally at a lower price), or discarded. Whichever of these options is
chosen, there is a significant penalty to manufacturers for turning
out poor quality products.

This traditional QA approach made sense in the days when


manufacturing operations made only one product and market
needs seldom changed. But in todays rapidly changing manufac-
turing environment, the static QA approach is unsatisfactory
because a dynamic environment requires dynamic monitoring
and control.

Walter A. Shewart pioneered an alternative approach to QA in


the 1920s and 1930s while working to improve the quality of tele-
phone equipment at AT&T Bell Laboratories. As noted in chapter
1, Shewart authored the book, Economic Control of Quality of Man-
ufactured Product, which presented a structure for controlling
product quality by using statistical analysis of the extent by
which each manufactured product varied from its specification.
Shewart argued that statistics could help manufacturers identify
nonrandom and unexpected faults in the manufacturing process
that lead to poor quality. Once identified, the faults could be

B O T T O M - L I N E A U T O M A T I O N 55
BottomLine05.book Page 56 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

repaired, and the process would make products with the expected
quality characteristics. Shewarts work gave birth to a new field
that became known as statistical quality control (SQC), a key
moment in the quality trend that continues to the present day
(see figure 3.2). He successfully implemented an SQC program at
Western Electric Company in New York.

The basic concept behind SQC is


Economic Control of Quality of that manufacturing machinery can
Manufactured Products
World War II
be combined into a process that,
Agricultural Walter A. Western Electric Co. W. Edwards Joseph
Statistics Shewart Rochester, NY Deming Juran when it is working well, will pro-
1920 1930 1940 1950 1960
duce products within an accept-
able range of the products desired
SQC Fever
characteristics. As long as every
6 Lean
Process
Japanese Manufacturing Success Imitation TQM Performance Enterprise machine in the process is working
Focus
1970 1980 1990 2000 within its designed limits the
resulting product will be within
Figure 3.2 The Evolution of specification. It is very important to note that even when all
Quality Improvement
machines are working well, however, the resulting products will
not all be identical. There is no such thing as an absolutely perfect
manufactured product that exactly matches design specifications.
The physical world just doesn't work that way. That is why quality
characteristics are often quoted in terms of tolerance bands rather
than as absolutes.

For example, if a drill press is part of a manufacturing process


for making nuts, two of the key actions of the drilling operation
will be to make sure that the hole is in the middle of the piece
and that it has the correct diameter. Though the specification for
the holes diameter may be one-quarter inch, if we actually mea-
sured the hole in any given nut it could vary, in a very small way,
around one-quarter inch. However, such variance is not consid-
ered to violate the products basic specification. The expectation
is that if the actual diameters of all the manufactured nuts are
mathematically averaged, and assuming that the drill press is
working properly, then the average would be one-quarter inch,
even though no single part would exactly meet that specification.
In other words, if the standard deviation of the holes diameters

56 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 57 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

from the specification of one-quarter inch is small enough, a large


percentage of the parts are considered to be within tolerance.

If a products specifications are seen as absolute values, then


no single manufactured piece will be without error. Machines
cannot produce a product, and measuring devices cannot mea-
sure any products characteristics, to such an ideal specification.
What this implies is that some degree of error in manufacturing is
acceptable, and in fact expected. These errors are normally distrib-
uted about the mean, which should equal the specification with a
reasonably small standard deviation. This acceptable deviation
corresponds to the tolerance. If we measure a product in this way
and our measurements exhibit this expected and tolerable degree
of error, then the manufacturing process is said to be in statistical
controlthe machinery is working as well as we can expect. As
W. Edwards Deming points out: statistical control does not
imply absence of defective items. It is a state of random variation,
in which the limits of the variation are predictable.1

When SQC is discussed in terms of manufacturing, the focus


of the quality control is on the manufacturing process and based
on statistics generated by the indicators of the quality of the prod-
uct. The product can only be as good as the process that manu-
factures it. If each part of the manufacturing process is working at
the expected level, the manufacturing line, as designed, is operat-
ing as well as it can be expected to. If this is not good enough, the
processing equipment must be changed to introduce a technol-
ogy that will enable the specification to be met. Because of this
focus on the process, the use of SQC tools to provide the informa-
tion that allows the operators of a manufacturing process to oper-
ate at specification and within its limits of variation is known as
statistical process control (SPC).

Quality problems in manufacturing arise when the products


exhibit unexpected variations from the quality specification.
When this occurs, some of the manufactured pieces will be of
unacceptable quality because they do not meet the customer's
valid requirements. If the quality errors occur randomly, then the
existing process equipment is probably working close to its limits.

B O T T O M - L I N E A U T O M A T I O N 57
BottomLine05.book Page 58 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

If the quality errors do not occur randomly, the cause is probably


in the process and needs to be isolated and fixed.

Returning to the example of the drill press, suppose that with


time the drill bit becomes overheated, causing its shank to bend
slightly. The holes that the drill produces will become larger than
normal, and the result will be nuts of unacceptable quality. A sta-
tistical profile of the resulting parts would show that the distribu-
tion is no longer normal, with the same characteristics clustered
around the desired specification, and that the average and/or
standard deviations are larger than expected. This would lead us
to conclude that an unusual situation in the drilling process is
causing this unexpected result. Identifying and repairing the
cause in the manufacturing process should result in a return to
quality production, at least until another unexpected problem
occurs.

The statistics, therefore, are used not to identify the existence of


variation in the quality characteristics, since variations exist in every
manufactured product, but to identify unexpected errors caused by an
unusual circumstances in the manufacturing process. If the cause can
be identified, the process can be repaired and returned to produc-
ing parts with acceptable quality characteristics. Shewart felt that
we should use statistics at every stage of the manufacturing pro-
cess to identify and fix process errors as they occur, thus minimiz-
ing the amount of off-spec product manufactured. This
statistical approach allows manufacturers to control the quality of
product as it is made rather than waiting to check until the end of
the process, as with QA.

One other key aspect of Shewart's work his realization that it


was impractical to measure each manufactured part as it is made,
so to make the statistical approach more pragmatic he suggested
sampling groups of parts randomly and then applying sample sta-
tistics to these groups. He also suggested using statistics other
than just the average and standard deviation to provide a more
complete profile of what was actually happening in the process.

Although Shewart's approach seems to be very clear and


extremely practical, manufacturers were very slow to adopt it.
Very few plants seem to have used SQC in any systematic way

58 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 59 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

before the 1950s. Perhaps this is because there was such a great
demand for manufactured goods and such small supply that
product quality was not as much of an issue as product volume.
Also, perhaps the dynamics of manufacturing and the market-
place didnt call for much change. Whatever the reason, like so
many other novel ideas, Shewarts SQC was not widely imple-
mented in its originators lifetime.

SQC IN MANUFACTURING
The circumstances in Japan after World War II created an ideal
environment for implementing new concepts in manufacturing.
Much of the Japanese manufacturing capability had been
destroyed, and certainly the Japanese economy was in ruins.
Rebuilding Japan was a top priority both within Japan and for
other countries, such as the United States.

The United States invested heavily in the rebuilding effort


with both financial and human resources. One of the most
renowned of these human resources was a statistician by the
name of W. Edwards Deming. Deming was a talented scientist
who had studied Shewart's SQC method while working with him
at Western Electric Company. He found in Japan an ideal test bed
for implementing SQC in actual manufacturing environments.

Deming offered this new SQC approach as a way to help Japan


reestablish a manufacturing base, and in fact a dominant position
in the world. As he said, I think that I was the only man in 1950
who believed that the Japanese could invade the markets of the
world and would within five years.2 And the rest, as they say, is
history.

The implementation of SQC techniques did not go as


smoothly as many would have us think. Although the statistical
approaches were found to be very revealing, getting manufactur-
ing people to act on this information was not easy. The laborers
in these plants did not necessarily see it as part of their job to
study statistical charts, and engineering staff tended to be too
scant and focused on other issues to do the job on their own. It

B O T T O M - L I N E A U T O M A T I O N 59
BottomLine05.book Page 60 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

became clear that quality was more than a technical issue; it was also a
people issue and an organizational issue. Statistics alone cannot
assure reduced variation, Deming wrote; it takes everyone, manage-
ment and labor working together.

One approach to getting workers to use the SQC information


was to develop teams around each work cell in a manufacturing
operation and give them responsibility for managing the product
characteristics that impacted the product qualitythe quality
indicatorsfor their cells operation. This approach achieved
some success and justified the SQC approach, but it did not
accomplish all of the desired results.

One of Shewarts concepts was that identifying and elim-


inating the causes of nonrandom, unexpected error in
the manufacturing process on an ongoing basis would
lead to continuous improvement of quality. He built a
simple model based on a cycle for continuous improve-
Act Plan
ment, often referred to as P-D-C-A for Plan-Do-Check-Act
(figure 3.3). The model suggested that quality improve-
ments would be the inevitable result if the statistics were
Check Do used to identify process problems and plan for their cor-
rection, then execute that plan, check to see if the plan
worked, and act on the results. Since the goal is continu-
ous improvement, once the cycle is complete you cycle
Figure 3.3 PDCA through it again and again, improving the process, perhaps just a
little, each time around.

Deming referred to this P-D-C-A model as the Shewart Cycle,


although most others today refer to it as the Deming Cycle. In
fact, it appears that Deming was the first to implement this
approach on any large scale through his work with quality circles.
As a result, continuous improvement became one of the catch-
phrases of the quality movement.

This concept of continuous improvement required a signifi-


cant change in management philosophy, which was dominated at
the time by Scientific Management and are still common today.
This change included delegating responsibility and authority for
managing the quality indicators right down to the frontline
workers in the manufacturing operation. These are the people

60 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 61 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

with the most pragmatic and detailed understanding of their seg-


ment of the operation and are, therefore, in the best position to
identify and correct the conditions in the process that cause qual-
ity problems. Scientific Management was based on treating the
laborers as repetitive machines, and not insightful personnel.

Deming was enormously successful in implementing statistical


techniques for quality control in manufacturing plants. The basic
quality circle approach, however, didnt seem to go far enough to
motivate people to work spontaneously toward continuous
improvement, as he had hoped. One major reason for this was
that delegating the appropriate level of authority to the quality
circle required a management approach that was somewhat revo-
lutionary at the time. The bottom-up perspective implicit in the
idea of quality circles did not spontaneously prompt a corre-
sponding change in management philosophy.

Four years after Deming went to Japan, another pioneer in the


quality movement, Joseph Juran, lectured in Japan on managing
for quality improvement. Juran also had a keen understanding of
Shewarts techniques and had also worked with him at Western
Electric Company. Rather than focusing on statistics, Juran
focused more on the organizational aspects of a manufacturing
operation, and in particular the top-down management accep-
tance and directives that would lead to the continuous improve-
ment of quality and make it an inherent part of the
organizational philosophy.

Juran believed that if quality improvement concepts were to


reach their full potential, a fundamental change had to occur in
manufacturing operations, starting at the highest management
levels. He realized that the quality implementation had to occur
at the front lines of the operation, that the people on the front
lines had to have the culture and structures necessary to do their
job, and that there had to be a system of recognition to match.
Juran proposed that organizations would manage effectively for
quality by allocating the responsibility for quality management to
the frontline workers, encouraging teamwork to solve the quality
problems, and then recognizing a job well done. As obvious as
this approach may seem to us, it was diametrically opposed to tra-

B O T T O M - L I N E A U T O M A T I O N 61
BottomLine05.book Page 62 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

dition. Traditionally, management ordered and labor did,


period. To bring labor integrally into the management process
was revolutionary.

To prove his theories, Juran set up a total management


approach for quality. His approach included an aggressive effort
on managements part to plan the quality structure, a structured
approach to implementing P-D-C-A, and an effort to change the
way people worked to incorporate the quality concepts in day-to-
day activities. The structure for managing quality represented an
entirely new management and organizational opportunity for a
plant or factory. Many operations actually set up two separate
organizational chartsone for the traditional functional opera-
tion and one for the quality operation, even in reality though the
two structures are literally inseparable.

Juran proposed the formation of a quality council at the


executive level to plan and administer the quality improvement
process. Under the quality council were one or more levels of lead
teams who oversaw the quality program in each functional
segment of the operation. At the lowest level were the quality
teams that actively cycled through the P-D-C-A model to improve
quality.

The quality teams worked with a structured


problem-solving approach to identify and cor-
rect quality problems in a well-defined, system-
Quality
Planning atic process. The quality problems were defined
as those issues that negatively impact valid cus-
tomer requirements. As the number of cor-
rected problems rose, so too did the quality of
the product and organization, because more
Quality Quality valid customer requirements were being met.
Improvement Control
After the quality planning was done by the
quality council and the quality improvement
was implemented by the various teams, the
Figure 3.4 Juran Trilogy third part of the Juran Trilogy (see figure 3.4) was day-to-day
quality control. This entailed making P-D-C-A a normal part of
everyday work. The result was a change in culture to a true focus

62 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 63 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

on the customer's valid requirements by everyone throughout an


organization.

Through this vigorous approach, Juran showed that improved


quality involved both effective tools, such as the statistical tools,
and a management approach and structure to effectively imple-
ment the tools. As James Gagne has noted, Juran recommends
using statistical process control but warns that it can lead to a
'tool oriented' approach.3 Moving to high quality is not just a tech-
nical issue; it is even more a management issue. Manufacturers who
take this view show remarkable results and become global manufactur-
ing leaders.

Incidentally, it would be a mistake to leave the impression that


Dr. Deming only worked with the statistics and did not address
the management requirements of a quality program. In fact, he
developed many of the same management conclusions as Dr.
Juran.

The importance of the quality movements transition from a


statistical approach to a management philosophy cannot be over-
stated. Today, many companies refer to their quality program as
Total Quality Management (TQM) rather than SQC. The manage-
ment implications of an organization moving to a TQM approach
are enormous. The traditional top-down, militaristic management
approaches that have been so ingrained in the manufacturing cul-
ture are directly opposed to a true TQM approach.

The TQM approach totally redefines the relationship between


management and labor, and in fact, almost eliminates the tradi-
tional connotation of labor. Frontline workers must become part
of the management team, as must the entire organization, for a
TQM structure to be most effective. Each individual must be
viewed by the operation, and must view him- or herself, as an
important manager of quality improvement.

The concept of quality has changed over the past fifty years. It
has evolved from fixed product characteristics to a whole corpo-
rate philosophy based on continuous improvement. Those
manufacturers who have taken seriously the total quality

B O T T O M - L I N E A U T O M A T I O N 63
BottomLine05.book Page 64 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

improvement approach have moved into an elite category of


world-class manufacturers.

QUALITY FOR TECHNOLOGY


Over the past few decades, Japanese companies have moved
into dominating positions in a number of manufacturing sectors.
As this happened, companies from the other industrialized areas
of the world have tried to emulate the Japanese success factors.

Customers with firsthand knowledge of the continuously


improving Japanese quality began putting pressure on traditional
manufacturers to provide products with the same quality. Imme-
diately, traditional manufacturers wanted to be seen as stepping in
line with the new and successful Japanese approaches. They
wanted to demonstrate to their customers that they were using
the same approaches as the high-quality competitors.

Quality management, as Juran proposed it, is not overly visible


on initial inspection, but the tools of SQC are highly conspicuous.
As a result, during the 1980s an epidemic of SQC fever started
to sweep the industrialized world. Statistical charts started pop-
ping up on office walls and corridors across the globe as proof of
corporate commitment to continuous improvement.

It is not clear how much of an improvement in quality most


manufacturers attained. As a matter of fact, many didnt even
appear to be using or even looking at the statistical charts their
efforts generated. The stated objective was to implement a quality
improvement program, but in many cases attention was paid to
window-dressing, demonstrating to customers that a quality pro-
gram was in place. Implementing a true and comprehensive qual-
ity management program did not seem to be much of a
consideration.

What happened was exactly what Juran had feared. Manufac-


turers focused on the tools, not on a comprehensive program to
continuously improve quality. Perhaps it was this mindset that
led Dr. Hans Bujaria, president of Multiface, Inc., to point out in a
discussion on SPC that seven years and billions of dollars later,

64 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 65 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

many users are still waiting for appreciable return on their invest-
ment in SPC.4

This SQC fever is similar in many respects to the technology-


for-its-own-sake mentality that governed the application of com-
puter technology to the control of manufacturing operations, as
described earlier. This time the technology was based on statis-
tical techniques and tools. Obviously, without a firm commit-
ment to quality improvement, SQC fever was not destined to
succeed. Worse yet, the absence of appreciable results tended to
prove to the skeptics that it was not really worth investing in SQC
and that the quality movement was more of a fad than a trend.
The fact is, since most were treating it as a fad their words became
a self-fulfilling prophecy.

SQC FEVER IN THE PROCESS


INDUSTRIES
The original SQC statistical tools and techniques were devel-
oped for discrete parts manufacturing operations. Even Shewarts
early work was almost entirely focused on the discrete manufac-
turing segments. The operations at Western Electric Company,
where Shewart originally tested his concepts, manufactured such
items as electromechanical switching relays for telephone sys-
temsa classic discrete manufacturing operation.

In discrete manufacturing operations, products tend to be


made by machining and assembling pieces and parts. These oper-
ations are naturally discrete since the same basic functions are
performed on parts that are physically separate or distinct. Apply-
ing sample statistical techniques in this type of environment is
relatively straightforward. The statistics developed for basic SQC
were based on discrete manufacturing concepts in which individ-
ual samples were randomly selected, measured, and analyzed.
Shewart applied random sampling techniques because there were
too many parts to effectively measure the whole population.

Perhaps the best way to examine the impact of SQC fever is to


investigate their application to the manufacturing segments in

B O T T O M - L I N E A U T O M A T I O N 65
BottomLine05.book Page 66 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

which SQC did not originate: the process industries. In the pro-
cess industries, product is more often produced through combina-
tions of movements of fluids flowing together through vessels
and pipes. Intuitively, these manufacturing processes are clearly
considerably more continuous in operation. Nevertheless, many
engineers who applied SQC techniques to the process industries
did so by artificially cutting continuous flows into discrete seg-
ments, thus pretending that the fluid, non-distinct nature of the
process did not exist. This approach has been shown to be techni-
cally questionable by a number of statisticians. To demonstrate its
shortcomings, lets examine simple examples in which statistical
tools are applied to both continuous and discrete processes.

In the last chapter, SQC concepts were introduced through the


example of the discrete operation of drilling a hole in the center
of a part that was to become a nut. The two quality indicators for
this operation were the centering and diameter of the hole. As
long as the hole was consistently and reasonably centered and
reasonably close to the correct diameter, then from the perspec-
tive of this one manufacturing operation the process was under
statistical control. Note that centeredness and diameter are truly
reflective of the characteristics that the customer values in the
product since if the hole does not have the correct diameter the
nut will not work as it should, and if it is not centered, the user
will have difficulty using the nut. In either case, the nut will not
be viewed as having an acceptable level of quality.

Now consider a simple continuous process operation that


involves the flow of a liquid through a pipe. Suppose the parame-
ter of interest in this operation is the temperature of the liquid. A
temperature sensor is inserted through the pipe into the liquid. To
apply the same statistical approach as we applied to the nut
would require us to take a random sample of the liquid and
develop averages and standard deviations from groups of samples.
This has been accomplished traditionally by taking a snapshot of
the temperature at different points in time using time-based sam-
pling techniques. Although one could question the randomness
of any sample selected using these techniques, it is not necessary
to do so for the purposes of this discussion. The averages are cal-

66 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 67 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

culated from a group of the snapshot temperatures across some


number of time samples. Using the same logic Shewart used in
developing his SQC method, our process is in statistical control as
long as the averages are normally distributed about the desired
temperature specification, with no obvious pattern in the distri-
bution.

The heat in the liquid exhibits thermodynamic characteristics


and, essentially, physically averages itself across the liquid in the
pipe. When the sample is taken and averaged, an average of an
average is being calculated. By the Central Limit Theorem of sta-
tistics, the average of an average is always normally distributed
regardless of the structure of the initial distribution. The result is
that the traditional statistical chart derived from the sampling of a con-
tinuous process will not reflect manufacturing process problems in the
same way as it would in a discrete manufacturing process.

Repetitive patterns are a second characteristic that is tradition-


ally looked for in the control charts because it might indicate an
abnormal process condition. Most process plants have quite a
high level of deterministic control, primarily through the use of
feedback controllers. Feedback controllers inevitably introduce
oscillation into the process to bring the controlled variable back
to the desired set point. Feedback controllers actually force repeti-
tive patterns or oscillations into the variables. Therefore, in pro-
cess operations even the existence of a repetitive pattern cannot
be considered to be giving the same information as it does in dis-
crete operations.

The point of all this is not that SQC techniques have no appli-
cation in process plantsthey do. Rather, it is that these statisti-
cal techniques and approaches cannot be applied in the same way
to these two very different types of manufacturing operations.

Temperature was chosen as the quality indicator in the previ-


ous example primarily because it is typical of the type of indica-
tors chosen when SQC tools are applied in process plants. It is
important to compare the relative meaning of the selected quality
indicators in both the discrete and continuous examples. As was
already pointed out, the centeredness and diameter of the hole in
the nut directly reflect the customer's true requirements for the

B O T T O M - L I N E A U T O M A T I O N 67
BottomLine05.book Page 68 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

product, and continuous improvement along both of these indi-


cators means better quality.

Temperature, however, is a process variable, not a product vari-


able. What does it mean to continuously improve temperature?
How can temperature be made better? This question clearly
doesnt have the same meaning as it would for other valid cus-
tomer requirements. The process variable temperature in the
liquid stream does not have the same connotation as the quality
indicators of centeredness and diameter do for the nut. The
concept of continuous improvement has a different meaning in
process manufacturing than it does in discrete manufacturing.

Good deterministic decision-making and control in which the


corrective action is directly calculable is superior to stochastic
decision-making and control in which the corrective action must
be inferred from statistical data. The center and diameter of the
hole in the nut are not, and typically cannot be, deterministically
controlled as the drilling operation is underway because the tech-
nology to accomplish this is either not available or too costly.
However, most process variables can be, and are, deterministically
controlled. If a variable is deterministically controlled using feed-
back or feedforward control techniques, does it make sense to also
apply stochastic analysis to the variable? Control experts agree
it does not.

The real issue is abandoning the unproductive focus on the


tools designed to help people continuously improve quality and
focusing instead on the desired outcomes. For the past decade,
the technology-driven approach to quality has been almost as
illogical as the technology-driven approach to automation dis-
cussed in chapter 2. Industry must not abandon the quality
movement; rather, it must work to make it effective.

QUALITY FOR THE BOTTOM LINE


The significant question now becomes: What are the true
quality indicators in a process plant? In other words, how should
the concept of continuous improvement be applied in a process

68 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 69 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

environment? Since process plants utilize a significantly higher


level of direct deterministic control than discrete factories, the
answer involves a further level of abstraction in the concept of
quality and valid requirements.

To develop effective quality indicators for process operations,


the Foxboro team interviewed dozens of operations managers and
production managers from various segments of the process indus-
tries. After we explained the dilemma of identifying, selecting and
measuring quality indicators, we asked them a very simple ques-
tion, What in your plant would you most want to improve on a
continuous basis? The answer was equally straightforward:
plant performance. In fact, this response was so basic that it was
almost discarded as unhelpful, but at second glance, it proved to
be much more significant. As Peter Drucker wrote in his book
Innovation in Entrepreneurship, an innovation, to be effective, has
to be simple. It is vitally important that we do not overlook the
obvious.5

The next question on our survey was, What is your plants


performance now? Typically, the answer was that they really
didnt knowoperations and production manager were not even
sure how to measure the key parameters of performance in their
plants. Using the same basic thought process employed by
Shewart, Deming, and Juran, it is painfully apparent that one can-
not continuously improve on performance when the basic param-
eters of performance have not been established!

As basic as the concept of continuous improvement is to plant


performance, it is also the most significant issue in process opera-
tions, and it is becoming more significant in the non-process sec-
tors as well as their level of sophistication increases. The quality
indicators of plant performance are, in fact, very similar to tradi-
tional quality indicators in discrete manufacturing in that they
represent true value to both the manufacturer and the customer
even if at a higher level of abstraction than traditional quality
indicators. These performance quality indicators hold the key to
the success of quality control in process environments and in any
performance-based operation.

B O T T O M - L I N E A U T O M A T I O N 69
BottomLine05.book Page 70 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

The shift in focus to a performance-based approach to quality


improvement is well aligned with some of the more intensive
quality improvement programs that have become popular in
major organizations in recent years. One good example is the 6
(or Six Sigma) program, which has been driven through a num-
ber of major global companies such as Motorola, General Electric,
Allied Signal, and Invensys. The 6 program is based on the same
basic tool set as the TQM programs it replaces but has a much
more aggressive performance focus. The name 6 itself denotes
this shift in emphasis over previous approaches. The Greek letter
or sigma represents standard deviations from a target in statisti-
cal analysis. A standard deviation of 3 means that 99.9 percent
of all parts will be within an acceptable range, resulting in 1,349
defects per million parts made. Most good manufacturing pro-
cesses have been around the 3 level. A standard deviation of 6
means that the defect rate has been reduced to 3.4 parts per mil-
lion. This is a very aggressive target, but adopting it clearly dem-
onstrated the seriousness of the effort to gain performance
improvement through these programs. It is also important to
note that most of the jargon surrounding programs such as 6
was still oriented very much toward discrete statistics and discrete
parts.

When Dr. Juran structured the TQM approach, one of his guid-
ing principles was that performance improvement teams should
be charged with selecting their own quality indicators, develop-
ing a method for measuring them, and then developing and exe-
cuting a strategy for continuously improving them. His idea was
that the people on the front lines can identify the issues that
need correcting much more effectively than the managers who
are removed from the details of the process. His concept was that
each time a team or individual goes through a P-D-C-A cycle a lit-
tle improvement would be realized. If enough people and teams
went through enough improvement cycles the total improve-
ment would become quite significant. In essence, his approach
was to just trust the teams or people responsible for improvement
to select the right quality indicators and improvement strategies.

70 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 71 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

This laid-back approach to quality improvement did not sit


well with many manufacturing executives, especially given the
tough global marketplace where exceptional levels of perfor-
mance improvement were required for survival. These executives
essentially agreed that a quality improvement program was a
good thing, but they wanted to ensure that the quality indicators
and improvement projects selected would have an immediate and
strong impact on the economic performance of their manufactur-
ing operation.

The traditional approach to TQM was based on the idea that


people basically understand how the organization needs them to
perform and will therefore select the most appropriate quality
indicators to drive company performance. This is not the case in
most manufacturing operations. As Robert Kaplan and David
Norton pointed out in their Harvard Business Review article Mea-
suring Corporate Performance, Executives also realize that tra-
ditional financial accounting measures like return-on-investment
and earnings-per-share can give misleading signals for continuous
performance improvement and innovation.6 Without valid per-
formance information and direction for the quality improvement
teams, many executives have questioned the validity of the qual-
ity improvement projects and quality indicators that were
selected.

This debate prompted an analysis of the difference between


quality indicators and performance measures. Performance mea-
sures are quantitative metrics designed to indicate how every
group, process, and person in an organization is performing with
respect to the organizations strategic objectives. Quality indica-
tors are measures of the quality of delivery of parts, products, or
services from an individual or group to their direct customer (fig-
ure 3.5). The customer can either be internal or external. The dif-
ferences between performance measures and quality indicators
have been much discussed, but lack of valid performance mea-
sures has rendered much of the discussion moot. Dr. Juran
appears to have had a clear idea of the difference, but he assumed
that organizations would have effective performance measure-
ment systems in place. He assumed that if teams or individuals

B O T T O M - L I N E A U T O M A T I O N 71
BottomLine05.book Page 72 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

truly understood how their behav-


Performance
iors impacted the organizations
Measures
strategic objectives and understood
their job within the organization
Empowered then they would naturally select
Team Quality
appropriate quality indicators and
Indicators
improvement projects. In theory,
this seems quite reasonable, but in
Work Flow practice the lack of effective perfor-
mance measures created a
Figure 3.5 Quality significant gap between expectations and perceived results. With
Indicators or Performance
Measures a poor performance measurement system, even if the quality
improvement teams were accomplishing a great deal of good for
the organization, it would be tough to quantify and document
the value.

These debates over quality indica-


Performance
Measures tors and performance measures
(quality indicators) coincided with the trend toward
performance-based quality initia-
Empowered tives, such as 6 programs. This
Team combination of drivers led to a
convergence of performance mea-
sures and quality indicators (figure
Work Flow 3.6). In discrete manufacturing
environments, this convergence
Figure 3.6 Quality caused considerable confusion and upheaval because these firms
Indicators and Performance
Measures Converge traditional approach to quality indicators was often totally inde-
pendent of performance measurement considerations.

In the process industries, this convergence injected some logic


back into the quality improvement programs. Process manufac-
turing environments had traditionally had great difficulty with
the concepts behind quality indicators, but they had a much
stronger historical focus on economic performance. Even though
most process manufacturing operations did not have effective
performance measurement systems in place, the idea of using per-
formance measures as the quality indicators was a natural fit. All
that was missing was an effective performance measurement sys-

72 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 73 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

tem with well-defined performance measures for each process


unit in the manufacturing plant. In addition, converting the
quality improvement program into an individualized, day-to-day
program rather than just a quality team approach as Dr. Juran rec-
ommended requires that the performance measures be available
in a time frame that was appropriate to the individuals jobs or
tasks. In process operations, that time frame is typically measured
in minutes or seconds because the operators are overseeing the
production of products in process units that can change that
quickly. Providing the performance measures daily or weekly
would not correspond to the work being done and would limit
the effectiveness of a continuous improvement process.

The quality movement has come a long way over the past
decade. Many very positive results have been realized through the
focus on continuous improvement of quality throughout manu-
facturing operations. The movement began in discrete manufac-
turing operations and has enjoyed the most success in discrete
environments. The recent drive toward performance-based qual-
ity programs has led to a convergence of quality indicators and
performance measures. This has provided great incentives in pro-
cess manufacturing environments to reinvigorate their continu-
ous improvement initiatives. This new convergent approach will
drive bottom-line quality and economic performance improve-
ments. The only thing missing in most plants is a real-time per-
formance measurement system that will provide the new class of
performance-based quality indicators to measure and drive con-
tinuous improvement.

THE CONVERGENCE OF LEAN


MANUFACTURING AND QUALITY
IMPROVEMENT
Since the 1990 publication of The Machine that Changed the
World by James P. Womack, Daniel T. Jones, and Daniel Roos, the
concepts of lean manufacturing and the lean enterprise have
gained a great deal of attention in the manufacturing sector. The
book chronicles the success Toyota and several other automobile

B O T T O M - L I N E A U T O M A T I O N 73
BottomLine05.book Page 74 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

manufactures realized by changing the perspective of their manu-


facturing operations from mass production to lean production.
The mass production perspective encourages the manufacturing
operation to make as much product in a given time as possible
while lean production adopts a demand-pull perspective toward
manufacturing. Once this change in perspective occurred, many
traditional paradigms in manufacturing that had once been unas-
sailable suddenly collapsed. The results were phenomenal. The
quality of the products made in these lean manufacturing envi-
ronments continually improved while the real costs of manufac-
turing went down. Unfortunately, most of the successful
examples of lean techniques realizing positive results have been
in the automobile or other discrete manufacturing environments.
However, experts in lean concepts intuitively believe that the lean
approach should and will work in other business areas and envi-
ronments.

LEAN MANUFACTURING CONCEPTS


Six basic principles form the foundation of the lean
enterprise approach as it is described in Lean Transforma-
Continuous
Flow tions by Bruce A. Henderson and Jorge L. Larco (see figure
Production
Six 3.7). Systematically applying these principles has led to
Empowered
Sigma
Teams
Quality significant improvements in discrete manufacturing
Lean
Enterprise environments. A brief summary of each of the principles
Workplace
Visual follows:
safety, order,
Management
cleanliness
Pursuit
of 1. Continuous Flow Production This principle aims at
Perfection
converting a traditional isolated work cell approach to
production to one that follows a more ongoing or
Figure 3.7 Lean Enterprise uninterrupted flow. This principle applies to the pro-
duction operations within an organization as well as
across the supply chain.

2. Six Sigma Quality This principle is representative of the


drive to achieve zero defects in the products produced by
a manufacturing process. Six sigma is a statistical value
that implies that 99.9996 percent of all products produced

74 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 75 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

are within acceptable quality limits. The Six Sigma Qual-


ity initiative is a performance-based extension of the Total
Quality Management (TQM) approaches introduced by
Deming and Juran.

3. Empowered Teams As part of the TQM approach, both


Deming and Juran encouraged organizations to create and
use quality teams that are empowered to manage the con-
tinuous improvement in quality indicators they are
responsible for. In recent years, the concept of empowered
teams has been extended into more of a general manage-
ment philosophy for creating improvement in many
areas.

4. Visual Management This principle involves presenting


the key performance indicators to the empowered teams
as well as to all levels of management so as to gain an
ongoing assessment of the improvement being realized.
The visual management approach encourages the use of
easy-to-understand graphics to present this data.

5. Workplace Safety, Order, Cleanliness This principle is


focused on the concepts that a well-organized work area
can use to encourage improvement. It is based both on
the idea of providing an environment that demonstrates
to employees that they are valued members of the organi-
zation and of promoting orderliness to encourage excel-
lent work habits.

6. Pursuit of Perfection This principle tends to be an overrid-


ing principle of all lean operations: they should continu-
ally strive for perfection. Though it is clearly an
unattainable objective, striving for it helps to drive the
continuous improvements that are required to be world
class.

Reasonable reflection on these six principles usually leads to


ready acceptance. They amount to a collection of common sense
ideals. Unfortunately, common sense often has to be systemati-
cally applied in order to be applied at all.

B O T T O M - L I N E A U T O M A T I O N 75
BottomLine05.book Page 76 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

LEAN PROCESS PLANTS


The primary principles behind the lean enterprise concepts, as
with the principles behind Total Quality Management, had their
genesis in discrete manufacturing operations. There is a very tight
correlation between these principles and the conditions that have
evolved in assembly-line or similar operations. Though there are
few examples of formal lean production programs being imple-
mented in process plants, the basic principles definitely apply.
The following is a short overview of how each of the lean princi-
ples may apply in process manufacturing environments:

1. Continuous Flow Production Most pure process manufac-


turing environments are continuous flow by nature and
design. Batch process environments can be made more
continuous by reducing cycle time and time-between-
cycles as much as possible. This is similar to the reduction
in switchover time in discrete environments. If the time-
between-cycles in batch process environments can be
minimized, the traditional trade-off between production
and flexibility can be significantly reduced. Developing
continuous flow through the supply chain in a process
environment is a very challenging task. With lean tech-
niques in discrete manufacturing operations, the work-
flow is regulated by the takt time of the operation. Takt is
a Japanese work meaning the beat to which to adjust the
pace of your manufacturing operation.7 Considering the
concept of a takt time in what is already a continuous
flow operation is difficult. The key is to treat the entire
supply chain like a single continuous flow operation by
providing common performance measures across the vari-
ous organizations within the supply chain. This will
enable each component in the chain to manage the flow
of production to the downstream group to optimize pro-
duction and cost.

2. Six Sigma Quality Applying the principles of quality


management in process plants has been a challenge ever
since Walter Shewart introduced statistical control in the
late 1920s. The statistical control tools were designed

76 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 77 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

around sample statistics in discrete component environ-


ments. Process manufacturers have been struggling to
apply them in continuous process environments ever
since. Part of the problem in both discrete and continuous
process environments has been defining the quality indi-
cator to be measured. In discrete operations, quality indi-
cators such as defects per millions of parts made are quite
reasonable, but how does such a measure translate into
process environments? Additionally, Juran encouraged
organizations to separate the performance measures and
quality indicators of an empowered individual or team.

As TQM concepts made the transition to a more aggres-


sive performance-oriented approach through the intro-
duction of structured, performance-based TQM
approaches like 6, the distinction between performance
measures and quality indicators began to blur. This blur-
ring was actually beneficial to the application of 6 tools
in process plants because, if performance measures were
available in real time, the quality indicators for plant per-
formance would be readily available. The tools for contin-
uous improvement in process environments are different
than the traditional statistical control tools and include
such tools as advanced process control, multivariable pre-
dictive control, and process optimization. Nevertheless,
the objective is identical. The availability of real-time
plant performance measures will make possible the effec-
tive application of 6 concepts in process plants.

3. Empowered Teams The teams process plants need to gen-


erate continuous improvement include process engineers,
control engineers, maintenance personnel, operations
personnel, and production personnel. The best way to
empower any team is to provide it with the toolkit and
clear objectives and measures for improvement necessary
to realize the improvement. The real-time performance
measures provide this for the teams, and the performance
targets provide the clear objective functions.

B O T T O M - L I N E A U T O M A T I O N 77
BottomLine05.book Page 78 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

4. Visual Management Visualizing plant performance in


process plants has been a problem since the earliest days
of process manufacturing. Performance dashboard dis-
plays at the system workstations that present the perfor-
mance measures in real time are finally making visual
management in process plants a reality. Visual manage-
ment works. Process manufacturers can finally take
advantage of this high-powered communication and per-
formance-enabling approach.

5. Workplace Safety, Order, Cleanliness In process plants,


unlike discrete manufacturing operations, the operator is
often removed from the process and located in a control
room. Control room environments are typically cleaner
and more orderly than plant environments, but most con-
trol rooms could be significantly improved. One of the
key ways to improve the workplace environment in pro-
cess plants is to incorporate more electronic technology to
make it easier for operators to access the information they
need, with less clutter. Process automation systems are
designed to enable this.

6. Pursuit of Perfection Facilitating the pursuit of perfection


in process plants is the essence of continuous economic
and quality performance improvement. To pursue perfec-
tion and drive continuous improvement, there must be
an understood target and a metric system that defines
improvement. The real-time performance measures, asso-
ciated performance CRT dashboard displays, and embed-
ded performance targets help process manufacturers to
define perfection and drive toward perfection. Perfection
is no longer an abstract ideal in process plants. It is a well-
defined set of objectives that plant personnel can work
toward.

Lean production and lean manufacturing have been limited to


discrete manufacturing operations. The availability of real-time
performance measures, performance dashboards, continuous
improvement tools, and advanced automation system technology

78 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 79 Friday, August 19, 2005 9:54 AM

C H A P T E R 3 : Q U A L I T Y A N D T H E B O T T O M L I N E

can combine to make lean manufacturing approaches available to


process manufacturers.

NOTES
1. Gagne, James, Quality Performance Means More at Dow.
Midland, MI: Dow Chemical U.S.A., page 3.

2. Dobyns, Lloyd, Ed Demings Wants Big Changes, and He


Wants Them Fast. Smithsonian, August 1990, page 77.
3. Gagne, Quality Performance Means More, page 5.
4. Bhote, Kevin R., Americas Quality Health Diagnosis: Strong
Heart, Weak Head. Management Review (American Manage-
ment Association), May 1989, page 36.
5. Drucker, Peter, Innovation and Entrepreneurship: Practices and
Principles. New York: Harper & Row, 1985, page 135.
6. Kaplan, Robert S., and David P. Norton, The Balanced Score-
card Measures that Drive Performance, Harvard Business
Review on Measuring Corporate Performance. Boston: Harvard
Business School Publishing, 1998.
7. Henderson, Bruce A., and Jorge L. Larco, Lean Transformation.
Richmond, VA: Oaklea Press, 1999, page 58.

B O T T O M - L I N E A U T O M A T I O N 79
BottomLine05.book Page 81 Friday, August 19, 2005 9:54 AM

CHAPTER 4

Cost
Accounting
and the
Bottom Line

THE COST ACCOUNTING TREND


The cost accounting trend is somewhat different than the
technology and quality trends in that its origin is more tightly
linked to the financial operations within manufacturing compa-
nies rather than their manufacturing operations. This was also a
much more difficult trend for our research team to gain an under-
standing of because the team members are automation focused
and had much less interest in and knowledge of accounting prin-
ciples and techniques.

At the beginning of the research program, the executives being


interviewed discussed cost accounting as a critical success factor
in gaining the benefit of automation. Nevertheless, the research
team did not grasp the significance of this factor, perhaps because
team members felt that it was obvious that proving the economic
value of automation or any other program would require an effec-
tive accounting approach. Perhaps the team just assumed that the
accounting systems companies currently used possessed all the
information required. What the team missed early on was that
the executives being interviewed were trying to point out that an

B O T T O M - L I N E A U T O M A T I O N 81
BottomLine05.book Page 82 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

effective accounting approach did not yet exist to effectively cap-


ture the economic value resulting from automation or any other
manufacturing process improvement.

As did the technology and quality


Technology trends, the cost accounting trend has
evolved through three major phases
Bottom
Quality Line over time and is still evolving (figure
Improvement
Accounting
4.1). As the research project began,
for the
Cost
Management &
Bottom Line traditional cost accounting systems
Accounting Cost
Manufacturing
Accounting & had been in place and operating in
Manufacturing
one form or another for decades
Figure 4.1 The Cost without any appreciable modification. Some of the executives
Accounting Trend
interviewed seemed to feel that these systems were merely a fact
of life that had to be dealt with. Many expressed serious concern
over the lack of effective and timely data. We refer to this initial
traditional phase of the accounting trend as the Cost Accounting
& Manufacturing period.

Very shortly after the research team began to really understand


that traditional cost accounting had serious limitations when
applied to manufacturing operations, new approaches to account-
ing in manufacturing environments designed to overcome these
limitations were gaining recognition and acceptance. These new
approaches were initially labeled cost management systems or
activity-based costing systems. We refer to the period of transi-
tion from traditional cost accounting to these new accounting
systems as the Cost Management & Manufacturing period. It
became very clear soon after the team recognized these new
approaches that many of the executives interviewed were driving
their accounting systems in this direction.

As the early cost management and activity-based costing sys-


tems started to come on line, they seemed to offer significant ben-
efits over the traditional cost accounting systems. However,
serious limitations in these approaches and significant implemen-
tation difficulties were soon identified. This has led to a second
transition that is currently taking shape in many manufacturing
companies, a transition that should lead into a new concept and
approach for accounting in manufacturing operations. We refer

82 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 83 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

to this emerging period as the Accounting for the Bottom Line


phase.

COST ACCOUNTING AND


MANUFACTURING
Corporations are in business, for the most part, to make a
profit. The way corporations measure whether they are profitable
is through cost accounting systems. Years ago, these systems were
comprised of hundreds of bookkeepers and accountants manually
performing calculations. Today, computers do the bulk of the
work. Other than that, the basic structure of cost accounting sys-
tems has changed very little over the years.

The executives we interviewed expressed great concern over


the usefulness of the information they could gain from their cost
accounting systems. They accepted that the cost accounting sys-
tems in their operations were necessary for reporting required
financial data but did not feel that they provided much data that
was relevant to their manufacturing operations. They appeared to
believe that most cost accounting systems are historical, oriented
toward financial reporting, and inadequate to measure opera-
tional performance.1

Cost accounting systems evolved very little between about


1900 through the 1960s. The primary objective of these systems
was to report to management and the financial community on
the financial performance of companies in a way that was consis-
tent across all companies and operations. This consistency of
reporting was necessary to provide effective communication and
comparisons of the financial performance of different companies
and operations. Over this period, standard cost accounting princi-
ples and practices became very well defined and accepted.

From a manufacturing perspective, cost accounting systems


had very little impact on the day-to-day operations. This may
seem odd since a good percentage of the direct costs of most man-
ufacturing companies occur within their manufacturing opera-
tions. But the consensus was that the main purpose of the old

B O T T O M - L I N E A U T O M A T I O N 83
BottomLine05.book Page 84 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

cost-accounting system [was] to help the financial department


monitor operations and value inventory.2 In other words, the
primary customers of cost accounting systems were the finance
department and not the manufacturing operation, and the out-
puts from these systems were focused to meet the needs of this
constituency.

Despite the strong, widespread emphasis on manufacturing


performance since 1970, cost accounting systems, oddly enough,
did not evolve to provide performance-based information for
manufacturing operations. This lack of progress was closely asso-
ciated with the technology trend we discussed in chapter 2. As
digital computers gained acceptance for the management of busi-
ness information management during the 1960s and 1970s, one
of the first initiatives of many corporations was to automate their
cost accounting systems. This made reasonable sense since previ-
ously these systems were very manpower-intensive and extremely
costly.

In the early years of the computer movement, there were very


few professionals with computer programming expertise and no
standard software packages. As a result, companies typically hired
college graduates trained in technical fields such as mathematics
or physics and invested in training them in programming. Pro-
grammers with any degree of experience were usually promoted
to the role of systems analyst and became responsible for design-
ing the computer program-based systems, such as the cost
accounting system.

Computerized cost accounting systems were generally made


up of a large number of separate programs, each designed to per-
form a subset of the entire cost accounting process. Systems ana-
lysts would assign the responsibility for creating the programs to
several different programmers. The system analysts job was to
ensure that all the programs of the cost accounting system would
effectively work together to produce the desired outputs. Struc-
tured programming techniques had not yet been developed for
the early programmers, and in any case the programmers had
been trained to design programs to optimize scarce computer
resources, such as memory space or computer time. As a result,

84 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 85 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

the programs were very difficult for programmers who had not
written them to understand. To make matters worse, the average
tenure of the new programmers in any company was short
because once they had even a little experience they could earn
more money by joining a different company. Most companies
cost accounting systems were therefore comprised of many
unstructured programs written by several different programmers
who had since left the company. The new programmers hired to
replace them had great difficulty analyzing and modifying the
programs, and computerized cost accounting systems went years
with only minimal modification. The unwillingness and inability
of companies to modify and expand their cost accounting sys-
tems actually became one of the main reasons cost accounting
did not change appreciably from the 1960s through the 1980s.

The way in which manufactur-


Financial
ing performance was measured Community
through traditional cost accounting Cost
Accounting
systems was fairly universal across Cost per System
Unit Product
all industries (see figure 4.2). The (Variance Report)

quantity of each product manufac- Raw


Product
Material
tured over a given period was deter- Manufacturing Storage
Storage
mined and entered into the cost
accounting system. Cost accountants then determined the total Figure 4.2 Cost Accounting
and Manufacturing
cost to make the product and run the plant over this period, and Performance
then calculated the cost-per-unit-product-made. This statistic, cost-
per-unit-product-made for each product or product line manufac-
tured, has been the essential method for measuring manufactur-
ing performance since the very earliest days of manufacturing.

This information has typically been transmitted back to manu-


facturing management through biweekly or monthly variance
reports. The idea behind a variance report is that the accountants
would establish cost targets, or standard costs, for each product
made at the beginning of the fiscal year based on their most opti-
mistic prediction. The cost-per-unit-product calculated each
month by the cost accounting system was compared to the stan-
dard cost, and from this the variance from standard was calcu-
lated. If the calculated cost was less than the standard cost, the

B O T T O M - L I N E A U T O M A T I O N 85
BottomLine05.book Page 86 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

manufacturing operation was determined to be doing a good job.


If the actual cost was below the standard cost on an ongoing
basis, the manufacturing manager for that operation might
receive a healthy raise or even a promotion. If the calculated cost
was greater than the standard cost, the manufacturing manager
had to be prepared to do some explaining.

The primary objective of the manufacturing operations within


these corporations has been to make product at the required vol-
ume and the lowest possible cost. Clearly, this is a key element in
any cost-based structure. Years ago, the bulk of the costs incurred
by most corporations came during the manufacture of the prod-
uct. Since overall cost, including manufacturing cost, was the pri-
mary focus of the cost accounting systems, the variance reports
became the basis for calculating the cost-based manufacturing
performance measures. Measures of an individuals or organiza-
tions performance are referred to as performance measures. In the
case of traditional cost accounting, the performance measures
have been cost-per-unit-product-made.

Since the financial community measures the performance of a


corporation by its profits, the traditional cost accounting
approach for measuring manufacturing performance would seem
to be the most reasonable accounting method. Indeed, it was rea-
sonable seventy years ago when most manufacturing operations
made few products, typically on separate manufacturing lines,
and when most of the cost of an operation was directly assigned
to the products made (direct costs). Overhead costs, such as gen-
eral administration costs, comprised only a small percentage of
any products total cost. Also, historically manufacturing opera-
tions were viewed as necessary, but not strategic, operations
within the corporation. Gaining competitive advantage through
flexible manufacturing or repeatable quality was not considered
to be of prime importance. Within this context, the primary
driver of manufacturing operations should probably have been
cost, and the cost-per-unit-product-made metric for measuring
performance was probably reasonable.

The situation in manufacturing operations today is vastly dif-


ferent. Many operations make multiple products, even on the

86 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 87 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

same manufacturing lines. Moreover, Overhead costs have risen


dramatically and now exceed direct labor costs It is not uncom-
mon to find that direct touch labor accounts for only 8 12 per-
cent of total cost at many factories.3 Finally, in todays difficult
competitive environment, manufacturing is often viewed as stra-
tegic and essential to the execution of the companys strategic
plan.

To compensate for the large per-


centage of an operations total cost Direct Costs + Allocated Costs
Cost per Unit Product =
that cannot be directly assigned to Quantity of Product Produced
products, cost accountants today use
allocation algorithms that apportion the overhead costs that can- Figure 4.3 Traditional
Product Costing
not be directly assigned to the products manufactured (figure 4.3).
An allocation process assigns costs that have been accumulated
in a cost pool to a particular management reporting objective.4
These allocation objectives may be based on any number of fac-
tors, such as market demand for each product, volume of product
sold, or even price. Overhead (indirect labor, depreciation,
energy costs, maintenance, supplies etc.) is lumped into cost
pools and then absorbed into the cost of the products according
to some factor usually the number of standard direct labor
hours in the product.5

The way the allocation is accomplished is not the key factor.


The important point is that manufacturing managers are being
measured by the fraction, cost-per-unit-product-made. They have
little or no control over the value of the numerator (cost) of this
fraction. The largest percentage of the costs indicated by the
numerator is allocated to costs incurred by the operation that the
manufacturing manager can do little to impact. Even if a manu-
facturing manager does an excellent job controlling cost, this
might impact the numerator of his or her performance metric by
as little as one tenth of one percent, and because of allocation
algorithms it could impact another manufacturing managers per-
formance measure by as much, or even more.

As an example, I remember sitting in a meeting of manufactur-


ing executives a few years ago as a manager told a story involving
the allocation of cost in his facility. It seems that one of the prod-

B O T T O M - L I N E A U T O M A T I O N 87
BottomLine05.book Page 88 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

ucts manufactured in his facility had been determined to be a real


loser. The cost accounting system was consistently showing it to
have negative margins. Since some valued customers really liked
the product, the corporation decided it could not just pull it off
the market.

A team was formed to study the problem and determine how


to phase out the product without negatively impacting the cus-
tomer base. After a few months of team meetings, the product
began to show positive margins in the variance report. Corporate
management was pleased and congratulated the team on their
good work, but the team was astonished because they knew they
hadnt done anything that would have caused this change. They
decided to wait another month to see if the improvement in mar-
gins was just a fluke, or if something had really changed. The next
month the cost-per-unit for this product on the variance report
was even lower. The team investigated to see what had changed
to cause this favorable result. They checked raw material costs,
energy costs, and every other cost they could think of, but they
could not find any appreciable factor that would have made this
difference. They checked for changes in product volume and sell-
ing price, but still found no appreciable change. Finally, when
they were about to abandon their search, one team member
talked to a company cost accountant and learned that the over-
head allocation algorithms for the plant had been changed to
make the allocations more representative. The change in alloca-
tion approach was the cause of the performance improvement for
the product. The manufacturing manager and the manufacturing
operation for this product had absolutely nothing to do with the
favorable trend in measured performance. They appeared to be
performing considerably better because less overhead cost was
being allocated to their operation. I cant help but wonder if the
manager got promoted as a result.

The calculation and aggregate reporting of overhead by tradi-


tional cost accounting systems has hindered managements
understanding of what comprises overhead costs and the cost
implications of non-productive activities.6 Also, as a result of the
overhead allocation and the increasing percentage of cost that is

88 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 89 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

overhead, the numerator of the cost-per-unit-product fraction


generated by traditional cost accounting systems for manufactur-
ing performance is almost totally beyond the control of manufac-
turing personnel. Therefore, the only part of this measure that is
within their control is the denominator: the amount of product
made over a given period. Manufacturing managers who are
interested in performing well according to the performance mea-
sures of traditional cost accounting systems will always make as
much product as possible at all times regardless of the manufac-
turing strategy. In doing this, they can elevate the denominator
and correspondingly lower the overall value of the fraction. As a
result, the actual manufacturing strategy for most companies in
the traditionally industrialized areas of the world over the past
few decades has been to make as much product as possible regard-
less of market demand.

In recent years there have emerged several manufacturing


strategies and action plans that have been the center of much
attention in the manufacturing world. One of the most signifi-
cant of these is just-in-time (JIT) manufacturing, introduced in
chapter 1. JIT is a manufacturing action plan with wide-ranging
implications, one of the most important of which is that manu-
facturing operations are only supposed to manufacture to meet
existing demand. That is, they should only make the amount of
product that the market needs at any given point in time. This
allows inventories to be significantly reduced, resulting in a corre-
sponding reduction in the cost to store, maintain, and manage
the inventory. This all makes perfect sense, unless you happen to
be a manufacturing manager who is interested in driving the per-
formance measure cost-per-unit-product-made down because you
know this is how your performance will really be judged. Notice
that the implementation of JIT essentially takes the denominator
of this fraction out of the control of manufacturing management,
and that JIT is essentially opposed to traditional, volume-based
manufacturing strategies. In a JIT manufacturing environment, a
manufacturing manager who performs in line with the JIT action
plan can be accused of real performance problems because the
only controllable aspect of performance, product volume, has
been taken away.

B O T T O M - L I N E A U T O M A T I O N 89
BottomLine05.book Page 90 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

A story a manufacturing manager shared with me about JIT


helps to illustrate this point. A JIT program was initiated in this
manager's corporation. The success of the program was to be mea-
sured by the JIT team representatives walking around and record-
ing the amount of in-process and product inventory that each of
the manufacturing operations had and then developing a
monthly JIT report. Corporate management, claiming they were
committed to JIT, would reprimand the managers of the opera-
tions that were not meeting the JIT objectives. But the primary
measure of the manufacturing performance remained cost-per-
unit-product-made.

The JIT team members walked around with their clipboards for
the first two weeks of every month recording inventory levels. It
took the final two weeks for them to compile their reports, which
they did in offices some distance away from the manufacturing
operation. The manufacturing manager told me that for the first
two weeks of every month he had the best JIT operation the com-
pany had ever seen. But as soon as the door closed behind the JIT
folks, production was cranked up as high as possible. The manu-
facturing team hid product in closets, cars, and wherever else they
could store it. In this way, the operation was seen as being good at
JIT and still measured up to the cost accounting-based perfor-
mance measures. Something is obviously wrong when the manu-
facturing people have to work around the system to be successful.

It may be worth investigating the failure point of the tradi-


tional cost accounting systems with respect to manufacturing per-
formance. One principle that appeared to be shared by all of the
manufacturing executives we interviewed is that people will per-
form to their measures. This is true regardless of what the mea-
sures are. If the measures are appropriate, they will drive
appropriate behaviors. If the measures are bad, they will drive
inappropriate behaviors. As in the JIT example, manufacturing
organizations have performed to those measures even when they
are not appropriate. For cost accounting to drive the correct oper-
ational behaviors, the performance measures generated must be
aligned with the operations objectives.

90 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 91 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

As programs like JIT have been attempted, the failure of the


traditional cost accounting-based performance measures has
become more obvious. Manufacturing management has had very
little control over these measures except to drive production vol-
ume up as much as possible. This has led to the development of
huge manufacturing storage facilities such as warehouses or tank
farms, which actually add cost to the overall operation. Since that
cost is typically allocated so no specific product or product line is
affected more than any other, this strategy makes sense within
the manufacturing measures of performance.

As the globalization of the marketplace accelerated and the


competitive environment for many manufacturing companies
became very difficult, many companies have tried to make manu-
facturing more of a strategic weapon. Unfortunately, the cost-per-
unit-product measures have often opposed the strategic initiatives
that would align the manufacturing operations with the strategic
plan. For example, if manufacturing flexibility is seen as a key
strategic differentiator, the drive to increase the denominator of
the performance measure would tend to hinder manufacturing
flexibility. If a trade-off must be made between a stated program
and a measure, people will typically make sure they perform to
the measure. If the measures of performance do not represent the
strategy, the strategy will probably fail. Manufacturing managers
are being asked to make important decisions in spite of available
cost accounting information, not because it is relevant.7 Clearly,
cost accounting systems and the resulting manufacturing perfor-
mance measurement systems were in critical need of improve-
ment.

COST MANAGEMENT SYSTEMS


As the weaknesses of the traditional cost accounting systems
began to become apparent, initiatives to rectify these systems
were attempted. One of the first involved trying to revise the allo-
cation algorithms for overhead costs to make them better reflect
the manufacturing operations. Manufacturers quickly found out
that this approach was futile. It is an illusion that changing the

B O T T O M - L I N E A U T O M A T I O N 91
BottomLine05.book Page 92 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

basis of allocation will solve all cost management problems.8


Perhaps the most visible and widely accepted approach to modi-
fying cost accounting was based on the concept of activity-based
costing and went under the general banner of cost manage-
ment systems (CMS).

The primary objective of the CMS approach is to fix cost


accounting so that more accurate cost data can be determined.
The basis for the fix is that too much overhead cost has to be allo-
cated to provide the cost information with any accuracy. Manage-
ment could no longer accept an environment where cost
accounting contributes to overhead rates so high as to obscure
true product costs.9 The theory behind this movement was that
if more of the costs in an operation could be directly assigned to a
product or product line, less would have to be allocated (figure
4.4). The resulting cost information should more accurately
reflect the true cost of the operation and the products.

From a cost accounting perspec-


Objective:
Objective: Increase
Increase the
the Percentage
Percentage of
of Direct
Direct Costs
Costs tive, there is considerable merit to
the idea and approach of activity-

Product A
based costing since it should lead
Product B to more accurate cost profiles for
Assigned Costs ..
Product C
the products and any other cost-
Product X based activities throughout an
Traditional Overhead Cost
operation.

Result:
Result: Cost/Unit
Cost/Unit Product
Product isis More
More Accurate
Accurate The way to directly assign more of

Figure 4 4 the overhead costs in the


Figure 4.4 Cost operation is to keep much more accurate records on how tradi-
Management Systems
Activity-based Costing
tional overhead costs are incurred and then determine a way to
directly assign some of these costs to products or product lines.
Traditionally, the cost of functions such as product development,
marketing, and sales has been considered as overhead. Any costs
incurred by these functions have gone into the general overhead
pool for allocation. In reality, a large percentage of the time in
these functions is spent on a particular product or product line,
and the cost associated with this time can and should be directly
assigned. This is a key aspect of activity-based costing. Costing
systems based on this approach can significantly reduce the

92 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 93 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

amount of overhead that has to be allocated and can result in


much more accurate cost information.

As cost management systems gained in popularity during the


1980s and 1990s, many who had started to realize that the cost
accounting-based performance measurement systems were flawed
saw activity-based costing as the solution. Obviously, more accu-
rate cost information should result in more accurate calculations
of cost-per-unit-product-made. In theory, the resulting cost-based
performance measures might be more helpful in providing manu-
facturing with the information it requires managing performance
on an ongoing basis.

One shortcoming of this point of view is that, regardless of


how accurate the statistic is, manufacturing management and
operations personnel have absolutely no control over the tradi-
tional overhead costs, even if they are directly assigned instead of
allocated. A manufacturing manager has little, if any, direct con-
trol over product development costs, sales and marketing costs, or
general administrative costs. The fact that these costs are more
accurately assigned to products or product lines does not mean
that they provide effective manufacturing performance measure-
ments.

A second issue, stemming from the tough, competitive global


environment, is that manufacturing has began to be considered
as an important competitive weapon in the execution of com-
pany strategies. In a cost-based strategy, cost-per-unit-product-
made may be a reasonable metric, but if the corporate strategy is
based on flexibility, quality, or production, cost-based metrics fall
far short. Japanese manufacturers have moved into a strong posi-
tion in manufacturing over the past twenty years. Certainly, they
measure the performance of their manufacturing operations:
Japanese companies put much more emphasis on measuring the
non-financial aspects of factory performance.10 It is critically
important that we learn from their success and not fall victim to
the quick fix. Cost management systems represent a significant
advance in the accuracy of cost accounting, which is important
for financial management, but manufacturing performance mea-
surement requirements and cost accounting requirements may

B O T T O M - L I N E A U T O M A T I O N 93
BottomLine05.book Page 94 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

not be as tightly coupled as they have been in the past. Cost


accounting systems employ performance measurements that
often conflict with strategic manufacturing objectives, and they
cannot adequately evaluate the importance of non-financial mea-
sures such as quality, throughput and flexibility.11 Moreover,
Managers need up-to-date, concise information formatted to
assist them in making the right decisions.12 Perhaps it is time to
look for a whole new performance measurement system, one that
will work to move companies to world-class manufacturing
through world-class performance.

ACCOUNTING FOR THE BOTTOM


LINE
It should be fairly obvious at this point that in order for an
accounting system to have strong bottom-line impact on manu-
facturing operations it must produce effective and accurate manu-
facturing performance measures. Financial performance
measures should indicate whether the companys strategy,
implementation and execution are contributing to bottom-line
improvement.13 Traditionally, corporate and manufacturing per-
formance measures were viewed as byproducts of cost accounting
systems. They were treated as almost an afterthought. For manu-
facturers to compete in the competitive global environment, this
must change. Enhanced competitiveness depends on starting
from scratch and asking: given our strategy, what are the most
important measures of performance?14

The effectiveness of any companys performance measurement


system must be treated as a critical factor to the success of the exe-
cution of its competitive strategy. One critical success factor to
developing a good manufacturing performance measurement sys-
tem is to make certain the system is measuring the performance
parameters that really matter. Many companies do not institute a
measurement system that supports the companys strategy.15 A
performance measurement system that is not aligned with the
companys strategy is a formula for failure. It is very important
that the performance measures represent the corporate strategy

94 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 95 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

and the resulting strategy of the manufacturing operation. The


only reasonable way to establish this linkage is to clearly define
the corporate strategy and ensure that the manufacturing perfor-
mance measures are derived directly from it: Performance mea-
sures should help establish congruence between organizational
and company objectives.16

As we discussed in chapter 1, Dr. Thomas Vollman, a


Strategy
professor at the International Institute for Management
Development, developed a simple model to help visualize
and systematize this activity that is sometimes referred to
as the Vollman Triangle (figure 4.5). It can be very useful
in analyzing and developing effective corporate perfor-
Action Measure
mance measurement systems so we will describe it again
here. At the top of this triangular model is the plant's manufactur- Figure 4.5 Vollmann
Triangle
ing strategy. The strategy has to be clearly represented and com-
municated to all key people in the operation if their performance
is to support the strategy. Out of the strategy should come an
action plan indicating how the strategy should be implemented.
The performance measures should be developed to represent and
measure the success of the action plan: Performance measures
should provide the link between the activities of the business and
the strategic planning process.17 If the performance measures
reflect the strategy and are available at all levels of the organiza-
tion, the companys personnel will tend to perform to the mea-
sures and, thus, the plants manufacturing strategy will be
effectively executed.

Cost is only one aspect of performance greater emphasis is


given to providing non-financial information reporting for deci-
sion support.18 Japanese manufacturers have demonstrated an
understanding of the concepts portrayed by Vollmans model. If
the action plan requires financial measures, then financial mea-
sures should be developed. If the action plan requires non-finan-
cial measures, then non-financial measures should be developed.
To effectively execute most manufacturing strategies usually
requires a mix of financial and non-financial measures. If the
right measures are utilized, the action plan will be implemented
successfully because the employees will understand how they are

B O T T O M - L I N E A U T O M A T I O N 95
BottomLine05.book Page 96 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

measured and will act appropriately. As Harvards Peter Drucker


points out, the scarcest resources in any organization are per-
forming people.19 Perhaps this is because most of the people in
most organizations just dont know how to perform because they
have no way of determining how their actions impact perfor-
mance.

The Vollman Triangle can be viewed as a continuous improve-


ment model for manufacturing strategy similar to the Plan-Do-
Check-Act model of Total Quality Management. It represents a
cycle that starts with the strategy. From the strategy an action
plan is developed, and performance measures should be con-
structed to measure the effectiveness of the action plan. Once the
performance measures are under control, or market conditions
dictate that a new strategy is required, the Vollmann Triangle
cycle can be utilized over again, resulting in continual improve-
ment of manufacturing performance in line with the competitive
strategy.

In traditional industrial plants, manufacturing strategies and


action plans are often developed, but they are measured and eval-
uated via the traditional cost-per-unit-product-made measures. If
these measures happen to be accurate and match the action plan,
there is some chance that the correct strategy will be imple-
mented. But if the measures do not match the action plan, which
is most often the case, the probability of the strategy succeeding is
almost zero. Many manufacturers have seen their manufacturing
strategies fail for exactly this reason.

It is critically important that the measures of manufacturing


performance precisely match the plants manufacturing strategy
and action plan so that the operation, and every person in it, is
taking informed actions based on fulfilling the strategy. There
needs to be an explicit link between the performance of individu-
als with the performance of the company as a whole.20 Any
measurement is meant to enable us to do, to take purposeful
action based on knowledge rather than opinion or guesswork.21
The purposeful action desired in any manufacturing facility
should be directed at making the strategy work. Thus, the plants

96 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 97 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

performance measures must match the plants manufacturing


strategy.

One of the clas- Strategy January


sic challenges to 1 2 3 4
5 6 7 8 9 10 11
12 13 14 15 16 17 18
executing compet- 19 20 21 22 23 24 25
26 27 28 29 30 31
Action Measure
itive strategy has
Strategy Strategy Strategy
been gaining
alignment
throughout an Action Measure Action Measure Action Measure

entire company. A
company can have
the most effective
competitive strat-
egy in a market,
but if its management team cannot determine how to take the Figure 4.6 Vollmann
Decomposition
high-level corporate strategy down to an actionable and measur-
able plan at all levels in the operation, the strategy will not be
effectively implemented. Once again, Dr. Vollmann has helped to
provide a fairly straightforward methodology for driving the strat-
egy through an organization. As we discussed in the foreword to
this book, this structured methodology is referred to as a Voll-
mann Decomposition (figure 4.6). The concept behind the Voll-
mann Decomposition is rather straightforward, but experience
has demonstrated that it requires significant corporate discipline
and strong leadership to execute the decomposition down
through all appropriate levels of the organization.

The Vollmann Decomposition is accomplished by top-down


repetition of the Vollmann Triangle model. The executive team of
most high-performing companies typically develops or adjusts
the corporate strategy (see figure 4.7) on a regular, although infre-
quent basis. For most companies this happens about once per
year. The corporate strategy should include an overall strategy
statement that provides the short-term vision for the company.
Experience has demonstrated that good strategy statements
should encapsulate the vision of the company, define the com-
panys current and desired states (if different), and define the cus-
tomer value the company expects to provide. Since the strategy

B O T T O M - L I N E A U T O M A T I O N 97
BottomLine05.book Page 98 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

statement should typically be rather con-


Strategy Statement cise, it is important to also develop a set
Visionary
Current and desired state Strategy Statement
of strategic objectives that define the spe-
Customer value proposition
Corporate value proposition Strategic Objectives
cific goals and objectives of the company
Short (2 to 3 sentences) 1.
for the coming year. The objectives
Simple and Clear 2.

Strategic Objectives 3.
should provide an added level of specific-
4.
Decompose strategy statement ity to the corporate strategy. Good objec-
5.
Reflect movement to desired state
Simple 6. tives must be simple, understandable,
Few ( < 7 total)
Short (1 sentence)
clearly stated, customer focused, and
Customer focused actionable. Strategic planning teams
Actionable
Figure 4.7 seem to have a tendency to want to
Figure 4.7 Strategy develop a large number of strategic objectives. This can be a big
mistake because too many objectives can result in presenting a
confused strategy. As a rule of thumb, keeping the objectives to
seven or fewer can help strategies stay clear and simple.

As we have seen, the


Action Plan second step in the
Action Plan Structure 1.
Objective Action Responsibility Start Complete Status
Vollmann Triangle
Objective Action Supports
Action Statement
2.
model is to develop a
3.
Tied to strategic objectives corporate action plan
Identifies gap (as is - desired) 4.

Measurable (figure 4.8). A good


5.
Responsibility action plan is com-
6.
Schedule
7. prised of a set of mea-
Starting time
Completion time surable corporate
Progress milestones
(if necessary) action statements
Progress that are derived from
the strategic
Figure 4.8 Action Planning objectives. The action plan should define a person or organiza-
tion responsible for executing each action statement, and it
should define a schedule. The action plan can be a very good
management tool if it is treated as a living document and the
status of each action step is updated throughout the action plan-
ning time period.

The final step is to identify the corporate performance mea-


sures from the action statement. Since each action statement
should be constructed to be measurable, if the action statements
are developed correctly the performance measures should be obvi-

98 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 99 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

ous. Sometimes multiple action statements are developed to drive


the same performance measures. In these situations, the number
of corporate performance measures could be fewer than the num-
ber of action statements. Also, not all of the performance mea-
sures defined in the corporate plan are necessarily impacted by all
operations in the organization. Therefore, operations, such as
manufacturing, may focus their plans on only a subset of the cor-
porate performance measures.

The strategy, action plan, and performance measures at the


corporate level are typically too abstract and too far removed
from any department or employee to be used effectively in guid-
ing their day-to-day-performance. A structured decomposition is
required to bring these down to the level where they drive depart-
ment and individual performance.

The sample decomposition model in figure 4.6 shows the


decomposition structure for a corporation comprised of three
major operating divisions. The first level of decomposition is
from the corporate performance measures to the divisional level.
Each divisional executive team should evaluate the measures of
performance defined by the corporate strategy and develop its
strategy for driving those measures in the desired direction. From
the divisional strategy should come a divisional action plan and
divisional performance measures. The divisional performance
measures will not necessarily be the same as the corporate perfor-
mance measures, but should be indicative of the contribution the
division can make to the corporate strategy. The key critical com-
ponent of this structured approach is that, if it is done correctly,
as the divisional metrics start to move in the correct direction the
corporate measures should also start to improve, by design. If this
does not happen, the result could be a seriously dysfunctional
organization with different operations moving in very different
strategic directions.

If a particular division is comprised of four manufacturing


plants, each plant management team should analyze the divi-
sional performance measures and develop a strategy for driving
those measures in the correct direction. From this plant-level
strategy should come the plant action plan and plant perfor-

B O T T O M - L I N E A U T O M A T I O N 99
BottomLine05.book Page 100 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

mance measures. This level of the decomposition is sometimes


referred to as the plant manufacturing strategy since the ulti-
mate reason for the existence of each plant is to manufacture
products. The decomposition should continue down through the
operation until each group and person within the operation has a
clear understanding of how their performance is measured by the
company. Following a structured decomposition such as this will
ensure that all people and departments within the company are
working toward the same objectives. It will also ensure that as
each individual in the organization drives his or her performance
measures in the desired direction, the corporation, as a whole,
will benefit.

One interesting and important element of the decomposition


that is critical to the successful implementation of strategy is the
timing entailed by each layer in the decomposition. As was
pointed out, the corporate strategy, action plan, and measures are
typically derived annually due to external business planning
issues. The divisional level analysis should be done according to
the competitive and market dynamics that impact the competi-
tive environment of that division. This may be more frequent
than the corporate plan. In many of todays markets this could be
necessary on a semiannual or even quarterly basis. The plant-level
analysis must typically be done more frequently than the divi-
sional level analysis because of the dynamics of the local markets
and local supply and demand changes. Plant-level plans may
need to be executed as frequently as quarterly or even monthly to
be effective. As the decomposition continues down through the
plant the time-constant requirements for the performance mea-
surement process become shorter and shorter. Fortunately, major
strategy changes are not typically done below the plant or divi-
sional level, but action planning and performance measurement
systems must continue to be decomposed and must be able to
handle very short time constants to be effective. In process manu-
facturing operations the time constants can be as short as minutes
or even seconds. This type of time requirement is totally out of
the scope of traditional cost accounting systems or even the more
recent cost management systems.

100 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 101 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

In reviewing the Vollmann E


c
Decomposition approach with o
n
o
the executives taking part in the m
i
c
research project, our project team s

found that many of them had


started to drive a very similar The
The Gap
Gap
strategy decomposition in their
A
u
operations, although few were as t
o
m
rigorous and highly structured as a
t
the Vollmann Decomposition. i
o
n
The team also found that the
decomposition tended to terminate at the plant management Figure 4.9 The Automation
Economic Gap
level for virtually all of the plants investigated. The plant manage-
ment teams we interviewed tended to understand what their per-
formance measures were, but most of the engineering,
maintenance, and operations personnel within the plant did not.
The research team ended up referring to this phenomenon as the
automation economic gap (figure 4.9).

Although most plant automation systems were installed to


help the corporation manage the performance of its manufactur-
ing operations, the performance measures did not tend to make it
far enough down to even determine the performance impact of
the automation investment. A number of plant operation and
engineering teams were interviewed throughout the research pro-
cess. They were often asked if they knew what their individual or
group performance measures were. Most seemed to believe they
knew what their measures were, but when we asked them to iden-
tify them the measures they identified were seldom aligned with
the management-level performance measures.

One example of this was particularly interesting. At one chem-


ical plant of a major global chemical company, a plant operations
and engineering team was interviewed and found to be using very
advanced automation techniques, including multivariable predic-
tive control and nonlinear optimization. When asked what the
objective function of their nonlinear optimization was, they
responded that they were optimizing to reduce the manufactur-
ing cost within the defined constraints of the process. This

B O T T O M - L I N E A U T O M A T I O N 101
BottomLine05.book Page 102 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

seemed quite reasonable until the plant manager was later inter-
viewed and told us that the primary plant objective derived from
the corporate strategy was to increase production. The engineer-
ing team was doing an absolutely excellent job of optimizing the
process to the wrong performance measure. The more successful
they were, the less successful the plant management team was at
meeting its primary objective. This is a clear example of the gap in
action.

It is interesting to note that the concepts surrounding the Voll-


mann Decomposition approach and activity-based costing are
fairly well aligned. Identifying the activities for activity-based
costing systems is often done through a similar decomposition
approach: Functions can be decomposed into processes that rep-
resent the ongoing set of activities.22 In a manufacturing envi-
ronment, the manufacturing processes would be considered
activities.23 Done correctly, the financial performance measures
would be the key activity-based accounting values at every level
right down to the process units. One key difference between tradi-
tional activity-based costing and the financial performance mea-
sures derived from a Vollmann Decomposition is that the
economic measures may include more than just cost. A Vollmann
Decomposition may also result in nonfinancial measures. In fact,
driving the performance measurement system from the corporate
strategy right down to the manufacturing process units would
lead to a new type of cost management system approach. This
new approach may be best referred to as activity-based account-
ing, and it must be capable of being accomplished in real time as
dictated by the time constant of the processes being managed.

DYNAMIC PERFORMANCE
MEASURES
Suppose accurate performance measures could be developed
that truly matched the plants manufacturing action plan. Sup-
pose a system were put in place such that these measurements
were generated and reported to manufacturing personnel on the
same basis as traditional variance reports, lets say at the end of

102 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 103 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

every other week. By the time the manufacturing people knew


what their performance was, the product they were making at the
time the measure was taken would probably already be at their
customers site. There is not much that they will be able to do
with the performance information to improve the product or the
operation. Also, even if the performance measures indicate that
the manufacturing operation had done a good job, they probably
will not even remember what they were doing back when the
measures were taken that caused them to do well, and they will
have great difficulty replicating their success. Performance mea-
sures have to be much more timely than traditional variance
reports to be truly effective and ultimately useful.

To be most useful the frequency of reported information


should follow the cycle of the production process being mea-
sured.24 The manufacturing performance measures must be pre-
sented to plant personnel in a time frame that will allow them to
respond to the information effectively. They must be real time:
Performance data should be cost-effective, available and timely.
They should be reported in a timely basis and in a format that
aides decision making.25 If the people and operations who are
actually manufacturing the product can get the measures of their
performance as the product is being produced, they will be able to
perform better and as a result the entire plant will also perform
better. As completely obvious as this may seem, very few manu-
facturing operations have put this into practice.

Measurements of manufacturing performance that reflect the


plants manufacturing strategy and are reported to the manufac-
turing personnel on a real-time basis are called dynamic perfor-
mance measures. The word dynamic has two meanings in this
phrase. First, it is intended to reflect the real-time aspects of the
measures. Second, it reflects the fact that as the manufacturing
strategy changes the values being measured has to change to
reflect the new strategy.

The computer-based technologies currently available in manu-


facturing operations provide the promise that dynamic perfor-
mance measures are technically feasible. As Harvards Peter
Drucker points out, computers have now made activity-based

B O T T O M - L I N E A U T O M A T I O N 103
BottomLine05.book Page 104 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

cost accounting possible; without them it would be practically


impossible.26 Computer-based control systems are often directly
connected to the manufacturing process via process sensors that
measure the process flows, levels, temperatures, pressures, and
many other process variables. In process plants there are typically
hundreds and even thousands of sensors measuring dozens of dif-
ferent types of process values and making these values available
to the control system on an immediate, or real-time, basis. To
construct dynamic performance measures requires making use of
this sensor-based data to model the performance measures of the
plant on a real-time basis, right in the computerized process con-
trol system. This approach has been tried in numerous process
plants, and it has been determined that if there is enough process
instrumentation to enable the control of the process there is
almost always enough to enable the modeling of the performance
measures. In this way, real-time performance measures can be
made to represent the performance of every process unit and the
entire manufacturing process on a continuing real-time basis, at
exactly the same time as the product is being made. Since the
measures are based on live process values, there is no need to
construct a model of the plant as has been attempted so many
times in the past. The information provided by the sensor-based
measures represents how the plant is actually operating. As the
plant dynamics change, and they will, the measures will immedi-
ately reflect the new dynamics.

The specific performance mea-


sures must directly relate to the
Plant
Plant job responsibilities of the people
to whom they are presented.
Each performance measure
Area
Area Area
Area Area
Area
should be totally within the
accountability of the person or
group performing the activity to
Unit
Unit Unit
Unit Unit
Unit Unit
Unit Unit
Unit
be measured.27 In most plants,
the first-level operators have
Figure 4.10 Process Plant responsibility for a subset of the entire plant operation. The oper-
Decomposition
ators are typically assigned an area of responsibility according to

104 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 105 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

the natural decomposition of a plant into areas, and areas into


units, and so on (figure 4.10).

The plant performance


measures must likewise
decompose into the same
areas of responsibility. The
decomposition must be
done in a top-down manner
from the plant-level strate-
gies right to the unit-level
performance measures (fig-
ure 4.11). Once the decom- Manufacturing
position has been
accomplished down to the unit level, the unit-level performance Figure 4.11 Plant Level
Vollmann Decomposition
measures can be modeled off of the process instrumentation data
for that unit. Since this data is typically available at the natural
time constants of the processes being measured, the models can
run in the automation system in real time. When these measures
are modeled in real time they are dynamic performance measures.
This approach will make the first-level dynamic performance
measurements available to the plant operators for that section of
the plant in the same time frame in which they perform their
operations activities. This type of information in this type of time
frame is exactly what the plant operators require to understand
how to perform better in their jobs.

A hierarchy of dynamic per-


Plant
Plant
formance measures can then be
DPMs
DPMs
developed by recomposing the R
e
plant performance measurement Area Area Area c
Area Area Area o
DPMs
DPMs DPMs
DPMs DPMs
DPMs m
hierarchy according to the (fig- p
o
ure 4.12). The implementation of s
Unit
Unit Unit
Unit Unit
Unit Unit
Unit Unit
Unit e
DPMs
DPMs DPMs
DPMs DPMs
DPMs DPMs
DPMs DPMs
DPMs
dynamic performance organiza-
tional structure and domain of
Manufacturing Process
responsibility of every individual
in the organization measures must be performed bottom-up, in Figure 4.12 Dynamic
Performance Measure
the reverse direction as the initial decomposition analysis: Oper- Hierarchy
ations need both bottom-up and top-down know-how about

B O T T O M - L I N E A U T O M A T I O N 105
BottomLine05.book Page 106 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

the plant.28 Area supervisors, for example, will have responsibil-


ity for the dynamic performance measures that are aggregated
from the dynamic performance measures of each of the operators
reporting to them. This hierarchy of performance measures pro-
vides needed bidirectional access. The lower-level performance
measures provide the input for calculating the higher level. It also
provides the measures of performance to the people who have the
responsibility and authority to manage the performance at each
level of the operation in a totally consistent manner across the
entire organization. As the dynamic performance measures are
aggregated, the time constants of higher-level measures will typi-
cally be longer than the measures at the lower level. This charac-
teristic matches the time constants of the job functions being
performed.

The visual presentation of dynamic perfor-


mance measures to all manufacturing per-
sonnel should be simple and straightforward.
For example, the graphic approach com-
monly used to present statistical data in a
SQC environment is well accepted and sim-
ple, and traditional process gauges have
proved to be easy for process operations per-
sonnel to deal with. Figure 4.13 shows an
example of a display that incorporates both
of these concepts. This display is referred to
Figure 4.13 Performance as a performance dashboard because it encompasses all the
Dashboard
information necessary for an operator to drive the performance of
the plant in a form similar to an automobile or airplane dash-
board. The modeled performance measures can be displayed
along with their current performance targets and an indicator
that shows which direction of movement means good perfor-
mance. Notice that the dashboard in the figure only displays the
value of four performance measures for the operator. It is critical
to keep the number of measures down so that there is no ambigu-
ity about where people should be focusing their energy.29 Also,
very often the performance measures have a tendency to fight
each other. As an action causes one measure to improve, the same
action may cause another measure to move in the wrong direc-

106 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 107 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

tion. Performance measures must, therefore, be prioritized


according to those strategic success factors deemed most criti-
cal.30 In the dashboard display shown in figure 4.13, the mea-
sures are prioritized by their placement on the display: left to
right with the left-most measures being higher priority.

The performance dashboards can follow the dynamic perfor-


mance measures up through the organization (figure 4.14). The
dashboard displays can be relatively consistent with the measures
that are displayed to reflect the area of operational responsibility
at each organizational node. This structure ensures that all per-
sonnel in the operation are working in concert and serves to
encourage teamwork because all people at each operational node
are working to improve the same measures.

Dynamic performance measures


represent the application of an Plant Management
Dashboards
important new management tool

Measurement Conformance
that has been missing from tradi-
tional manufacturing systemsa Operations
Management
tool to implement real-time activity- Dashboards

based accounting models right down Operation


Dashboards
to the plant floor. It expands the
scope of traditional cost accounting
well beyond just financial measures Manufacturing
to a new and more wide-ranging
concept of performance measures. Using this real-time activity- Figure 4.14 The
Performance Measure
base accounting, management can finally use the accounting sys- Hierarchy
tem as a real-time decision support system throughout their orga-
nizations. This new approach can help every person in the
organization work to drive the organizations bottom line in con-
cert with its strategic plan. This changes the focus of accounting
from after-the-fact cost accounting to accounting for the bottom
line.

This real-time activity-based accounting tool in itself can


become the first step toward world-class manufacturing but, by
itself, is clearly not enough. Management must view this as
merely a tool that allows them to fill the economic-automation
gap that has been a major barrier to performance improvement

B O T T O M - L I N E A U T O M A T I O N 107
BottomLine05.book Page 108 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

and strategy deployment for decades. The advent of real-time


activity-based accounting models merely provides them with a
facility to help them drive performance and execute competitive
strategy. Management must be willing to exert leadership in using
this new technology to drive their manufacturing operations to
new, world-class levels of operation and performance.

DYNAMIC PERFORMANCE
MEASURES AND ACCOUNTING
This chapter has explained how Dynamic Performance Mea-
sures (DPMs) were a direct result of the evolution of process
requirements that traditional cost accounting systems did not
address. Once DPMs were installed and operating in process
plants, the logical question then became whether and how they
should associate with and link into these traditional cost account-
ing systems. Many DPMs are direct models of the cost or profit
associated with a unit or area of a manufacturing process. One of
the objectives of accounting systems is to resolve cost and profit
issues, but the first DPMs to be installed were typically totally
independent of their plants accounting systems.

This separation of the DPMs from the accounting systems was


initially not of much concern. The DPMs were being installed to
generate information that had not traditionally been available, so
making them available to accounting systems did not seem
imperative, or even particularly desirable. DPMs were typically
installed in the section of a process plant in which a drive was
underway to perform activities that would improve economic
performance. The DPMs provided the operations, engineering,
and maintenance personnel with direct, real-time feedback on the
performance of their section of the plant, and also allowed them
to determine the economic impact of specific improvement activ-
ities. The problem was that when an activity had been completed
and its economic value determined, the plants management
would often ask the finance department to provide confirmation
of the economic improvement. The only tool the finance depart-
ment had to do this was the installed accounting system. As

108 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 109 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

explained earlier in this book, the cost accounting system lacked


the information necessary to confirm the improvement. If the
finance department could not confirm the improvement, as far as
plant management was concerned the improvement had not
occurred.

This is not a new issue. Over many years, for example, a con-
siderable amount of intellectual power has been focused on opti-
mizing the operations of oil refineries. Sophisticated techniques
such as linear and nonlinear programming have been effectively
optimized both refinery planning and scheduling and unit opera-
tions. I clearly recall how very impressed I was with the optimiza-
tion team of one refining company as they discussed and
presented the advanced methods they were using in their daily
operations to ensure that their refineries were operating at the
highest potential. During our conversation, there was quite a bit
of debate among the refinery team members, which demon-
strated their unprecedented level of commitment, experience,
and capability in applying highly sophisticated mathematical
models to their operations. As the discussions concluded, I
expressed my admiration for the team and asked if they knew
how much economic value they had generated for their company
in the past year. They responded with a value well over $20 mil-
lion. I then asked if their CFO agreed with this assessment. After a
bit of a pause, the leader of the group stated that over the many
years he had worked for the company the finance department
had never been able to confirm the value the group had gener-
ated. He went on to say that management would not accept as
accurate the value the optimization team said it was generating in
the refineries unless the finance department could confirm it.

In another case, DPMs were installed in a section of a pharma-


ceutical plant. Plant personnel determined baseline values for the
measures by operating the DPMs so the operators werent aware
of them and periodically storing the DPMs values in a plant his-
torian. After several weeks, the plant personnel had determined
the average weekly value of the measures for the baseline period
and began using them as the starting baseline for measuring plant
performance. After this baseline period, the plant personnel

B O T T O M - L I N E A U T O M A T I O N 109
BottomLine05.book Page 110 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

undertook several initiatives that they felt would positively


impact their sections economic performance. Among other activ-
ities, they installed advance process controls, trained operators,
and determined and disseminated best practices for their opera-
tions. The DPMs showed that the plant teams activities had
improved the plants economic performance by approximately
$1.5 million annually. When this information was conveyed to
the plant manager, he asked the plants financial manager to con-
firm the results. Fortunately, the finance team in this plant was
more open minded than most and worked diligently to detect the
improvement. Several months later, the finance team confirmed
that there had been a marked economic improvement in plant
operations, on the same order as the operations team had
reported, but they could not determine where the improvement
came from. Again, it was fortunate for this plant that this general
confirmation of improvement was sufficient to convince the
plant management team that the operations team had accurately
estimated the value that the performance improvements had con-
tributed. Nevertheless, cases like these were demonstrating the
need to somehow bring the DPMs and accounting systems
together.

Converging DPMs with cost accounting systems would appear


to be fairly straightforward. However, many of the same limita-
tions in accounting systems that had led to the need for DPMs in
the first place hampered efforts to bridge the gap between tradi-
tional accounting and DPMs. Much of the functionality of
accounting systems has been defined by governmental reporting
and auditing requirements, not by the operating requirements of
the plant. Because of this, the accounting systems have not
evolved to be well aligned with operational information require-
ments. A plant-centric view of a traditional cost accounting sys-
tem is shown in figure 4.15. Although this is an overly simplified
view, it demonstrates the gap between accounting and DPMs. To
perform the plants accounting functions, its accounting staff
inputs the plants total energy consumption, raw material con-
sumption, and production volume over a given period (typically a
month) into the accounting system along with other direct and
indirect costs. The plant accounting system stores this informa-

110 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 111 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

tion in its database, analyzes it according to predetermined rules,


produces several reports, and provides summary information to
the plants enterprise resource planning (ERP) system for further
analysis and reporting.

One of the reports gener-


ated by the accounting sys- Enterprise Resource Planning System
Plant Financial Measures
tems is the monthly variance
Total Material Consumed Plant Level
report, which compares the Total Product Made
Total Energy Consumed Cost Accounting
expected standard cost per System
Variance Report
unit of product produced Standard vs Actual
Cost per Unit Product
against the actual cost per Process Plant
unit of product produced.
This type of variance report Energy
has been the most frequently Products
used system for measuring Materials
manufacturing performance
over the past century.
However, this measurement approach has many basic flaws, Figure 4.15 Current Plant
Accounting Model
which have been addressed extensively in numerous articles and
books. But two of these flaws are particularly relevant to the effort
to link these traditional accounting systems with the DPM infor-
mation. First, the resolution or granularity of the data in tradi-
tional accounting systems typically stops at the total plant level.
DPMs, on the other hand, provide cost and profit data at the
plant-unit and plant-area levels, and the traditional accounting
database provides no location to which information about these
levels can be linked. The second disconnect involves timing. Cost
accounting systems use a cost window of months, weeks, and in
some advanced cases days. DPMs, by contrast, are calculated in
real time to provide the timely information required to support
operational decision-making. As plants implemented the early
DPM systems and attempted to link them into the accounting sys-
tems, these two issues became significant barriers.

The advent of activity-based costing (ABC) systems offered


plant personnel a ray of hope that DPMs could be linked to their
plants finance systems. (Activity-based accounting systems were
introduced earlier.) As we have seen, this relatively new approach

B O T T O M - L I N E A U T O M A T I O N 111
BottomLine05.book Page 112 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

to manufacturing accounting was developed to try to compensate


for some of the deficiencies in cost accounting for manufacturing.
Although the concepts of ABC have been around for many years,
ABC systems only really began to take shape and gain popularity
in the late 1980s. ABC systems provide an alternative to the tradi-
tional general ledger-based accounting practices because they
focus on the activities related to product manufacturing within
an operation. ABC systems, for example, develop accounting
views for each defined activity, which provides accounting data of
much higher resolution or granularity. The manufacturing pro-
cess equipment installed in a process plant defines the plants pri-
mary manufacturing activities. Although most ABC systems
developed in the past decade are targeted to discrete manufactur-
ing operations, somesuch as the Prism and Protean systems
developed at Marcam and the Ross System iRenaissance pack-
agewere specifically designed with the characteristics of process
manufacturing in mind.

To model the activities


Enterprise Resource Planning System in process plants these
Plant Financial Measures
Total Material Consumed
ABC systems have a
Plant Level Total Product Made
Total Energy Consumed Cost Accounting production model (see
System
Allocation-Based figure 4.16) that pro-
Total Production Model
Plant vides a database view
Metrics
Process Plant capable of matching
the equipment layout
Energy of the process. This
Products appears to be well
Materials
aligned with the DPM
approach since the
Figure 4.16 Production basic activity level of a production model can be a process unit,
Model-based Accounting
System
which is also the basic level for DPMs in process plants. The prob-
lem, however, was that the process manufacturing plants in
which the first ABC systems were implemented still used informa-
tion on total plant resources (energy cost, raw material costs, and
so on), so allocation algorithms had to be devised to allocate these
total costs to the individual activities. For this reason, any
improvement made to the operation of a single process unit had
to be input into the ABC system as an improvement on plant

112 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 113 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

totals and then allocated to all individual process units in accor-


dance with the predetermined allocation algorithms. This cum-
bersome allocation approach made it very difficult to rationalize
these ABC process systems with the DPMs.

Adding a produc-
Plant Level
tion model to plant Cost Accounting System

accounting systems
was a major step for-
ward, but most early Plant Plant Plant
Area Area Area
implementers did
not take full advan- DPMs
Unit Unit Unit Unit Unit
tage of their new sys-
In Process Plants Unit Unit Unit Unit
tems. Perhaps one of Process Units
Define the Activities
the reasons was that
DPMs did not exist in Energy
Products
the plants where the
Materials
first ABC systems
were installed. Process Plant
Without DPMs, the only reasonable approach to determining the Figure 4.17 Performance-
based Plant Accounting
costs and profits for the activities in a process plant was to allocate Model
backward from the plants total costs. With DPMs, the cost and
profit of each activity (process unit) in a process manufacturing
operation can be directly determined from DPM models, which
use sensor-based data (figure 4.17). These calculated DPMs can be
used to directly assign values into the ABC systems production
model without having to rely on allocation algorithms. By
directly assigning the values, the cost and profit information in
the ABC system becomes much more representative of what is
actually occurring in the manufacturing process.

The alignment of DPMs with process ABC systems provides a


new level of operational information that can be tightly aligned
with a plants finance system. Unfortunately, ABC systems have
not been implemented in manufacturing to any large extent thus
far. There are several reasons for this. First, as the interest in and
activities surrounding ABC systems began to grow in the late
1990s, the Y2K scare was commanding the attention of most IT
and finance departments. Considerable cost and effort was

B O T T O M - L I N E A U T O M A T I O N 113
BottomLine05.book Page 114 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

expended on migrating to a standard ERP system that was Y2K


compliant. The focus on new approaches, such as ABC, took a
back seat. Second, ABC was often presented as an alternative to a
traditional accounting system rather than a supplemental system
that could complement ERP systems. Once manufacturing com-
panies had invested all their time and effort into upgrading their
ERP systems the last thing they wanted to hear was that they
should replace them with a new type of system! Third, imple-
menting ABC systems is much more complex than implementing
traditional accounting systems. The individual activities have to
be identified and measured, which adds a level of complication
not required by traditional ERP systems. Largely because of these
factors manufacturing companies have not moved to ABC sys-
tems as readily as was once expected.

In more recent years,


Financial Reporting however, interest has
Systems
grown significantly in
ABC and its use as a
decision-support and
management toolkit
Planning and
Budgeting
(often referred to as
Activity-Based Operational and Strategic
Management Performance Measurement activity-based man-
Systems System
Actual Utilization agement or ABM) as
and Efficiencies
well as in new perfor-
mance measurement
Kaplan, Robert S.; Cooper, Robin; Cost & Effect, Using Integrated Cost Systems to Drive Profitability and Performance; Harvard Business Press, 1998.
systems. This may be
Figure 4.18 Model for Cost indicative of a transition in cost accounting that will help to bring
and Performance
Measurement (Stage IV)
a much more holistic approach to the effective management of
process manufacturing operations. In their groundbreaking book
Cost and Effect, Robert S. Kaplan and Robin Cooper provide an
excellent and compelling perspective on the evolution of cost
accounting. They partition that evolution into four stages, evolv-
ing from the homegrown systems that dominated cost accounting
throughout the past century through standard ERP-based systems
all the way to what they predict will be the cost accounting
approach of the future (they refer to it as Stage IV). Their assess-
ment is that todays ERP-centric view of a single accounting sys-
tem will not suffice. ERP-based accounting systems are absolutely

114 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 115 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

necessary for fiduciary reporting, but they are not sufficient for
supporting operational decision-making. Kaplan and Cooper do
not view ABC/ABM systems as alternatives to ERP systems but as
supplementary operational support systems. Their model of the
Stage IV system (figure 4.18) projects the optimal structure for the
next generation of performance-based accounting.

In the Stage IV model, ERP systems perform the functions of


financial reporting that are necessary to meet fiduciary require-
ments, but they do not provide the information needed for opera-
tions decision support. They predict that to support a
performance- based accounting approach along with the ERP sys-
tems, manufacturers will require not only ABM systems but an
Operational and Strategic Performance Measurement system.
DPMs provide the real-time, sensor-based operational and strate-
gic performance measurements. These can be directly connected
into the ABM system to assign costs and profits to the activities in
the ABM database. They can also be used to provide a continuous
learning opportunity for the operators, engineers, and mainte-
nance personnel in the plant. Other performance measures that
cannot be modeled from process sensors will be required to com-
plete the Operational and Strategic Performance Measurement
system. The ABM systems operate as entities independent of the
ERP systems. They take input from the Operational and Strategic
Performance Measurement system and use it to directly deter-
mine cost and profit at a subplant level. ABM systems are ideal for
short- to medium-term production planning and scheduling and
can provide performance targets for the Operational and Strategic
Performance Measurement system.

In Kaplan and Coopers Stage IV model the communication


path they predicted extends from the ABM system and the Opera-
tional and Strategic Performance Measurement system up to the
ERP system. This is exactly the opposite of the most common
practice in manufacturing-related cost accounting today. The cur-
rent common practice is to try to extract information from the
ERP system to help drive the product-costing, ABC, and perfor-
mance measurement (variance reports) systems. However, the
Stage IV model makes more sense in that its premise is that cost,

B O T T O M - L I N E A U T O M A T I O N 115
BottomLine05.book Page 116 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

profit, and other critical operational information can be most


accurately determined at the point at which it is realized the
manufacturing process. Thus, the accounting systems should be
driven from this base information and not the other way around.
If organizations implement a Cost and Performance Measurement
System in the way stipulated by the Stage IV model their account-
ing information will not only be more accurate. It will also ensure
that the organization is aligned with its strategy as long as the
strategic performance measurements are used to help drive opera-
tions. This Stage IV model is clearly of greater scope than the
DPM approach, but the two approaches are totally complemen-
tary. Kaplan and Cooper present the Stage IV model as an ideal
that may become reality in the future. But with the advent of
DPMs in process manufacturing it is already available today.

The combined DPM, ABM, and ERP approach appeals natu-


rally to the common sense of managers striving to optimize the
performance of their process plants. However, I should point out
that very few companies have started to move in this direction.
Implementing an effective ABM system is a very complex under-
taking. Fortunately, in process plants the definition of the pro-
cesss basic activities dont vary as much as they do in discrete
manufacturing. This gives process manufacturing environments a
level of stability once the ABM systems are implemented. Still,
few companies have had the drive and patience to implement
ABM systems, and the same is true for Operational and Strategic
Performance Measurement systems and DPMs. The Vollmann
Decomposition process presented in this book makes it possible
to identify a companys strategic performance measures, which
are the measures that should primarily be used for operations
decision support. Identifying operational performance measures
is a pure accounting task.

There is a degree of flexibility entailed in the performance


measurement system that can further complicate the mainte-
nance of the system. If market conditions or an organizations
market strategy change, its strategic performance measures
should change accordingly. Typically, this involves changing the
priority of the measures, but it could involve more drastic alter-

116 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 117 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

ations, such as adding new strategic performance measures. It is


this complexity that seems to have retarded industrys adoption
of these systems. What will overcome this barrier is the realiza-
tion that the economic value these systems can contribute to
organizations can more than outweigh the cost and effort of
implementing and maintaining them. The case studies presented
in this book provide preliminary evidence of this.

Process manufacturers have also been slow to adopt ABM sys-


tems because most commercially available ABC/ABM systems
have been designed to meet the requirements of discrete manu-
facturing operations. Fortunately, as we have noted, a few sys-
tems, such as Marcams Prism, Protean and Ross System
iRenaissance systems, have recently emerged that are specifically
designed for the requirements of process plants. These process-
based systems may help to drive the acceptance of the ABM and
ABC approach in process plants. They are also well aligned with
DPMs. As more of these systems are installed and operational the
economic value of a plants activities will become clear to their
finance departments as well as to the operations, engineering,
and maintenance departments. This clarity and agreement on the
economic value-added will enable the lean process organizations
to focus their limited resources on the areas that offer the most
potential positive impact on the plant and company. The result
will be process manufacturing plants performing at a higher oper-
ational and economic level than was ever before possible.

NOTES
1. Berliner, Callie, and James A. Brimson, Cost Management for
Todays Advanced Manufacturing. Boston: Harvard Business
School Press, 1988, page 39.

2. Ness, Joseph A., and Thomas G. Cucuzza, Tapping the Full


Potential of ABC, Harvard Business Review on Measuring Cor-
porate Performance. Boston: Harvard Business Press, 1998,
page 54.

B O T T O M - L I N E A U T O M A T I O N 117
BottomLine05.book Page 118 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

3. Berliner and Brimson, Cost Management for Todays Advanced


Manufacturing, page 25.
4. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 93.
5. Henderson, Bruce A., and Jorge L. Larco, Lean Transformation.
Richmond, Virginia: Oaklea Press, 1999, page 203.
6. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 34.
7. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 19.
8. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 26.
9. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 1.
10. Hiromoto, Toshiro, Another Hidden Edge Japanese Man-
agement Accounting. Harvard Business Review, July-August
1988, page 26.
11. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 2.
12. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 1.
13. Kaplan, Robert S., and David P. Norton, The Balanced Score-
card Measures that Drive Performance, Harvard Business
Review on Measuring Corporate Performance. Boston: Harvard
Business Press,1998, page 135.
14. Eccles, Robert G., The Performance Measurement Mani-
festo, Harvard Business Review on Measuring Corporate Perfor-
mance. Boston: Harvard Business Press, 1998, page 27.
15. Meyer, Christopher, How the Right Measures Help Teams
Excel, Harvard Business Review on Measuring Corporate Perfor-
mance. Boston: Harvard Business Press. 1998, page 108.
16. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 15.
17. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 164.

118 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 119 Friday, August 19, 2005 9:54 AM

C H A P T E R 4 : C O S T A C C O U N T I N G A N D T H E B O T T O M L I N E

18. Berliner and Brimson, Cost Management for Todays Advanced


Manufacturing, page167.
19. Drucker, Peter F., The Information Executives Truly Need,
Harvard Business Review on Measuring Corporate Performance.
Boston: Harvard Business Press, 1998, page 19.
20. Peters, Thomas J., and Robert H. Waterman, In Search of Excel-
lence: Lessons from Americas Best Run Companies. New York:
Harper & Row, 1979, page 7.
21. Drucker, Peter F., Innovation and Entrepreneurship: Practices and
Principles. New York: Harper & Row, 1985, page 161.
22. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 6.
23. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 6.
24. Kaplan, Robert S., One Cost System Isnt Enough. Harvard
Business Review, January February 1988, page 62.
25. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 167.
26. Drucker, The Information Executives Truly Need, page 23.
27. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 165.
28. Zuboff, Shoshana, In the Age of the Smart Machine. New York:
Basic Books, 1988, page 95.
29. Davila, Antonio, and Robert Simons, How High Is Your
Return on Management?, Harvard Business Review on Measur-
ing Corporate Performance. Boston: Harvard Business Press,
1998, page 87.
30. Berliner and Brimson, Cost Management for Todays Advanced
Manufacturing, page 163.

B O T T O M - L I N E A U T O M A T I O N 119
BottomLine05.book Page 121 Friday, August 19, 2005 9:54 AM

CHAPTER 5

The
Convergence

The tools exist here, now, today to forge


an entirely new paradigm for managing process Technology
plant performance. This paradigm is firmly rooted
Bottom
in the convergence of the historical trends the Line
Quality
executives we interviewed identified: technology, Improvement

quality, and accounting. Further, it can be forged


Accounting
into a dynamic management structure designed to
drive performance in process manufacturing and
enable the manufacturing operations, along with all other key Figure 5.1 The
Convergence: Survival in a
organizational operations, to be utilized as key components of a Global Economy
companys strategy (figure 5.1).

Before describing this new management paradigm, its appro-


priate to take a few moments to look at the relationship between
the dynamic performance measures and the three historical
trends we have discussed in this book. Dynamic performance
measures provide the next step in the logical progression of each
of these three trends and are, therefore, critical to their conver-
gence. It will therefore be helpful to review each of these trends
while we examine how dynamic performance measures provide
the next logical step in the progression of each.

B O T T O M - L I N E A U T O M A T I O N 121
BottomLine05.book Page 122 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

THE TECHNOLOGY TREND AND


DYNAMIC PERFORMANCE
MEASURES
Manufacturing managers have made significant capital expen-
ditures in automation technology over the past thirty years, with
little or nothing to show for it. Even with programs such as CIM,
millions of dollars were spent with no discernible performance
improvement. Many manufacturing managers have indicated
that their plants are not running any better today then they were
three decades ago, before the existence of computer-based auto-
mation systems. This is unfortunate and totally unacceptable.
Manufacturing managers should be seeing real returns on all of
their capital investments, and automation systems should be
leading the pack.

The only way to get more payback out of automation invest-


ments is if the new technology provides increased functionality
that improves the economic performance of the plant. The
dynamic performance measures provide the information that will
continually communicate the performance impact of any capital
or operational investments. Automation suppliers have invested
tremendous amounts of money to increase the functionality of
their automation systems and software in order to provide manu-
facturers with improved capability to, in turn, drive the economic
performance of their manufacturing operations. The expanded
performance-enabling functionality embedded in the automation
systems is seldom utilized because of the replacement automa-
tion and project team approaches that are almost universally
used when implementing automation projects in process manu-
facturing environments. As pointed out earlier, the fact that there
has been no measurement system in place indicates the economic
impact from automation has actually shielded the lack of perfor-
mance improvement from management. For years, manufactur-
ing management has tended to approve capital budgets for
automation projects on faith, assuming that they must be deliver-
ing reasonable returns. In the difficult competitive environment
that globalization has spawned over the past decade, manage-
ments patience seems to be wearing thin. It can no longer afford

122 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 123 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

to approve capital expenditures without a clear understanding of


the economic value-added generated as a result.

When dynamic performance


measures are implemented in these
manufacturing operations, the
automation-with-no-measure-of- Life-cycle Benefit Continuous
Improvement
return approach is bound to
change. The dynamic performance $
measures provide the performance
score cards that will continually
report on whether the capital invest- Life-cycle Cost

ments being made are providing


economic returns. This fact should serve to drive the engineering Figure 5.2 Improving
Economic Performance
and operations staff to get more performance from the automa-
tion investments by utilizing the advanced capabilities of these
systems. Even the very limited data collected so far has demon-
strated that the potential for improvement is enormous. But the
improvement does not happen merely by acquiring a new, up-to-
date automation system. As with any performance-enabling tools,
the advanced features of the system must be used, and used effec-
tively, to realize performance improvements. With the reductions
in the size of plant engineering and operations staffs, the highly
skilled personnel required to implement and manage the
advanced automation features may not be available. This is caus-
ing automation suppliers to change the focus and scope of their
market niches beyond just hardware and software companies by
expanding their service offerings to fill the resource gap. Automa-
tion still offers one of the best opportunities for managing plant
costs and production. Coupling advanced automation solutions
with dynamic performance measures provides a proven way to
continually increase the economic benefits of automation and
improve the returns on the automation investments (figure 5.2).

The effective implementation of dynamic performance mea-


sures requires the use of computer-based technology. As Robert
Eccles has noted, Information technology has played a critical
role in making a performance measurement revolution possi-
ble.1 For these measures to accurately reflect the current status of

B O T T O M - L I N E A U T O M A T I O N 123
BottomLine05.book Page 124 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

the manufacturing operation, the system that models them must


be real time, and it must utilize the data about the manufacturing
process provided directly by process sensors to generate the
appropriate performance measure values at each level in the oper-
ation. The technology in real-time systems has become sophisti-
cated enough to make the implementation of dynamic
performance measure systems not only feasible but also actually
quite simple, and very desirable. As the technology trend contin-
ues to evolve, organizations ability to use automation technology
to implement dynamic performance measures is driving the
change from technology for its own sake to technology for the
bottom line.

THE QUALITY TREND AND


DYNAMIC PERFORMANCE
MEASURES
The quality movement has been growing very rapidly in the
industrialized world for the past few decades. One of the key indi-
cators of this is the increased use of SPC packages throughout
manufacturing. It is unusual today to go into a plant that has not
implemented SPC to some extent in its operation.

Many manufacturers in the process industry segments have


tried to utilize the SPC packages that are popular in the discrete
manufacturing segments. Although these implementations pro-
vide tidy and appealing SPC charts, the statistical techniques they
employ do not translate directly to continuous process variables
and they consequently produce suspect results. As Paul Badavas
and Albert Epperly note, For nondynamic processes such as parts
manufacturing, the SPC discipline applies directly to the measure-
ments and actuators. While for dynamic processes, such as distil-
lation columns, SPC applies after Traditional Process Control
(TPC) has driven the variables in question to steady state. The
upper-level functions also profit from the statistical validity of
data.2

This is not to say that the principles of SPC /SQC do not apply
in process plantsthey do. The point is that applying SPC/SQC in

124 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 125 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

these environments requires careful and close investigation to


achieve the desired continuous improvement.

Process plants have traditionally employed a much higher


level of deterministic process control than that found in discrete
factories. Because it has been prohibitive to provide deterministic
control over discrete work cells, statistics were applied, and SPC/
SQC evolved. This statistical approach was very successful at
bringing the level of control in some discrete operations close to
that of continuous process plants. But higher levels of control can
still be realized. It is advisable to continually ask ourselves what
aspects of the operation require the next level of focus for contin-
uous improvement. In other words, what is the next level of qual-
ity indicators in these manufacturing operations?

If you walk into a process plant that has implemented a rea-


sonable level of process control, and if you ask the operations and
production managers what they need to continuously improve,
their answer will often be quite simple: plant performance.
Although this may seem an obvious response, it is actually very
significant: Quality-related metrics have made the performance
measurement revolution more real.3 The key to success in
todays marketplace is the performance of the entire plant, thus
the plant quality teams need to find the next level of quality indi-
cators (i.e., that are the indicators of the plant performance) at
each point in the operation. These indicators are the dynamic
performance measures for the operation. Most executives have
not yet realized, however, that such teams need new perfor-
mance-measurement systems to fulfill their promise.4 Until
management recognizes this issue and takes the leadership neces-
sary to drive the dynamic performance measures through their
operations and use them to empower their manufacturing quality
teams and plant personnel, the returns on the quality invest-
ments will certainly not be what they could be.

Christopher Meyers observes that A well-designed financial


control system can actually enhance rather than inhibit an orga-
nizations total quality management program.5 The financially
based dynamic performance measures provide the components of
the well-designed financial control system that is required to

B O T T O M - L I N E A U T O M A T I O N 125
BottomLine05.book Page 126 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

achieve effective quality management in process plants. Although


the financial metrics may be important in a quality management
system, nonfinancial metrics are also very important. Many plant
operations should have nonfinancial measures of performance,
and these should be key components of the continuous improve-
ment programs of quality teams and individuals in process envi-
ronments.

A reasonable conclusion from this is that dynamic perfor-


mance measures provide a new and effective approach for devel-
oping quality indicators in process operations and as such provide
the key to the next logical step in the quality trend. Utilizing the
dynamic performance measures as integral elements of a plants
total quality management program will make the program much
more effective and facilitate the transition to quality for the bot-
tom line.

THE ACCOUNTING TREND AND


DYNAMIC PERFORMANCE
MEASURES
Cost accounting systems have been the basis for measuring
manufacturing performance for many years. Every month, manu-
facturing managers receive their variance reports informing them
whether they did better or worse than expected in terms of the
cost-per-unit of the products their operations makes. If the cost is
consistently smaller than expected, they do well in the organiza-
tion. If it gets larger, they do not do as well.

Today, much of the manufacturing world recognizes that the


traditional cost-based performance measurement system is not
effective or relevant for manufacturing operations. There seems to
be a growing awareness that new measurement systems are
required: Activity-based costing represents both a different con-
cept of the business process, especially for manufacturers, and a
different way of measuring.6 But even activity-based costing
does not go far enough to meet the optimal manufacturing per-
formance requirements.

126 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 127 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

It is generally accepted that the new measurement systems


must take into consideration both the financial and nonfinancial
aspects of an operation. Certainly, whatever the measures of per-
formance are, they must match the plant's manufacturing strat-
egy. People will perform to their measures. If the measures match
the strategy, the strategy will be implemented; if the measures do
not match the strategy, the strategy cannot possibly be imple-
mented.

To ensure effectiveness, the performance measures must also


be provided in a much more timely manner than those of tradi-
tional cost accounting systems. They must be provided in a time
frame directly relative to the manufacture of the products. In
other words, at least at the lowest levels of the manufacturing
operation, they have to be provided in real time. The manufactur-
ing personnel will not be able to improve their performance if
they do not know how they are performing as they do their jobs.
This real-time performance measurement approach is absolutely
critical to their success.

Dynamic performance measures are real-time measurements of


manufacturing performance that match the manufacturing strat-
egy of the operation. The implementation of dynamic perfor-
mance measures in a manufacturing operation results in either
the separation of the manufacturing performance measurements
from the cost accounting systems performance measures or the
fundamental restructuring of the way in which activity-based
accounting is done. Either approach can provide excellent results.

Since traditional cost accounting has a set of performance


measures built in that may not align with the dynamic perfor-
mance measures, the most effective approach is probably the sec-
ond: viewing dynamic performance measures as a fundamental
restructuring of accounting for manufacturing. This new account-
ing approach can be referred to as real-time, activity-based
accounting and can change accounting from an after-the-fact
cost monitoring system to a performance-based, decision support
tool. If the Vollmann Decomposition approach is used to develop
the dynamic performance measures, this performance-based deci-
sion support tool will help drive consistency of purpose and per-

B O T T O M - L I N E A U T O M A T I O N 127
BottomLine05.book Page 128 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

formance throughout the operation. Real-time activity-based


accounting provides a new accounting paradigm to promote
accounting for the bottom line. Thus, dynamic performance mea-
sures provide the next logical step in the accounting trend that
has been evolving over decades.

THE CONVERGENCE
When the same basic technology and/or approach provides
the next logical step in the progression of multiple trends it is
indicates that the trends are converging. This is exactly the situa-
tion we see with respect to dynamic performance measures and
the technology, quality, and accounting trends. Traditionally,
these three trends have been viewed and dealt with as though
they were quite independent. The reality is that the trends and
the drivers behind them have been very closely aligned, but since
in most manufacturing organizations different departments have
focused on each trend separately their alignment has often been
overlooked.

A convergence of this type may appear to be innocuous on


first inspection. But as Peter Drucker has pointed out, a charac-
teristic of knowledge-based innovations is that they are almost
never based on one factor, but on a convergence of several kinds
of knowledge, not all of the scientific or technological.7 In
todays market, knowledge-based innovations typically provide
maximum impact on businesses, even though they may initially
seem rather subtle. This is the case with dynamic performance
measures and the convergence of the three trends: the implica-
tions of this convergence on manufacturing operations are very
far reaching. They are leading to a fundamentally new paradigm
for the management and operations of manufacturing facilities.
For manufacturing operations to gain maximum benefit from the
convergence they must implement the dynamic performance
measurement structure and move to a new management struc-
ture, approach, and attitude.

128 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 129 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

A NEW MANAGEMENT PARADIGM


As we have seen, a new management paradigm is starting to
evolve that takes advantage of the availability of dynamic perfor-
mance measures. It is a naturally occurring management structure
that results from the convergence of the three industrial trends
discussed in this book (figure 5.3).

The objective of this new structure is fundamentally simple: to


take advantage of all key manufacturing resources by combining
them so as to maximize the performance of the entire operation.
Fortuitously, the convergence of the three trends does exactly
that.

This is fundamentally the


same objective as the lean Technology
manufacturing approach so
effectively implemented in Bottom New
Quality Line Management
discrete manufacturing oper- Improvement Paradigm
ations. Many discrete manu-
facturing companies have Accounting
revolutionized their
manufacturing performance by effectively applying the six lean Figure 5.3 The
Convergence
principles we discussed in chapter 3 (see figure 5.4).

How to use the lean enterprise approach to achieve the


same level of revolutionary performance in process manu- Continuous
facturing environments has been far from obvious. Flow
Production
Six
Though the characteristics of process manufacturing have Empowered
Sigma
Teams
Quality
made it very difficult to effectively apply all six lean princi- Lean
ples in process operations, the availability of dynamic per- Enterprise
Workplace
Visual
formance measures makes lean process manufacturing safety, order,
Management
cleanliness
operations a possibility. Pursuit
of
Perfection
The following is a short overview of how each of the
lean principles may apply in process manufacturing
environments and how dynamic performance measures actually Figure 5.4 Lean Enterprise
enable the application of some of the principles.

1. Continuous Flow Production As we pointed out earlier,


most pure process manufacturing environments are con-
tinuous flow by nature and design. In fact, this is the one

B O T T O M - L I N E A U T O M A T I O N 129
BottomLine05.book Page 130 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

lean principle that is a natural aspect of process manufac-


turing and thus that discrete manufacturing has had to
emulate in order to become lean.

2. Six Sigma Quality As we discussed in chapter 3, applying


the principles of quality management in process plants
has been an ongoing challenge. Using the dynamic per-
formance measures as the quality indicators for plant per-
formance helps to drive process manufacturing operations
in the right direction. Clearly, it is difficult to think about
process manufacturing in terms of defects per million, but
the concepts of continuous improvement certainly apply.
The statistic six sigma is not the key to this principle; a
measurable way to drive the operation toward perfection
is. Utilizing the dynamic performance measures as the
quality indicators in a six sigma environment provides
that drive for perfection.

3. Empowered Teams The concept of empowered teams


stems directly from the quality trend and TQM. To
empower teams effectively, management must be able to
measure their performance on an ongoing basis. The
teams themselves also require their own ongoing perfor-
mance measures. This has been the primary dilemma in
empowering teams in manufacturing environments and
ensuring their success. In process manufacturing plants
that have implemented the dynamic performance mea-
sures, the basis for empowering the teams exists. The
dynamic performance measures, established from the
manufacturing strategy, are the measures to which the
empowered teams should work. If the dynamic perfor-
mance measures are decomposed out of the strategy and
the empowered teams use them to measure their perfor-
mance, management can feel comfortable that the fruits
of the empowered teams efforts will be very positive.

4. Visual Management The dynamic performance measure


performance dashboards offer a very effective methodol-
ogy for providing visual management throughout manu-
facturing operations. The fact that the dynamic

130 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 131 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

performance measures are derived from actual process


sensors means that the timeliness of the performance
dashboard updates can be made to match the time con-
stants of the jobs being performed. The dashboards of the
front-line operators can update in seconds. This allows
them to discern the performance impact of their actions
as soon as they take the actions. As the dashboard displays
for upper levels of management are developed, the time
constants of the data update may be longer than that of
the operators, but they should be right in line with the
time constants of their activities. This provides the simple
and clear visual management approach necessary to drive
the lean enterprise concepts through process operations.

5. Workplace Safety, Order, Cleanliness In process plants, the


operator is often removed from the process and located in
a control room. As a result, the concepts behind this
principle may not translate as directly in process
environments. As we noted in chapter 3, the control room
environments are typically cleaner and more orderly than
the manufacturing plant environments, but most control
room environments could be significantly improved.
Providing the information the operators need to be
effective in a simple, clear, and orderly way on the
operator workstations helps to implement this principle
of lean production.

6. Pursuit of Perfection The fundamental concept of pursuit


of perfection implies that there is a set of identifiable and
measurable objectives that essentially define what perfec-
tion is relative to current conditions. DPMs provide the
basis for the definition of perfection in process manu-
facturing operations and help convert the pursuit of per-
fection from a nebulous concept to very specific and
highly visible goals. Providing DPM dashboard displays
across the entirety of a process plant helps to align all
operations, engineering and maintenance personnel to
the key performance objectives and helps to drive contin-
uous improvement in the pursuit for perfection.

B O T T O M - L I N E A U T O M A T I O N 131
BottomLine05.book Page 132 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

The Lean Production approach and concept have most


commonly been associated with discrete manufacturing
operations. Effective implementation of Lean principles
can help to boost the performance of any manufacturing
operation. Applying Lean Production techniques in pro-
cess plants has been a daunting challenge, but with the
advent of Dynamic Performance Measures the path has
been cleared. The combination of DPMs and Lean Produc-
tion provides a highly focused, performance-based pro-
cess manufacturing environment in which every
employee is enabled to perform to a common mission.
Manufacturing may be a mature discipline but the combi-
nation of new technologies and management approaches
can help to drive new and higher levels of performance
even in the most traditional manufacturing operations.

NOTES
1. Eccles, Robert G., The Performance Measurement Mani-
festo, Harvard Business Review on Measuring Corporate Perfor-
mance. Boston: Harvard Business Press, 1998, page 32.

2. Badavas, Dr. Paul C., and Dr. Albert D. Epperly, Statistical


Process Control Integrated with Distributed Control Sys-
tems, 1988 NPRA Computer Conference, October 1988.
3. Eccles, The Performance Measurement Manifesto, page 31.
4. Meyers, Christopher, How the Right Measures Help Teams
Excel, Harvard Business Review on Measuring Corporate Perfor-
mance. Boston: Harvard Business Press, 1998, page 100.
5. Kaplan, Robert S., and David P. Norton, The Balanced Score-
card Measures that Drive Performance, Harvard Business
Review on Measuring Corporate Performance. Boston: Harvard
Business Press, 1998, page 136.

132 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 133 Friday, August 19, 2005 9:54 AM

C H A P T E R 5 : T H E C O N V E R G E N C E

6. Drucker, Peter F., The Information Executives Truly Need,


Harvard Business Review on Measuring Corporate Performance.
Boston: Harvard Business Press, 1998, page 4.
7. Drucker, Peter F., Innovation and Entrepreneurship: Practices and
Principles. New York: Harper & Row, 1985, page 111.

B O T T O M - L I N E A U T O M A T I O N 133
BottomLine05.book Page 135 Friday, August 19, 2005 9:54 AM

CHAPTER 6

The
Economic
Optimization
of Industrial
Assets

INTRODUCTION
Chapter 5 presented the concept of a new management para-
digm for bottom-line improvement that emerged through the
convergence of three historically separate performance improve-
ment approaches. This new paradigm is based on empowering all
personnel by adopting strategically aligned performance mea-
sures that apply from the plant floor up through the business
management levels. The first step in moving to this new manage-
ment paradigm is to install and operate DPMs in the automation
systems of the industrial operations that seek to become perfor-
mance-based operations.

The original intent of many of these installations was to deter-


mine both whether the DPMs could be developed and imple-
mented in a wide range of industrial operations and whether they
provided the same level of value in different types of manufactur-
ing processes. One positive outcome of this activity, aside from

B O T T O M - L I N E A U T O M A T I O N 135
BottomLine05.book Page 136 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

validating the DPM approach, was that historical DPM informa-


tion from many plants in a variety of industries became available
that could be analyzed using the standard information analysis
tools provided with process historians. Since DPM information
was typically collected in process historian databases along with
other plant operational data, over an extended period large quan-
tities of performance and operational data were available for anal-
ysis and comparison.

The availability of this information presented a unique oppor-


tunity since this was the first time in industrial history that such
information was available in this format and quantity. This avail-
ability made it possible to search for relationships that had never
before been detectable. Using historical database-viewing tools,
plant personnel could review the data and for the first time iden-
tify and even quantify relationships between the business and the
operational performance of a plant as well as the business value of
interactions between operational measures. This also allowed the
testing of new operational activities to determine whether value
was actually created or destroyed by implementing activities in
the plant that were designed to drive value, such as operator
training, advanced controls, and Six Sigma improvement projects.
The results of analyzing these historical databases, though still in
its infancy, have provided valuable information that is supporting
new approaches to generating higher business value from plant
assets.

To outline what the analysis of several plants historical DPMs


and operational data uncovered requires us to briefly discuss both
the functional roles and the performance measurements that are
applied in manufacturing operations, plant assets, and traditional
approaches to optimizing plant assets. Next, we will present a
new approach to optimizing business value from plant assets that
is based on relationships that are discovered when we analyze
these historical databases.

136 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 137 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

PRIMARY PLANT FLOOR


FUNCTIONS IN INDUSTRIAL
OPERATIONS
Operations, maintenance, and engineering are the three pri-
mary functions that impact the day-to-day and minute-to-minute
operations of industrial plants. These disciplines have changed lit-
tle over the years in terms of the basic activities that personnel in
these functions perform. However, as automation and informa-
tion technologies have evolved, one of the largest impacts to
these groups has been head-count reductions, which gave the
remaining personnel more responsibilities. With this wider scope
of responsibility came a skills shift. Operators in particular
became much more dependent on the automation technology
and in some cases lost the ability to operate the plant manually.

Engineering, which supports operations, maintenance, and


other plant functions, has probably been the group most
impacted by the evolution of automation technology. Engineers
apply intellectual property and technology to drive efficiency
improvements throughout the plant. This group is not linked as
closely to the minute-to-minute operations of the plant as are
operations and maintenance, but the technology that these engi-
neers deploy has a huge impact on both the operations and main-
tenance staffs.

Fundamentally, operators in industrial plants are responsible


for running to produce products. Their primary objective has
been to best utilize the plants equipment and other assets, such
as raw material and energy, so as to maximize plant production.
This is a bit of an oversimplification in that not every plant is
driven to maximize production all the time. However, making as
much product as possible over any given period has been the gen-
eral objective for manufacturing operations throughout the pro-
duction-driven era. Before automation and instrumentation
technologies evolved, each plant required many operators, each
of whom was responsible for a very small section of the opera-
tion. Each operator would keep an eye on one or more gauges and
manually adjust valves accordingly. With the advance of technol-
ogies, especially in the area of automatic process control, a single

B O T T O M - L I N E A U T O M A T I O N 137
BottomLine05.book Page 138 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

operator could manage a much larger number of control loops,


first through control panels and later through CRTs. As a result,
this group was impacted by headcount reductions.

The primary objective of maintenance personnel has tradition-


ally been to maximize the availability of process equipment so
operators can use it to make products. The science behind plant
maintenance has evolved a bit more slowly than operations, but
major advancements have been realized in the past decade. The
initial impact on maintenance of the introduction of automation
technology into manufacturing operations was to give them
another set of high-technology equipment to keep operational.
Today, maintenance technology provides maintenance personnel
with advisories, equipment condition analyses, and predictive
analyses. It also automatically performs functions, such as work-
order management, designed to simplify maintenance activities.

Since engineerings role is to support operations and mainte-


nance, most of the business managers responsible for industrial
operations that the author interviewed stated that improvements
realized in operations and maintenance should be the measure by
which engineering is evaluated. Engineers establish the potential
while operators and maintenance are tasked with reaching the
potential. When the author then asked business managers to
identify and define the most appropriate performance measures
for operations and maintenance, they concluded that the opera-
tors should be primarily measured on the asset utilization of the
plant assets under their control and the maintenance personnel
by the asset availability of the resources in their domain of respon-
sibility. Since engineers apply technology that helps operators
improve asset utilization and maintenance personnel improve
asset availability, engineers should be measured on both mea-
sures.

The majority of managers interviewed indicated that engineers


should be measured on the incremental business value that their
improvement activities generate. Until DPMs were developed, the
business value such activities generated was not easily discernible.
This may help to explain the drastic job losses suffered by the
engineering departments of industrial operations over the past

138 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 139 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

decade. When the author queried plant personnel about these


measures, they generally agreed that asset availability and asset
utilization are the correct performance measures for maintenance
and operations, respectively. However, there was neither agree-
ment on how to define nor how to calculate these measures. In
other words, there seems to be general agreement in concept on
defining plant-level operational performance measures, but no
consistent way to measure them. As a result, asset utilization and
asset availability were seldom truly measured. While operators
and maintenance personnel work to keep the plant assets run-
ning and producing products, they have very few measures avail-
able to help them improve their performance. To a large degree,
therefore, the performance of operations, maintenance, and engi-
neering remains, for the most part, effectively unmeasured.

PERFORMANCE MEASURES FOR


PRIMARY PLANT FLOOR
FUNCTIONS
Asset utilization and asset availabil-
ity are not measured directly in most
manufacturing plants. However, since
the histories of plant operational data
in several industrial operations are
available, the general relationship
between these two measures can be dis-
cerned by using statistical analysis
tools, such as factor analysis and corre-
lation, on other operational measures
stored in these historical databases.
These tools reveal that asset availability
and asset utilization tended to be
inverse functions as they approach their maximum ranges (see Figure 6.1 Asset Availability
and Asset Utilization Are
figure 6.1). This is interesting, because in most plants operations Inverse Functions as They
and maintenance are essentially managed as independent func- Approach Their Maximum
Ranges
tions, and asset availability and asset utilization are treated as
independent variables. The inverse nature of their relationship

B O T T O M - L I N E A U T O M A T I O N 139
BottomLine05.book Page 140 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

implies that an increase in one will likely drive a decrease in the


other. In other words, the negative impact of one on the other
increases as operations and maintenance perform their jobs more
effectively.

The inverse nature of the relationship between asset utilization


and asset availability is really very much in line with common
sense. It can be practically demonstrated by considering the avail-
ability and utilization of an automobile. An automobile is a com-
plex mechanical asset set that is analogous to an industrial plant.
If you wanted to maximize the availability of your car, you would
leave it in the garage. The asset availability of this automobile
would then approach 100 percent, but the asset utilization would
be zero. Conversely, to maximize utilization of the car you would
drive it continuously as fast as possible. The result would be to
approach maximum asset utilization, but asset availability would
certainly be negatively impacted. Although these are extreme
cases, it does not take much extrapolation to see that, as the car
approaches maxi-
mum utilization,
the availability will
decrease over time.
The same general
relationship exists
for all complex
mechanical asset
sets, although the
exact ratio of the
inverse relation-
ship will be unique
to the specific
assets.
In industrial
plants, this inverse
Figure 6.2 An Inverse relationship may occasionally put operators and maintenance
Relationship Puts
Maintenance Personnel and personnel at odds (figure 6.2). The better the operations and
Operators at Odds maintenance teams perform, the more likely it is they will experi-
ence conflict. The evolution of sophisticated control and mainte-
nance technologies to improve the efficiency and effectiveness of

140 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 141 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

maintenance and operations has increased the probability that


both asset availability and asset utilization will achieve their max-
imums.

The probability is higher in other words that a schism between


operations and maintenance may develop. These conflicting mea-
sures of operational and maintenance performance mean that, in
the quest for optimal plant performance, it is more essential than
ever to analyze plant-level performance measurement systems
and perhaps even to reevaluate plant-floor workflows.

PLANT ASSET ANALYSIS


Since both maintenance and operations directly interact with
all aspects of the plant asset base, it may be helpful to briefly dis-
cuss what the asset base is comprised of. The authors discussions
with teams of managers, engineers, operators, and maintenance
personnel in more than 500 industrial operations made it very
clear that the semantics associated with plant assets and their
management is unclearthe terminology used is often poorly
defined. This lack of semantic clarity can confuse issues and hide
important relationships. Even the phrase asset management is so
broad that different people use it that it can mean very different
things. At an organizations plant-floor level, for example, asset
management is often analogous to equipment maintenance. But
business managers use the exact same phrase to mean the deploy-
ment of any of a broad range of available assets in order to meet
critical business objectives. Fleet managers use the phrase asset
management to refer to the management of fleets of vehicles.
Clearly, when analyzing plant performance, it is very helpful to
clarify the terms being used. To clarify the discussion throughout
this chapter, therefore, we will define commonly used terms
according to the meaning that has proved useful, not necessarily
the meaning that has been accepted across industry. Where no
common agreement on the specific definition of a word could be
found, a useful definition was developed and used consistently
throughout. As much as possible, the definitions developed were
true to the most common usage of the term, but often a degree of

B O T T O M - L I N E A U T O M A T I O N 141
BottomLine05.book Page 142 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

added precision was required to make them useful in quantitative


analysis.

What actually can be included as an asset in an industrial


operation? Some managers limited the definition of asset to the
physical equipment in the plant, while others had a much
broader view. When the author presented the narrow definition
to managers, many argued that it was too limiting. In contrast,
when a broader definition was presented, even people who ini-
tially defined asset more narrowly seemed to understand why a
broad definition was useful. Therefore, a broad definition of
industrial assets was adopted. The resulting definition encom-
passed all resources, either tangible or intangible, used in the pro-
duction of the products in an industrial plant, including:

plant equipment
energy
raw materials
products
humans
automation systems (distributed control systems, pro-
grammable control systems, safety systems, human
machine interface software, etc.)
advanced automation software (advanced process
control, multivariable predictive control, production
planning, production scheduling, batch management,
etc.)
information systems and advance information tech-
nology software (enterprise resource planning, supply
chain management, customer relationship manage-
ment, enterprise asset management, etc.)

The industrial operations assets can be partitioned into two


distinct groups. We will refer to the first grouping as the plants
primary assets, which are the basic assets required to produce the
plants products. These include the raw materials, products, phys-
ical equipment (vessels, piping, etc.), as well as energy. The physi-
cal equipment in process manufacturing environments includes
the instruments, valves, and pumps, but not the automation
equipment. This distinction between primary and physical assets

142 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 143 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

was made because the instruments, valves, and pumps are funda-
mentally essential to produce the products in a process environ-
ment while, in contrast, most processes can operate manually
without an automation system.

The second class of plant


assets is referred to as support-
ing assets. These assets are
deployed to improve the effi-
ciency and effectiveness of the
primary assets (see figure 6.3).
The supporting assets include
automation systems, informa-
tion systems, and human
assets. Human assets can be
categorized as both primary
and supporting assets because
humans are absolutely essential
to the manufacturing of the products, but they also help improve Figure 6.3 Supporting
Assets Improve the Efficiency
the efficiency and effectiveness of the primary assets. Human and Effectiveness of the
assets can also be categorized as supporting assets because the Primary Assets
general impression given by the plant personnel the author inter-
viewed is that the human asset base had just begun to realize its
potential in helping to improve the efficiency and effectiveness of
industrial operations. As a result, human assets would best be cat-
egorized with the supporting assets.

Effective plant asset management must encompass all asset


sets. It is very difficult to develop effective models for the relation-
ship between the primary and supporting assets with respect to
overall plant performance. This is because the primary assets per-
formance depends both on their availability and utilization and
on the availability and utilization of the supporting assets that are
used to control and manage primary assets. Considerable effort
and technology have been invested to improve the availability of
supporting assets, using such techniques as redundancy and fault
tolerance. These approaches can be implemented in computer-
based systems in a fairly cost-effective manner. But the existence
of supporting assets is justified only by their ability to improve

B O T T O M - L I N E A U T O M A T I O N 143
BottomLine05.book Page 144 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

the performance of the primary assets. Until approaches like reli-


ability-centered maintenance were introduced, plants made mini-
mal efforts to apply redundancy at the primary asset level. One
key reason is that the cost of implementing these techniques at
the primary asset level can be considerably greater than imple-
menting the supporting assets. Each set of assets should be evalu-
ated with respect to its asset availability and asset utilization as
well as its impact on the plants performance. Clearly, primary
assets should run considerably better when supporting assets are
available and operating.

The utilization of supporting assets is based on the direct


impact these assets have on the primary assets in their domain of
influence. If the capabilities that exist in the supporting assets
(automation, IT, humans) are not put to use, the potential utiliza-
tion impact on the plant can be limited. In fact, many supporting
assets are significantly under utilized. In part, such underutiliza-
tion of capital supporting assets (automation and IT systems) can
be traced to the capital budgeting, procurement, and acquisition
processes practiced by most manufacturing companies. In the
typical capital procurement process (see chapter 2), the manufac-
turer defines the system requirements in a request for proposal
(RFP). Often, these RFPs are developed in such a way that the sys-
tem that is being replaced is specified function for function, and
the RFP is then issued to automation suppliers for the purpose of
making competitive bids. Typically, the inherent capability and
functionality of modern automation systems is significantly
greater than that of the system being replaced because automa-
tion suppliers continually improve their systems. Unfortunately,
most of this additional capability does not get into the RFP
because the system being replaced did not have the new func-
tions. Also, the project team building the RFP would not have
been aware of the additional capability when constructing the
document. Automation system suppliers are certainly aware that
their systems can offer considerably more benefit than has been
requested in the RFP but they also know that, if the cost to realize
this benefit increases the system price they may lose the order on
price. So, suppliers typically do not include these advanced capa-
bilities in their proposals, and the new automation systems

144 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 145 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

merely become direct functional replacements for the systems


being replaced. Replacing old technology with new technology
that performs the exact same functions seldom improves perfor-
mance significantly.

Exacerbating this situation is that most automation projects


are implemented by project teams convened by the manufactur-
ers for this purpose. Studies have shown that project teams pri-
mary measures of success are on-time and on-budget delivery of
their projects. Project teams are seldom, if ever, measured on how
well the resulting system impacts plant performance. Project
teams are therefore not motivated to incorporate any new func-
tionality that may impact either schedule or budget. As a result,
the new capability is left out.

When this situation is made clear to the company acquiring


the new systems, they often point out that they plan to take
advantage of this new capability at a later date, after the new sys-
tem is up and running. Unfortunately, in most cases, they never
seem to get around to implementing this new advanced function-
ality because, once the project team leaves and moves on to
another project, the plant engineers left behind to maintain the
plant are often too busy just trying to keep the plant running.
They may not have the skills even if they have the time. The net
result is that the impact of the supporting assets on the utilization
and availability of the primary assets is very often much less ben-
eficial than what it could or should be. Huge opportunities for
improvement are lost.

As we have said, the significance of the availability of support-


ing assets is in their utilization and availability impact on the pri-
mary assets. It may be important to evaluate the automation and
IT system components in terms of mean-time-between-failure
and mean-time-to-repair in order to obtain a head-to-head com-
parison of alternate supporting asset technologies. However, it is
also important to understand that these issues are only important
if they impact the availability and utilization of the primary
assets. Therefore, the focus when measuring any asset perfor-
mance optimization activity should be its impact on primary
assets in the plant.

B O T T O M - L I N E A U T O M A T I O N 145
BottomLine05.book Page 146 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

HUMAN ASSETS
Most discussions of supporting assets have focused on technol-
ogies and systems, but human assets can be discussed in similar
terms. During the Industrial Revolution huge numbers of unedu-
cated and unskilled laborers had to be harnessed to provide value
in manufacturing plants. This fostered the attitude, still prevalent
today, that the functions performed by process operators should
be minimized. Engineers invest huge amounts of time and money
to develop advanced software applications that perform the func-
tions for which operators historically might have been responsi-
ble. Although many of these software solutions are very
sophisticated and generate real improvements, engineers often
overlook the best uses of the human asset base. As a result, many
operators merely wait for an alarm and respond to it, spending
much of the rest of their time reading the newspaper. This has led
the operators to run the plants so they stay away from constraints
in order to reduce alarms. But to improve plants operating perfor-
mance, operators must shift from running away from these con-
straints to moving toward them.

This reluctance to use valuable human talent is a waste and


leads to lower motivation and morale among operators. Operators
are no longer the uneducated, unskilled workforce they once
were. In fact, they are often quite well educated and have devel-
oped a craftsmans sense for the plants operation that no one else
in the operation could possibly have. In forward-leaning manu-
facturing companies, operators are more empowered to operate
for performance improvements. Given the complexity and
dynamics of industrial operations, it is impossible to mathemati-
cally model every possible aspect of the operation. Humans are
the only plant resource that have the ability to analyze and
respond to unexpected events or highly complex situations that
may not have been previously encountered or understood. Indus-
trial operations staff must therefore view their human resources as
highly valuable assets that, if they are valued and utilized effec-
tively, have the potential to drive huge incremental improve-
ments.

146 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 147 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

TRADITIONAL APPROACHES TO
OPTIMIZING PLANT ASSETS
The science and technology of industrial plants have been
steadily evolving for decades. New technologies and methodolo-
gies that drive higher efficiencies in both asset availability and
asset utilization become available almost monthly. As a result,
today sophisticated mathematical approaches are available to
support both maintenance and operations in industrial plants.

The traditional separation of operations and maintenance has


led to the independent evolution of technologies in each area. On
the whole, the technologies that support improvements in the
efficiency of plant operations have evolved more rapidly than
those that improve maintenance efficiency. However, over the
past decade the maintenance technologies have begun to catch
up. Some of the technologies that have been targeted at opera-
tions efficiency are systems like programmable logic controllers,
distributed control systems, safety shutdown systems, and human
machine interface systems. Moreover, recent advanced opera-
tions-focused software has included process historians, advanced
process control, multivariable predictive control, batch manage-
ment, production planning, production scheduling, and supply
chain management. On the maintenance side, the recent technol-
ogies include software such as preventive maintenance, predictive
maintenance, reliability-centered design, condition monitoring,
plant asset management, and enterprise asset management. The
traditional separation of operations and maintenance has been
unfortunate for industry because many of the technologies
designed to make operations more efficient could also be used to
improve the maintenance efficiency. For example, effective pro-
cess control strategies can be implemented that actually reduce
equipment wear. The reverse relationship can also be true.

The traditional approach to deploying these maintenance and


operations improvement technologies has been dictated by
industrial plants hugely complex systems of circuits, piping, and
vessels. These plant systems are so complex that the control sys-
tem required to get an entire plant operational has itself become
very complex and difficult. Therefore, creating a plant-wide con-

B O T T O M - L I N E A U T O M A T I O N 147
BottomLine05.book Page 148 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

trol system is seldom, if ever, accomplished. This complexity,


coupled with the dynamic nature of industrial operations, has led
engineers to take very pragmatic approaches to both the manage-
ment and control of asset utilization and asset availability.

People traditionally attack very complex problems by parti-


tioning, or decomposing, the problem into workable compo-
nents, that is, solving each component and then working to
recompose all these individual solutions to cover the overall prob-
lem. This has certainly been the approach used with industrial
operations. The overall plant is logically divided into plant areas,
and then into process units, right down to the base components,
so engineers can
arrive at compo-
nents that are suf-
ficiently bounded
to be addressed by
the available sci-
ence (see figure
6.4). In the man-
agement of the
asset utilization of
plant operations,
this base decompo-
sition unit has
Figure 6.4 Overall Plant been the process loop level. Once a process control scheme is
Breakdown
developed for a loop, engineers use higher-level control science to
combine loops into more complex multi-loop strategies. By
investing considerable time and effort, these schemes can grow to
multivariable predictive control schemes and even to unit-level
optimization schemes. Unfortunately, this approach to recompos-
ing base control schemes has proved to require so much effort
that well over 90 percent of process control implementations in
industry today are probably still at the single-loop level. Very few
are composed even as far as the unit level. As the control schemes
grow more complex, it becomes increasingly difficult to compen-
sate for changing plant dynamics. This often causes plants opera-
tors to shut down higher-level control schemes because they lack
confidence in their operation.

148 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 149 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

This same point also applies to machine-oriented plant opera-


tions, such as packaging lines. Typically, programmable logic con-
trollers are set up to control a machine or line using a program
that sequences the machines operations. Although programma-
ble logic controllers can be networked together for coordinated
operations, for the most part they operate as separate work cells.

Controlling a plants asset availability through the mainte-


nance function has followed a different but analogous decompo-
sition. Some plants have maintenance specialists for different
equipment types who are focused on sets of equipment within
their area of specialty. To a large degree, the maintenance per-
formed in many industrial operations still follows a break-fix
approach. That is, when a piece of equipment fails, a mainte-
nance crew is dispatched to repair it. This can be the least efficient
approach because equipment failure can actually result in an
increased level of damage as well as longer and more costly main-
tenance procedures. To avoid such failures, many plants are peri-
odically shut down so preventive maintenance can be performed
on the equipment. Statistical failure rates and run times can be
used to determine prospective trouble areas so they can be dealt
with before a failure occurs. Predictive maintenance models have
also recently been implemented in which measurements of equip-
ment characteristics, such as vibration levels and temperature, are
compared to normal operating models for the purpose of directly
detecting and predicting an imminent failure. Maintenance per-
sonnel can then be alerted in time for them to react and resolve
the issues before the failure occurs. These approaches are realizing
favorable results, but the overall field of asset availability
improvement is still developing.

The science behind the improvement of asset availability in


industrial plants has evolved independently from the science
behind asset utilization improvements. This is most likely because
of the traditional functional separation between operation and
maintenance. In reality, however, the asset utilization and asset
availability of any industrial asset set are highly interdependent.
They are directly related to the business value that the asset set
generates. It is clear that the improvement approaches historically

B O T T O M - L I N E A U T O M A T I O N 149
BottomLine05.book Page 150 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

employed at industrial plants must change to address under-


exploited opportunities to create economic value in industrial
operations.

FUNDAMENTALS OF
MATHEMATICAL OPTIMIZATION
Linear and nonlinear optimization represent perhaps the most
sophisticated technological approaches used to drive improve-
ments into plant assets. Before considering the relative merits of
traditional optimization techniques in industrial operations, an
understanding of the mathematics behind both approaches is
essential. A brief overview may therefore be helpful to put plant
optimization activities into perspective.

Mathematical optimization, whether linear or nonlinear,


involves simultaneously resolving a set of functions so as to arrive
at the solution that best either maximizes or minimizes the value
of a single, important variable. The variable could be financial,
though it does not necessarily have to be. The mathematical func-
tion that defines the behavior of the important variable is called
the objective function since it defines the objective of the opti-
mization. The other mathematical functions involved in the opti-
mization are referred to as constraint functions because they
define the constraints on the objective function. For example, if
the optimization activitys objective is to maximize the produc-
tion value derived from a certain set of industrial assets, but the
temperature of those assets should not exceed 500 degrees, then
the objective function would define the production value that the
asset set produces. One of the constraint functions would be rep-
resented by the temperature function. The constraint functions
are relevant for any variable that may limit the value of the objec-
tive function.

To resolve a mathematical optimization problem, the objective


function and all appropriate constraint functions must first be
well defined in mathematical terms. Each constraint function is
typically represented by an inequality. For example, the tempera-
ture must be less than 500 degrees. The constraint functions

150 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 151 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

taken together define a region of feasible solutions. Every point in


the region of feasibility represents a solution to the problem that
does not violate the constraints. The objective function is ana-
lyzed to determine the range of feasible solutions as the objective
is either maximized or minimized, depending on the desired
state. The analysis basically solves the objective and constraint
functions simultaneously so as to determine the optimal point. If
all of the objective and constraint functions are linear, then the
operation is a linear optimization. If any of the functions is non-
linear, then the operation is a nonlinear optimization. Linear
optimization is easier to resolve than nonlinear because it deploys
proven methodologies, such as the simplex method. Nonlinear
optimization requires more sophisticated techniques such as gra-
dient vector analysis. It is difficult to determine whether the
result of the exercise is truly the global optimum across the entire
feasible solution space or merely a local optimum.

Although both linear and nonlinear optimization involve


fairly sophisticated mathematics and have yielded very good
results, they also have three very significant shortcomings. First, a
significant amount of expertise and effort is required to develop
them properly. The objective function and each of the constraint
functions must be very carefully developed and tested to ensure
that they truly represent the manufacturing process. Second, the
functions used are static, which is fine if the system being opti-
mized is static over time. But if the system has dynamic tenden-
cies, then the optimizer can get out of tune in short order. Finally,
there can only be one objective function. All of the other func-
tions used in the optimizer must be constraints. This is also all
right if the problem is truly a single objective problem. But many
problems involve multiple objectives. Unfortunately, in these
cases, the practitioner is required to select a single objective and
handle all the other objectives as if they were constraints. An
optimizer that has a single objective function can certainly be
used to improve multiple objective situations in which no optimi-
zation had previously been attempted. However, such an opti-
mizer can also be quite limiting. Single objective optimization
can be particularly difficult if the external conditions associated
with the optimization are variable because this requires different

B O T T O M - L I N E A U T O M A T I O N 151
BottomLine05.book Page 152 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

objective functions for different conditions. The objective func-


tion of an optimization can be changed, with the previous objec-
tive function reverting to a constraint function and one of the
constraint functions becoming the objective. However, the
amount of incremental work required to set up and manage such
an optimizer is significantly increased. The net result is that single
objective optimization approaches are often inadequate for
dynamic and highly complex problems, like those found in
industrial operations.

TRADITIONAL APPROACHES TO
INDUSTRIAL OPTIMIZATION
As manufacturers have strived to generate more value from
industrial assets, they have deployed both linear and nonlinear
optimization techniques. The vast majority of the optimization
projects have been in the more scientifically oriented industry
segments, such as refining and petrochemicals. Most of these
efforts have been directed at what are considered to be the most
important assets in the operation from either an operational or
business perspective. In spite of the huge expenditures that have
been made in advanced automation systems, the vast majority of
industrial assets are still only at the base level of control. Most
optimization activities that have been deployed have been at the
process unit level or below, and are focused on operations as com-
pared to maintenance. The objective functions of these solutions
have often been stated in financial terms, but seldom tied into
the plant accounting systems. Therefore, the results are seldom
validated by the accounting team. Most plant accounting systems
do not provide information down to the process unit level, and it
is very difficult for accountants to rationalize with just account-
ing information what the engineers believe the benefit of an opti-
mizer to be. This can lead to a further schism between
engineering and accounting personnel (see chapter 4).

The authors discussions with a very advanced team of optimi-


zation specialists working for a major refining and petrochemical
company revealed that they had completed or launched an

152 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 153 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

unusually high number of optimization projects. This teams


degree of knowledge and capability was extremely impressive,
and as they discussed their activities it became clear that they
were justifiably very proud of their accomplishments. The one
surprising aspect of this team was that they were comprised of
only about a dozen professionals, and it was clear that they were
extremely busy. When asked why there were so few on the team,
the team leader responded that they had had more the year
before, but corporate downsizing had cut the team in half. When
asked how much economic benefit they generated for the com-
pany, they responded with a number that amounted to almost $4
million per year per person. With results like this, there is no way
the team should have been downsized. When asked how much of
this benefit the companys chief financial officer could see, the
teams leader said, That is the problem; he does not understand
the value we generate. He is right. That is exactly the problem.

Frequently, engineering and operations teams attempt to


prove the value of an optimization activity by first establishing a
baseline of performance. They do this by conducting a pre-test of
the operating conditions while the optimizer is not running.
They then turn on the optimizer and conduct the same test to
measure the incremental improvements. Although this is a valid
approach for determining the impact of the optimizer on the
operational measures being affected, it is still very difficult to rec-
oncile the operational measures of improvement with the plants
accounting measures. If these measures can be reconciled, it is
only for a short period of time and for very specific plant operat-
ing conditions. Since most ERP systems calculate accounting mea-
sures on a monthly basis, and since a very large number of
performance-impacting events take place within any month of
operation, it is impossible to reconcile any operational perfor-
mance improvements with accounting performance. It is not
unusual to find industrial operations that have two somewhat
independent sets of measures both of which are attempting to
measure the performance of the same operation. The operations
and engineering teams often work KPIs, while the accounting
teams work to financial performance measures.

B O T T O M - L I N E A U T O M A T I O N 153
BottomLine05.book Page 154 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

The result of the optimization efforts that have been deployed


in industrial operations has been mixed, at best. The amount of
time, effort, and expertise that are required to effectively deploy
an optimizer in an industrial operation can be very significant.
The mathematical modeling of the objective and constraint func-
tions has to be done very accurately or the optimizer will not pro-
vide the desired results. Developing these functions accurately is
very challenging. The net result is that once an optimizer is devel-
oped, it is seldom modified. Industrial plants are very dynamic,
and many of their operating characteristics tend to change over
time. For example, scaling on pipes or wear in rotating equipment
can modify the way the plant runs. These changes may be slight
initially, but over time they can become significant. If optimizer
models are not adjusted to compensate for these changes, they
eventually get out of tune with the plant and will actually pro-
duce suboptimal performance. As this happens, operators lose
confidence in the optimizers and tend to turn them off and run
the plant without them. When this happens, the effort and exper-
tise employed to generate these advanced solutions are essentially
wasted.

Another issue impacting the effectiveness of optimizers is that


market conditions and business drivers can change on a fairly fre-
quent basis (see chapter 1). Globalization has caused market
dynamics to become more complex and changes in market condi-
tions are much more frequent. Business managers in industrial
segments such as power generation revealed to the author that a
decade ago they rarely had to alter the operating strategy of the
generating plants. Today, however, it is not unusual for them to
change the strategy multiple times in a single day. Typically, a sig-
nificant change in market strategy necessitates that the optimiza-
tion strategy adopt a new objective function. The previous
objective function might now actually become a constraint func-
tion, and one of the constraint functions may become the objec-
tive. This is not an easy modification to make in most
optimization strategies. Because of the amount of effort this adap-
tation of the optimizer causes, it typically does not happen. But if
the optimizer stays in operation, it will actually be driving an
undesired result.

154 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 155 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

To summarize, sophisticated optimization approaches require


considerable expertise, are very time and cost intensive to build,
and are very difficult to maintain and adjust. Once the optimizers
have been developed, they are seldom adjusted, and as a result
they optimize the utilization of the plants asset sets at the
expense of flexibility. With todays dynamic market conditions,
flexibility in operations is essential. Single-objective optimization
strategies are very limiting since the operating objective of most
industrial asset sets cannot be defined by only a single objective.
Most operations have multiple, prioritized objectives, not just a
single objective. The dynamics of all of the objectives may mean
changes are made frequently. Focusing on only one objective
clearly can limit the inherent value of the optimization.

Unfortunately, mathematics has not come up with a practical


approach to multiple objective optimizations. Some operations
have tried to employ objective switching approaches in which an
expert system is used to identify changes in strategy and then
switched to an optimization model that complies with the new
strategy. This is a very costly approach that, while certainly better
than optimizing to the wrong primary objective, still offers many
of the limitations we have already described.

Another critical issue associated with optimization is that


when process units are processed independently in industrial
operations, the procedure poses a performance risk to the overall
operation. Process units operate integrally with the overall plant.
Thus, locally optimizing a unit to a single objective may result in
the suboptimization of the plants overall performance. As opti-
mizers are implemented to maximize asset utilization and sophis-
ticated techniques, such as predictive maintenance, are
implemented to maximize asset availability, maintenance and
operations will tend to become more competitive. The inverse
nature of the relationship between asset utilization and asset
availability is more extreme as each approaches its maximum.

Finally, most industrial plants typically have two sets of dis-


connected performance measurement systems. The accounting
group develops and maintains a set of financial reporting perfor-
mance measurements. The operational and engineering teams

B O T T O M - L I N E A U T O M A T I O N 155
BottomLine05.book Page 156 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

develop and
maintain a set of
measures, often
called KPIs,
which are sel-
dom aligned
with the
accounting mea-
sures (see figure
6.5). Since the
implemented
optimizers are
developed by
engineering and
Figure 6.5 Operational operations, they work to meet objective functions that represent
Measures are Seldom
Aligned with Accounting
the KPIs, not accounting measures. The KPIs may or may not be
Measures valuable in identifying opportunities for improving business
value.

Some KPIs are actually presented so as to include a financial


perspective, but this typically is not an accounting-based finan-
cial perspective. This fact often puts the plants operational and
engineering personnel at odds with the plants accounting teams
that use traditional accounting techniques to measure the plants
performance. To establish KPIs (e.g., asset availability and asset
utilization) that are aligned with the accounting system, it is
essential to achieve the desired convergence and cooperation
among the organizations functions. In theory, an organization
that is aligned through its measures will help generate higher lev-
els of performance from the plants assets.

A new approach to the economic business optimization of


industrial assets is required. Traditional bottom-up, mathematical
model-based approaches are inadequate to meet todays dynamic
and challenging business requirements. Traditional approaches
must sometimes be abandoned in order to make real progress.

156 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 157 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

OPTIMIZING PLANT ASSETS FOR


BUSINESS PERFORMANCE
As we noted at the beginning of this chapter, the practice of
collecting financial business data within the same historical data-
base as operational data has enabled analysts to identify relation-
ships that have previously either been hidden or not quantifiable.
When analysts studied this information, they followed the tradi-
tional bottom-up perspective of plant operations, which led them
to identify relationships involving base-level primary plant assets
such as instruments, valves, and pumps. It was theorized that
relationships at this level would be most readily discerned and
that identifying these relationships might lead to more general
relationships across more complex asset sets and, perhaps, even to
entire industrial plants.

Although asset utilization and asset availability were not


directly measured in any of the operations that participated in the
authors analysis, enough operating data was collected that was
related to availability and utilization to provide some degree of
insight into their behavior. The terms asset availability and asset
utilization are not consistently well defined across industry, and
most useful definitions involve gross plant behavior over
extended periods of time. To make these measures more useful,
definitions had to be developed that were specific enough to lead
to a consistent measurement approach. Moreover, these defini-
tions had to be developed in such a way that they made possible
measurements over very short operating durations so operators
and maintenance personnel could have real-time feedback on the
impact that their minute-by-minute activities had on the relevant
measures. To accomplish this, the author conducted interviews
with business managers to gain their perspective on the informa-
tion that such measures should convey so as to be most useful to
their businesses. Business managers agreed that a measure like
asset availability should be the primary operational measure of
maintenance performance and that one like asset utilization
should be the primary operational measure of operator perfor-
mance. With this in mind, the managers communicated that,
from their perspective, the measure of asset availability should
indicate the currently maintained state of the equipment with

B O T T O M - L I N E A U T O M A T I O N 157
BottomLine05.book Page 158 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

respect to that equipments effectiveness to the manufacturing


process.

Historically, asset availability has often been determined based


on whether a piece of equipment or asset set was either opera-
tional or not operational. The immediate measure would either be
0 or 1, with the totalized asset availability being the percentage of
time that the asset or asset set is available (1) to the operation
over some extended time period. But the interviewed business
managers indicated that this was too simplistic a way to look at
the maintenance function. For example, a piece of equipment
could be operational, but due to lack of maintenance it might
only be able to work at 80 percent efficiency. In this case, the
asset availability of that piece of equipment should not be 0 or 1
but rather 0.8. This approach to the availability metric would pro-
vide a more accurate assessment of the performance of mainte-
nance and it would provide a definition of availability that could
be measured instantaneously in real time rather than over an
extended period. This approach led to the development of the fol-
lowing definition of effective asset availability (EA):

Maximum Usable Output of an Asset or Set of Assets in Current Maintained State


EA =
Theoretical Maximum Output

By this definition, EA ranges from 0.0 to 1.0. This measure of


effective asset availability combines the concepts of traditional
availability and effectiveness into a single, more useful measure.

When these same business managers were asked how to mea-


sure the performance of their plant operators, they responded
that the operators job is to utilize the plant assets under their
control so as to produce as much product as the assets will allow
in their currently maintained state. This, of course, implies that
the current manufacturing strategy is a production-driven strat-
egy, which historically has been the most common strategy
deployed in industrial operations. But then again the whole con-
cept of asset utilization implies a production-driven approach. As
a result of these discussions, we arrived at the following definition
for effective asset utilization (EU):

158 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 159 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

Current Output of an Asset or Set of Assets


EU = Maximum Usable Output of an Asset or Set of Assets in Current Maintained State

By this definition, EU also ranges from 0.0 to 1.0. Although


this definition is admittedly more specific than most traditional
definitions of asset availability and utilization, it is clear that
effective asset utilization and effective asset availability are highly
interdependent. The business managers interviewed agreed that
the two functions are highly interdependent and that the mea-
sures should therefore reflect this interdependence.

Using these more specific definitions of effective asset avail-


ability and effective asset utilization enabled the author to ana-
lyze the operating data collected in the plant historians to obtain
strong indicators of the behavior of EA and EU and to statistically
correlate that behavior to the financial data in the database. DPM
data was only established for the process unit level and up, and
the data did not typically decompose down to the base asset level.
Analyzing these data sets made it possible to determine that the
relationship between EA and EU is multiplicative relative to busi-
ness value. The multiplicative nature of this relationship passes
the common-sense test, and the implications are very significant.
Since effective availability and utilization both range from 0.0 to
1.0, they are almost always fractional values.
Multiplying fractions together results in a
very steep decline as the factors decrease,
which means that the impact that EA and
EU have on each other is extreme. The equa-
tion shown in figure 6.6 was developed to
describe the instantaneous relationship of
business value from a single asset (instru-
ment, valve, pump, etc.) with respect to EU
and EA. Figure 6.6 Describes the
Instantaneous Relationship
To expand the equation so it shows asset value over the longer of Business Value from a
Single Asset with Respect to
periods normally required for business analysis, the equation has EU and EA
to be totalized over the time period of interest. The resulting
equation is shown in figure 6.7.

B O T T O M - L I N E A U T O M A T I O N 159
BottomLine05.book Page 160 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

The EA and EU are discernible and fairly


measurable at this level, which represents one
positive aspect of working at the single asset
level in the plant asset base. Increases to more
complex asset sets make the measurement of
EA and EU much more difficult to evaluate.
The Cmax term can also be evaluated for base
assets. The connotation of AVmax at the base
level is very difficult, and perhaps even impos-
Figure 6.7 Expanded sible, to reconcile with accounting information. The problem is
Equation to Show Asset
Value Over Longer Periods
that the concept of the business value of an instrument, pump, or
valve with respect to a plants overall operation does not have
much meaning. Business value essentially does not effectively
decompose to the base asset level. Extrapolations of business
value down to the single asset level are required to resolve these
equations. Unfortunately, no valid method for accomplishing this
level of business value decomposition has been discovered. There-
fore, it is impossible to resolve either equation 6.6 or 6.7 in a true
business context. Because of this, the only viable approach to
dealing with base assets is the traditional approach of trying to
maximize EA and EU somewhat independently.

ASSET SET OPTIMIZATION


Asset sets within industrial plants are defined as being logical
collections of primary assets that combine to perform a logical
production function or related set of production functions. An
asset set could be comprised of at least two primary assets and
could include a process subunit, unit, train, or even an entire pro-
duction area of a plant. Moving from single assets to asset sets
generates a huge leap in the level of complexity of the measure-
ment of EA and EU. Upon initial inspection, it may appear as
though determining the EA and EU of an asset set would involve
combining the EAs and EUs of the primary assets that comprise
the asset sets in a fairly straightforward fashion. The fact is that
most industrial operations are extremely complex networks and
circuits that combine in very intricate ways to produce the prod-

160 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 161 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

ucts. This complexity means that advanced applications of sys-


tems theory, network theory, and circuit theory are required to
effectively and directly calculate the overall EA and EU of the
asset sets.

From an operations perspective, a number of performance


improvement technologies and strategies have been deployed to
drive asset and asset set utilization improvements. These technol-
ogy strategies include cascade control, feedforward control,
advanced regulatory control, multivariable predictive control,
batch management, and linear and nonlinear optimization. These
technology strategies involve increasingly complex levels of plant
assets, from simple loops all the way up to process units.
Although the science of control and optimization has expanded
significantly over the past few decades, most asset utilization
improvement strategies implemented today are still at the single-
loop or single-machine levels, and very few go all the way to the
process unit level. Higher-level control strategies and science are
still typically static. Moreover, as a plants dynamics change over
time, the implemented higher-level control strategies match the
plant operations less and less and are thus quite often turned off
by the operators.

From a maintenance perspective, the growth in technological


approaches and strategies has evolved along a slightly different
path than that of operations. Instead of growing in complexity
incrementally along production units, the science of improving
asset availability has developed much like maintenance strategies,
which are focused on improving both the uptime of specific
equipment types and the reliability in plant design. As discussed
earlier, many industrial plants maintenance strategies in the past
were primarily based on a break-fix approach. Equipment would
be periodically maintained and run until a breakdown occurred.
The maintenance team would fix the problem as quickly as possi-
ble so as to get the equipment back online. More sophisticated
preventive maintenance strategies evolved that were based on the
periodic maintenance procedures that maintenance personnel
conduct on plant equipment using historical maintenance
records. If the plant behaves as it always has, the preventive main-

B O T T O M - L I N E A U T O M A T I O N 161
BottomLine05.book Page 162 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

tenance programs can be fairly effective. Predictive maintenance


programs involve monitoring the current condition of the various
equipment types in the operation so as to determine whether the
current conditions measures make it possible to predict both the
timing of a breakdown and the problem that will cause it. This
information can be used to address the issue before the event hap-
pens. Often the amount of time and effort needed to avoid an
incident is considerably less than the time and effort needed to
fix a piece of equipment once it breaks down. Predictive mainte-
nance strategies thus typically provide a significant increase in
the availability of plant assets. So-called Reliability-Centered
Maintenance approaches involve analyzing the plants equip-
ment to determine both which equipment are highly critical to
the plants operation and how backup equipment can actually be
built into the process to minimize the probability of process avail-
ability loss even when a piece of equipment fails.

REQUIREMENT FOR PERFORMANCE


VECTORS FOR ASSET SETS
As the sophistication of both operation and maintenance strat-
egies increases, often involving larger sets of assets, it would be
useful to have more encompassing performance measurements to
make comparisons possible. In contrast to base assets, when assets
are combined into logical asset sets the business value starts to
have real definable meaning. For example, it is much more feasi-
ble from an accounting perspective to discuss the value contribu-
tion and associated costs for a cooking oven than it is for a valve.
This implies that plant accounting/business value is measurable
for many asset sets. Accounting departments measure the busi-
ness value of an asset set by using a number of different measures,
such as contribution margin, material cost, storage cost, energy
cost, human resource cost, and equipment cost. This is why
accounting departments report results as multiple values on vari-
ance reports and balance sheets. The point is that business value
is not a single value for a complex asset set. It is a set of values,
each of which measures different business objectives. The effec-

162 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 163 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

tive business management of any asset set involves balancing the


multiple business measures in order to maximize the overall
amount of money that the company is earning at any point in
time. Focusing on improving a single business measure could
actually cause the overall business value to decline. For example,
a strategy to maximize the production output of a set of assets
may result in a higher production value through that asset set of
$X, but the increased energy, material, and labor cost required
may cost $2X. Although the strategy may have worked for the
result it was designed to accomplish, it actually reduced the over-
all value of the asset set to the company. When it comes to man-
aging the multiple business objectives of industrial assets, single-
objective optimization is just not sufficient. The problem involves
vectors of business values, not just single values.

In reality, different business values will take on different levels


of importance at different times in a manufacturing operation.
Often the driving forces that lead to a change in priority are exter-
nal to the plant and sometimes even the company. For example,
if the demand for a product is very high and the production
capacity is low, production volume may be the most important
measure, with energy cost still important, but at a lower priority
level. If, on the other hand, market conditions change such that
demand is reduced below capacity and energy cost increases, the
priority of the measures may shift. Such a shift in priority would
reflect the manufacturing strategy the company needs to create
the most wealth at a specific point in time with a specific set of
manufacturing conditions. The business measures chosen and the
priority assigned to those measures must reflect the optimal man-
ufacturing strategy at any point in time. In the 1980s, the manu-
facturing strategies of industrial operations seldom changed.
With globalization and the resulting complexity of market
dynamics, industrial operations may have to change strategy and
reprioritize measures multiple times in a single day.

The DPMs for any asset set provide a prioritized business value
vector that is represented by multiple DPMs. In earlier chapters of
this book, we discussed prioritizing DPMs according to manufac-
turing strategy. This produces a prioritized business measurement

B O T T O M - L I N E A U T O M A T I O N 163
BottomLine05.book Page 164 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

system that effectively encompasses the complexities of the asset


sets being measured and can be represented as an ordered asset
value vector. This ordered vector matches the complex, multiple
objective nature of a manufacturing business. To develop an asset
value equation for asset sets that is similar to the equation for a
single asset shown in figure 6.6, the vector approach must be
applied across each of the factors of the equation. The economic
value factor, effective asset utilization factor, and effective asset
availability factor all must be represented as vectors. This makes
logical sense because each DPM must be evaluated with respect to
its relationship to economic
value, effective availability,
and effective utilization.
The resulting equation of
instantaneous asset set
value is shown in figure 6.8.
The Emax vector assumes
Figure 6.8 Instantaneous a slightly different form here than does the corresponding AVmax-
Asset Set Value
Cmax in the single asset equation (see figure 6.6). This is because a
single asset in economic terms will always be represented by some
maximum asset value minus the cost of operating at the maxi-
mum value. However, for asset sets, the equation becomes a vec-
tor analysis, and the economic value term may change its
connotation from profit to cost, for example, depending on the
vector element. If the vector elements represent energy consump-
tion, then the ei will represent the energy cost at the maximum
possible consumption level of the asset set.

It is important to note that the prioritized ordering of the


DPMs in the asset value vector must correspond to the vector
ordering of the other vectors. If, for example, DPM1 represents
the current energy cost, then e1 would represent the maximum
energy cost if the asset set was operating at top capacity, while u1
would be the current utilization relative to energy cost and a1 the
current availability relative to energy cost. Interestingly, the right-
hand side of the model shown in figure 6.8 becomes very difficult
to solve because of the inherent complexities of the EA and EU
vectors, as previously discussed. However, the left-hand side is
directly measurable through the prioritized DPMs. This model

164 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 165 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

provides a holistic way of relating the critical KPIs to the business


value of the operation through the DPMs.

The set of DPMs for


each asset set provides
the elements of the per-
formance (AV) vector.
These elements must be
compared to those of the
EA vector and the EU vec-
tor in order to under-
stand the relationships
between them. It is
important to recognize
that the E vector is a set
of constants for the asset
set. Figure 6.9 presents a
simplified, two-dimensional view of this vector relationship, Figure 6.9 A Simplified,
Two-Dimensional View of
although clearly this is not a two-dimensional problem. the Asset Value Sector
The model in figure 6.8 represents the instantaneous calcula-
tion of value from an industrial asset set. Just as with the base
asset case shown in figure 6.7, it is often useful to calculate the
total value
generated over
a time period.
This can be
accomplished by taking the integral over that time period as Figure 6.10 To Calculate
an Assets Total Value Over a
shown in figure 6.10. Time Period, Use the
Integral Over That Period

BUSINESS VALUE ANALYSIS


Typically, in trying to maximize an industrial operations busi-
ness value, the time period t should cover a business reporting
period. In other words, the asset should be optimized across a
business reporting period, not an arbitrary time period defined by
plant personnel. From a business perspective, the objective is to
generate as much business value as possible. The issue of business

B O T T O M - L I N E A U T O M A T I O N 165
BottomLine05.book Page 166 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

value is very much associated with the reporting period under


consideration and the inverse nature of EA and EU. For example,
over a five-minute period, both the EA values and EU values may
be driven close to their maximums with no adverse impact. But as
the timeframe expands, the probability of a negative impact on
effective availability from running close to maximum utilization
will increase. The business value from the asset set will start to
decline. From a business perspective, the issue is how to drive the
maximum value over a full financial reporting period. This may
actually require operating the assets at a lower target utilization in
order to improve the effective availability over the entire time
period, thus maximizing the total asset value for the period.
Understanding the multiplicative relationship presented in the
equation, even if the equation is not implemented, can help plant
operations personnel experiment with its utilization of equip-
ment and the resulting availability trends so as to move toward
maximized asset value. The key is to balance availability and utili-
zation so as to drive maximum value. However, this can only be
done if each of the values in the asset value vector, the DPMs, is
calculated.

Traditionally, the relationships between operational (KPI) and


business metrics have not been known and understood. When
business values begin to be collected and historized along with
the operational values, these relationships will begin to be
revealed. Effective asset utilization and asset availability are but
two operational concepts for which plant staff can develop mea-
sures and compare them to business value. Similar analyses can be
done for many more operational metrics. The behavior of each
asset set is unique, but general relationships can be revealed. The
first step is to develop, model, and collect the DPMs and the key
operational KPIs such as EU and EA. Once this information is col-
lected in the historical database, the specific relationships can be
developed, modeled, and analyzed. The actual business contribu-
tion of both the maintenance and operations functions can be
determined, and these can then be aligned to the ultimate objec-
tive of generating incremental business value for the company
(see figure 6.11).

166 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 167 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

Aligning met-
rics all the way
from the plant
floor through the
KPIs up to the
DPMs and finan-
cial measures is a
critical success fac-
tor in the manage-
ment of
performance and
the generation of
economic value
from industrial
assets.

Figure 6.11 Aligning


Metrics from the Plant Floor
to the Business Measures Is
Critical

A PRAGMATIC APPROACH FOR


OPTIMIZING BUSINESS VALUE
FROM PLANT ASSETS
The models discussed so far provide interesting information,
but implementing the full models for each asset set in a plant can
take considerable time and effort. However, there is a practical
approach to using these models that will allow companies to gain
most of the business benefit without having to implement all of
the models. It involves combining a level of engineering analysis
with human learning.

The first aspect of this practical approach is to extend the


accounting system down to plant floor business value and calcu-
late the performance measures in real time. This is accomplished
by installing DPMs throughout the plant to the process unit level
or below (see chapter 4). The DPMs must then be prioritized to
the plants current manufacturing strategy by applying a Voll-
mann decomposition process. Once the DPM models are
installed, the DPM data should be collected into a process histo-

B O T T O M - L I N E A U T O M A T I O N 167
BottomLine05.book Page 168 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

rian that is also being used to gather and store operational data.
The process historian software can be used to aggregate the data
so as to shift measures and their frequency, from daily to weekly
to monthly. This not only provides key time periods for data
analysis, but also puts the data into a more usable accounting
timeframe. Often the aggregated data can be inserted right into
the plant accounting system. However, two critical issues must be
kept in mind if this data is to become source data for the account-
ing system. First, the DPM data includes direct cost data; no over-
heads are allocated. In contrast, most accounting systems assume
some allocations for overhead costs. This disparity may need to be
resolved to make the information useful to the accounting team.
Second, the source of data that is directly used in the accounting
system may need to be audited.

Once a reasonable history of this performance and operational


data has been collected and stored in the historian database, the
engineering team can analyze the resulting information to iden-
tify the operational conditions that will create maximum value
from the assets. This process will enable engineering to develop
sets of operational initial conditions for different desired operat-
ing states. When the plant is to be operated to meet a certain set
of business objectives, this historical data can be used to drive the
plant to the initial operating state, which, in theory, should be
very close to the optimal operating point.

The problem with this approach is that plants are dynamic,


and the initial conditions derived using this method will repre-
sent the way the plant was operating at some time in the past.
However, its a safe assumption that some changes in the plant
have occurred in the intervening period. This is where the opera-
tor dashboard displays, explained in chapter 4, can be very useful.
The operators can use the prioritized DPM information presented
to them on these displays to essentially adjust for the changes
that may have occurred. This is similar to using a feedback trim
mechanism with a feedforward controller. The feedback trim
looks at the way in which the plant is actually operating while the
feedforward controller provides a model of the way in which the
plant operated at some point in the past. This combination

168 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 169 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

enables engineers, operators,


and maintenance personnel to
work together to arrive at the
optimal operating points for
the plant assets (see figure
6.12).

Even though the approach


weve just outlined is more
practical than building the full
asset value model for each
asset set in a plant, the initial
conditions and dashboards
should be developed for each anticipated manufacturing strategy. Figure 6.12 Developing
Initial Conditions from
The priority of the business value measurements will change with History
each strategy change. This requires new initial operating condi-
tions and a reprioritization of the DPM vector. It may be difficult
and time consuming to develop the initial conditions for each
potential strategy, but if they can be developed for the most com-
mon strategies that may suffice. Experience has shown that using
the operator feedback system through a dashboard will move the
plant close to optimal operation without having to set the initial
conditions, though this takes considerably longer. Engineers
should evaluate the economic value of reducing this time against
the cost of doing the analysis to determine initial conditions. It is
important to point out that traditional operational improvement
and optimization strategies at the base asset level, which is below
the asset value measurement context, can provide value and can
underpin this higher-level approach, producing better business
performance.

Using business measures and operational measures to develop


a business performance improvement system such as the one just
described yields several benefits and positive characteristics. The
results of this practical approach are directly measurable in busi-
ness value terms. Moreover, this system is very flexible and can
readily change if a strategy is changed. It essentially provides the
elusive multi-objective optimization approachwith each DPM
representing a different objective functionthat complex opera-

B O T T O M - L I N E A U T O M A T I O N 169
BottomLine05.book Page 170 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

tions require. It is interesting to notice that humans offer the only


efficient solution to achieving multi-objective optimization. The
prioritized DPMs presented on the dashboards give the operators
what they need to accomplish this. This system generates a new
level of understanding at the plant floor level (operators and
maintenance), at the engineering level, and at the business level
concerning what truly drives value. It provides business managers
with the tools they need to align personnel with strategic busi-
ness objectives and to realign them when the strategy changes.
And finally, all of this can be accomplished using technologies
available today.

BARRIERS TO DEPLOYMENT OF
BUSINESS PERFORMANCE
OPTIMIZATION
In practice, several manufacturing operations have encoun-
tered a few barriers to the successful deployment of this kind of
plant asset business performance optimization approach. Under-
standing and being able to identify these barriers may enable
companies to address them effectively. Almost none of the barri-
ers encountered involved technology limitations. Rather, the
most common barriers involved people and organizations either
not being willing or not being able to adjust to a new way of
doing business.

Almost every chief financial officer the author interviewed


appeared to understand and value the concepts of real-time per-
formance measurement aligned with accounting and strategy and
their application to optimizing business value, when presented
with them. However, many plant accountants were very uncom-
fortable with them. Accountants are trained intensively in the
process of accounting, which is broken down into formulas, equa-
tions, and reports. Thus, changes to the traditional accounting
perspective and approach can lead to huge resistance. Experience
has shown that the most effective way to overcome this resistance
is to get the CFO to sponsor the program and make it clear that
support for this change comes from the CFOs office.

170 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 171 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

Plant personnel are often very comfortable with their opera-


tional environment and have become used to not being mea-
sured. It is to be expected that they will resist the implementation
of a performance measurement system that makes the business
impact their activities. This resistance is really costly and counter-
productive because experience has shown that the plant floor
people add tremendous value when they are empowered to do so.
Once the performance measurement systems are installed and the
plant level teams become used to them, they tend to become very
strong advocates. The performance measurement systems typi-
cally make their jobs more satisfying and enjoyable. People tend
to like knowing that they are making a positive contribution, and
performance measurement systems clearly demonstrate this
impact.

For decades, technology has been blamed for creating islands


of information or automation. Now that the technology barriers
are being pulled down, islands still appear to exist, but they are
islands of organization. Functional groups within industrial orga-
nizations, such as engineering, accounting, strategic planning,
operations, maintenance, and production planning have become
comfortable working in isolation. Accountants and engineers sel-
dom interact, and often do not enjoy it when they do. Operators
and maintenance have often been in conflict. With a holistic
business value approach such as the one outlined here, the met-
rics actually work to encourage functional organizations to work
with other functional organizations with the common goal of
maximizing business value. Although functional organizations
may resist working together initially, experience has shown that
the existence of clear, common goals and associated metrics are
the key to getting them together. Once the disparate organiza-
tions become aligned to common goals and metrics, the resulting
performance improvements and organization convergence can be
very significant.

Separating information technology (IT) systems and automa-


tion systems can be a huge barrier to success. Although this may
at first glance appear to be a technology issue, it is not. The tradi-
tional separation of the business computer system from the plant

B O T T O M - L I N E A U T O M A T I O N 171
BottomLine05.book Page 172 Friday, August 19, 2005 9:54 AM

C H A P T E R 6 : T H E E C O N O M I C O P T I M I Z A T I O N O F
I N D U S T R I A L A S S E T S

automation system has been more of an organizational than a


technological partitioning. Many automation systems are based
on the same hardware and system software as the business sys-
tems. However, business systems have been the domain of the IT
department while automation systems have been the domain of
engineering. As the characteristics of these two classes of systems
converge over time, the inclination has been to bring them
together. But engineering and IT groups have, for the most part,
tried to keep them separate to safeguard their control over their
traditional domains. These systems must be bridged to build a
business performance optimization system.

The extreme technology focus of both the IT and engineering


organizations has to change. For decades, most decisions concern-
ing which computer technologies to procure have apparently
been made on technology features and functions rather than on
the business value the systems might provide. This could be
because organizations did not have the correct performance mea-
surement systems in place. Without contextualized performance
measures, people tend to gravitate toward what they like and are
good at. In the case of IT and engineering it is technology. Having
the latest and greatest technology is not as important as solving
business problems that add business value to the company. The
focus of these groups must therefore change from a technology
for its own sake focus to technology for business value.

Finally, the concerns over traditional capital approval and


spending processes we discussed earlier have become a huge bar-
rier to the innovative deployment of technologies. These pro-
cesses have resulted in a replacement technology mindset
rather than the required business solution mindset.

All the required technology is available today to begin the


drive toward new levels of industrial performance improvement.
The only real barrier is people. It is time to change the way things
have been done and create new levels of business value from our
high-potential plant assets.

172 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 173 Friday, August 19, 2005 9:54 AM

CHAPTER 7

Case Studies

INTRODUCTION
Implementing a DPM system correctly requires a blend of stra-
tegic, accounting, and process engineering skills. It would be great
if a nice fill-in-the-blank template existed to easily guide a practi-
tioner in how to correctly identify the strategic performance mea-
sures and generate those measures algorithms. Unfortunately, no
such template exists.

The most critical aspect of a DPM implementation is identify-


ing the correct measures. Even if the measures modeled and dis-
played on a DPM dashboard display are wrong, the plants
personnel will still tend to perform to those measures, which
could drive the plants performance in the wrong direction. Rigor-
ously following the Vollmann decomposition process will help to
ensure that the correct measures are identified, but the process
still relies on the analysts strategic, accounting, and engineering
skills. We discussed the Vollmann decomposition process in chap-
ter 4, but the best way to understand the process is to illustrate it
with case studies of actual implementations. Each of the follow-
ing case studies are based on real programs, and each illustrates
important aspects of DPM implementations.

B O T T O M - L I N E A U T O M A T I O N 173
BottomLine05.book Page 174 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

CASE STUDY 1:
DYNAMIC PERFORMANCE MEASURE
PROGRAM AT A MAJOR
PHARMACEUTICAL COMPANY
The pharmaceutical company in this case study has become a
leader in automation for automating its global bulk pharmaceuti-
cal processing facilities. Its management team has been driving a
major corporate transition to make the company more competi-
tive in the tough global marketplace while maintaining its strong
emphasis on quality. Major components of the transition include
refocusing the employees on those activities that provide the
highest level of economic value-added (EVATM the residual after-
tax profit after deducting the full cost of capital. EVA is a trade-
mark of Stearn Stewart). This shift has begun to influence all of
the firms employees, including its technology community. Com-
pany personnel have invested significant effort developing life-
cycle economic models for the firms automation investments
and determining the economic value-added (EVATM) it is realizing
from those investments. One of the major issues they face imple-
menting managements objectives is how to measure the EVATM
of the activities in the firms manufacturing operations.

In late 1998, a series of discussions occurred between the phar-


maceutical companys Global Automation Team and the Perfor-
mance Measurement Team, which consisted of two business
measurement consultants and two performance improvement
engineers of The Foxboro Company. The subject of the discus-
sions was how to measure the economic performance of the bulk
manufacturing operations accurately and in real time. The impe-
tus behind the implementation of the real-time performance
measures was to demonstrate which activities in the plants add
economic value and which do not. With this knowledge, the
pharmaceutical company can make more intelligent decisions
about where and how to invest its resources.

These discussions generated considerable interest among the


pharmaceutical companys automation personnel, who have
agreed to test the DPM concept to determine its applicability in
their bulk pharmaceutical operations. One of the companys large

174 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 175 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

bulk pharmaceutical plants has been selected for the implementa-


tion. It is a highly automated, best-in-class facility that performs
quite well and is therefore determined to be an ideal site for this
performance measurement program.

Overview of the Pharmaceutical


Companys Facility
The selected facility manufactures bulk active pharmaceuticals
for the global specialty pharmaceutical market, which are shipped
to other facilities for final packaging. The plants production tech-
nology is based on organic synthesis and product purification.
The site has production buildings in which the pharmaceuticals
are produced, a tank farm for supplying solvents to the produc-
tion buildings, and an environmental controls section, which
manages all aspects of the production waste streams. The environ-
mental controls section is comprised of two areas for recovering
solvents from the spent solvent slurry created from the manufac-
turing operations and two incinerators for disposing of the plants
liquid waste.

The facilitys environmental controls area is a critical part of


the site, and it was very clear to Foxboros Performance Measure-
ment Team that the pharmaceutical company was extremely con-
cerned about its environmental stewardship. It applied
considerable effort to minimizing the facilitys environmental
impact. In fact, the pharmaceutical company is so concerned
about environmental issues that it has decided not to increase the
number of incinerators on site, although this could constrain the
plants production capacity.

From the interview process, the Performance Measurement


Team has learned that the pharmaceutical companys manage-
ment has established several important goals for this production
facility. The objectives are to

continuously improve environmental stewardship


continuously improve safety
maximize productivity of the physical assets
operate at minimum inventory levels

B O T T O M - L I N E A U T O M A T I O N 175
BottomLine05.book Page 176 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

ensure that manufacturing costs are lower than those


of good generic manufacturers
be the best value supplier in all aspects of customer
service
ensure that the site is the customers first choice for
new products
aggressively understand and use compliance to drive
long-term improvement

Understanding a sites goals is critical to any performance-


improvement program. The plant had already worked very hard
to provide measurements that are in line with the sites goals. All
the personnel the Foxboro Performance Measurement Team
encountered at the plant appeared to be quite aware of the
measures and regularly referred to them in their daily reports.
This strong measurement focus created an ideal environment for
implementing, using, and accepting dynamic performance
measures.

Performance Measurement Analysis


at the Pharmaceutical Company
After several discussions, the plant personnel and the Perfor-
mance Measurement Team decided that the best initial course of
action would be for the Performance Measurement Team to visit
the plant. The objectives of the visit were (1) to present and dis-
cuss the dynamic performance measurement (DPM) process with
the plants management team, (2) to select a plant section, and (3)
to conduct a DPM analysis on the selected plant section. The visit
at the facility was scheduled over a two-day period in March
1999. The primary representatives of the pharmaceutical com-
pany who were involved in the analysis were the leader of the
facilitys Process Control Group and an engineering consultant
from its Engineering Tech Center.

The Performance Measurement Teams first step once on site


was to do a presentation for the plants management team on
DPM and the analysis process. The presentation was critical
because it created alignment between the Performance Measure-

176 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 177 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

ment Team and the plants management team and set expecta-
tions with respect to the analysis process and overall DPM
process.

As a result of the presentation some members of the team


raised a concern that in an FDA-regulated plant such as this plant,
the operators may not have the necessary freedom to drive the
improvements once the performance measurements are in place.
Once the firm and the FDA agreed on good manufacturing prac-
tices (GMPs) for the facility, they would have to be followed to
the letter. The concern was that the FDA would have to certify
any improvement before it was implemented. The joint team
acknowledged this limitation and agreed that the FDA regulations
could significantly limit the plants improvement potential. On
the other hand, the team determined that some improvement
might be possible even within the definition of the GMPs, and so
the program was worth a try.

The Performance Measurement Team then began to undertake


a structured series of interviews with plant personnel. These inter-
views are the primary on-site activities underlying the Vollmann
decomposition analysis. They were conducted in such a way that
the person being interviewed did not need to understand the
intricacies of the decomposition process. They only needed to
respond to a few key questions. The interviews were designed to
identify the strategic performance measures that reflect the plant
management teams strategic goals and objectives and to decom-
pose those measures down through the various plant sections to
the process units.

The leader of the Process Control Group was selected to repre-


sent the plant on the Performance Measurement Team. His partic-
ipation helped the team to synthesize the information it received
from other company personnel and to provide any additional
information it needed.

The Vollmann decomposition analysis process provides a top-


down approach to identifying and decomposing the facilitys stra-
tegic performance measures. Therefore, the first interview was
conducted with the sites general manager. During this interview,
the plants key performance measures were clearly identified. The

B O T T O M - L I N E A U T O M A T I O N 177
BottomLine05.book Page 178 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

primary measure was plant production capacity, which is the


measure of potential plant production in a given period of time.
This measure was a crucial piece of information because the plant
is but one of a number of the pharmaceutical companys plants
that were designed to manufacture bulk pharmaceuticals. When
its R&D group develops new products management determines
which of its plants will manufacture them, based on the plants
manufacturing capability and its available production capacity.
Thus, the economic profile of any plant is directly driven by its
production capacity.

As the interview with the general manager progressed, the


plant teams concern about the environmental impact of the
plant on the local community deepened. Local environmental
concerns are very strong throughout the areas in which the phar-
maceutical company operates, so some environmental measures
were determined to be essential considerations for this plant. Sev-
eral times during the interview, the GM also pointed out that the
plants environmental controls section was operating at capacity,
so they would not add more incinerators to expand the produc-
tion capacity. The Performance Measurement Team initially inter-
preted these comments to mean that the plants environmental
control and incineration sections were a major constraint on the
plants primary performance measure: production capacity.

The plant teams other interviews were conducted with several


plant operations managers as well as with engineering personnel.
The objective of these interviews was to understand the plants
layout and structure, to ascertain the organizational decomposi-
tion of the plants operations teams, and to gather information
for use in the DPM decomposition analysis. The team also toured
various plant areas throughout the day.

The interviews demonstrated that the plants management


team had a solid understanding of the concepts of DPM. The met-
ric program that was already in place at the plant, which had
already been decomposed to the level of the plants processing
sections, provided a good basis for the DPM program. Each man-
ager produced summary reports to determine the daily or weekly
improvement on each identified metric.

178 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 179 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

As the interviews progressed, several members of the plants


team suggested that the Performance Measurement Team should
focus the test DPM analysis and implementation on the plants
Environmental Controls area. This suggestion seemed to be in
line with the issues expressed in the interview with the sites gen-
eral manager. The solvent recovery and incineration sections were
suggested and selected for three important reasons:

1. The management team was keenly aware of the plants


impact on the local environment. Since a significant por-
tion of the plants environmental impact derives from the
recovery and incineration operations, these operations are
critical to plant management. The local office of the Envi-
ronmental Protection Agency (EPA) requires that emission
measurements be made continually to gauge the environ-
mental impact of the plant. With this in mind, the phar-
maceutical company has decided to limit the plants
incineration capacity to the two existing incinerators. If
the incineration capacity is exceeded, they will transport
the waste off site for incineration, at considerable
expense.

2. With the plants incineration capacity limited to the two


existing incinerators, its recovery and incineration capac-
ity is becoming a major production bottleneck.

3. The alternative areas, namely, the pharmaceutical build-


ings, strictly adhere to FDA cGMPs (current good manu-
facturing practices), and plant personnel feel they may
not have the necessary freedom to achieve improvements
in these areas of the plant.

The joint team agreed to focus the remainder of the analysis


on the facilitys Environmental Controls area, which, as noted,
consists of the solvent recovery and incinerator sections. This
focus was not intended to imply that these were the only areas of
this facility that would benefit from real-time performance mea-
sures. Nor does it mean that these plant sections would provide
the highest economic returns. Rather, this focus was intended to
establish an initial area of concentration that meets the require-

B O T T O M - L I N E A U T O M A T I O N 179
BottomLine05.book Page 180 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

ments of both the pharmaceutical company and the Performance


Measurement Team.

Process Overview and DPM Analysis


The facilitys solvent recovery section and incinerator section
are tightly linked from a processing and from a DPM perspective.
A spent solvent mixture containing the various solvents used in
the production process as well as some residual waste from the
dry ingredients is output from the production buildings. The
spent solvent slurry is transferred from the manufacturing build-
ings, through the tank farm, to one of the two solvent recovery
sections. The slurry is processed through distillation columns and
stills in the solvent recovery section in order to reclaim as much
as possible of the original solvents in a pure enough form to be
reused in the production processes.

The primary objective of the Environmental Controls section


is to process as much solvent slurry as possible so as to reduce the pro-
cessing cost and increase the overall production capacity of the plant,
within the plants environmental constraints. The Environmental
Controls section also produces very high quality solvents for
reuse in the processing steps of the production buildings.

There are three basic elements of the solvent recovery sections


mission: (1) to recover as much as possible of the different sol-
vents used in the plant, (2) to meet regulatory requirements, and
(3) to manage the supply of solvents to the operations. Increasing
the amount of recovered solvent reduces the amount that needs
to be processed through the incinerators. This has two effects on
the facility. The first is an increase in the availability of solvent for
operations. The second is a reduction in the amount of waste gen-
erated.

Each of the solvent recovery sections consists of a number of


distillation columns and stills, which are used to recover the sol-
vents from the solvent slurry stream (see figure 7.1). The distilla-
tion columns must be tuned to the specific characteristics of

180 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 181 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

the solvent slurry that is


Solvent Recovery Overview
being treated at the time.
Since the manufacturing oper- Goal: Maximize Spent Feed Rate
ation involves multiple prod- Spent Tank Farm
Solvent Flow
ucts and campaigns, the (Feed) Recovered
Solvent

operators may need to adjust Column


New
Solvent
7 Recovered
the tuning of each column Approved
Solvent

frequently. The start-up of the Steam

columns for each slurry stream Production


Buildings
is a manual operation that
typically takes from three to Primary
Waste
five hours, depending on the Secondary Waste EC
Spent Solvent

operator. Until a column is up


to full production and efficiency, the columns output goes back Figure 7.1 Solvent Recovery
Overview
into solvent slurry storage to be retreated.

The selected solvent slurry is introduced into the distillation


column from a spent solvent in the tank farm. The composition
of the feed is determined through lab analysis. The integrity of
the composition of the feed stream is extremely important to the
recovery process. Some solvent slurry feed streams need to be pro-
cessed many times to achieve a solvent of sufficient purity to be
reused in the operations. Solvents are recovered through distilla-
tion and transferred to a recovered solvent tank in the solvent
tank farm. Lab analysis is performed on the recovered solvent to
determine if it is at the quality level that has been approved for
use in the manufacturing operations. If approved, it is then trans-
ferred to a recovered/approved tank and can be reintroduced into
the manufacturing buildings as appropriate. Solvent that does not
meet quality standards is reintroduced into the recovery process.

There are two incinerators at the plant, a John Zink incinerator


and a Thermall incinerator. Both work on the same basic princi-
ples (see figure 7.2). Two waste streams are introduced into each
incinerator, a primary waste stream and a secondary waste stream.
The primary waste stream contains the nonrecovered solvent out-
put from the recovery process. It consists of solvents and residual
materials from the manufacturing operations and is highly flam-
mable. The primary waste is introduced as fuel into each

B O T T O M - L I N E A U T O M A T I O N 181
BottomLine05.book Page 182 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

incinerator and is used to


Diesel Fuel incinerate the secondary
Primary Waste
Flow waste stream. The secondary
Secondary Waste
Flow waste stream is comprised of
Atomized Air
production waste, such as
Residence Time the water-based residue gen-
(inferred)
1 second
erated when the process
equipment is cleaned. The
secondary waste has a high
Incinerator
water content that requires a
AT
Quench Water significant amount of heat to
Flow
Hot Gases
(1000 C)
incinerate. The temperatures
in the incinerators are
Figure 7.2 Incinerator maintained at just over 1,000C. A residence time of one second
Overview
at this temperature is required to incinerate the secondary waste.
Atomized air is also introduced into the incinerators to stimulate
combustion. Diesel fuel is used to help start up the incinerators
and to compensate for any shortcomings in the BTU content of
the primary waste.

Analytical instruments are used to determine the emission lev-


els in the flue gasses (HC, CO, HCL) and to ensure that they com-
ply with EPA limits. Quench water is used to pull some of the
impurities out of the flue gasses and accelerate cooling. When the
incinerators reach certain environmental limits (HC, CO, HCL),
they will trip, which causes them to automatically shut down.
Once an incinerator shutdown occurs it takes the operations staff
some time to restart the incinerator and return it to the opera-
tional limits. The primary operational objective of the incinera-
tors is to be operated at the highest rate of incineration possible
with the lowest possible environmental emissions.

The efficiency of the combined solvent recovery and incinera-


tion processes is the key to the sites performance effectiveness.
The Performance Measurement Team suggested implementing
and managing appropriate dynamic performance measures for
each of the solvent recovery and incineration sections. Although
there are a number of economic performance metrics that could
provide indicators for these important areas of the operation (e.g.,

182 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 183 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

power costs and recovered solvent value), optimizing these vari-


ables independently would not address the plant's chief concerns.
In fact, optimizing some of these variables could lead to operator
and engineering actions that actually decrease the economic
value of the overall facility. For example, it was suggested that one
appropriate measure might be calculated as the percentage recov-
ery divided by the cost of recovery. The objective would be to
increase this metric through a combination of actions: increasing
the numerator and decreasing the denominator. The best way to
increase this metric may actually be to reduce the denominator
(power cost)which is not a major economic consideration to
the plantrelative to production value. Optimizing this metric
could actually reduce the production capacity and production
value of the plant, which would have dire economic conse-
quences for the overall facility.

The next aspect of the DPM analysis was to develop an opera-


tional decomposition of the site to determine which control
rooms and operating areas would need to have DPM dashboard
displays presented to the operators. The actual DPM construction
is accomplished for each operational unit in a plant, but the dash-
board displays are constructed for an operators functional
domain of responsibility. It is therefore important to map the
operational measures to the operator displays. There is an operat-
ing superintendent for the overall Environmental Controls sec-
tion, but there are two control rooms under that superintendent.
One of the control rooms hosts the operating team for both
recovery operations, and the other control room is for both incin-
erators. This operational structure led the team to the conclusion
that two sets of DPM dashboards would need to be developed one
for the solvent recovery operators and the other for the incinera-
tor operators. Once this information was collected the Perfor-
mance Measurement Team left the site to compile the DPM
analysis report.

B O T T O M - L I N E A U T O M A T I O N 183
BottomLine05.book Page 184 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

DPM Analysis Report


The Performance Measurement Team developed the DPM
analysis report and sent it to the plant management team for
review within a month of the first site visit. After reviewing the
initial report, the plant team members explained that they under-
stood the logic of the Performance Measurement Teams perspec-
tive on production capacity. The plant management team
explained that the value of the products made at the plant was so
great that they would never allow the Environmental Controls
area to become the limiting factor on production capacity. Once
this was understood, it became clear that the value of increasing
the capacity of the Environmental Controls section was the cost
savings that would result from not having to send out the spent
solvent for external processing vis--vis the value of the increased
production capacity.

The type of interaction that took place between the Perfor-


mance Measurement Team and the site management team after
the initial review of the DPM analysis report is essential, and
changes at this point in the process are not unusual. The Perfor-
mance Measurement Team has found in its experience that most
manufacturing operations and engineering groups are not com-
fortable dealing with econometrics. So making some econometric
assumptions and presenting them back to the plant team is a use-
ful way to get the team to understand what information is
required and how important it is to select the correct values. Once
the initial proposal was sent to the plant, the plant teams
approach to the analysis changed. The plant team and the Perfor-
mance Measurement team appeared to begin working more effec-
tively together at this point because they had a common
understanding of what the programs intended scope was.

The DPM analysis report was corrected based on a second


round of discussions that occurred after the initial report was sub-
mitted. The plant team analyzed the second report and after a few
weeks agreed with the proposal to implement the DPMs in the
Environmental Control section of the plant. The Performance
Measurement Team then paid a second trip to the site to build the
DPM models, install them on the operating automation system,

184 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 185 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

and historicize them using the historical database running in the


automation system.

After the Performance Measurement Team made the requested


adjustments, the primary objective of the DPM implementation
was to enable the plant operators in the plants solvent recovery
and incineration sections to increase the capacity of the spent sol-
vent processing. The econometric calculations for the solvent
recovery and incineration sections were based on (1) the reduc-
tion in cost to the plant because it no longer had to pay for off-
site incineration of spent solvent and (2) the increased value of
the incremental recovered solvents.

The capacity of the spent solvent processing is determined by


the serial relationship between the solvent recovery operation
and the incineration operation. The first part of this relationship
is the solvent recovery sections ability to recover more solvent
from the spent solvent streams. The recovery of this additional
solvent means more solvents are available to the plant for the
manufacturing operation and less waste has to be incinerated.
This metric is calculated as a function of the increase in recovered
solvent.

The cost of the plant to process the waste solvent off site was
determined to be a constant value v1 per liter. The model is as
follows:

Recovered Solvent Baseline Recovered Solvent * v1

The second part of the serial relationship was the increased


capacity of the incinerators to handle waste. For each liter of
increased incineration capacity, one less liter had to be sent off
site for processing. This metric is calculated as follows:

(Incinerator Throughput Baseline Throughput) * v1

The total economic value is the sum of the reduction in waste


through recovery, the value of the incremental recovered solvent,
and the value of the incremental waste processed by the incinera-
tors. Hence,

Value = Value of the Increase in Environmental Controls Throughput


+ Value of the Increase in Returned Products from Solvent Recovery

B O T T O M - L I N E A U T O M A T I O N 185
BottomLine05.book Page 186 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

Two additional measures proved to be valuable: incinerator


uptime (averaged over a twenty-four-hour shift day starting at
8:00 am) and amount of waste product from columns. The incin-
erators are a significant bottleneck in the process. If they stop pro-
cessing solvents then the plants ability to dispose of waste is
reduced and the costs are significantly increased. Such an inter-
ruption in processing can be caused by either process and
mechanical trips or by interlocks initiated by environmental
monitoring instrumentation on the flues. Finally, reducing the
total waste by recovering solvent would provide greater through-
put of solvents in the Environmental Controls section, assuming
the incinerators are being used on a constant basis. These metrics
are not really econometrics, but they provide valuable informa-
tion that helps the operators manage their prime econometrics.

Models of the DPMs and supporting operating measures were


developed to operate in the automation system installed at the
pharmaceutical companys site. Automation systems are natural
platforms for implementing DPMs since they have direct access to
the process sensors and have software that enables algorithmic
models to be developed that operate in real time in the automa-
tion systems control packages.

Since the solvent recov-


ery and incinerator sec-
Target
tions of the operation are
Target
run by different operating
teams, two DPM dash-
Good Good board displays were devel-
Recovered Feed Percent Recovered/ Spent Feed
oped for the Environ-
mental Controls area at
the site. One display was
Target
developed for the solvent
recovery operators and

Recovery Section Rework


Good one for the incinerator
operators. Supporting
Figure 7.3 Solvent Recovery metrics were developed for each section to provide additional
Operator Dashboard
information that would help the operations groups maximize the
value of the primary metric. In each recovery section, the primary

186 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 187 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

metric was the increase in recovered feed of the entire facility. Sec-
ondary metrics were also provided for increased percentage of
recovered solvent to spent feed and reduced rework volume.

The primary metric of the incineration sections dashboard


display was increased incineration throughput. This throughput
is constrained by the level of contaminants released into the envi-
ronmentan absolutely critical constraint on the operation of
the plant. The operation will shut down if the limits are exceeded
and the cost of a shutdown is excessive. Therefore, a second sup-
portive metric is incineration uptime (see figure 7.4). Clearly,
more time available for incineration will result in higher levels of
throughput. A single incinerator trip could lead to an extended
shutdown period since it takes time and effort to get the incinera-
tors back on line and operating efficiently.

The Performance Measurement Team implemented the DPMs


for the plants Environmental Controls area during a second visit
to the site in September 1999. They also implemented the histo-
rian data collections for each of the DPMs and calculated the
monthly averages for each DPM into the historian. These
monthly averages were used to establish the baseline values for
the initial performance. The value of any performance improve-
ment activities could be measured by their incremental deviation
from the baseline operating value. The operators dashboard dis-
plays were not implemented at this point so a reasonable baseline
could be established for the metrics. The dashboards are actually
performance-improvement support tools, and it would be impos-
sible to derive valid baseline performance while the dashboards
are operating.

The baseline values


were established during a
Target Target
six-week period that
ended in the middle of
October. Once the base-
lines were determined, the Good Good
Incinerator Throughput Incinerator Up-Time
plant team decided that
the first performance improvement activity they would undertake Figure 7.4 Incineration
Operator Dashboard
would be to apply advanced process controls to one of the distilla-

B O T T O M - L I N E A U T O M A T I O N 187
BottomLine05.book Page 188 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

tion columns in the solvent recovery section. The advanced con-


trols were implemented, and the DPMs indicated that the annual
value of this activity would add a sustainable $400,000 per year
going forward. Without the DPMs it would have been difficult to
determine the value of the advanced controls.

The Performance Measurement Team made a third site visit in


January 2000. During this visit it implemented the DPM operator
dashboard displays and trained the operator teams in how to use
them. The plant team then undertook a series of other improve-
ment activities involving new operating practices and operator
training. Once again, it would have been very difficult to deter-
mine the value of these activities to the pharmaceutical company
before the DPMs were implemented. With the DPMs operating,
the plant team was able to obtain immediate feedback regarding
whether an activity was providing incremental value. By using
this information they were able to direct the operations staff to
focus on those activities that were providing the most value. The
net result was an incremental sustainable EVATM to the plant
operation of approximately $1 million per year.

The plants performance improvement program is ongoing.


The initial results of about $1.5 million of annual economic bene-
fits to the pharmaceutical company have been very positive, and
further improvements are expected to provide even better results.

CASE STUDY 2:
DYNAMIC PERFORMANCE
MEASURES AND CULTURE AT
DYNEGY MIDSTREAM SERVICES
The bulk of the initial research that led to the development of
the DPM approach and methodology was directed toward auto-
mation and information technologies, accounting systems, and
quality improvement. As DPM and performance improvement
programs were executed at various manufacturing plants, it began
to become evident that it was necessary to determine and imple-
ment the correct DPMs in an operation and build the DPM dash-
boards for real-time operations feedback. However, it also became

188 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 189 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

apparent that there were not enough components for an effective


continuous performance improvement process. Once the opera-
tions management teams agreed on the appropriate DPMs and
the DPMs were implemented, more often than not they were not
utilized as effectively as anticipated. The success or failure of the
DPM approach was directly tied to the culture of the manufactur-
ing organizations. Generally, DPMs were most effective in team-
oriented, empowered, accountability-based cultures.

The management team at Dynegy Midstream Services clearly


understood these cultural issues and drove a company wide trans-
formation that was designed to enable continuous economic per-
formance. Dynegy Midstream Services is a leading-edge natural
gas liquids subsidiary of Dynegy Inc., a $29 billion global provider
of energy and communication solutions. The management of
Dynegy Midstream Services strongly believes that its employees
are one of the primary critical factors for Dynegys success. This
belief is so strong that Dynegys corporate value has been defined
as We Believe in People. Dynegys success in its key markets is
evidence that its managements belief in employees is very effec-
tive. Through this focus, Dynegy Midstream Services manage-
ment team clearly views automation and information technology
as the keys to enabling its people to improve Dynegys perfor-
mance: As Tom Fiske notes, Dynegy Midstream Services is on the
leading edge when it comes to integrating technology, informa-
tion and corporate culture.1

The market for natural gas liquids had been undergoing tre-
mendous change in the years before the launch of Dynegys DPM
program in 1999. Previously, natural gas liquids had been an inte-
grated component of the business of the major oil companies.
When many of these companies decided to exit this market, a
number of new competitors emerged, among them Dynegy.
Rather than the traditional service focus of the oil companies,
these new businesses have a very strong focus on profitability.
This led to rapid consolidation and the rationalization of asset
portfolios. Basically, the natural gas liquids market transitioned
from a side business of giant companies into the mainstream
business of smaller, very aggressive companies. Survival in this

B O T T O M - L I N E A U T O M A T I O N 189
BottomLine05.book Page 190 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

segment requires operational excellence and unsurpassed eco-


nomic performance.

Stephen Furbacher, Dynegy Midstream Services president and


COO, together with his team became interested in the concept of
real-time performance measurement systems because they were so
well aligned with the direction in which Furbacher was driving
the business. Dynegy had a clear, well-communicated strategy
based on the following goals:

being the low-cost provider in its market segment,


optimizing its asset base to extract value across the liq-
uids value chain, and
enabling an entrepreneurial culture.

Dynegy presented an extremely interesting case in the utiliza-


tion of DPMs primarily because of its third strategic objective,
enabling an entrepreneurial culture. While conducting Voll-
mann decomposition analyses in several manufacturing opera-
tions, The Foxboro Companys Performance Measurement Teams
had encountered many corporate strategic objectives. However,
they had seldom encountered one as strong as Dynegys cultural
objective. On several occasions, business executives of other com-
panies manufacturing operations had told Performance Measure-
ment Teams that they felt their internal culture might hinder the
success of a proposed DPM program. In contrast, when Dynegy
Midstream Services DPM program was initiated the firm had
already recognized the importance of corporate culture to its stra-
tegic success.

Furbacher and his team had started a very aggressive program


to drive the companys business to a new operational and cultural
model. His management team agreed that small incremental
changes would not be sufficient; a radical change in business par-
adigm was required. The management team decided to move
from the firms traditional hierarchical, command-and-control
management structure to a flat, team-based, empowered organiza-
tion. As Fiske comments, The companys restructuring effort
included a transition from a six-strata hierarchical management
structure to a much flatter two-layer self-managed organization.
In what they describe as a benevolent dictatorship, the lower

190 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 191 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

worker management layer is free to determine how to achieve


the goals and vision of the organization, yet it is still the top layer
that defines these goals and vision.2

The functional organization transitioned to a performance-ori-


ented organization. Performance-based compensation plans were
implemented throughout the operation: The management pro-
vides, but also fully endorses, encourages, and empowers workers
to use information tools to fulfill their goals. The information
tools not only provide workers with valuable information to per-
form their jobs, but also help to form a culture of collaboration
and cooperation among individuals and groups.3 It is easy to see
why a metric-based DPM approach that uses automation and
information technologies to deliver empowering, real-time infor-
mation to employees was in tight alignment with the Dynegy
management directives.

One principle that was very important to the Dynegy manage-


ment team in moving to the new entrepreneurial culture was
that this new culture required their operations to adopt a very dif-
ferent management approach. Dynegys management team did
not put the onus of driving the entrepreneurial culture on its
workers; they put it on themselves. Team members determined
that they had to invest in new skills for their employees and
develop the support systems that make possible effective utiliza-
tion of the new skill set. They created accountability matrices and
allowed their employees to manage the business within their
domains of accountability. They also encouraged employees to
identify and focus on the big rocks that would generate the
most value for the corporation. Furthermore, they tied individual
incentives to the metrics of each individuals domain of authority
and responsibility, and they encouraged a multi-skill human
development program to broaden the perspectives and skills of
their employees. The management team recognized that it
needed effective real-time information so employees could con-
tinually discern their performance and improvement opportuni-
ties as they performed their day-to-day activities. Finally, the team
provided the front-line employees with the skills and tools to
manage their segment of Dynegys business and then entrusted

B O T T O M - L I N E A U T O M A T I O N 191
BottomLine05.book Page 192 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

them to do the job. As Fiske notes, Top management at Dynegy


fully endorses a culture that not only empowers but also expects
employees to respond to situations using the new information to
increase productivity.4

Dynegy worked with Foxboro to implement DPMs in critical


areas of its operations. As with every DPM analysis, the Perfor-
mance Measurement Team conducted a series of top-down inter-
views: Determining DPMs top down guarantees that they meet
the strategic objectives of the company.5 The results were very
positive. Dynegy found that the availability of real-time informa-
tion and its employees flawless execution were the critical success
factors. DPMs effect nearly everyone at Dynegy Midstream Ser-
vices directly or indirectly, Fiske comments. DPMs provide Dyn-
egy with a condensed set of accurate, reliable, contextual and
visual targets with associated actionable items that affect perfor-
mance and profitability targets. Most importantly, DPMs provide
critical and timely information to maintenance workers, opera-
tors, managers and top executives that allow them to make more
knowledgeable and faster decisions.6 Dynegy has effectively uti-
lized technology to enable and reinforce its cultural change. It
runs its plants as businesses with team-based structures that
extend right down to front-line plant operations. All personnel in
the Dynegy Midstream Services plants are aligned with the busi-
ness plan and are totally focused on executing it. DPMs help to
facilitate this alignment.

Dynegy Midstream Services has made significant progress in


moving its culture and operational performance forward, and the
DPM approach has been an important ingredient in their success.
In the first year of this program, they reduced operating expenses
and G&A by $58 million, reduced maintenance capital spending
by $6-8 million, and, on average, awarded $4,800 in incentive-
plan bonus payments to their hourly operations and mainte-
nance personnel. Since that first year, the savings have continued
to grow, and Dynegys safety record has improved dramatically.
The result has been continuing annual incentive bonus payments
to hourly employees.

192 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 193 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

Dynegys management team is quick to point out that these


improvements are not the result of any single initiative. Rather,
they come from a comprehensive program that combines tech-
nology and people in a new team-oriented culture in which the
teams perform many of the functions traditionally reserved for
management, such as doing performance reviews and setting
compensation levels. The DPMs are a part of the total program
that ensures that the teams are receiving reliable, actionable infor-
mation quickly enough to enable them to accomplish their activ-
ities in the most effective and efficient way. As Fiske confirms,
Dynegy Midstream Services is experiencing significant savings
from improved operational performance by actively promoting a
corporate culture that empowers workers to use DPMs to perform
tasks that directly influence profitability.7

As Dynegy has progressed through this process it has learned


several very important lessons that can help others embarking on
a similar performance-improvement program. First, the cultural
and people aspects of success were bigger than they had initially
believed. Getting the entire organization to align itself with cor-
porate strategy, vision, and value is critical to success. Real-time
performance information, such as DPMs, enables the empowered
teams to execute the strategy and reach established performance
goals. This is a good example of how technology should be
deployed as an enabler, not as an end unto itself. Using DPMs,
Dynegy has successfully coupled technology and information to
people, which is a crucial element in extracting maximum value
from their investments. By actively pursuing an atmosphere that
integrates technology, information, and culture, Dynegy is experi-
encing significant operational performance enhancements.8

LESSONS LEARNED
Several very important lessons can be drawn from the case
studies we have discussed as well as from the many other DPM
implementations undertaken in other process manufacturing
facilities. It may be useful to summarize some of these issues and
discuss their implications.

B O T T O M - L I N E A U T O M A T I O N 193
BottomLine05.book Page 194 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

One interesting phenomenon occurred at more than one site


in which DPMs were implemented, including one of the sites cov-
ered in our case studies. This was that even though the users
clearly identified sustainable economic value and the operational
use of the DPMs, after some time the plants operations teams
turned off the dashboard displays. This behavior was unexpected,
and evaluations were conducted to identify the reasons. In more
than one case, the plant team indicated that the economic value
improvement, which was over $1 million, was too small for them
to spend the time and effort required to manage it. There were
more important issues in the plants, namely, major capital
projects in other parts of the plants that required both the plant
engineering and operation teams attention. Although this expla-
nation was determined to be legitimate, the plants personnel
agreed that they still needed to man the operating station for the
plant sections that had DPMs. They also agreed that reviewing the
DPM displays did not consume significant resources.

After further discussion another important cultural issue was


identified. Plant personnel related that management tended to
reward improvements in economic performance, but not efforts
to sustain those improvements. So a few months after the
improvements were realized, the plants new level of economic
performance was used as the new performance baseline, and the
operations teams efforts were taken for granted. This is a major
cultural issue. Achieving an improvement in performance war-
rants recognition, but sustaining improvement warrants equal
recognition. The Dynegy Midstream Services management team
recognized the importance of a change in culture that supported
the effective utilization of DPMs. Their awareness represents one
example of how a culture change can help.

Another issue that arose in several instances was an inclination


by plant engineering teams to build dashboard displays that
looked exactly like DPM dashboard displays. They then would
populate them with a measure they could derive that had an asso-
ciated economic value. This practice was very counterproductive
in most cases. Not only did the measures identified not support
the plants manufacturing strategy; they encouraged behaviors

194 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 195 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

that actually reduced the economic performance of the plant.


Plant operators, and most employees, will typically strive to do
their best for their organization. If they are provided with a set of
measures that define what good is, they will try to perform to
them. Driving performance according to the incorrect measures
can lead to diminished economic performance.

The companies whose operations performed the best had


devised a reward system that was integrated with the performance
measures. Having a set of interesting displays is nice, but if com-
panies want their employees to drive performance the best way to
communicate this is with a reward system. Stronger rewards
imply a higher emphasis placed on improvement and typically
correlate with greater performance improvements. It is very
important, however, for the management team to really consider
what behaviors it wants and to structure the rewards to match
those behaviors. For example, if only the improvements in perfor-
mance are rewarded and not the sustaining of those improve-
ments, unusual behaviors can result.

Finally, when the company had put performance measure-


ment systems in place and made them operational and when it
had identified the improvements, the plant operating teams often
went to the firms financial department to validate the economic
value of the improvement. This proved to be a long and tedious
task because todays cost accounting systems do not make it easy
to isolate and identify economic improvements that occur within
an area or process unit of a plant. Most plants at which the DPM
projects were implemented used an enterprise resource planning
(ERP) system as the basis for their financial reporting. The ERP
systems did not provide the level of information necessary to con-
firm or deny the economic value of the improvements. This is a
very key issue. If the financial department does not recognize that
an economic improvement has occurred, in many companies the
improvement is assumed not to exist. In their book Cost and
Effect, Dr. Robert Kaplan and Dr. Robin Cooper address the short-
comings of todays cost accounting systems especially with
respect to manufacturing performance. They propose a new
model for cost accounting that combines ERP with activity-based

B O T T O M - L I N E A U T O M A T I O N 195
BottomLine05.book Page 196 Friday, August 19, 2005 9:54 AM

C H A P T E R 7 : C A S E S T U D I E S

management (ABM) systems as well as with operational and stra-


tegic performance measurement systems. Implementing the
approach they recommend would tie the DPMs into a compre-
hensive accounting structure that will make operational decision
support much more effective and align the firms financial report-
ing with the actual measures of plant-level performance. This will
lead to much more effective decision-making and economic man-
agement for manufacturing operations.

The advent of new approaches to performance measurement


such as DPMs are steps along a new path leading to improved
manufacturing performance. Many other issues need to be
addressed if manufacturing operations are to gain the full positive
impact of these approaches. Changes in corporate cultures and in
cost-accounting approaches are supportive of a total perfor-
mance-based organization. Implementing these changes will not
be easy. Those manufacturers that take these new approaches very
seriously and drive their manufacturing operations to new levels
of economic performance will be the leaders in the new manufac-
turing economy.

NOTES
1. Fiske, Tom, Improving Operational Performance by Inte-
grating Technology, Information, and Culture, ARC Insights,
March 14, 2001, page 1. ARC Insights is published by Auto-
mation Research Corporation of Dedham, MA.

2. Ibid, page 3.
3. Ibid, page 4.
4. Ibid, page 4.
5. Ibid, page 4.
6. Ibid, page 4.
7. Ibid, page 1.
8. Ibid, page 4.

196 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 197 Friday, August 19, 2005 9:54 AM

Bibliography
A Special Report on Factory Automation, Automation's Global
Game, IndustryWeek, June 20, 1988.

A Special Report on Factory Automation, Linking the Pieces of


CIM, IndustryWeek, March 23, 1987.

A Special Report on Factory Automation, The U.S. and Quality: A


New Culture, IndustryWeek, April 17, 1989.

A State-of-the-Art Report, Factories of the Future, IndustryWeek,


March 21, 1988.

Abel, Janice, and Peter Martin, Automation, Business, and Oper-


ating Advances Align into New Paradigm for Economic Perfor-
mance Improvement, Pharmaceutical Engineering, January/
February 2000.

Allen, L., and D. Stovicek, CIM Goes to School, Automation,


May 1990.

Anderson, Norman A., Instrumentation for Process Measurement and


Control, Radnor, PA: Chilton, 1980.

Andrews, W G., and H.B. Williams, Applied Instrumentation in the


Process Industries, Houston, TX: Gulf Publishing, 1979.

Babb, Michael, Process Control Systems in the 1990s Will Enter a


'New Age', Control Engineering, January 1990.

Badavas, Dr. Paul C., and Dr. Albert D. Epperly, Statistical Process
Control Integrated with Distributed Control Systems, 1988 NPRA
Computer Conference, Pittsburgh, PA, October 1988.

Badavas, Dr. Paul C., and Dr. Albert D. Epperly, Statistical Process
Control Integrated with Distributed Control in an Oven Industrial Sys-
tem, white paper, The Foxboro Company, 1988.

B O T T O M - L I N E A U T O M A T I O N 197
BottomLine05.book Page 198 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Baer, Tony, The Software Side of Quality, Managing Automation,


May 1990.

Balch, Burton, and Ted Miller, Statistical Process Control in Food


Processing, Intech, July 1990.

Beaverstock, Malcolm, and Lorne Byzyna, Measuring Perfor-


mance Dynamically, Control Engineering, June 1990, volume 11.

Behar, M. E., History in the Making, Instruments, March 1948.

Behar, M. E., On Its Twentieth Anniversary Instruments Honors


Instruments, Instruments, vol. 21, 1948.

Benassi, Frank, The Long Road to Quality, Managing Automation,


May 1990.

Berliner, Callie, and James Brimson, Cost Management for Todays


Advanced Manufacturing, Boston, MA: Harvard Business School
Press, 1988.

Bhote, Kevi R., America's Quality Health Diagnosis: Strong Heart,


Week Head. Management Review, May 1989.

Bowen, Earl K., Statistics with Applications in Management and Eco-


nomics, Homewood, II.: Richard D. Irwin, 1960.

Box, George E. P., William G. Hunter, and J. Stuart Hunter, Statis-


tics for Experimenters: An Introduction to Design, Data Analysis, and
Model Building, New York: John Wiley & Sons, 1978.

Brimson, James A., How Advanced Manufacturing Technologies


Are Reshaping Cost Management, Management Accounting,
March 1986.

Brown, Mark E., Kevin Parker, and Cory P. Senyard Jr., Computer
Integrated Manufacturing in the Chemical Process Industries,
Chemical Processing, October 1989.

198 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 199 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Bunnell, John, and Richard L. McAllister, Continuous Improve-


ment through Plantwide Integration, Control Engineering, June
1990, volume 2.

Bylinsky, Gene, Challengers Move in on ERP, Fortune, Novem-


ber 22, 1999.

Cargill, Carl F., Information Technology Standardization, Maynard,


MA: Digital Press, Digital Equipment Corporation, 1989.

Casti, John L., and Robert E. Larson, Principles of Dynamic Program-


ming - Part I-Basic Analytical and Computational Methods, New
York: Marcel Dekker, 1978.

Clark, Kim B., and Robert H. Hayes, Why Some Factories Are
More Productive Than Others, Harvard Business Review, Septem-
ber-October 1986.

Cooper, Robin, and Robert S. Kaplan, Measure Costs Right: Make


the Right Decisions, Harvard Business Review, September-October
1988.

Copi, Irving M., Symbolic Logic, Toronto: MacMillan, 1970.

Cox, Brad J., Object Oriented Programming, An Evolutionary


Approach, Reading, MA: Addison-Wesley, 1986.

Crosby, Philip B., Quality Is Free: The Art of Making Quality Certain,
New York: McGraw-Hill, 1979.

DeMarco, Tom, Structured Analysis and System Specification, New


York: Yourdon, 1978.

Deming, W. Edwards, Out of the Crisis, Cambridge, MA: MIT, Cen-


ter for Advanced Engineering Study, 1986.

Dixon, J. R., A. J. Nanni, and T. E. Vollmann, The New Performance


Challenge: Measuring Operations for World Class Competition, Busi-
ness One-Irwin, 1990.

B O T T O M - L I N E A U T O M A T I O N 199
BottomLine05.book Page 200 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Dobyns, Lloyd, Ed Deming Wants Big Changes, and He Wants


Them Fast, Smithsonian, August, 1988.

Drucker, Peter F., The Emerging Theory of Manufacturing, Har-


vard Business Review, May-June 1990.

Drucker, Peter F., Innovation and Entrepreneurship: Practices and


Principles, New York: Harper & Row, 1985.

Drucker, Peter F., The Practice of Management, New York: Harper &
Row, 1954.

Drucker, Peter F., Robert G. Eccles, Joseph A. Ness, Thomas G.


Cucuzza, Robert Simons, Antonio Davilla, Christopher Meyer,
Robert S. Kaplan, and Davis P. Norton, Harvard Business Review on
Measuring Corporate Performance, Boston: Harvard Business School
Publishing, 1998.

Elwart, Steven P., and Peter G. Martin, New Software Structures


Extend Control Capabilities, Control Engineering, volume 2, June
1990.

Epperly, A. D., and P. C. Badavas, Statistical Quality Control, Sta-


tistical Process Control: Part 1, Quality Improvement; Part 2, Par-
ticipative Management; Part 3, Statistical Tools; Part 4, Problem
Solving, State-of-the-Art Newsletter, Sec. 3 Nos. 5-8, Foxboro, MA:
The Foxboro Company, 1987.

Ferguson, Marylyn, The Aquarian Conspiracy, Los Angeles: J. P.


Tarcher, 1980.

Fiske, Tom, Improving Operational Performance by Integrating


Technology, Information, and Culture, ARC Insights, Automation
Research Corp., Dedham, MA, March 14, 2001.

Franson, Deborah L., A Roundtable Discussion on Automation


Systems on the 1990's, Control Engineering, volume 2, June 1990.

200 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 201 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Gagne, James, Quality Performance Means More at Dow, Mid-


land, MI: Dow Chemical U.S.A.

Ghosh, Asish, and Howard P. Rosenof, Batch Process Automation:


Theory and Practice, New York: Van Nostrand Reinhold, 1987.

Goldstein, Mark; William Pat Patterson, and John Teresko, Link-


ing the Pieces of CIM, IndustryWeek, March 23, 1987.

Gordon, Maureen, E., Manufacturing Automation Protocol


(MAP), State-of-the Art Newsletter, Foxboro, MA: The Foxboro
Company, January 1987.

Gould, Lawrence, The CIM Task Force: Purchasing CIM Is Also a


Distributed, Hierarchical, Pursuit, Managing Automation, Decem-
ber 1990.

Grant, Eugene L., and Richard S. Leavenworth, Statistical Quality


Control, 5th ed., New York: McGraw-Hill, 1980.

Harbor Research, Monthly Outlook, CIM Failures, Outlook, Har-


bor Research Corp., July 1987.

Harbor Research, Monthly Outlook, CIM Failures, II, Outlook,


Boston, MA: Harbor Research Corp., August 1987.

Harbor Research, Performance Measurement: Impact on Com-


petitive Performance, Outlook, Boston, MA: Harbor Research
Corp., volume 6, number 4, 1990.

Harrold, Dave, Enterprise Integration Requires Understanding of


the Plant Floor, Control Engineering, February 2000.

Harry, Mikel J., Ph.D., The Vision of Six Sigma: A Roadmap for Break-
through, Tempe, Arizona: Six Sigma Academy, 1994.

Hays, William L., Statistics for the Social Sciences, New York: Holt,
Rinehart and Winston, 1973.

B O T T O M - L I N E A U T O M A T I O N 201
BottomLine05.book Page 202 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Henderson, Bruce A., and Jorge L. Larco, Lean Transformation,


Richmond, Virginia: The Oaklea Press, 1999.

Hickey, J., Demands & Requirements Companies Will Face in the


1990's, Presentation, May 1990.

Hidden, A. E., and E. Lowe, Computer Control in the Process Indus-


tries, London: Peter Peregrinus Ltd., 1971.

Hillier, Frederick S., and Gerald J. Lieberman, Operations Research,


San Francisco: Holden-Day, 1974.

Hiromoto, Toshiro, Another Hidden Edge-Japanese Management


Accounting, Harvard Business Review, July-August 1988.

Horn, George, Japan: Should We Copy to Compete? Automa-


tion, May 1990.

IBC Technical Services, Computer Integrated Manufacture in the Pro-


cess Industries-Conference, London: IBC Technical Services Ltd.,
1990.

Jasany, L.C., Knowledge (and Power) to the People, Automation,


July 1990.

Johnson, H. Thomas, and Robert S. Kaplan, The Rise and Fall of


Management Accounting, Management Accounting, January 1987.

Jones, Daniel T., Daniel Roos, and James P. Womack, The Machine
That Changed the World: The Story of Lean Production, New York,
Harper Perennial (Harper-Collins), 1991.

Juran, Dr. Joseph, Juran on Leadership for Quality - An Executive


Handbook, New York: Free Press, 1989.

Juran, Dr. Joseph, Juran on Quality Leadership, videotape, Juran


Institute, 1988.

202 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 203 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Juran, J. M., Managerial Breakthrough: A New Concept of the Man-


ager's Job. New York: McGraw-Hill, 1964.

Kaplan, Robert S., Accounting Lag: The Obsolescence of Cost


Accounting Systems, Harvard Business School 75thAnniversary Col-
loquium on Productivity and Technology, Cambridge, MA: Harvard
Business School, March 1984.

Kaplan, Robert S., One Cost System Isn't Enough, Harvard Busi-
ness Review, January-February 1988.

Kaplan, Robert S., Yesterday's Accounting Undermines Produc-


tion, Harvard Business Review, July-August 1984.

Kaplan, Robert S., and Robin Cooper, Cost & Effect, Using Inte-
grated Cost Systems to Drive Profitability and Performance, Boston:
Harvard Business School Press, 1998.

Kaplan, Robert S., and David P. Norton, Translating Strategy into


Action: The Balanced Scorecard, Boston: Harvard Business School
Press, 1996.

King, Bob, Better Designs in Half the Time: Implementing QFD-Qual-


ity Function Deployment in America, Lawrence, MA: Goal/QPC,
1987.

Kompass, E. J., The Road to Plantwide Information, Control


Engineering, June 1990, volume 2.

Kuhns, William, Lewis Mumford, Siegfried Giedion, Jaques Ellul,


Norbert Wiener, R. Buckminster Fuller; Marshall McLuhan,
Harold Adams Innis, The Post-Industrial Prophets Interpretations of
Technology, New York: Harper & Row, 1973.

Lareau, Albert F., Bringing Information to Automation, Control


Engineering, June 1990, volume 2.

MacGregor, John F., On-Line Statistical Process Control, Chemi-


cal Engineering Progress, October 1988.

B O T T O M - L I N E A U T O M A T I O N 203
BottomLine05.book Page 204 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Mandl, Vladimir J., Teaming Up for Performance, Manufacturing


System, June 1990.

Martin, Peter G., A New Blueprint for Automation, Chemical


Plants and Processing, September 1989.

Martin, Peter G., A Software Structure for Open Industrial Sys-


tems, Proceedings of the Fourteenth Annual Advanced Control Con-
ference, Purdue University and Control Engineering, September
1989.

Martin, Peter G., Automation Payback, Industrial Computing,


Research Triangle Park, North Carolina: ISAThe Instrumenta-
tion, Systems, and Automation Society, August 1999.

Martin, Peter G., Computer Control of Batch Processes, Mea-


surement and Control, issue 106, September 1984.

Martin, Peter G., Designing a Batch Control System, I&CS,


October 1987.

Martin, Peter G., Dynamic Performance Management: The Path to


World Class Manufacturing, New York: Van Nostrand Reinhold,
1993.

Martin, Peter G., IEEE 802 Standards, State-of-the-Art Newsletter,


Foxboro, MA: The Foxboro Company, July 1986.

Martin, Peter G., Offene Industriesysteme fur die Verfarhren-


stechnik-Schritt in ein neues Automatisierungszeitalter, Verfahr-
enstechnik, Nr. 10, 1989.

Martin, Peter G., OIS: A Blueprint for Automation, Control Engi-


neering, February 1989.

Martin, Peter G., Open Communications in the Process Indus-


tries, paper presented at ISAThe Instrumentation, Systems, and
Automation Society, Houston Channel Section, April 1990.

204 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 205 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Martin, Peter G., Open Industrial Systems, Measurement and


Control, October 1989.

Martin, Peter G., Reference Model of Open System Interconnec-


tion, State-of-the Art Newsletter, Foxboro, MA: The Foxboro Com-
pany, November 1985.

Martin, Peter G., The Move to Open Communications in the


Process Industries, paper presented at ISAThe Instrumentation,
Systems, and Automation Society, Calgary Section, April 1990.

Martin, Peter G., The Move toward Open Industrial Systems,


Proceedings of the Advanced Control Symposium, Texas A&M Univer-
sity, February 1989.

Martin, Peter G., The Move toward Standard Operating Sys-


tems, State-of-the-Art Newsletter, Foxboro, MA: The Foxboro Com-
pany, January 1986.

Maskell, Brian H., Performance Measurement for World Class


Manufacturing, Part I, Manufacturing Systems, July 1989.

Maskell, Brian H., Performance Measurement for World Class


Manufacturing, Part II, Manufacturing Systems, August 1989.

McClenahen, John S., Automation's Global Game, Industry-


Week, June 20, 1988.

McLuhan, Marshall, The Mechanical Bride: Folklore of Industrial


Man. Boston, MA: Beacon Press, 1952.

McLuhan, Marshall, and Quentin Fiore, The Medium Is the Mes-


sage: An Inventory of Effects, New York: Bantam Books, 1967.

McSween, Terry, and Victor Zaloom, Jr., Creating a Positive Work


Environment, Chemical Engineering, June 1990.

B O T T O M - L I N E A U T O M A T I O N 205
BottomLine05.book Page 206 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Miller, Jeffery G., Alfred J. Nianni, and Thomas E. Vollmann,


What Shall We Account For? Management Accounting, January
1988.

Miller, Jeffery G., and Thomas E. Vollmann, The Hidden Fac-


tory, Harvard Business Review, September-October 1985.

Murrill, Paul W., Fundamentals of Process Control Theory, Research


Triangle Park, NC: ISAThe Instrumentation, Systems, and Auto-
mation Society, 1981.

Naisbitt, John, Megatrends: Ten New Directions Transforming Our


Lives, New York: Warner Books, 1982.

Passino, Ralph, Computers Spur Profitability at General Chemi-


cal, Chemical Processing, November 1990.

Pastor, E., Networks: The Backbone of Integration, Automation,


May 1990.

Peat, F. David, Artificial Intelligence: How Machines Think, New


York: Simon & Schuster, 1985.

Peters, Thomas J., and Robert H. Waterman Jr., In Search of Excel-


lence: Lessons from America's Best-Run Companies, New York:
Harper & Row, 1979.

Quak, Al, Batch Control Systems, State-of-the-Art Newsletter, Fox-


boro, MA: The Foxboro Company, May 1986.

Qualtec, Inc., Managing Quality Improvement, Participant Workbook,


Miami, FL: Florida Power & Light Co., 1988.

Qualtec, Inc., Team Leader Training Course, Participant Workbook,


Miami, FL: Florida Power & Light Co., 1987.

Rheingold, Howard, Tools for Thought.- The History and Future of


Mind-Expanding Technology - The People and Ideas Behind the Next
Computer Revolution, New York: Simon & Schuster, 1985.

206 B O T T O M - L I N E A U T O M A T I O N
BottomLine05.book Page 207 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Rohan, Thomas M., Factories of the Future, IndustryWeek,


March 21, 1988.

Rossi, Dennis A., Consider Integrated Solution for Total Quality,


Intech, January 1990.

Schonberger, Richard J., Customer Chains: Links to Survival,


IndustryWeek, April 16, 1990.

Semler, Ricardo, Managing without Managers, Harvard Business


Review, September-October 1989.

Sentell, Gerald D., Satisfying Customers: Quality, Culture, Para-


digms and Management, Alcoa, TN: Tennessee Associates.

Shaw, William T., Computer Control of Batch Processes, Cock-


eysville, MD: EMC Controls, 1982.

Shewart, W. A., Economic Control of Quality of Manufactured Prod-


uct, New York: Van Nostrand Company, 1931.

Shinskey, F G., Process-Control Systems, New York: McGraw-Hill,


1979.

Skinner, Wickham, The Productivity Paradox, Harvard Business


Review, July-August 1986.

Southard, Robert K., Local Area Networks - An Overview, Part I,


Manufacturing Systems, June 1990.

Special Edition, Manufacturing Automation Protocol, Control


Engineering, October 1985.

Special Factory Automation Edition, I&CS, March 1986.

Special 1991 Bonus Issue, The Quality Imperative, What It Takes


to Win in the Global Economy, BusinessWeek, October 25, 1991.

B O T T O M - L I N E A U T O M A T I O N 207
BottomLine05.book Page 208 Friday, August 19, 2005 9:54 AM

B I B L I O G R A P H Y

Spiegel, Murray R., Probability and Statistics, New York: McGraw-


Hill, Schaum's Outline Series, 1975.

Spiegel, Murray R., Statistics, New York: McGraw-Hill, Schaum's


Outline Series, 1961.

Stalk, George, Jr., Time-The Next Source of Competitive Advan-


tage, Harvard Business Review, July-August 1988.

Sweeny, Allen, Accounting Fundamentals for Nonfinancial Executives


and Managers, New York: McGraw-Hill, 1972.

The Foxboro Company, Introduction to Process Control, Foxboro,


MA: 1986.

The MAP Process Industries Initiative Working Group, AW in the


Process Industry, Dearborn, Michigan, MAP/TOP Users Group of
the Society of Manufacturing Engineers.

Walsh, Susan, Perspectives on the Age of the Smart Machine,


Managing Automation, December 1989.

Waterman, Robert H., Adhocracy.- The Power to Change: How to


Make Innovation a Way of Life. Knoxville, TN: Whittle Direct
Books, 1990.

Watts, Patti, Section Editor, Special Section: In Search of Qual-


ity, Management Review, May 1989.

Williams, Theodore J., A Reference Model for Computer Integrated


Manufacturing (CIM): A Description from the Viewpoint of Industrial
Automation, Research Triangle Park, NC: ISAThe Instrumenta-
tion, Systems, and Automation Society, 1989.

Zuboff, Shoshana, In the Age of the Smart Machine, New York: Basic
Books, 1988.

208 B O T T O M - L I N E A U T O M A T I O N
209
Index Term Links
ABC 12 111115 117
action planning 98 100
activity-based costing (ABC) 12 111115 117
asset availability 139141
asset management 141
asset utilization 139141
benefits, lifecycle 4647 49
Bujaria, Dr. Hans 64
business value analysis 165
CIM 7 2729 122
CMS 12 82 9192
computer integrated manufacturing (CIM) 7 2729 122
continuous flow production 74 76 129
cost accounting 1112 46 8186 88 104 107112
114115 126127 195
cost management systems (CMS) 12 82 9192
cost-per-unit-product-made 1113 8587 8990 93 96
costs, lifecycle 4547
craftsmanship 3537 4142
DCS 2526 29 31
DEC 5 22
decomposition, Vollmann 97 101102 105 116 127 173 177
190
Deming, W. Edwards 89 57 59
Digital Equipment Corporation (DEC) 5 22
distributed control systems (DCS) 2526 29 31
DPMs 52 102113 115117 121128 131132 135
166170 176 182 184 186189 193194
Drucker, Peter 4 69 96 103 128
dynamic performance measures (DPMs) 52 102113 115117 121128 131132 176 182
184 186189 193194
empowered teams 75 77 130 193
globalization 13 54 91 122
Henderson, Bruce A. 74
human assets 146
industrial optimization 152 154156
industrial revolution 3537
JIT 14 8990
Jones, Daniel T. 73
210
Index Term Links
Juran, Joseph 9 61
just-in-time (JIT) 14 8990
Larco, Jorge L. 74
lean manufacturing 7374 7879 129
lifecycle benefits 4647 49
lifecycle costs 4547
lights out manufacturing 27 3335 43
management information systems (MIS) 3031
mathematical optimization 150
Meyers, Christopher 125
MIS 3031
net present value (NPV) 45 48
NPV 45 48
optimizing plant assets 147 149150
business performance 157 170
PDCA 60 96
performance dashboard 78 106
performance measurement 5 7173 94 114116 125 175176 190
196
performance measures 7173 7678 86 89 94 106
115116 127 139141 174 177 179 195
performance vectors 162165
plant asset analysis 141146
project team approach 51
pursuit of perfection 75 78 131
QA 55 58
quality
for manufacturing 53 55
for technology 54 64
for the bottom line 54 68
quality assurance (QA) 55 58
quality control 15 57 6162
quality improvement 5 9 56 6165 7073 188
quality indicators 10 60 6673 75 77 125126
replacement automation 50 122
Roos, Daniel 73
scientific management 3839 6061
Shewhart, Walter A. 8
Six Sigma 10 70 130 136
211
Index Term Links
Six Sigma quality 7476 130
SPC 5354 57 63 124
SQC 810 5660 6365 124
statistical process control (SPC) 5354 57 63 124
statistical quality control (SQC) 810 5660 6365 124
strategy 14 95 98 100102 108 127 190
193
Taylor, Frederick 38
Taylorism 38 42
technology
for manufacturing 1718 2122
for technology 17 21
for the bottom line 17 43 124
The Foxboro Company 24 174
total quality management (TQM) 3 63 7071 75 77 96
125126
TQM 3 63 7071 75 77 96
125126
visual management 75 78 130
Vollmann decomposition 97 101102 105 116 127 173 177
190
Vollmann triangle 13 9597
Vollmann, Dr. Thomas 13 97 101102 105 116 127 173
177 190
Womack, James P. 73
workplace
cleanliness 75 78 131
order 75 78 131
safety 20 75 78 131
Zuboff, Shoshana 3536 42

You might also like