Вы находитесь на странице: 1из 9

BENCHMARKING UK GOVERNMENT

PROCUREMENT PERFORMANCE IN
CONSTRUCTION PROJECTS
Robin Holt and Andrew Graves
Robin Holt, Management Research Officer, took his PhD in Government at LSE, and is now investigating links
between social theory, management practice and construction. Andrew Graves, MIT Professor of Automotive
Management and Director of Cross-Industry Research at the University of Bath, is responsible for research on
developing ``lean production'' in the car, aerospace and construction industries.

Abstract Benchmarking is introduced as a practice of description of a pilot study, designed to begin a


non-financial assessment that promotes continual performance benchmarking process amongst key central government
improvement. Its relevance to and possible consequences for the clients, we present a refined set of metrics that support a
public sector are discussed in relation to a case study in construction second phase of research and then discuss the findings.
procurement. A pilot study investigating the achievements of In conclusion we argue that unless good performance
government clients in construction procurement has identified a need
indicators are used in conjunction with one another, and
for better client ``ownership'' of project risk and opportunity. The
are ``owned'' by practitioners, any performance
article argues that benchmarking can provide the vehicle for this
``ownership''. In conjunction with the clients and HM Treasury, a improvement is at best sporadic and often temporary.
second stage of project assessment has just been completed and the Benchmarking as a process
methods and results described. The aim is to realize consistent, Benchmarking operates at two levels: a method for
relevant and feasible metrics, co-operatively authored by client
measurement and a focus of activity. As a method for
practitioners and academics with reference to private construction
measurement, benchmarking assesses relative
organizations, that will then be used for purposes of on-going
self-assessment at both project and strategic levels. performance (in terms of productivity, or profit, or
speed) to establish metrics that reveal good and bad
performance. The aim is to convey the existing delivery
Keywords Benchmarking, Public sector, Performance,
Risk management, United Kingdom of value as against potential delivery. Metrics only make
sense through their capture of narrowly defined aspects
Introduction of value. For example, a metric of safety is the number of
The UK Government is increasingly interested in policy reportable accidents; but the validity of this metric stems
reform that seeks to tie the realization of the public not from itself, but the quality of life it helps promote for
interest to specific performance improvements in service operatives; the benefits realized from an absence of
provision (Grout, 1997). The definition and attainment disruption and so on. Metrics must also be consistent
of these service levels is, we argue, determinable through with one another ± their aggregated portrayal of cost,
a benchmarked procedure of iterative improvement function and quality must be comprehensive so that, for
where links between policy and results can be made example, cost metrics are understood in terms of their
explicit (Coe, 1999). This article describes how the impact upon function and quality. Metrics can frustrate
measurement and consequent development of such the process of benchmarking if their definition fails to
service provision has been undertaken with specific promote genuine comparison. In the context of
regard to the procurement of construction projects. We construction, for example, different procurement
address criteria informing public sector perspectives that options incur the transfer of different risks. More
bear on the development and use of metrics and that traditional contracts require the client to retain design
potentially threaten the widespread and prolonged responsibility, whereas recent ``one stop shop'' relations
adoption and adaptation of a benchmarking process.
Informed by these criteria, we discuss the development The current issue and full text archive of this journal is available at
of our benchmarking method and its intent. From the http://www.emerald-library.com/ft

Measuring Business Excellence 5,4 2001, pp. 13-21, # MCB University Press, 1368-3047
1 3
transfer this risk to the constructor. As a consequence, transport network) (Ammons, 1999). As such, through
variations in cost must be identified at agreed points in benchmarking, the UK Government can assess and
the procurement process and be related to a particular refine what it means by the delivery of ``value for
procurement option. Under ``divided'' procurement money''.
variations are more prevalent than under ``one stop Attending to the exact scope of any metrics requires,
shop'' or ``turnkey'' arrangements and do not necessarily first, that the ``worldview'' of those undertaking
impinge on performance. Metrics are also required that benchmarking is acknowledged, in order that what is
allow for the inclusion of stakeholders. In public measured is appropriate and meaningful to practitioners.
construction projects these include: users; taxpayers; In this case, the scope of the metrics needs to be alive to
local residents; environmental and economic pressure the prevailing and emerging influences upon public
groups; planning authorities; future generations; managers within construction, not least of which is the
national and international political interests; financial growing use of performance measurement itself. These
interests; and the construction suppliers (Kouzmin et al., influences can be summarized as follows.
1998).
With regard to benchmarking being a focus for Departing from the culture of regulation
business activity, it looks at measuring the impact of The growth of procedural constraints experienced by
initiatives that seek to attain, sustain and exceed both the public managers, designed as controls that make
performance and increased managerial discretion
representational capability of more auditable, can stymie
these metrics. Benchmarking is initiative and incur considerable
not a list or even a procedure, it ``Benchmarking the direct and indirect cost in terms
is a responsive activity that of initiation, monitoring and
requires practitioners to delivery of construction compliance (Hood et al., 1998).
understand the circumstances of Benchmarking risks being seen
achievement as much as the level services can extend as yet one more mode of
of achievements. Any from the analysis of a regulation and surveillance. To
organisation is required by the overcome this problem, the
benchmarking process to ask specific process to . . . research adopted the use of
what was learnt through the ``good'' practice and not `best'
metrics, and so address their a general social vision.'' practice. Best practice is static
adaptation, rather than just their and suggests performance is
adoption (Ammons, 1999). measured as the ability to
Bestowing such a strategic role imitate. Good practice is more dynamic, inviting
upon metrics must be accompanied by a dynamic consideration by offering itself as something to be taken
interaction with practitioners, so as to invest in them a up and adapted. In this way the idea that benchmarking
progressive ``ownership'' of emerging metrics (Hinks and is primarily a sub-optimal strategy of ``catch-up'' (Cox
McNay, 1999). and Thompson, 1998; Davis, 1998) is avoided.
The public sector and construction context Benchmarking is seen as setting horizons through an
The concern of current UK Government policy is to active and on-going awareness of where others have set,
achieve value for money in the delivery of services such reached and re-configured theirs.
as construction, of which in the UK domestic market of Having identified good practice, the recipient must
£55bn (FT, 1999) the Government procures 40 percent not only recognise the benefits of using metrics but also
(Graves et al., 1998a). This recent emphasis upon be in a financial and cultural position to use them.
delivering ``value'' enhances the scope for using Public clients of different asset strength, maturity or
``evidence'' as a decision tool in public policy because ``worldview'' are often able to appreciate the good
metrics can be used to discover ``what works'' (Davies et practices of others without having the time, resources or
al., 1999), and what could work better, not just in terms context to apply them (O'Dell and Grayson, 1998). This
of price, but in terms of out-turn cost linked to function source of inertia may prove particularly powerful when
and quality. Benchmarking the delivery of construction considered in conjunction with the current status of
services can extend from the analysis of a specific process performance reporting in the public sector. Bespoke
(for example, the use and relative success of using management systems of reactive design (Kouzmin et al.,
different types of contract forms in procurement of 1998) and partial coverage of collected information
construction services) to cover a general social vision (Hyndman and Anderson, 1998) serve to dilute the
that looks to use metrics as milestones for general quality argument for the radical level of change necessary to
of life issues (for example, the creation of an effective realise the potential of benchmarking.

Measuring Business Excellence 5,4 2001


1 4
A dynamic culture and manufacturing processes (Mohamed, 1996). We
The reliance of public sector bureaucracies on explicit, have already argued that, to achieve a workable balance
codified information is also a source of frustration for of support metrics must be relevant, applicable,
would-be benchmarking practitioners. Any metrics used approachable and practical. Metrics must also embrace
will tend to orient themselves to what is most readily the structural variations of their context. This means the
captured in terms of existing data, rather than to the performance measurement capacity of benchmarking is
tacit knowledge of relationships and personal experience tied intimately with the mapping and management of
(Nonaka et al., 1998). As a process to catalyse change, risk.
benchmarking has to be carefully managed to avoid any
Research methodology
resulting undue restriction of scope. The desire to avoid
Informed by the foregoing discussion, our research
uncomfortable investigations and outcomes by limiting
embraces an iterative strategy to implement performance
the scope of investigation and reported evidence would
improvement using benchmarking within central
promote benchmarking on the basis of what was feasible
government. The process has undergone two cycles of
rather than on rational decision (Dorsch and Yasin,
development: the pilot and Stage two studies. These
1998; Leicester, 1999). One example might be the
have progressed using the following generally
tendency to look at capabilities, as opposed to satisfying
characterised methodology:
utility preferences (Sen, 1999). The measurement of
^ Client anonymity has been preserved to promote
capabilities (such as road space provision) is typically far
confidence through a non-threatening environment.
easier than utility preference (in this case, the means for
^ The project selection criteria addressed: stage of
satisfying travel need, that
completeness;
may include roads amongst a
contemporaneity; and
host of alternatives), because
value thresholds.
benchmarking is reliant upon ``. . . benchmarking has ^ Data anomalies evident
measurement, and metrics
themselves tend to to be carefully managed to in the samples for each of
the two studies were
concentrate upon explicit, avoid any resulting undue ``cleaned'' using
quantifiable data, the policy of
follow-up interviews.
using metrics itself might restriction of scope.''
privilege a capability above a Clients were asked to make
utility view. project selections without
regard to the perceived
Political influence outcome or history of projects. The widespread variation
Political realities face those implementing in performance (discussed later) was in line with UK
benchmarking. The multiple reality facing the new Audit Commission's findings of local authority
public manager (Dunleavy and Hood, 1994), where construction procurement and the experience of
efficiency and the public interest need to be reconciled, benchmarking in the private sector. This provided a
needs to be understood to pre-empt conflict in the metric of confidence that the aim for representative
design and implementation of metrics and promote their selection was at least moderately successful. Key
appropriate interpretation (Holt and Rowe, 2000). characteristics of the two mutually exclusive data
Political cycles form a persistent reality and any samples are shown in Table I.
benchmarking strategy must inform itself of prevailing
political norms and emerging shifts in power. This Pilot study
responsive nature of a civil service requires any The pilot study began with a focus group of the
medium- to long-term strategy be malleable and Government Construction Client Panel (GCCP),
under-determined. Metrics and their application should comprising representatives from 13 government clients.
not, therefore, form a dominating strategy. The group identified hypotheses that represented
procurement risks and proposed metrics to inform those
The dimension of complexity hypotheses. The approach was pragmatic: the number
Construction is a complex activity. The structural and scope of metrics were geared towards assured data
variations between projects in terms of potential capture; and the themes addressed by metrics were
influence (political, social, economic, environmental, limited principally to information of established interest
skill availability) and scope (planning, material use and to the client. The pilot study focused on project output
re-use, budget, urgency, type (road, rail, water, metrics to test the hypotheses. It was assumed that
building), life-span, flexibility) impact heavily upon the capture, analysis and presentation of the information
value of information comparison both between would stimulate dialogue and an exchange of
construction projects and between construction projects perspectives as to why particular projects might have

Measuring Business Excellence 5,4 2001


1 5
Table I Ð Comparison of pilot and stage two data samples
Data sample characteristic Pilot Stage two
Number of projects 60 65
Total value of projects £500 million £500 million
Number of client departments/agencies 6a (8b) 7a (10b)
Project type
Building 100% 52%
Infrastructure ± 48%
Project status
New build 62% 48%
Refurbishment 38% 52%
Average service commissioning date for projects June 1995 January 1998
Notes: a Clients submitting four or more projects; b
All clients submitting projects

excelled or gone awry. Informed by revealed trends and Pilot study findings and discussion
feedback from the process of data capture itself, a The evaluated metrics demonstrated considerable
process of client-informed metric refinement could be variation in performance and suggested a lack of
initiated. consistency in approach, both within and between
A key role of the pilot study was to demonstrate the clients. The two main findings were as follows.
post-project availability of data and the feasibility of
capture through a generic format. We felt that the A culture of limited expectation
delicately balanced task of building a data model capable Notions of possible ``role fuzziness'' and confusion
of population without excessive investigation yet surrounding the performance of product and service
sufficient for robust and challenging interpretations were endorsed by the juxtaposition of generally high
should not be under-stated. The need to balance levels of client and end-user satisfaction with poor cost
achievability and usefulness is an important lesson and time predictability. Given the UK Government's
drawn from benchmarking experiences in other insistence upon delivering performance improvements,
industrial sectors. the pilot study provided a steady reminder that value
assessment was so judgmental that even poor cost and
Stage two study programme performance did not meet with a rating of
The second stage of work continued to promote close unsatisfactory performance. The pilot study revealed
contact with the GCCP in developing a second an extensive reality gap between performance and
questionnaire. A joint responsibility was assumed for its performance assessment by practitioners. This gap
promotion within each client between the research team could be the result of the public managers looking
and client representatives. Stage two constituted both foremost to rules of accountability, judging their
reinforcement of the successes of the pilot and an ``success'' in terms of their having fulfilled their
opportunity to establish a maturing model. Feedback required functional role. The results describing this
from the pilot study was taken from the sampled clients as ``gap'' are shown in Figure 2.
to the validity and reliability of the metrics and the
appropriateness of the benchmarking study, given the Underestimating risks
criteria detailed in the earlier discussion. This feedback The pilot study found that clients accepted two thirds
was used, in conjunction with further focus group of realised programme risk. When compared to the
sessions of the GCCP and expert consultations with cost and programme overrun figures in Figure 2, this
construction research groups, to re-configure metrics so suggests an underestimation of risk. Figure 3 shows
as to better reveal existing and potential project risks and how in the majority of projects investigated, the risk
so increase the relevance and ``ownership'' of the was not spread through suppliers using partnering or
benchmarking practices. The risk identification and other teamworking agreements, but retained by the
control model that emerged is shown in Figure 1. It is client.
divided into four areas, each with associated metrics, The case for reviewing exposure to risk was enhanced
reflecting the extended concern of the Stage two study by the fact that one third of the projects were procured
over the pilot study. The metric count illustrates how the on the basis of ``one stop shop'' style contracts such as
focus on outputs as seen by project funders and end users Design and Build, a mechanism that is intended to
in the pilot study was supplanted by a balanced determine much of the client's risk in terms of a
measurement of risk, practice usage, and the success of all premium built into the tender price paid to suppliers
project participants/stakeholders in the Stage two study. (rather than explicit determination at the end of

Measuring Business Excellence 5,4 2001


1 6
Figure 1 Ð Linking metrics with themes under the risk-based evaluation framework

g
y
y V

y g

Figure 2 Ð Client perception of success measured against programme and cost in construction phase

Measuring Business Excellence 5,4 2001


1 7
Figure 3 Ð Use of partnering or teamworking set against cost growth in the construction phase

construction based on the actual risks encountered). Stage two study findings and discussion
This analysis lends further support to an approach to The research highlighted variations in the application of
benchmarking that is underpinned through more good practices. There remained a ``reality gap'' between
effective risk mapping to promote shared ownership of performance and performance assessment. Figure 5
risk. Where risk management did occur, and it was shows that construction cost overrun amongst the clients
limited as is as shown in Figure 4, it was not consistent remained high.
in delivering good project performance, further Despite this, and in line with the findings of the pilot
suggesting that the type of risk management systems in study, supplier satisfaction ratings made by the clients
place required review. remained consistently high, as shown in Figure 6.
Perhaps the most significant finding was that ``bolting
Towards stage two on'' specific good practices did not result in performance
Analyses were undertaken to investigate relationships improvement (see Figure 7).
between metrics, in particular between each metric and Where improvement was experienced, however, was
cost and time. This failed to establish strong trends and in those projects where clients used a ``basket'' of good
supported the notion that construction projects bear a practices; combining strategic relationship assessment
complex structural identity in which the specific with teambuilding with safety audits and so on (see
demographic, financial and political context has a large Figure 8).
influence. This general point of learning, combined with
Conclusions
the issues of unpredictable cost and time, was seminal in
Overall, the public sector client was found to have both a
moving us towards articulating the project to include
strategic and operational role in achieving performance
many more non-financial metrics, hence shifting the
improvement in value for money service delivery, but
emphasis in Stage two towards risk identification and its
that this had been underutilised. It is suggested that
control.
there needs to be greater ``ownership'' of projects from

Figure 4 Ð The use of risk management set against cost growth in the construction phase

Measuring Business Excellence 5,4 2001


1 8
Figure 5 Ð Construction cost increase from budget

Figure 6 Ð Supplier assessment of client

public sector clients and risk sharing with suppliers. that the gaining momentum of benchmarking will
Through benchmarking, public sector clients can better continue, gathering the support of new clients in the
assess the relevance of their existing aims and purposes; UK and internationally. There remains, however, a
and can appreciate how best to pursue those aims. It is suspicion that metrics amount to a regulatory audit (if
also suggested that public sector management needs to not now, then at sometime) and that those insisting
expand its scope to deliver whole life value for money. upon value for money criteria are doing so for reasons
The development of benchmarking to include of political expediency as much as a genuine desire to
whole life issues, value and the embodiment of both create an open, innovative public management
service-oriented and product-oriented approaches to culture. MBE
procurement have a long way to go. We are optimistic

Measuring Business Excellence 5,4 2001


1 9
Figure 7 Ð Plotting outcome success against the use of mature realtionships

Figure 8 Ð Plotting outcome success against a ``basket'' of good practices

References Dorsch, D. and Yasin, M. (1998), ``A framework for benchmarking in


Ammons, D. (1999), ``A proper mentality for benchmarking'', Public the public sector'', International Journal of Public Sector
Administration Review, Vol. 59 No. 2, pp. 105-14. Management, Vol. 11 No. 2/3, pp. 90-115.
Coe, C. (1999), ``Local government benchmarking: lessons from two Dunleavy, P. and Hood, C. (1994), ``From old public administration
major multigovernment efforts'', Public Administration Review, to new public management'', Public Money and Management,
Vol. 59 No. 2, pp. 110-21. Vol. 14 No. 3, July-September, pp. 9-16.
Cox, A. and Thompson, I. (1998), ``On the appropriateness of Graves, A., Rowe, D., Sheath, D. and Sykes, M. (1998a), Constructing
benchmarking'', Journal of General Management, Vol. 23 No. 3, the Government Client, HMSO, London.
pp. 1-20. Graves A., Rowe, D., Sheath, D. (1998b) Pilot Benchmarking Study,
Davies, H., Nutley, S. and Smith, P. (1999), ``Editorial'', Public Money HMSO, London.
and Management, January-March. Grout, P. (1997), ``The economics of the private finance initiative'',
Davis, P. (1998), ``The burgeoning of benchmarking in British local Oxford Review of Economic Policy, Vol. 13 No. 4, pp. 53-66.
government'', Benchmarking for Quality Management and Hinks, J. and McNay, P. (1999), ``The creation of a management by
Technology, Vol. 5 No. 4, pp. 261-70. variance tool'', Facilities, Vol. 17 No. 1/2, pp. 31-53.

Measuring Business Excellence 5,4 2001


2 0
Hood, C., James, O., Jones, G., Scott, C. and Travers, T. (1998), Leicester, G. (1999), ``The seven enemies of evidence based policy'',
``Regulation inside government: where new public management Public Money and Management, January-March, pp. 5-7.
meets the audit explosion'', Public Money and Management, Mohamed, S. (1996), ``Benchmarking and improving construction
April-June, pp. 61-8. productivity'', Benchmarking for Quality Management &
Holt, R. and Rowe, D. (2000), ``Total quality, public management and Technology, Vol. 3 No. 3, pp. 50-8.
critical leadership in civil construction projects'', International Nonaka, I., Reinmoeller, P. and Senoo, D. (1998), ``The art of
Journal of Quality & Reliability Management, Vol. 17 No. 4/5, knowledge: systems to capitalise on market knowledge'',
pp. 541-53. European Management Journal, Vol. 16 No. 4, pp. 673-84.
Hyndman, N. and Anderson, R. (1998), ``Performance information, O'Dell, C. and Grayson, C. (1998), ``If only we knew what we know:
accountability and executive agencies'', Public Money and identification and transfer of internal best practices'', California
Management, April-June, pp. 23-30. Management Review, Vol. 40 No. 3, pp. 154-74.
Kouzmin, A., LoÏffler, E., Klages, H. and Korac-Kakbadse, N. (1998), Sen, A. (1999), Commodities and Capabilities, Oxford University Press,
``Benchmarking and performance measurement in public Oxford.
sectors'', International Journal of Public Sector Management,
Vol. 12 No. 2, pp. 121-44.

Measuring Business Excellence 5,4 2001


2 1

Вам также может понравиться