Is it time for a cost management revolution? Or at least, an evolution?

Is it time for a cost management revolution? Or at least, an evolution?


I recently attended the Major Projects Association event: “The Unresolved Issue of Estimating and Forecasting”, (which was a great event) where I asked a question to the panel “whether  Cost Managers were the right people to produce cost forecasts at the early stage of infrastructure projects, or whether instead, we should utilise a more data-scientific approach instead”.


I won't provide the panel response to my question as I am keen to hear your thoughts - please feel free to leave your answer in the comments.

However, what I will do is share some of my thoughts that led me to ask the question in the first place.


Early project forecasts

The typical approach used to produce a forecast at the earliest project stage (Strategic Outline Case etc) is a “base cost + uplift” approach. The base cost is usually produced using a “should cost model” approach (in my experience, a bottom-up parametric type model), and then an uplift (often Optimism bias) is applied to mitigate the challenges associated with low project definition and high uncertainty associated with early project forecasts (see fig 1).

Fig 1. Diagram showing the typical accuracy of estimates over a project lifecycle

In my opinion, there are several issues around Optimism Bias as per HMT Greenbook. Firstly, as stated in a Greenbook, Optimism Bias it is not a financial contingency tool/application (although often regarded as one) and secondly, it is a blunt tool, (generic values) and may not be representative of your project – given the limited options available


Whereas, Reference Class Forecasting (RCF) offers a robust statistical and project-oriented methodology for calculating uplifts, employing a systematic top-down approach rooted in historical data and outcomes from analogous, previously executed projects and in my mind, should be employed where possible over HMT Optimism Bias factors, which HMT also recommends this (please feel free to reach out on all things RCF)


“5.45 Ideally adjustments should be based on an organisation’s own evidence base for historic levels of optimism bias. In the absence of robust organisation-specific estimates generic values are provided” HMT Greenbook

 

Another key factor that needs consideration when producing an early project stage forecast, is the availability and even accessibility of project-specific data (ie ground conditions, benchmark data etc) to those producing cost forecasts.


The absence of data (ie scope) presents an opportunity for those producing cost models to be influenced by their heuristics and cognitive biases, potentially undermining their capacity to produce accurate and robust forecasts. However, some may argue that a key benefit of cost professional expertise is that they can “fill in the blanks” by providing their own insight based on their experience. I do think that there is merit to the argument and therefore value in the expertise, but we also have to remember:

1, project maturity is low,

2 we are humans, and we can be overconfident in our abilities and

3, perhaps it is not their place or role to fill in the gaps (potentially baking in new/not-needed assumptions into the project)


I think the issues raised in the paragraph above are further exacerbated for linear infrastructure ie transport routes, new transmission networks, etc over a single site/location asset ie a prison/hospital etc


My hypothesis, feel free to challenge,  is that when you start to add scale both in terms of size, number of assets, multiple sites, and multiple complexities and trying to capture these elements in multiple cost models, then there is a potential for “information overload” that may “shutdowns the brain” when trying to figure out the complex independencies, the various characteristics and traits, increase level of optimism bias and then we do what we humans do best – we take mental shortcuts (heuristics) and try to bring order to the scale and complexity – and perhaps this can lead to mistakes such as the underestimation of projects


Data-led approach for early project forecasting

The “base cost + uplift model”, mentioned above is one approach. Another, as recommended by the Infrastructure Projects Authority, is the use of top-down benchmarks using data from previously completed projects, for early project-stage forecasting. This data will include any cost escalation incurred in the project, therefore providing a “de-bias” and more representative forecast and starting position.  Importantly, this approach can provide decision-makers with robust (top-down) models that can be spun up quickly to cover an array of project and scope options i.e. what the cost of an additional 50km of track, or 4 prisons instead of 3 to support and underpin robust and multiple option testing for the economic case (option appraisals), helping to reduce/mitigate the "linear project overload" issue (as impacts of independencies have been realised in the data).

The diagram below (fig 2) sets out the level of effort/input between data scientists and cost professionals at the project stage. Both roles should be utilised throughout the project lifecycle, but the effort of resources may depend on the project lifecycle/maturity. This approach also may provide tension and validation of the models between data-scientists and cost professionals, both bringing their divers expertise and insight into the knotty problem of robust early cost forecast.

Fig 2. Model depicting the effort of roles for producing forecasts across UK Government Project lifecycle


But why data scientists?

My reflection is that data scientists are better equipped to produce data-driven forecasts and equipped to navigate issues around variability, and bias. They are able to define the appropriate metrics based on the approach, and importantly the validity of the modelling and forecasting methods.

Why this is important?  The levels of uncertainty, presence of noise, and underlying biases within the data are significant at early project stages and this requires careful consideration and deliberation over the approach and the tools used (given the project maturity and understanding at the early stage - ie Montecarlo may not be appropriate).

Data Scientists are not only able to identify these potential obstacles in forecasting but they can look into quantifying what is unknown by providing confidence levels in regard to predictions and furthering insights on the deliverability and performance of the project and its range of likely outcomes (leveraging top-down and completed project data).

Lastly, through “regression analysis” you can quantify the relationships between identified cost drivers and project costs by calculating coefficients.  If you have data on potential cost drivers ie access constraints, project location (brownfield vs. greenfield), site security, procurement methodology, regional influences, and project complexity rating etc you can investigate their impact on the overall project cost. Allowing you to test “Does it make the boat go faster”, or “Does it make the project cheaper


The resulting coefficients indicate the extent of cost variation when comparing two scenarios: one with the inclusion of a specific cost driver and one identical in all aspects except for the exclusion of that driver. A positive coefficient signifies that utilising the cost driver leads to cost escalation, while a negative coefficient implies that the cost driver results in cost reduction.


I am not advocating the move away from cost professionals at the early project stage and replacing them with data scientists (a lot of my friends are cost managers). My article title talks about an “evolution” and not a revolution. I think there is huge value in having both professions working openly and collaboratively for the benefit of the project and project team (right tools for the right job – at the right stage) and in the future,  perhaps with changes in training and curriculum, we may see a convergence of these skills and expertise in a new role Cost-data-engineer, Cost-data-scientists? You read it here first…!


Lastly, I  hope that the above article sets out the value and benefits of data for organisations and their projects. In my next blog, I will set out some of the challenges, considerations, and opportunities for collecting data.


In the meantime, if you would like to know more or would like a discussion on the above, please feel free to contact me aleister.hellier@oxfordglobalprojects.com. Follow me for more exciting blogs and the second part of this blog (I know you can't wait!)


Martin Paver

Leading the transformation of data-driven project delivery | Recognised in DataIQ100 for 2 years running.

1y

I’ve been wrestling with this challenge for 6 years and I’m not personally persuaded by your hypothesis yet Aleister. There are a number of factors at play. 1️⃣ The data isn’t where it should be, so we need project professionals to be invested in improving it. 2️⃣ RCF is helpful to set an envelope, but doesn’t it need dynamic connected data at a work package level to drive day to day actions? 3️⃣ if we look forward 5 years from now, won’t the roles be combined? If not, what will happen to all the project professionals? Isn’t this a spectrum of skills rather than binary roles? 4️⃣ If a data person designs the team it will have a lot of people in their image. Likewise for traditional project roles. It’s the enlightened teams who take a holistic approach rather than squeezing more performance out of what could be ropey data. People will sit on different parts of the spectrum, just like we used to have Excel ninjas. Let’s give them the opportunity and skills to develop and shape their own destinies.

Like
Reply
Colin Warren

Consultant/Director Warren Geotechnical Associates. RoGEP Advisor. Expertise ground investigation, landslides & tunnels. Lead Geologist Eurotunnel, Technical Advisor for HS1 & HS2, Tideway & Folkestone Warren Landslide

1y

Always a believe there is a place in the estimation team for both parties - experienced cost managers and data people. Need to good information related to the reference design, ground information and proposed construction techniques/programme and of course a reasonable contractor willing to work as one team .

Like
Reply
Ian Heptinstall

Teacher & Coach in Projects and Procurement

1y

My first reaction was why is 'cost managers' v 'data-driven scientific' a choice? Then in reading it seems your experience is cost managers are only detailed estimators. So maybe this is an infrastructure issue? I would have hoped that the cost professional was already data-fuelled. And if not, is that not the issue to address, rather than bring yet another specialist to the table? The use of cost driver coefficients was a common part of the cost managers expertise back when I worked in the chemical industry. In fact most project managers and engineers were pretty fluent in then too. As John Hollmann has highlighted. Maybe if this is not the case for infrastructure projects, that offers an alternative intervention - make sure your cost managers are data-fuelled, and understand a range of different estimating methods, not just one.

David Jones FRICS FACostE IEng

Global Head of Costing at Sodexo and Costing BPO

1y

Most decent estimators will have the skills mentioned in the article, the best estimators will learn the additional skills they need.

Simon Reynolds

Projects & Programs: Infrastructure, Built Environment, Transport, Energy

1y

Aleister Hellier A revolution is needed, and it needs acceptance of complementary (but not combined) methods: - You can't do bottom up early on - the bottom is not visible. It's middle up. - Base plus OB uplift is fine. Provided you don't tinker with the base first by inserting benchmark rates and removing risk sums before adding the OB uplift. - Point estimate plus contingency is fine for the base. So is doing the whole base in one risk model. Do both ! - OB percentages are questionable. But so is most cost data gathering. - RCF is fine too. You can do it in addition to base plus uplift. - Top down using benchmarks kind of is RCFish..? Except that what gets done is high-up-in-the-middle down instead of top only. And then you have the temptation to cherry pick benchmarks, ending up with a botched multi-line estimate instead of a truly top level only single number. Benchmarking is too easily abused. - There's nothing wrong with using two or three rationally strucured methods. Hacking together bits of each is wrong though, yet it's done all the time. - The biggest difficulty faced by estimate reviewers is getting straight answers on methodology. Most estimators blend different methods, maybe unconsciously.

To view or add a comment, sign in

More articles by Aleister Hellier

Insights from the community

Others also viewed

Explore topics