Estimating uncertainty in prison population projections


The situation:


Prison population projections are the end result of a number of criminal justice system interconnected models. The uncertainty around each of these models had not been estimated, but understanding the uncertainty of prison projections was essential for capacity planning.

The analytical approach:

Asking the right question:

Court demand and sentencing and forecasts were used within a prison population microsimulation model, with policy forecasts overlaid.

  • How confident are we that the prison population will not exceed a particular threshold?
  • Was there sufficient resource to estimate uncertainty using a traditional bottom-up approach (i.e. combining the uncertainty from all constituent models)?

Defining and identifying uncertainty:

  • Court demand projections
  • Sentencing assumptions (custody rate)
  • Impact estimates of future policy changes
  • Unknown impacts such as exogenous system shocks or policy changes not known at time of projection
  • Estimation error (such as sampling bias) from simulation model

Understanding and measuring uncertainty:

  • A top-down approach was chosen as the resource requirement for a traditional bottom-up approach would have been prohibitive.
  • The variance of all the previous forecast errors after 1 time period were fitted to a chosen probability distribution (in this case a normal distribution was chosen). This distribution was used to estimate the uncertainty of the new forecast after 1 time period.
  • This process was repeated for the variance at each subsequent time period.
  • The rate of change of variance was used to extrapolate the variance for periods where no forecast error existed.

Presenting and communicating uncertainty analysis

  • We presented our historical forecast errors, explaining their main drivers and how each contributes to our current uncertainty range. This helped our key customers to better understand the uncertainties inherent in the system we model, and therefore why we can’t produce more precise forecasts.
  • The other aspect was helping customers think about how they should use the uncertainty estimates. For example: for the prison estate, assessing the likelihood of prison population reaching different levels together with an assessment of risk appetite and contingency measures we could use. This allowed for an evidence based approach for how much capacity we should plan for, not simply planning to the central estimate.

Estimating changes to staff travel times following an office relocation


The situation:


Alternative government office locations were being considered and their on staff travel times.

The analytical approach:

Asking the right question:

Google maps API was used to estimate journey times from each home address to each potential location.

  • 90 minutes was considered to be a ‘reasonable’ travel time.
  • The distribution of journey times was an output of interest, but the key focus was “What proportion of staff will have a journey time of over 90 minutes?”

Defining and identifying uncertainty:

Missing Data:
  • Only first half of staff postcodes were available.
  • Not known who could feasibly travel by car.
Assumptions:
  • Staff begin and end work at 9am and 5pm.
  • Journey time assumes all transport is running to schedule.
  • Assumptions for transfer times, walking speeds etc.

Understanding and measuring uncertainty:

  • Through sensitivity analysis on each parameter, the use of partial postcodes was found to be the most significant source of uncertainty and so Dominant uncertainty was chosen.
  • A Monte Carlo approach would have been a more complete analysis but was not viable in the available time.
  • A sample of postcode districts were reviewed to calculate: - The journey time from the centre of the postcode; - The shortest possible journey from within that postcode; - The longest possible journey from within that postcode.
  • The differences in these times were analysed across the sample to form distributions.
  • Producing an uncertainty estimate of 90% of journey times to be within ± 25 minutes of the calculated durations.

Presenting and communicating uncertainty analysis

  • A cumulative distribution of journey times along with uncertainty bounds clearly presented how many journey times were estimated to be within the crucial threshold of 90 minutes.
  • Bar graph ranges were presented to the senior move managers, to show how journey times compared between location – the central estimate wasn’t plotted to avoid the reader putting too much significance on it.
  • Prose sentences were used to clearly present the key results.
  • Unquantified uncertainties were communicated e.g. the assumption of no delays.
  • The communication of the uncertainty made clear the case for obtaining exact postcodes, allowing the analysis to be repeated with less postcode uncertainty.

Predicting staff pay costs


The situation:


Each department has to manage their pay costs to ensure they stay within their budget, making decisions to recruit (or not) in the face of uncertainty over staff turnover, pay rises, etc.

The analytical approach:

Asking the right question:

Use a microsimulation to model staff members over time, with departures, promotions, and new recruits modelled stochastically.

  • Initially ‘What will our pay costs be at the end of the year for a given recruit plan?’
  • Redefined as ‘how likely is it that our costs will exceed our budget?’, allowing a risk-based approach to decision-making, and also allowing uncertainty in the budget to be incorporated.

Defining and identifying uncertainty:

Unknowns:
  • The pay budget including additional sources of funding not known until the start of the year
Assumptions:
  • Number of staff who will join or leave the department in future months
  • Variation in starting salaries, pension etc. for new staffs
  • Variation in pay awards based on staff performance

Understanding and measuring uncertainty:

  • A Monte Carlo approach was used due to the large number of sources of uncertainty, its suitability in microsimulation, the (assumed) lack of correlation between sources, and availability of time. Given that the model was likely to be reused many times, it was worth investing the time to build a full probabilistic uncertainty assessment into the model.
  • Most numerical assumptions in the model (including dates) can be entered as a distribution (exact, uniform, or triangular, along with some bespoke distributions for specific purposes).
  • This resulted in a full probability distribution of the range of possible outcomes.

Presenting and communicating uncertainty analysis

  • All major uncertainties were quantified, and presented (with any unquantified sources noted elsewhere).
  • The key output was simply the likelihood of remaining within budget, using positive and negative framing (e.g. “We have an 70% chance of staying within budget, and a 30% chance of exceeding our budget”)
  • As the budget holder was an analyst, we chose to present the full probability density profile to illustrate the range of possible over-/under-spends.
  • Alongside, simple prose descriptions of the 90% prediction intervals to describe the likely range of the over-/under-spend
  • A further graph was used to show how performance against budget was predicted to evolve over the year, using multiple line graphs to show the 90% prediction intervals.