Jasper Slingsby
Today we’ll complete the forecast cycle (through data assimilation) and spend a little time discussing decision support.
Recap:
The iterative ecological forecasting and the scientific method are closely aligned cycles or loops.
Bayes Theorem provides an iterative probabilistic framework that makes it easier to update predictions as new data become available, mirroring the scientific method and completing the forecasting cycle.
Today we’ll unpack this in a bit more detail.
Most modelling workflows are set up to:
(a) fit the available data and estimate parameters etc (we’ll call this analysis), which they then often use to
(b) make predictions (which if made forward through time are typically forecasts)
They usually stop there. Few workflows are set up to make forecasts repeatedly or iteratively, updating predictions as new observations are made.
When making iterative forecasts, one could just refit the model and entire dataset with the new observations added, but there are a few reasons why this may not be ideal…
Why not just refit the model?
The alternative is to assimilate the data sequentially, through forecast cycles, imputing observations a bit at a time as they’re observed.
Sequential data assimilation has several advantages:
Assimilating data sequentially is known as the sequential or operational data assimilation problem and occurs through two steps (the main components of the forecast cycle):
The two main components of the forecast cycle are the forecast step (stippled lines), where we project from the initial state at time 0 (\(t_0\)) to the next time step (\(t_{0+1}\)), and the analysis step, where we use the forecast and new observations to get an updated estimate of the current state at \(t_{0+1}\), which would be used for the next forecast to \(t_{0+2}\).
While the first step has to be analysis, because you have to fit your model before you can make your first forecast, the forecast step is probably easier to explain first.
The goals of the forecast step are to:
In short, we want to propagate uncertainty in our variable(s) of interest forward through time (and sometimes through space, depending on the goals).
There are a number of methods for propagating uncertainty into a forecast, mostly based on the same methods one would use to propagate the uncertainty through a model as discussed in the previous lecture.
Approach | Distribution | Moments |
---|---|---|
Analytical | Variable Transform | Analytical Moments (Kalman Filter) |
Taylor Series (Extended Kalman Filter) | ||
Numerical | Monte Carlo (Particle Filter) | Ensemble (Ensemble Kalman Filter) |
Explaining the different methods is beyond the scope of this module, but just a reminder that there’s a trade-off between the methods whereby:
In short, if your model isn’t too taxing, or you have access to a large computer and time to kill, SMC is probably best (and often easiest if you’re already working in Bayes)…
The two main components of the forecast cycle are the forecast step (stippled lines), where we project from the initial state at time 0 (\(t_0\)) to the next time step (\(t_{0+1}\)), and the analysis step, where we use the forecast and new observations to get an updated estimate of the current state at \(t_{0+1}\), which would be used for the next forecast to \(t_{0+2}\).
NOTE: I’m only explaining the general principles for Bayes. There are frequentist approaches, but I’m not going to go there.
This step involves using Bayes Theorem to combine our prior knowledge (our forecast) with new observations (at \(t_{0+1}\)) to generate an updated state for the next forecast (\(t_{0+2}\)).
This is essentially sequential Monte Carlo (SMC), also known as a particle filter.
The procedure of particle filtering from Kim et al. (2018). Here, values are sampled from the posterior at time \(k-1\) (= forecast for time \(k\)) to be the prior in the analysis at time \(k\). These are then weighted by the likelihood (i.e. the new data) to update the initial conditions used to generate a new posterior (forecast for time \(k+1\)), which is then resampled to provide the prior in time \(k+1\), and so on.
This is better than just using the new data as your updated state, because:
Fortunately, Bayes deals with this very nicely:
Comparison of situations where there is (A) high forecast uncertainty (the prior) and low observation error (data), versus (B) low forecast uncertainty and high observation error on the posterior probability from the analysis step. Note that the data and prior have the same means in panels A and B, but the variances differ.
Lastly, just a note that I’ve mostly dealt with single forecasts and haven’t talked about how to deal with ensemble forecasts. There are data assimilation methods to deal with them, but we don’t have time to cover them.
The methods, and how you apply them, depend on the kind of ensemble. Usually, ensembles can be divided into three kinds, but you can have mixes of all three:
Probably the hardest part of the whole ecological forecasting business… people!
It is also a huge topic. Here I just touch on a few hints and difficulties.
First and foremost, the decision at hand may not be amenable to a quantitative approach.
Secondly, there’s also the risk of external factors making the forecasts unreliable, especially if they are not controlled by the decision maker and/or their probability is unknown (e.g. fire, pandemics, etc).
One way to try to deal with external factors is by developing scenarios with different boundary conditions.
A reminder of the distinction between predictions versus projections:
You’ll be working with an organized team that is a well-oiled machine at implementing Adaptive Management and Structured Decision Making and you can naturally slot into their workflow.
The advantages of Adaptive Management and Structured Decision Making are that they are founded on the concept of iterative learning cycles, which they have in common with the ecological forecasting cycle and the scientific method.
Conceptual relationships between iterative ecological forecasting, adaptive decision-making, adaptive monitoring, and the scientific method cycles (Dietze et al. 2018).
The iterative ecological forecast cycle integrates nicely with Adaptive Management…
Focused on the process of coming to a decision, not the process of management, but very useful in the first iteration of the Adaptive Management Cycle.
Could easily be the topic of a whole course in itself, e.g. this online course by the US Fish and Wildlife Service.
It is valuable when there are many stakeholders with disparate interests.
It tries to bring all issues and values to light to be considered in a transparent framework where trade-offs can be identified and considered.
It directly addresses the social, political or cognitive biases that marginalise some values or alternatives.
But…
The beauty for the forecaster in this scenario is that a lot of the work is already done.
Can focus on estimating (forecasting) consequences and evaluating trade-offs among alternatives (steps 4 and 5), rather than having to do the whole process from scratch.
Often one has to forecast multiple state variables, which may or may not be related to each other.
Decision-makers may also have to consider trade-offs among qualitative as well as quantitative consequences under different decision scenarios.
What’s missing?
Quantify and propagate uncertainty!
“It is better to be honestly uncertain than confidently wrong.”
Sensitivity analysis:
How robust is the decision to uncertainty in the model and assumptions?
How wrong does your model have to be before the decision changes?
Communicate uncertainty in the model and forecasts clearly to decision-makers.
Be transparent about the limitations of the model and the assumptions made.
Use uncertainty visualizations to help decision-makers understand the range of possible outcomes.
Frame uncertainty in multiple ways - e.g. 5% vs 1 in 20