estarzinger
Well-Known Member
Whilst agreeing with the sentiment am confused by your terminology.
features.... low pressure systems or land?
GS track?
weather features - lows, ridges, etc
Great Circle track (sorry about that - should have been GC)
Whilst agreeing with the sentiment am confused by your terminology.
features.... low pressure systems or land?
GS track?
.............
It would be very helpful to me to have some indication of what the met office thinks the accuracy/probability is.
First, the GFS, the usual source of GRIBs is not the Met Office, it is NOAA.
well, first, I did not mention gribs or gfs. and second, I am American, and for me NOAA IS the met office.
The reliability factor, or whatever, will vary with location and length of the forecast period. That is feasible in principle but impossible in practice for too many reasons to rehearse here.
Incorrect. If we are talking about gribs . . . They could run with various perturbation in the initial conditions and in the model coefficients (a 'super ensemble') and in each grib square for each period provide some indication in the variance. They also could do exactly as you suggest below - and indicate in each grib square at each time how stable the data has been over the past several model runs. Those two approaches are both very possible/practical to do, and would provide useful information. If we are not talking about gribs but about human assisted forecasts, the forecaster could again put in a confidence indication.
My only advice, if you are wanting some indication of confidence over the next few days, is to look at D + 8, D + 7, D + ^, D + 5 on successive days. If they are consistent, there is a good chance that the next 5 days will be well predicted. If the differ greatly, the next few days will be uncertain. All, of course in general terms.
Yes, of course, and also looking several different models gives some further indication of 'confidence'.
....
.....
All that makes it possible but nigh impossible just now.
I really don't understand /agree at all that this would be 'nigh impossible'. Rather it seems technically simple, with the question being more whether NOAA thinks an analytic confidence measure is worth the incremental cost..
I really don't understand /agree at all that this would be 'nigh impossible'. Rather it seems technically simple, with the question being more whether NOAA thinks an analytic confidence measure is worth the incremental cost..............
.
...............................
One thing that hasn't been mentioned here is that the process of ingesting new data is incredibly complex; it isn't as simple as "change the value here for the one measured". You have to take account of the fact that you've got measurements at a spot locations and extrapolate them over entire grid-cells; that is not an easy thing to do. How do you weight the values you've measured at spot locations so they represent the whole grid square? It's the inverse of the reason why the forecast is right over long distances but wrong locally.
The current ensembles only run a few times (I've seen 5 or 7), with limited variations suggested by expert knowledge in the initial parameters. Getting a global confidence parameter would be extremely costly...........................
.
............The UK runs its ensembles 24 times to get a useful sample. These are “degraded,” insofar as, for example, the grid is 60 km as opposed to 25 km. .............
A couple of minor corrections.
The UK runs its ensembles 24 times to get a useful sample. These are “degraded,” insofar as, for example, the grid is 60 km as opposed to 25 km. See http://www.metoffice.gov.uk/research/modelling-systems/unified-model/weather-forecasting.
But, you are quite correct; the models do require enormous computer resources. The most powerful computers are puny compared to the atmosphere.
PS I forgot to say that they run a 24 mode; ensemble.
If you have, for example, a strong NW airstream behind a cold front, the average gradient might correspond to, say, 30 KN. Within that there will be areas and times when the wind is above or below 30 kn. The computer model cannot resolve these to better than about 130 to 150 km size.
Frank - it is great that you join in these threads as it increases all our understanding of this complex subject.
I have done some modelling and analysis in the past, and one thing that always stood out was less the centre of the error bars, but much more, what thoseresults that were outside the error bars could portend. Effectively they provided a lot of the areas of risk, for both size and frequency. Thus looking at the average forecast from the model, I am always wondering what the models have been suggesting outside the error bars.
I think I'm 100% clear on the significance of the resolution.
The bit I don't get is why the model is typically 1 force over above 15 knots of wind instead of higher or lower and throughout the wind strengths.
You even say above that the different areas can be above or below.
http://www.ybw.com/forums/archive/index.php/t-281075.html
If it was as easy as you suggest, I guess that somebody would have done so. Quick thoughts.
Option 1 assumes that the error distribution would be meaningful. I see no reason why that should be the case.
What?! If the model has predicted the exact same mslp for grid xx/yy for all the last 40 model runs, I think we can be quite confident, certiantly more so than if it was predicting very different mslp over each of the last 10 model runs. You yourself already agreed that this exact process was useful to do manually. Why would it not be useful to do globally automatically?
Option 2 overlooks my point that you would have to, say, one ensemble for W Europe, another for the E US sea board, one for the W sea board etc/
Why? Why not do a global ensemble? All we are trying to do is spot areas/features that are more sensitive to perturbations. You might be able to make the perturbations in a more smart or accurate way by area, but a global ensemble would provide useful data.
Option 3 is the most sensible but are there enough models? GFS, NOGAPS. CMC. UK Met, JMA. ECMWF. Is that a big enough sample.
Yes of course it is. Again, we already know it provides useful data when we do the comparison manually. Why would it be less so to do it automatically/globally? It certainly was useful to see the spaghetti map of hurricane sandy. BTW, you could easily incorporate regional grib models into this option.
But, none is really a starter because none would meet national needs. Who would pay for running these various options? .
Well,we can at least agree that is the key question. I personally think that this confidence data would be valuable to users. The answer to who would pay is ultimately the US defense department (and thus the US tax payer).
PS. At presemt, the UK, I cannot speak for NOAA can only run the global model with a 25 km grid. That stretches their computer to its limits. Their next target will be to get more computer power so that they can get a more detailed analysis so reducing the scope for analysis errors/uncetainties.
Yes, understood, but do realize that my approach #1 & #3 would NOT require huge amounts of computer resources. Option #2 would, except they already run ensemble analysis and calculate the mean. It would again NOT add vast resources to also calculate a weighted variance.
It wouldn't be an "incremental cost" - it would be a straight multiplier for each model run. And to make the results statistically valid (Monte-Carlo simulation style) would require an enormous number of runs, to ensure that global peturbations were taken into account.
Not for my options #1 & #3. Those do NOT require extra model runs. and my option #2 might also NOT if one can simply use the ensembles that are already being run.
One thing that hasn't been mentioned here is that the process of ingesting new data is incredibly complex; it isn't as simple as "change the value here for the one measured". You have to take account of the fact that you've got measurements at a spot locations and extrapolate them over entire grid-cells; that is not an easy thing to do. How do you weight the values you've measured at spot locations so they represent the whole grid square? It's the inverse of the reason why the forecast is right over long distances but wrong locally.
Well they have obviously already solved this problem in the ensembles they already run. And to do further perturbations you can vary directly the grid level data and not the spot location data.
we are probably a few generations of super-computers away from being able to do it.
Again, not for my options #1 & #3, perhaps or perhaps not for #2.
My last comment on the topic since we are obviously not going to agree on this
I think I'm 100% clear on the significance of the resolution.
The bit I don't get is why the model is typically 1 force over above 15 knots of wind instead of higher or lower and throughout the wind strengths.
You even say above that the different areas can be above or below.
Effevtively it is a smoothing effect. You smooth out the peaks and the valleys.
I don't get why that smoothing, consistently leads to an underestimate. Surely it would lead to a value somewhere in the middle. Yes, you're missing the peaks, but you're also missing the valleys.
Then remeber that any grid can only describe features of about 5 grid length size ie of size ~ 130 km. It is all too easy for the strongest winds to be not well represented.
Yes, I can see that, they may well miss the strongest localized winds due to the limitations of the resolution. Why don't they also miss the lightest winds and come up with a value somewhere in the middle?
It's the fact that all the errors in the model are one way that I'm trying to comprehend.
You say: "Within that there will be areas and times when the wind is above or below 30 kn. The computer model cannot resolve these to better than about 130 to 150 km size."
I can't reconcile that with:
"Add on at least one wind force for GFS above F3 seems to be a good rule of thumb."
For example: Imagine the grib at a location predicts 20kts. A local high wind area might have been missed giving real wind of 30kts, but surely it's equally likely that a low wind are might have been missed giving real 10kts.
Frank will correct me if I've got this wrong, but I think it is because the distribution of wind strengths is assymetric around the mean. In other words, the light winds observed will be nearer the mean than the strong winds.