The polar vortex and the media hype of the cold weather

IFYes, it was cold. It was very cold. But the media and many in the forecasting profession made what was a perfectly normal weather event into a major story of extreme and rarely seen cold. The Baltimore Sun said “Baltimore’s coldest day in nearly two decades.” And the Washington Post said “Polar vortex delivering D.C.’s coldest day in decades, and we’re not alone.” USA Today did a nice map showing all the cities with new record lows.

A more balance picture of the event was given by Dr. Jeff Masters over at WunderBlog where he pointed out “As notable as this week’s cold wave was–bringing the coldest air seen since 1996 or 1994 over much of the nation–the event failed to set any monthly or all-time record low minimum temperature records at airports and cooperative observing stations monitored by NOAA’s National Climatic Data Center.”

The problem the media got into was claiming that this event was the coldest in decades when in fact that only applied to a specific date in January. A big deal was made when Dulles Airport outside of Washington DC hit a new record low of 1 degree. A little searching at weather.com show that the record low for the airport for the month of January is -18. I’m sorry but a low of 1 degree pales in comparison. It is cold at 1 degree but it is nowhere near how cold it is at -18 degrees. That is the number that should be in the comparison.

IFFor fun I uploaded two images of a nearby stream. Decide which one was taken this morning. If you picked the first image you are wrong. That one was taken during the cold spell of late January 2004.

But I did not have to go back even ten years to find a similar cold spell comparable to what we had the last few days. I used my weather records where I have lived for the past 30 years. On of my hobbies is addition to statistics is weather. I have a Davis weather instruments station out in the yard that records the weather near the house, and not to far from the stream in the picture in 15 minute intervals.

In the heart of the current cold wave I recorded a low of 4.8 on the morning of the 7th, a high of 18.1 that day, and a low of 4.7 the next morning. Just five years ago on January 16th and 17th it was every bit as cold. On the morning of January 16, 2009 the low was 10.9. The high temperature that day was 17.7. However the next morning the low was -0.2. So in the end one night was about 5 degrees warmer and the next night was 5 degrees colder. The daytime highs were very close to each other. jan 2009 vs 2014 lowsThe graph below captures to temperature comparison over a three day span at both points in time. To me they are very much comparable. I don’t see any metric that can claim that the current cold wave is any worst than that seen just five years ago.

Now this is just one measurement at just one place on the map. But it is indicative of the comparisons that can be made. Claiming that what we just saw over the last few days is the coldest in decades just is not justified. In only holds up when the media focuses on a specific day in January. When considered as part of the overall weather pattern of a winter weather if fails completely. It makes for great media hype, but it does not reflect reality.

Many years ago I stopped at the side of the road in the middle of the summer to buy some freshly picked sweet corn. It was in the upper 90s that day. A business man stopped by all decked out in his nice suit. He said to the old farmer “sure is hot today.” The farmer replied “it’s summer, it’s supposed to be.” Cold weather is normal. Very cold weather is normal. The media likes to tell us every day what are the “normal” highs and lows. That is the new name for average high and low. The average are not the norm. The norm is reflected in the full distribution of temperatures that are seen during the winter. We should be thinking in terms of variance. Thinking in terms of the average is deceptive.

Accuweather’s 45 day forcasts – How do we evaluate them?

Last August Accuweather extended their long range forecasts from the previous 30 days out to 45 days. They are now claiming they can predict the weather a month and half into the future with the degree of accuracy that they are worthy of publication.

I had serious doubts as to just how good those forecasts are as I had watched the 30 day forecasts as I wondered what the weather would be like on the days my daughter would make her trips home form Penn State (the home of Accuweather) during the Christmas holiday season. I looked at the 30 day forecasts they were publishing at the time and found them lacking.

In early October the crew at the Capital Weather Gang at the Washington post published a rather simple analysis of the accuracy of the those new forecasts. They were not impressed.

A group of meteorology students at Penn State did a more in depth evaluation. They concluded that the climatologically normal actually made for a better forecast of the daily high temperature than did the Accuweather forecasts. They also concluded that this was true not just 45 days into the future but even just 10 to 15 days into the future.

Accuweather unsurprisingly defended the forecasts. But I think the defense was well summed up by the reporter for the Washington Post when he wrote: “The thrust of Myers’ rationale for providing 45-day forecasts is customer demand and satisfaction.” Read the comments. The arguments mostly go along the lines of there is a demand for the data, people are reading the data, they are coming back to our website, we have the data, and therefore we have an obligation to provide the data. The arguments are based solely on demand for the data. They are not based on the accuracy or quality of the data.

Accuweather did also made a couple of relevant points. They argued that the students only looked at the forecast of the high temperature. That is a very good point. As the spokesman said looking at precipitation and the forecast for the low temperature really should have been included. The spokesman also said the looking at only a 93 day window was too short. That point can be debated. He also thought the looking at only 15 cites was too small a sample. Give the consistent results across the 15 cites this does not seem to me to be a very serious criticism of the students work. Interestingly the original piece in the Washington post looked at the high and low temperate and precipitation in their analysis and still found the forecasts lacking.

So how should on evaluate a long range weather forecast How do we conclude that it is good for bad?

The students at Penn State used average absolute deviation of the foretasted high temperature relative to the actual high. The Washington Post original evaluation used the same absolute deviation, but extended it to both the high and low temperatures and added in a component that incorporated participation into their measure of accuracy. The Washington Post graphics for the three cites they looked at are informative. They looked at San Francisco CA, Denver CO, and Mobile Al forecasts. Denver was the city with the worst performance. The differences illustrate the problems with using a single measure for all three cities. First participation rate and frequency vary by city. In the desert southwest I can make a fairly accurate forecast of participation by just claiming it will never rain. I cannot do anything similar in Florida. Also important is how much temperatures vary. The less variability there is in the temperatures a given city the more accurate my forecasts are likely to be when I use temperature deviation as my evaluation criterion. I simply have to have a model that gives forecasts within a smaller temperate range for those cities with the smaller variability. In the Washington DC metropolitan area summer high temperatures vary from about 60 to 100. In the middle of winter they can vary from 10 to 70. The range is much large in the winter than the summer. So if like the Penn State students I use as my benchmark the average absolute deviation from the climatologically normal my summer measurement us very likely to look better than my winter measurement. This is simply because the average absolute deviation in the winter most likely with be larger than it is in the summer.

Another question to consider in developing an evaluation criterion is what are the users expectations. An error of a few degrees in the the forecast is likely inconsequential. The failure to forecast a major snowstorm is a big deal. We all expect forecasts to change over time because we recognize that forecasting is an inexact science. Perhaps a better measure would incorporate how much the forecast varies around the actual weather that occurs. A forecast that is 20 degrees above normal one day and 20 degrees below normal a week later for the same date is a problem regardless of the final weather on the forecast day. So I would want a measure the incorporated how much the forecast changed over time and how close to the forecast date did the prediction get to the final weather for that data and remain there.

So what evaluation criterion would I suggest? First off I would not penalize the forecast for being off by a few degrees. That much error would be expected by the users. For errors of more than a few degrees I think I still prefer a quadratic loss function. Part of the reason for that is that I think these types of forecasts tend to avoid the extreme. I have not seen forecasts of 20 degrees above or below the norm. This is a conservative response by the forecaster, but missing those extreme weather events to me is a big error. Next I would measure the error over time as the date approaches. So the final error measure would be based on the 45 forecasts for the given day. I would use a weighting system for the error measure. This is because the forecast claims to be good out to 45 days. So an error at 45 days should be considered worst than an error 5 days out.

I would penalize the forecaster for changing the forecast. I would do this as a good forecast should not need to be changed. So any changes are a self admission that the original forecast has come up short. As some changes in the forecast should be expected I would make the penalty proportional to the size of the change – perhaps using a variance measure on the 45 high and low temperate forecasts. I would view greater variability in the forecast for a given date as an indication of the poorer quality of the forecast. I would also want to incorporate a measure of the error on the precipitation forecast. This also need should include components measuring the actual error as well as how much and how often the forecast changes over time.

I am not sure at this time how I would wrap all of these measures together. At this point with all I’d like to do in the evaluation criterion I am thinking of seems much too complex. I would want to do separate measures for each forecast area. There are enough climate differences across the country the developing and appropriate way to combine an error measure would be problematic.

With all that in mind I think I’ll start tracking Accuweather’s forecast for the Superbowl game. It is 26 days out and in their media hype they are already giving very specific forecasts. They claim

It is important to note that long-range forecasts are not intended to give an exact forecast so far ahead of time. However, they can give an accurate representation of the kind of weather pattern that will play out. The trend for this year’s winter is based on a number of factors, including the potential for a split jet stream.

While saying there will be cloudy with a high of 45 with no snow on game day. I’ll be looking at just how much that forecast change over the next three plus weeks.

Virgin Births

It is the time of year when someone wants to always bring up the idea of the virgin birth of Jesus as described in the Christian Scriptures. This time it is a study published in the Christmas edition of the British Medical Journal that claims reports of virgin births are also a modern day phenomenon. The study, titled US researchers ponder modern day virgin births goes into considerable detail about the sex-education of the virgins and how many took a chastity pledge.

The study was conducted by researchers at the University of North Carolina at Chapel Hill using data from the U.S. National Longitudinal Study of Adolescent Health (Add Health). They found a reported 45 such births reported over a 15 year time span. This was an under one percent characteristic.

I liked the Reuters head line piece that noted near the end that the mothers in question were more likely to have boys than girls and be pregnant during advent. They did say that these observations were not statistically significant.

In fairness to the authors they do admit such measures are subject to some degree of respondent bias.

My questions – why does this kind of thing show up in the literature at all? Rather than tout measures of virgin births they need to be discussion issue of data quality. A little Google search revealed a paper using the same data source that did just that titled: An Exploratory Study about Inaccuracy and Invalidity in Adolescent Self-Report Surveys. They discuss two groups of respondents – those who make mistakes in answering the questionnaire and the jokesters.

These virgin birth are likely nothing more than data errors. If respondents is able to answer the survey questions in such a way that they can report a birth prior to any sexual activity they will do so. Some will do it in error and some will just be having a bit of fun at the expense of the survey authors. A good survey, if automated, should have consistency checks in place while the survey is conducted that will flag and prevent the errors. The editing process should also catch the inconsistencies.

But to publish a paper discussing the detailed characteristics of the group is something I expect to see on April 1st, not in the week before Christmas.

  • Subscribe to Blog via Email

    Enter your email address to subscribe to this blog and receive notifications of new posts by email.

  • December 2024
    S M T W T F S
    1234567
    891011121314
    15161718192021
    22232425262728
    293031  
  • Recent Posts