{"id":6172,"date":"2012-12-07T16:23:09","date_gmt":"2012-12-07T21:23:09","guid":{"rendered":"https:\/\/hwt.nssl.noaa.gov\/ewp\/internal\/blog\/?p=6172"},"modified":"2017-06-23T14:25:49","modified_gmt":"2017-06-23T19:25:49","slug":"warning-verification-pitfalls-explained-part-4","status":"publish","type":"post","link":"https:\/\/inside.nssl.noaa.gov\/ewp\/2012\/12\/07\/warning-verification-pitfalls-explained-part-4\/","title":{"rendered":"Warning Verification Pitfalls Explained &#8211; Report Density Can Inflate POD"},"content":{"rendered":"<p>I&#8217;m back after a too-lengthy absence from this blog.\u00a0 I&#8217;ve been thinking about some experimental warning issues again lately, and have a few things to add to the blog regarding some more pitfalls of our current warning verification methodology.\u00a0 I hinted on these in past posts, but would like to expand upon them.<\/p>\n<p>Have you ever been amazed that some especially noteworthy severe weather days can produce record numbers of storm reports?\u00a0 Let&#8217;s take this day for example, 1 July 2012:<\/p>\n<p><a href=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/120701_rpts_filtered.gif1_.png\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/120701_rpts_filtered.gif1_.png\" alt=\"120701_rpts_filtered.gif\" title=\"120701_rpts_filtered.gif\" width=\"582\" height=\"408\" class=\"alignnone size-full wp-image-14613\" \/><\/a><\/p>\n<p>Wow!\u00a0 A whopping 522 wind reports and 212 hail reports.\u00a0 That must have been an exceptionally-bad severe weather day.\u00a0 (It actually was the day of the big Ohio Valley to East Coast derecho from last July, a very impactful event).<\/p>\n<p>But what makes a storm report?\u00a0 Somebody calls in, or uses some kind of software (e.g., <a href=\"http:\/\/www.spotternetwork.org\/\">Spotter Network<\/a>), to report that the winds were <em>X<\/em> mph or the hail was <em>Y<\/em> inches in diameter from some location and at some time from within the severe thunderstorm.\u00a0 But the severe weather event is actually impacting an area surrounding the location from which the report was generated, and has been and will occur over the time interval representing the lifetime of the storm.\u00a0 It is highly unlikely that a hail report represented only a single stone falling at that location, or that the wind report represented a single wind gust local to that single location, and there were no other severe wind gusts anywhere else nor at any other time during the storm.\u00a0 Each of these reports represent only a <em>single sample<\/em> of an event that covers a two-dimensional space over a time period.<\/p>\n<p>If you recall from <a href=\"https:\/\/hwt.nssl.noaa.gov\/ewp\/internal\/blog\/2011\/10\/11\/warning-verification-pitfalls-explained-%E2%80%93-part-2-2x2-x2\/\">this blog entry<\/a>, the official Probability Of Detection (POD) is computed to be the number of reports that were within warning polygons over the total number of reports (inside and outside polygons).\u00a0 It&#8217;s easy to see that to effectively improve a office&#8217;s overall POD for a time period (e.g., one year), they only need to increase the number of reports that are covered by the warning polygons issued by that office during that time period.\u00a0 One way to do this is to <a href=\"https:\/\/hwt.nssl.noaa.gov\/ewp\/internal\/blog\/2011\/10\/11\/warning-verification-pitfalls-explained-%E2%80%93-part-3-lets-go-fishing\/\">cast a wide net<\/a>, and issue larger and longer-duration warning polygons. \u00a0 But another way to artificially improve POD is to simply increase the number of reports within storms via aggressive report gathering.\u00a0 Let&#8217;s consider a severe weather event like this one:<\/p>\n<p><a href=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/hail1.jpg\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/hail1-1600x1174.jpg\" alt=\"hail1\"  title==\"hail1\" width=\"1024\" height=\"751\" class=\"alignnone size-large wp-image-14614\" srcset=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/hail1.jpg 1600w, https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/hail1-800x587.jpg 800w, https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/hail1-768x564.jpg 768w\" sizes=\"auto, (max-width: 767px) 100vw, (max-width: 1200px) 60vw, 720px\" \/><\/a><\/p>\n<p>Look at all those (presumably) severe-sized hail stones.\u00a0 We can make a report on each one, at each time they fell.\u00a0 After about an hour of counting and collecting (before they all melted), this observer found 5,462 hail stones that were greater than 1&#8243; in diameter.\u00a0 Beautiful &#8211; the Probability Of Detection is going to go way up!\u00a0 We can also count all the damaged trees as well to add hundreds of wind reports.\u00a0 Do you see the problem here?\u00a0 Are you getting tired of my extrapolations to infinity?\u00a0 Yes, there are literally an infinite number of severe weather reports that can be gleaned from this event (technically, there is a <em>finite<\/em> number of severe-size hail stones fell in this storm, but who&#8217;s really counting that gigantic number?).\u00a0 But let&#8217;s scale this back.\u00a0 Here&#8217;s a scenario in which a particular warning is verified two different ways:<\/p>\n<p><a href=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/adding-reports5.png\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/adding-reports5.png\" alt=\"adding-reports5\" title=\"adding-reports5\" width=\"580\" height=\"827\" class=\"alignnone size-full wp-image-14615\" \/><\/a><\/p>\n<p>Each warning polygon verifies, so no false alarms.\u00a0 For the scenario on the top, there is one hit added to all reports for the time period (maybe a year&#8217;s worth of warning), but for the bottom scenario, there are <em>seven<\/em> hits added to the statistics.<\/p>\n<p><span style=\"text-decoration: line-through;\">But wait, doesn&#8217;t the NWS Verification Branch filter storm reports that are in close proximity in space and time when computing warning statistics?\u00a0 Wouldn&#8217;t those seven hits be reduced to a smaller amount?\u00a0 They use a filter of 10 miles and 15 minutes to avoid my hypothetical over-reporting scenario.\u00a0 But that really doesn&#8217;t address the issue entirely.\u00a0 One can still try to fill every 10 mile and 15 minute window with a hail or wind report in order to maximize their POD.\u00a0 But if you think about it, that&#8217;s not really a bad idea.\u00a0 In essence, you are filling a grid with a 10 mile and 15 minute resolution with as much information known about the storm as possible.\u00a0 But this works <em>only <\/em>if you also call into every 10-miles\/15-minute grid point inside and outside every storm.\u00a0 Forecasters rarely do this (and realistically can&#8217;t), because of workload issues, and because only one report within a warning polygon is all that is needed to avoid that warning from being labelled a false alarm (again, <a href=\"https:\/\/hwt.nssl.noaa.gov\/ewp\/internal\/blog\/2011\/10\/11\/warning-verification-pitfalls-explained-%E2%80%93-part-3-lets-go-fishing\/\">cast the wide net<\/a> so that one can increase their chance of getting a report within the warning).<\/span><\/p>\n<p><span style=\"text-decoration: underline;\">CORRECTION (1\/10\/2013)<\/span>:\u00a0 I just learned that the 10 mile \/ 15 minute filtering was only done in the era of <em>county-based<\/em> warning verification, and is not done for storm-based verification.\u00a0 Therefore, my arguments against the current verification methodology where hit rates and POD can be stacked by gathering more storm reports is further bolstered.\u00a0 More information is in the <a href=\"http:\/\/www.nws.noaa.gov\/directives\/sym\/pd01016001curr.pdf\">NWS Directive on forecast and warning verification<\/a>.<\/p>\n<p>If we knew exactly what was happening within the storm at all times and locations at every grid point (in our case, every 1 km and 1 minute), we&#8217;d have a very robust verification grid to use for the geospatial warning verification methodology.\u00a0 But we really don&#8217;t know exactly what&#8217;s is happening everywhere all the time because it is nearly impossible to collect all those data points.\u00a0 The <a href=\"http:\/\/ewp.nssl.noaa.gov\/projects\/shave\/\">Severe Hazards Analysis and Verification Experiment (SHAVE) <\/a>is attempting to improve on the report density in time and space.\u00a0 But their resources are also finite, and they don&#8217;t have the staffing to call into every thunderstorm.\u00a0 Their high-resolution data set is very useful, but limited to only the storms they&#8217;ve called.\u00a0 What could we do to broaden the report database so that we have a better idea of the full scope of the impact of every storm?\u00a0 One concept is <em>proxy verification<\/em>, in which some other remotely-sensed method is used to make a reasonable approximation of the coverage of severe weather within a storm, like so:<\/p>\n<p><a href=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/use_swath_instead1.png\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-content\/uploads\/sites\/22\/2012\/12\/use_swath_instead1.png\" alt=\"use_swath_instead1\" title=\"use_swath_instead1\" width=\"552\" height=\"383\" class=\"alignnone size-full wp-image-14616\" \/><\/a><\/p>\n<p>This set of verification data will have a degree of uncertainty associated with it, but the probability of the event isn&#8217;t zero, and is thus, useful.\u00a0 It is also very amenable to the geospatial verification methodology already introduced in this blog series.\u00a0 More on this later&#8230;<\/p>\n<p><em>Greg Stumpf, <\/em><em>CIMMS and NWS\/MDL<\/em><\/p>\n<p><em><br \/>\n<\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>I&#8217;m back after a too-lengthy absence from this blog.\u00a0 I&#8217;ve been thinking about some experimental warning issues again lately, and have a few things to add to the blog regarding&#8230; <\/p>\n<p class=\"link-more\"><a href=\"https:\/\/inside.nssl.noaa.gov\/ewp\/2012\/12\/07\/warning-verification-pitfalls-explained-part-4\/\" class=\"more-link\">Read more \u00bb<\/a><\/p>\n","protected":false},"author":333,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[5],"tags":[],"class_list":["post-6172","post","type-post","status-publish","format-standard","hentry","category-experimental-warning-thoughts"],"_links":{"self":[{"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/posts\/6172","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/users\/333"}],"replies":[{"embeddable":true,"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/comments?post=6172"}],"version-history":[{"count":2,"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/posts\/6172\/revisions"}],"predecessor-version":[{"id":14617,"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/posts\/6172\/revisions\/14617"}],"wp:attachment":[{"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/media?parent=6172"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/categories?post=6172"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/inside.nssl.noaa.gov\/ewp\/wp-json\/wp\/v2\/tags?post=6172"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}