NCAR uses AI to tackle the persistent challenge of predicting hail and tornadoes

LinkedIn +

The National Center for Atmospheric Research (NCAR) is using AI to run experimental forecasts for hail, tornadoes and intense winds – storm hazards that can cause serious damage but that are notoriously difficult for weather models to accurately predict.

The Neural Network Convective Hazard Forecasts, which provide the probability of hail, tornadoes or winds for a particular location at a particular time, are updated twice daily and freely available online.

NCAR began running these research forecasts in the spring of 2020 as part of the Hazardous Weather Testbed Spring Experiment at NOAA. When the spring severe weather season wrapped up at the end of June, the research team analyzed the forecasts, comparing them to more traditional techniques for forecasting storm hazards.

The scientists found that the neural network significantly improved the accuracy of forecasts based on traditional model output, especially in situations where those traditional forecasts tended to perform the worst, including in the eastern and western parts of the USA and for night-time storms.

“We were able to show a significant improvement in most situations,” said NCAR scientist Ryan Sobash, who is leading the project. “Not only did the neural network more skillfully predict when and where severe storm hazards were likely, it was also able to better predict whether the hazardous event would be dominated by hail or by winds.”

The system is up and running again this spring with some tweaks. The project is being funded by NOAA and the National Science Foundation, which is NCAR’s sponsor. The forecasts are run on the Cheyenne supercomputer at the NCAR-Wyoming Supercomputing Center.

From one predictor to 40

In order for a weather model to spin up a thunderstorm, it has to be run at a high enough resolution to capture the fine-scale atmospheric phenomena – including updrafts and downdrafts – that drive the storm’s creation. This typically requires spacing of 4km (2.5 miles) or less between grid points inside the model.

At that resolution, the model can begin to simulate the storm itself, but it is unable to produce many of the hazards associated with the storm, which happen at even smaller scales, including hailstones and tornadoes. Because of that, forecasters have relied on particular outputs in the model data, or proxies, to determine the likelihood that a severe storm will produce such hazards.

One of the most frequently used proxies is updraft helicity, a measure of a storm’s rotation. Storms with stronger rotation tend to be more severe and more capable of producing hail and tornadoes. This proxy has worked relatively well for supercells and other rotating storms, but it doesn’t capture some of the severe weather that can be produced in straight-line storms, such as derechos.

By contrast, the neural network used in NCAR’s new forecasts can ingest about 40 different factors, including updraft helicity but also the storm’s location, time, dew point, windspeeds, surface pressure, and much more.

The neural network uses patterns in how those predictors relate to one another that it gleaned from its training data set – in this case nearly 500 past forecasts from NOAA’s High-Resolution Rapid Refresh model along with the accompanying reports of actual storms – to calculate the probability that a storm will produce hail, tornadoes or strong winds. The forecasts produced by the neural networks show the likelihood of a storm hazard forming within either 40km (25 miles) or 120km (75 miles) of individual grid points in the model.

By taking into account factors beyond storm rotation, the neural network is better able to predict hazards associated with straight-line wind storms than forecasts based solely on updraft helicity, and it also improves on forecasts for regions where supercells are not as likely to form, especially areas outside the Midwest.

“The success of our neural network forecast suggests that machine learning could be a useful tool for operational forecasts,” Sobash said. “The forecasts are being run again this spring as part of the Hazardous Weather Testbed and we look forward to getting more feedback from the operational forecasters about how they might incorporate this kind of information into their existing forecasting process.”

Read the full study here: A Comparison of Neural-Network and Surrogate-Severe Probabilistic Convective Hazard Guidance Derived from a Convection-Allowing Model

Share this story:

Comments are closed.