An Assessment of How Domain Experts Evaluate Machine Learning in Operational Meteorology

David R. Harrison, Amy McGovern, Christopher D. Karstens, Ann Bostrom, Julie L. Demuth, Israel L. Jirak, Patrick T. Marsh

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

As an increasing number of machine learning (ML) products enter the research-to-operations (R2O) pipe-line, researchers have anecdotally noted a perceived hesitancy by operational forecasters to adopt this relatively new tech-nology. One explanation often cited in the literature is that this perceived hesitancy derives from the complex and opaque nature of ML methods. Because modern ML models are trained to solve tasks by optimizing a potentially complex combi-nation of mathematical weights, thresholds, and nonlinear cost functions, it can be difficult to determine how these models reach a solution from their given input. However, it remains unclear to what degree a model’s transparency may influence a forecaster’s decision to use that model or if that impact differs between ML and more traditional (i.e., non-ML) methods. To address this question, a survey was offered to forecaster and researcher participants attending the 2021 NOAA Hazardous Weather Testbed (HWT) Spring Forecasting Experiment (SFE) with questions about how participants subjectively perceive and compare machine learning products to more traditionally derived products. Results from this study revealed few differences in how participants evaluated machine learning products compared to other types of guidance. However, comparing the responses between operational forecasters, researchers, and academics exposed notable differences in what factors the three groups considered to be most important for determining the operational success of a new forecast prod-uct. These results support the need for increased collaboration between the operational and research communities.

Original languageEnglish
Pages (from-to)393-410
Number of pages18
JournalWeather and Forecasting
Volume40
Issue number3
DOIs
StatePublished - Mar 2025
Externally publishedYes

Keywords

  • Artificial intelligence
  • Decision making
  • Forecasting
  • Forecasting techniques
  • Operational forecasting

Fingerprint

Dive into the research topics of 'An Assessment of How Domain Experts Evaluate Machine Learning in Operational Meteorology'. Together they form a unique fingerprint.

Cite this