List of Metrics Available - NOAA-OWP/wres GitHub Wiki

The metrics available within WRES are listed below. They are organized by input data type and output data type. Transformations between data types are possible. For example, an ensemble forecast may be transformed into a single-valued forecast by calculating an ensemble average (e.g., mean). Similarly, a single-valued prediction may be transformed into a dichotomous prediction using an event threshold.

When declaring use of the metric in a YAML file, use the name provided in the table below in lower case and replace any dashes, '-', with spaces, ' ', so that, for example, "Time-to-Peak Error" is declared as "time to peak error".

* To calculate the difference between a predicted and baseline dataset for this metric, simply add the word difference to the metric name.

+ When the Mean Square Error Skill Score is declared without an explicit baseline, a default baseline of climatology is used (i.e., the variance of the observations). In this case, the score is equivalent to the Nash-Sutcliffe Efficiency.

ǂ These metrics reveal the paired values directly, rather than summarizing the paired values with statistics. As such, these metrics generate voluminous outputs and should be used with caution, particularly when the evaluation contains a large number of paired values. In such cases, it is not recommended to generate svg graphics, as these graphics are encoded with uncompressed text, which will be extremely large. The spaghetti plot contains the paired values from all ensemble forecast traces, so the outputs associated with this metric are particularly large.

A B C
1
INPUT TYPE (PAIRED VALUES) METRIC NAME OUTPUT TYPE
2
SINGLE VALUED Bias Fraction* SCORE
3
Coefficient of Determination*
4
Pearson Correlation Coefficient*
5
Index of Agreement*
6
Mean Absolute Error*
7
Mean Error*
8
Median Error*
9
Root Mean Square Error*
10
Root Mean Square Error Normalized*
11
Sum Square Error*
12
Volumetric Efficiency*
13
Sample Size*
14
Kling Gupta Efficiency*
15
Mean Square Error*
16
Mean Square Error Skill Score+
17
Quantile-Quantile Diagram DIAGRAM
18
Box Plot of Errors
19
Box Plot of Percentage Errors
20
Scatter plotǂ
21
SINGLE VALUED TIME-SERIES Time-To-Peak Error Summary Statistics SCORE
22
Time-To-Peak Relative Error Summary Statistics
23
Time-To-Peak Error PAIRED VALUES
24
Time-To-Peak Relative Error
25
Time Series Plotǂ
26
DISCRETE PROBABILITY Brier Score* SCORE
27
Brier Skill Score
28
Relative Operating Characteristic Score*
29
Reliability Diagram DIAGRAM
30
Relative Operating Characteristic Diagram
31
DICHOTOMOUS Critical Success Index* SCORE
32
Threat Score*
33
Equitable Threat Score*
34
Frequency Bias*
35
Peirce Skill Score*
36
Probability of Detection*
37
Probability of False Detection*
38
MULTICATEGORY Contingency Table (2x2) TABLE
39
Peirce Skill Score SCORE
40
ENSEMBLE FORECAST Sample Size SCORE
41
Continuous Ranked Probability Score*
42
Continuous Ranked Probability Skill Score
43
Rank Histogram DIAGRAM
44
Box Plot of Errors by Observed Value BOXPLOT
45
Box Plot of Errors by Forecast Value
46
ENSEMBLE FORECAST TIME SERIES Spaghetti Plotǂ PAIRED VALUES
⚠️ **GitHub.com Fallback** ⚠️