This function compares two comparators based on the subset of forecasts for which both comparators have made a prediction. It gets called from pairwise_comparison_one_group(), which handles the comparison of multiple comparators on a single set of forecasts (there are no subsets of forecasts to be distinguished). pairwise_comparison_one_group()
in turn gets called from from get_pairwise_comparisons() which can handle pairwise comparisons for a set of forecasts with multiple subsets, e.g. pairwise comparisons for one set of forecasts, but done separately for two different forecast targets.
scores: An object of class scores (a data.table with scores and an additional attribute metrics as produced by score()).
compare: Character vector with a single colum name that defines the elements for the pairwise comparison. For example, if this is set to "model" (the default), then elements of the "model" column will be compared.
name_comparator1: Character, name of the first comparator
name_comparator2: Character, name of the comparator to compare against
metric: A string with the name of the metric for which a relative skill shall be computed. By default this is either "crps", "wis" or "brier_score" if any of these are available.
one_sided: Boolean, default is FALSE, whether two conduct a one-sided instead of a two-sided test to determine significance in a pairwise comparison.
test_type: Character, either "non_parametric" (the default) or "permutation". This determines which kind of test shall be conducted to determine p-values.
n_permutations: Numeric, the number of permutations for a permutation test. Default is 999.
Returns
A list with mean score ratios and p-values for the comparison between two comparators