-
Notifications
You must be signed in to change notification settings - Fork 1
scores_metrics
Scale-dependent-errors
Percentage errors
Scaled errors
Goodness of Fit
Correlation and Synchrony
Information/Entropy measures
Cross-validation
The errors are on the same scale as the data, i.e. different data sets cannot be compared
Mean absolute error (MAE)
Minimizing MAE leads to prediction of median.
Mean square error (MSE)
Strongly penalizes large wrong predictions
Root mean square error (RMSE)
Minimizing RMSE leads to prediction of mean
Percentage errors are unit free and therefore allow comparison of different data sets.
Mean absolute percentage error (MAPE)
Problems:
- cannot be used for zero values
- puts more weight on negative than positive errors
Symmetric absolute percentage error (sMAPE)
Used to overcome the problems of MAPE
Alternative to percentage errors when comparing different datasets
Mean absolute scaled error (MASE)
- scale invariance
- symmetric
- less than one if it arises from a better forecast than the average naïve forecast and conversely it is greater than one if the forecast is worse than the average naïve forecast
https://otexts.com/fpp2/accuracy.html
https://scikit-learn.org/stable/modules/model_evaluation.html#
Check if a hypothesis is correct. Often referred to as explained variance scores.
Coefficient of determination (
Describes the variance (of y) which is explained by the model prediction.
Chi-square test
Statistical test that measures the significance of the hypothesis given the data. Chi-square is obtained by
The sum of square errors follows the chi-square distribution if the errors are independent and normally distributed. The probability of the calculated
F-test
The F-value expresses how much of the model has improved compared to the mean (null hypothesis) given the variance of the model and data.
For regression, it can be generalized in order to compare the fit quality of a complex model as compared to a simpler version of the same model.
Here, we assume model 1 with k_1 number of parameters and model 2 with k_2 number of parameters, where k_1 > k_2. The F-test is obtained as follows:
-
Compute the sum of squares of residual errors for each model
-
The F-statistics of the two regression models is obtained by
where n is the number of data points.
-
The RSS are random variables described by a probability distribution. The sum of independent and standard normal variables follow the pdf of the chi-square distribution. This happens to be the case in the equation above. The ratio of the two scaled chi-square distributions is described by the F-distribution.
-
We can evaluate the probability of occurrence p and compare it to our threshold value (some small number we choose). If we can conclude that model 1 is able to explain the variance in the data better than model 2.
https://towardsdatascience.com/fisher-test-for-regression-analysis-1e1687867259
Pearson correlation
Pearson correlation measures how two continuous signals co-vary over time. The linear relationship between these signals are given from -1 (anticorrelated) to 0 (incorrelated) to 1 (perfecly correlated).
The Pearson correlation coefficient for two random variables X_1 and X_2 is:
For time-series on can calculate a
- global correlation coefficient: a single value
- local correlation coefficient: determine correlation in a rolling window over time
Caution:
- outliers can skew the correlation
- assuming the data is homoscedatic, i.e. constant variances
Anomaly correlation coefficient (ACC)
In climate science and meteorology correlating forecasts directly with observations may give misleadingly high values because of the seasonal variations. It is therefore established practice to subtract the climate average from both the forecast and the verification. The anomaly correlation coefficient is obtained by
In case the climate average is known, it often replaces .
https://www.jma.go.jp/jma/jma-eng/jma-center/nwp/outline2013-nwp/pdf/outline2013_Appendix_A.pdf
Time Lagged Cross Correlation (TLCC)
TLCC is a measure of similarity of two series as a function of displacement. It captures directionality between two signals, i.e. leader-follower relationship.
Idea: Similar to convolution of two signals, i.e. shifting one signal with respect to the other while repeatedly calculating the correlation.
Windowed time lagged cross correlations (WTLCC) are an extension of TLCC where local correlations coefficients are computed for each lag-time which is then plotted as a matrix.
Dynamic Time Wrapping (DTW)
DTW computes the path between two signals that minimize the distance between the two signals. DTW computes the euclidean distance at each frame across every other frames to compute the minimum path that will match the two signals.
Properties:
- deal with signals of different length
- requires interpolation of missing data
Instantaneous phase synchrony
For time series with oscillating properties the instantaneous phase synchrony measures the phase similarities between signals at each timepoint. The phase between signals is refered to as angle which is obtained by a Hilbert transformation of the signals. Phase coherence can be quantified by subtracting the angular difference from 1.
http://jinhyuncheong.com/jekyll/update/2017/12/10/Timeseries_synchrony_tutorial_and_simulations.html
https://towardsdatascience.com/four-ways-to-quantify-synchrony-between-time-series-data-b99136c4a9c9
Information measures on time series amounts to constructing empirical distributions and apply Shanon Information Measures on them.
Mutual Information (MI)
MI is a measure of the mutual dependence of two random variable. Applied to two time series and , we first construct the empirical distributions > and the joint . The mutual information is defined as
which is essential the KL-divergence of the distributions.
https://elife-asu.github.io/PyInform/timeseries.html
Leave-one out error (LOOE)
Bootstrap
Wasserstein metric
Granger causality
Statistical hypothesis test to determine whether one time series is useful in forecasting another.
- Euclidian distance (ACC)
- correlation maps
- density metrics
- saliency maps: highlights which changes in the input would most affect the output.
- heat maps: highlights which inputs are most important for the prediction
Latitude weighted RMSE
Skill metric used in climate science to compare time-series from different spatial points:
with latitude weighting factor
- rang-index
- PIT diagram
- reliability diagram
- KL-divergence
- Wasserstein metric