Measures in the CrossValidation Report
Applies to: SQL Server 2019 and earlier Analysis Services Azure Analysis Services Fabric/Power BI Premium
Important
Data mining was deprecated in SQL Server 2017 Analysis Services and now discontinued in SQL Server 2022 Analysis Services. Documentation is not updated for deprecated and discontinued features. To learn more, see Analysis Services backward compatibility.
During crossvalidation, SQL Server Analysis Services divides the data in a mining structure into multiple crosssections and then iteratively tests the structure and any associated mining models. Based on this analysis, it outputs a set of standard accuracy measures for the structure and each model.
The report contains some basic information about the number of folds in the data and the amount of data in each fold, and a set of general metrics that describe data distribution. By comparing the general metrics for each crosssection, you can assess the reliability of the structure or model.
SQL Server Analysis Services also displays a set of detailed measures for mining models. These measures depend on the model type and on the type of attribute that is being analyzed: for example, whether it is discrete or continuous.
This section provides a list of the measures that are contained in the CrossValidation report, and what they mean. For details on how each measure is calculated, see CrossValidation Formulas.
List of Measures in the CrossValidation Report
The following table lists the measures that appear in the crossvalidation report. The measures are grouped by test type, which is provided in the lefthand column of the following table. The righthand column lists the name of the measure as it appears in the report, and provides a brief explanation of what it means.
Test Type  Measures and Descriptions 

Clustering  Measures that apply to clustering models 
Case likelihood: This measure usually indicates how likely it is that a case belongs to a particular cluster. For crossvalidation, the scores are summed and then divided by the number of cases, so here the score is an average case likelihood. 

Classification  Measures that apply to classification models 
True Positive/True Negative/False Positive/False Negative: Count of rows or values in the partition where the predicted state matches the target state, and the predict probability is greater than the specified threshold. Cases that have missing values for the target attribute are excluded, meaning the counts of all values might not add up. 

Pass/Fail: Count of rows or values in the partition where the predicted state matches the target state, and where the predict probability value is greater than 0. 

Likelihood  Likelihood measures apply to multiple model types. 
Lift: The ratio of the actual prediction probability to the marginal probability in the test cases. Rows that have missing values for the target attribute are excluded. This measure generally shows how much the probability of the target outcome improves when the model is used. 

Root Mean Square Error: Square root of the mean error for all partition cases, divided by the number of cases in the partition, excluding rows that have missing values for the target attribute. RMSE is a popular estimator for predictive models. The score averages the residuals for each case to yield a single indicator of model error. 

Log score: The logarithm of the actual probability for each case, summed, and then divided by the number of rows in the input dataset, excluding rows that have missing values for the target attribute. Because probability is represented as a decimal fraction, log scores are always negative numbers. A number closer to 0 is a better score. Whereas raw scores can have very irregular or skewed distributions, a log score is similar to a percentage. 

Estimation  Measures that apply only to estimation models, which predict a continuous numeric attribute. 
Root Mean Square Error: Average error when the predicted value is compared to the actual value. RMSE is a popular estimator for predictive models. The score averages the residuals for each case to yield a single indicator of model error. 

Mean Absolute Error: Average error when predicted values are compared to actual values, calculated as the mean of the absolute sum of errors. Mean absolute error is useful for understanding how close overall the predictions were to actual values. A smaller score means predictions were more accurate. 

Log Score: The logarithm of the actual probability for each case, summed, and then divided by the number of rows in the input dataset, excluding rows that have missing values for the target attribute. Because probability is represented as a decimal fraction, log scores are always negative numbers. A number closer to 0 is a better score. Whereas raw scores can have very irregular or skewed distributions, a log score is similar to a percentage. 

Aggregates  Aggregate measures provide an indication of the variance in the results for each partition. 
Mean: Average of the partition values for a particular measure. 

Standard Deviation: Average of the deviation from the mean for a specific measure, across all the partitions in a model. For crossvalidation, a higher value for this score implies substantial variation between the folds. 
See Also
Feedback
https://aka.ms/ContentUserFeedback.
Coming soon: Throughout 2024 we will be phasing out GitHub Issues as the feedback mechanism for content and replacing it with a new feedback system. For more information see:Submit and view feedback for