# Linear Models Visualizations

The following tables and options are available for Linear Models visualizations.

**Model Information table**

This table contains information on how the model was fitted, so you can make sure that the model you have is what you intended. It contains information on input settings such as model selection methods, as well as summary measures of prediction
accuracy. R^{2} gives the squared correlation between observed and predicted values, which in a linear model with an intercept gives the proportion of variance in the target variable accounted for by the model, which ranges from 0 for
a model with no predictive ability to 1 for a perfectly fitting model. Adjusted R^{2} shrinks or penalizes this proportion based on the number of parameters in the model, in order to facilitate comparisons among models with different numbers
of predictors. The Corrected Akaike Information Criterion (AICc) measure can be used to compare models with different numbers of parameters when fitted to the same target variable with the same data. Smaller values are preferred. This measure
is a function of the target variable values, so unlike R^{2} measures it cannot be used to compare models for different targets or different sets of data.

**Records Summary table**

This table shows you how many records were used to fit the model and whether any records were excluded due to missing data. If frequency weighting is in effect, it shows information about both unweighted and weighted numbers of records.

**Predictor Importance chart**

This chart displays bars representing the predictors in descending order of relative importance for predicting the target, as determined by a variance-based sensitivity analysis algorithm. The values for each predictor are scaled so that they add to 1. Hovering over the bar for a particular predictor shows a table with its importance value and descriptive statistics about the predictor.

**Tests of Model Effects table**

This table gives a standard analysis of variance (ANOVA) table for each term in the linear model, including effects representing multiple parameters for categorical predictors. The Sig. column provides the probability of observing an F statistic
as large or larger than the one observed in a sample if sampling from a population where the predictor has no effect, and can be used to identify "statistically significant" predictors. In very large samples predictors may be identified
as statistically significant even though in practical terms they are not important. If specified in creating the model, columns containing effect-size estimates known as η^{2} and partial η^{2} and confidence intervals
for these effect-estimates are provided. These measures are similar to R^{2} in being based on proportions of variance in the target variable associated with the predictors.

**Parameter Estimates table**

This table displays the parameter estimates (also known as regression coefficients, beta coefficients or beta weights) for the fitted linear model, along with measures of sampling variation, tests of statistical significance and confidence intervals.
These coefficients combine to form the linear prediction model, which typically consists of a constant or intercept coefficient plus each regression coefficient multiplied by its predictor variable value, to produce the model predictions. As with
tests of model effects, individual predictors may be statistically significant without being practically important, so effect size estimates can be specified in creating the model, and if so partial η^{2} estimates and confidence
intervals for them will also be displayed in this table. If a regularization method (Lasso, ridge regression or Elastic Net) has been used to fit the model, only the regression coefficients will be displayed.

**Observed by Predicted chart**

This chart shows a scatterplot of predicted values against observed target values. The plotted points may represent averages of binned values. In a perfect-fitting model, the points would all fall exactly on the 45-degree line from lower left to upper right. Vertical departures from this line show the residuals or prediction errors for individual data points or averages of binned values. Points lying particularly far above or below this line are outliers that may warrant attention.

**Residuals by Predicted chart**

This chart is a scatterplot of residuals or prediction errors vs. predicted values from the linear model. The plotted points may represent averages of binned values. If the prediction model is capturing all of the systematic variation in the target, you would expect to see these residuals randomly scattered around the horizontal prediction line. Patterns in the residuals indicate that the model is not capturing all of the systematic variation in the target and may indicate the need for additional predictors, a different functional form for the model (such as one including nonlinear terms) or unequal variances, all of which can cause statistical inferences to be incorrect.

**Normal P-P Plot of Residuals**

The chart plots the percentiles of the cumulative distribution of observed residuals against the percentiles from a cumulative normal distribution. This allows you to assess a common linear model assumption of normality of prediction errors. Systematic deviations from the 45-degree line indicate departures from normality such as skewness and kurtosis. Outliers may also be seen here as well.

**Residual Histogram**

This chart shows a binned representation of the residual values with the vertical axis indicating relative frequencies. This allows you to assess the shape of the distribution of residuals, which under the normality assumption should look more or less like a standard "bell curve," symmetric around a single peak in the center, with frequencies decreasing as you move away from the peak in either direction.

## Next steps

Like your visualization? Why not deploy it? For more information, see Deploy a model.