SUR Regression

Example Variables(Open Data):

The Fennema-Sherman Mathematics Attitude Scales (FSMAS) are among the most popular instruments used in studies of attitudes toward mathematics. FSMAS contains 36 items. Also, scales of FSMAS have Confidence, Effectance Motivation, and Anxiety. The sample includes 425 teachers who answered all 36 items. In addition, other characteristics of teachers, such as their age, are included in the data.
You can select your data as follows:
1-File
2-Open data
(See Open Data)

The data is stored under the name FSMAS-T(You can download this data from here ).
You can edit the imported data via the following path:
1-File
2-Edit Data
(See Edit Data)

Example Variables(Compute Variable):

The three variables of Confidence, Effectance Motivation, and Anxiety can be calculated through the following path:
1-Transform
2-Compute Variable
Items starting with the letters C, M, and A are related to the variables Confidence, Effectance Motivation, and Anxiety, respectively.
(See Compute Variable)

Introduction to SUR(Seemingly Unrelated Regressions) Regression:

This method is a generalization of a linear regression model that consists of several regression equations. Each equation is a valid linear regression on its own and can be estimated separately, which is why the system is called seemingly unrelated, although some authors suggest that the term seemingly related would be more appropriate since the error terms are assumed to be correlated across the equations.
Consider a system of m equations, where the ith equation is of the form
Where is a vector of the dependent variable, is the coefficient vector and ui is a vector of the disturbance terms of the ith equation.
We can write
or more simply as .
We assume that there is no correlation of the disturbance terms across observations.
We explicitly allow for contemporaneous correlation , where and indicate the equation number. However, we explicitly allow for contemporaneous correlation. The covariance matrix of all disturbances is Where is the (contemporaneous) disturbance covariance matrix, is the Kronecker product, is an identity matrix of dimension , and is the number of observations in each equation.
The equations can be consistently estimated by seemingly unrelated regression (SUR). This estimation can be obtained by .
The covariance matrix of these estimators can be estimated by .
Since the (true) disturbances ( )of the estimated equations are generally not known, their covariance matrix cannot be determined. Therefore, this covariance matrix is generally calculated from estimated residuals ( ) that are obtained from a first-step OLS estimation. The residual covariance matrix is calculated
Where and are the number of independent variables in equation and , respectively.

Path of SUR Regression:

You can perform SUR(Seemingly Unrelated Regressions) regression by the following path:
1-Exploratory Analysis
2-Regression
3-SUR

A. SUR Regression window:

This window includes the Select Dependent Variable button, the results button, and the buttons for saving the residuals of model. Defining models is done in two steps:
1- Define dependent variables
2-Select independent variables for each dependent variable defined in the first step.

B. Select Dependent Variables:

You can select the dependent variables through this button. After opening the window, you can select it by selecting the desired variables.
For example, the variable Confidence and Effectance.Motivation are selected in this data

C. Select Independent Variables panel:

In the SUR Regression window, after selecting dependent variables, a button is activated to select independent variables for each dependent variable.

D. Select Independent Variables:

You can select the independent variables through this button. After the window opens, you can select them by selecting the desired variables.
For example, for the Confidence variable, the variables Effectance Motivation and Anxiety are selected in this data and for the Effectance Motivation variable the variables Confidence and Anxiety are selected.

E. Run Regression:

You can see the results of the OLS regression in the results panel by clicking this button.
Results include the following:
-Model Summary
-Parameter Estimates
-Collinearity Diagnostics

Model Summary:
In the first part, the general indexes of the models are as follows:
Where is sample size and is the number of dependent variables (equations).

Where is number of dependent variables (equations) and is the number of independent variables in equation .

Where is SSR of equation .

detRCov:
This index is i.e. The determinant of matrix. Where is the covariance matrix of residuals and |.| is the determinant of a matrix.

OLS-R2:
This index is i.e. mean of OLS ( Coefficient of determining models after fitting by OLS method.).

McElroy-R2:
The goodness of fit of the whole system can be measured by the McElroy’s values

In the second part, the indexes of each model are presented:
N:
N value is equal to the sample size i.e.

Where is sample size.

DF:
This index is obtained from the following equation:

Where is the number of independent variables in equation .

SSR:
This index is obtained from the following equation:


MSE:
This index is obtained from the following equation:

Where is the number of independent variables in equation .

RMSE:
This index is obtained from the following equation:


R2:
The goodness of fit of each single equation can be measured by the traditional values

Where is the value of the jth equation and is the mean value of .

Adj R2:
This index is obtained from the following equation:

The covariance matrix of the residuals used for estimation:
Since the (true) disturbances ( ) of the estimated equations are generally unknown, their covariance matrix cannot be determined. Therefore, this covariance matrix is generally calculated from estimated residuals ( ) that are obtained from a first-step OLS estimation.
The residual covariance matrix is calculated

Where and are the number of independent variables in equation and , respectively.
The covariance matrix of the residuals:
The residual covariance matrix is calculated

Where and are the number of independent variables in equation and , respectively.
The correlations of the residuals:
The components of this matrix are defined as follows:


-Parameter Estimates
For each equation, the coefficients presented in this table are obtained from the following relationships:

*Estimate:


*Std. Error:


*t value:


*Pr(>|t|):
P-Value=Pr(|t|>tn-p-1)

*Collinearity Diagnostics:
In the Collinearity Diagnostics table, the results of multicollinearity in a set of multiple regression variables are given.
For each independent variable, the VIF index is calculated in two steps:
STEP1:
First, we run an ordinary least square regression that has xi as a function of all the other explanatory variables in the first equation.
STEP2:
Then, calculate the VIF factor with the following equation:
Where is the coefficient of determination of the regression equation in step one. Also,

A rule of decision making is that if then multicollinearity is high (a cutoff of 5 is also commonly used)

F. Save Regression:

By clicking this button, you can save the regression results. After opening the save results window, you can save the results in “text” or “Microsoft Word” format.

G. Run Diagnostic Tests:

In SUR Regression, For each equation, we shall make the following assumptions:
1-The residuals are independent of each other.
2- The residuals have a common variance
3- It is sometimes additionally assumed that the errors have the normal distribution. For example, if the residuals are normal, the correlation test can replace the independence test.
The following tests are provided to test these assumptions:

* Test for Serial Correlation:
-Box Ljung:
The test statistic is:

Where


The sample autocorrelation at lag and is the number of lags being tested. Under , .
In this software, we give the results for q=1.

*Test for Heteroscedasticity:
-Breusch Pagan:
The test statistic is calculated in two steps:
STEP1:
Estimate the regression:

Where ,
STEP2:
Calculate the test statistic:
Under , .

*Normality Tests:
-Shapiro Wilk:
The Shapiro-Wilk Test uses the test statistic

Where the values are given by:


is made of the expected values of the order statistics of independent and identically distributed random variables sampled from the standard normal distribution; finally, V is the covariance matrix of those normal order statistics. is compared against tabulated values of this statistic's distribution.

-Jarque Bera:
The test statistic is defined as

Where


Under , .

-Anderson Darling:
The test statistic is given by:

Where F(⋅) is the cumulative distribution of the normal distribution. The test statistic can then be compared against the critical values of the theoretical distribution.

-Cramer Von Mises:
The test statistic is given by:

Where F(⋅) is the cumulative distribution of the normal distribution. The test statistic can then be compared against the critical values of the theoretical distribution.

G.1. Result of Diagnostic Tests:

The test results for each face model are broken down as follows:
*****Model with Dependent Variable "model dependent variable name ”**********
The results show that in all tests, the P-value is greater than 0.05 (Except in Heteroskedasticity, the first model). Therefore, at 0.95 confidence level, the normality, homogeneity of variance, and independence of the regression model residues(Except in Heteroskedasticity, the first model) are confirmed.

H. Save Diagnostic Tests:

By clicking this button, you can save the results of the diagnostic tests. After opening the save results window, you can save the results in “text” or “Microsoft Word” format.

I. Add Residuals & Name of Residual Variable:

By clicking on this button, you can save the remnants of the regression equations with the desired name(Name of Residual Variable). The default software for the residual names is “SURResid”.
This message will appear if the residuals are saved successfully:
“Name of Residual Variable Residuals added in Data Table.”
For example, “SURResid Residuals added in Data Table.”.