Abstract

In a standard linear regression model the explanatory variables, , are considered to be fixed and hence assumed to be free from errors. But in reality, they are variables and consequently can be subjected to errors. In the regression literature there is a clear distinction between outlier in the - space or errors and the outlier in the X-space. The later one is popularly known as high leverage points. If the explanatory variables are subjected to gross error or any unusual pattern we call these observations as outliers in the - space or high leverage points. High leverage points often exert too much influence and consequently become responsible for misleading conclusion about the fitting of a regression model, causing multicollinearity problems, masking and/or swamping of outliers etc. Although a good number of works has been done on the identification of high leverage points in linear regression model, this is still a new and unsolved problem in linear functional relationship model. In this paper, we suggest a procedure for the identification of high leverage points based on deletion of a group of observations. The usefulness of the proposed method for the detection of multiple high leverage points is studied by some well-known data set and Monte Carlo simulations.

Highlights

  • The linear functional relationship model (LFRM) is an extension of a linear regression model (LRM) which allows for sampling variability in the measurements of both the response and explanatory variables

  • Our main objective is to propose a new method for the identification of high leverage points in linear functional relationship model

  • After obtaining a method of finding the fixed-X values, we propose three different identification rules based on robust measures of leverages

Read more

Summary

Introduction

The linear functional relationship model (LFRM) is an extension of a linear regression model (LRM) which allows for sampling variability in the measurements of both the response and explanatory variables. In regression the model is poorly fitted because of the presence of outliers It is a common practice over the years to use residuals for the identification of outliers. Residuals are estimates of the true errors that occur in the Y-space We anticipate at this point that fitting of the LFRM could be even more complicated because here outliers could occur in the X-space more frequently than the linear regression model. Imon (2009) pointed out that in the presence of high leverage points the errors become heteroscedastic, they might produce big outliers as well Another way to deal with outliers is to use M-estimators (Mahdizadeh et al, 2020; Zamanzade et al.; 2020; Zamanzade et al, and Zamanzade et al, 2018).

Measures of Leverages
Estimation of the Fixed-X in LFRM
Identification of High Leverage Points in LFRM
Conclusions
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call