Regression Estimators under Joint Multicollinearity and Autocorrelation Conditions: The Two-Stage Kibria-Lukman Estimator as an Enhanced Approach

Authors

  • Ayanlowo E. Ayanlola Department of Basic Sciences, Babcock University. Ilishan-Remo, Ogun State. Nigeria Author
  • Oladapo D. Ifedayo Department of Mathematical Sciences, Adeleke University, Ede, Osun State. Nigeria Author
  • Obadina, G.Olugbenga Department of Statistics, Olabisi Onabanjo University, Ago-Iwoye, Ogun State, Nigeria Author
  • Madu, N. Peter Federal University of Agriculture Abeokuta (INHURD) Author
  • Odeyemi, A. Sunday Department of Statistics University of Fort Hare Alice, Eastern Cape, South Africa Author

DOI:

https://doi.org/10.62054/ijdm/0201.17

Keywords:

Multicollinearity, Autocorrelation, Two-Stage Kibria-Lukman (KL) Estimator, Mean Squared Error (MSE), Biased Regression Estimators

Abstract

Multicollinearity among predictors and autocorrelation in residuals present significant challenges to the reliability and accuracy of linear regression models. These issues cause traditional Ordinary Least Squares (OLS) estimators to yield inflated variances and biased parameter estimates, ultimately leading to unreliable statistical inferences. To address these limitations, various biased estimators have been developed. This paper investigates the performance of several such estimators, including the Ridge, Liu, Kibria-Lukman (KL), and the newly proposed Two-Stage Kibria-Lukman (Two-Stage KL) estimator. The Two-Stage KL estimator integrates the Prais-Winsten transformation, which corrects for autocorrelation, with the KL estimator’s biasing mechanism to reduce the inflated variances caused by multicollinearity. Using extensive Monte Carlo simulations, we evaluate the performance of these estimators in settings characterized by varying levels of multicollinearity (predictor correlation values, ρX , of 0.8, 0.9, and 0.99) and autocorrelation (residual autocorrelation values, ρ, of 0.6, 0.8, and 0.9), across sample sizes ranging from 25 to 500. The simulations reveal that OLS is highly sensitive to these conditions, with Mean Squared Error (MSE) values reaching as high as 738.6690 in extreme multicollinearity (ρX=0.99) and autocorrelation (ρ=0.9) at a sample size of 50. In contrast, the Two-Stage KL estimator consistently achieves the lowest MSE values, reducing the error to 265.3667 under the same conditions. For moderate multicollinearity (ρX=0.8) and autocorrelation (ρ=0.8), and a sample size of 50, OLS yields an MSE of 1.254, while the Two-Stage KL estimator reduces this to 0.764, outperforming both Ridge and Liu estimators, which record MSEs of 0.953 and 0.902, respectively. In empirical testing using the Portland cement dataset, which is known for its multicollinearity, the Two-Stage KL estimator provides the lowest MSE of 0.0486, compared to OLS (0.0638), Ridge (0.0581), Liu (0.0554), and KL (0.0522). These results demonstrate that the Two-Stage KL estimator effectively mitigates the effects of both multicollinearity and autocorrelation, offering a robust solution for regression models where these conditions co-occur. The integration of the Prais-Winsten transformation with the KL biasing approach allows the Two-Stage KL to maintain low error rates, even in high-dimensional and high-correlation settings.

References

Arkorful, B. (2023). Regularized and Robust Regression Methods of Linear Model with Multicollinear Predictors and Autocorrelated Errors (Doctoral dissertation, University of Cape Coast).

Dertli, H. I., Hayes, D. B., and Zorn, T. G. (2024). Effects of multicollinearity and data granularity on regression models of stream temperature. Journal of Hydrology, 639, 131572.

Hoerl, A.E., and Kennard, R.W. (1970). Ridge Regression: Biased Estimation for Non-Orthogonal Problems. Technometrics, 12, 55–67.

Hwang, T., and Vogelsang, T. J. (2024). An Estimating Equation Approach for Robust Confidence Intervals for Autocorrelations of Stationary Time Series.

Kibria, B. M. G., and Lukman, A. F. (2020). A New Ridge-Type Estimator for the Linear Regression Model: Simulations and Applications. Hindawi Scientifica.

Oyewole, O., and Obadina, O. (2020). Monte Carlo Approach for Comparative Analysis of Regression Techniques in the Presence of Multicollinearity and Autocorrelation Phenomena. Fudma Journal of Sciences, 4(1), 770-778.

Prais, S. J., and Winsten, C. B. (1954). Trend Estimators and Serial Correlation. Cowles Commission Discussion Paper.

Shrestha, N. (2020). Detecting multicollinearity in regression analysis. American Journal of Applied Mathematics and Statistics, 8(2), 39-42.

Downloads

Published

2025-04-02

Data Availability Statement

The research data used in this study was generated through simulation based on predefined statistical parameters and models. As such, it does not represent real-world observations but is reproducible using the methodological details provided in the manuscript. Readers can recreate the dataset by following the simulation procedures described in the methodology section.

How to Cite

Regression Estimators under Joint Multicollinearity and Autocorrelation Conditions: The Two-Stage Kibria-Lukman Estimator as an Enhanced Approach. (2025). International Journal of Development Mathematics (IJDM), 2(1), 217-227. https://doi.org/10.62054/ijdm/0201.17

Similar Articles

11-20 of 37

You may also start an advanced similarity search for this article.