透過您的圖書館登入
IP:18.117.111.1
  • 學位論文

Gram-Schmidt轉換過程最小化之演算法及其應用在具有共線性之迴歸分析

Gram-Schmidt Transformation Minimization Algorithm and Its Applications to Regression Analysis with Multicollinearity

指導教授 : 陳正剛
若您是本文的作者,可授權文章由華藝線上圖書館中協助推廣。

摘要


迴歸分析是最常被使用的統計方法,然而,在迴歸分析中常會遇到的問題就是共線性,共線性是來自於獨立變變數之間的高度相關所引起的,也就是說,在資料向量中,存在有微小角度的問題存在。 文獻上,有兩種用來垂直化向量群的方法,一個是知名的葛蘭-史密特垂直化過程(Gram-Schmidt Process),另一個是由R. M. Johnson學者在1966年所提出來的,我們稱之為R. M. Johnson method,然而,Gram-Schmidt Process沒有一個有意義的機制來決定向量群垂直化的優先順序,而經由R. M. Johnson method所轉換出來的垂直向量群也無法具有解釋能力,特別是在具在有高相關的資料向量的情況中。 本篇研究中,我們嘗試去發展一個演算法來決定向量群垂直化的優先順序並且達到資訊轉換最小化,稱之為Gram-Schmidt轉換過程最小化演算法(Gram-Schmidt Transformation Minimization algorithm, GSTM algorithm),它把在向量投影過程中的資訊轉換最小化,但是在執行GSTM algorithm之前,有一些前置處理需要先進行,進行完之後,再針對這些資料向量執行GSTM algorithm,並在這些垂直向量群執行迴歸分析,最後再針對這些分析結果做解釋。 我們發現此演算法不僅克服在迴歸分析中共線性的問題和最小化在向量投影過程中的資訊轉換,也使得分析的結果更具有解釋能力。

並列摘要


Regression analysis is the most used statistical method. However, we may encounter the multicollinearity problem in regression analysis. Multicollinearity is due to high correlation among independent variables, namely, small angles among data vectors of the independent variables. In the literature, there are two methods to orthogonalize vectors. One is the well-known Gram-Schmidt Process and the other is a method proposed by R.M. Johnson in 1966, referred to as the R.M. Johnson method. However, the Gram-Schmidt Process has no meaningful mechanism to determine the sequence order of vector orthogonalization; while the results transformed by the R.M. Johnson method can not be interpreted meaningfully, especially in a case with highly correlated data vectors. In this research, we attempt to develop an algorithm to determine the sequence order of the Gram-Schmidt Process with minimized transformation, called the Gram-Schmidt Transformation Minimization (GSTM) algorithm. It minimizes information subtraction during the vector projection processes. But before performing the GSTM algorithm, some procedures need to be done first. After those procedures, we perform the GSTM algorithm on data vectors, and with the orthogonalized data vectors, we perform regression analysis. Finally, we interpret the analysis results in regression analysis by the GSTM algorithm. We find that this proposed algorithm not only overcomes the multicollinearity problem in regression analysis and minimizes information subtraction during the vector projection processes but also makes the analysis results more interpretable.

參考文獻


1. James W. Longley, An Appraisal of Least Squares Programs for the Electronic Computer from the Point of View of the User, Journal of the American Statistical Association, Vol. 62, No. 319, (Sep., 1967), pp. 819- 841.
2. Gilbert Strang, Linear Algebra and its applications, pp.174-185.
3. R.M. Johnson, The Minimal Transformation to Orthonormality, Psychometrika, 1966.
5. Bowerman, B. L., O’Connell, R. T. & Richard, T. (1993). Forecasting and Time Series: An Applied Approach. Belmont, CA Wadsworth.
8. Feng-Jenq Lin, Solving Multicollinearity in the Process of Fitting Regression Model Using the Nested Estimate Procedure, Department of Applied Economics National I-Lan University

延伸閱讀