• Stepwise Procedures In Discriminant Analysis

  • CHAPTER THREE -- [Total Page(s) 5]

    Page 1 of 5

    1 2 3 4 5    Next
    • CHAPTER THREE

      RESEARCH METHODOLOGY

      STEPWISE METHODOLOGIES IN DISCRIMINANT ANALYSIS

      Stepwise discriminant analysis is just like discriminant analysis. The only difference is the method of variables’ selection. It shares the same assumptions as well as tests with discriminant analysis. This is why deep knowledge on discriminant analysis is very necessary before one can apply the stepwise method.

      Here, the independent variables are selected step by step (one at a time) on the basis of their discriminatory powers without the alteration of the model. Unlike in multiple regressions analysis, where independent variables are selected and the model altered as a result.

      (1) The forward selection is based on Wilks’ ^ . At the first step we

      calculate ^ (xi) for each individual variable and choose the one with

      minimum         (xi) or maximum associated F.

      With    m         =          degrees of freedom for error

      n          =          degrees of freedom for hypothesis

      P          =          Number of variables (dimension)

      The partial ^ -Statistics and the partial F are distributed as ^ 1,n,m-P+1 and Fn, m - P+1, respectively. From the foregoing, in step one, the partial ^ (xi) and the accompanying partial F are found for each individual variable and the one with maximum partial F and minimum ^ (xi) is selected. x1 shows the variable selected at step 1 which is the variable with maximum F and minimum (xi).

      At step two, we calculate ^ (xi/x1) for each of the P - 1 variables not entered at the first step, with x1 indicating the variable first entered. We then choose the next variable with maximum associated F or minimum ^ (xi/xi) or the variable which adds to the maximum separation to the one entered at step 1. The variable entered at step 2 is denoted by x2.

      The above process is repeated for other available variables until the F falls below some predetermined threshold value. If the partial F is less than a second threshold value, Fout, the variable with the smallest partial F will be removed. There is also a very small possibility that the procedure will cycle continuously if Fout = F1n, though using a value of Fout slightly less than Fin, removes this possibility.

      THE F-DISRTIBUTION

      The F-Distribution is named after R.A. Fisher who originally developed it (Oyeka, 1996). This is also used in testing of hypothesis involving population means and variances. Let a random variable U be distributed as Chi-square with r1 degrees of freedom, and another random variable V, independent of U, be distributed as Chi-Square with r2 degrees of freedom. Then;

      u/r

      F = yt1 is said to have an F-distribution

      /l2

      with r1 and r2 degrees of freedom.

      F-distribution is asymmetric and is generally, strongly skewed to the right. This is because the shape of F-distribution varies considerably with Changes in its degrees of freedom, r1 and r2 .

      The cumulative probability distribution of F for the three commonly used probability values (^ = 0.05, 0.1 and 0.01), can be obtained from the Table of the F-Distribution. In the Table of the F-Distribution, the values in the main body of the table are Fi ri, r2, where â– => is the proportion of the area under the curve to the right hand tail of the F-Distribution with r1 = degrees of freedom of the numerator and r2 = degrees of freedom of the denominator. The F-ratio is obtained at the intersection of the column labeled r1 and the row labelled r2 in the Table of the F-Distribution.

      F-Distribution is also related to Normal, x2 and t distributions based on the assumption that the underlying population is normally distributed.

      The F-variable is a ratio of two independent Chi-square variables, each divided by its degrees of freedom as;

      Sample size from population one (1)

  • CHAPTER THREE -- [Total Page(s) 5]

    Page 1 of 5

    1 2 3 4 5    Next
    • ABSRACT - [ Total Page(s): 1 ] Abstract Several multivariate measurements require variables selection and ordering. Stepwise procedures ensure a step by step method through which these variables are selected and ordered usually for discrimination and classification purposes. Stepwise procedures in discriminant analysis show that only important variables are selected, while redundant variables (variables that contribute less in the presence of other variables) are discarded. The use of stepwise procedures ... Continue reading---

         

      APPENDIX A - [ Total Page(s): 1 ] ... Continue reading---

         

      APPENDIX B - [ Total Page(s): 1 ] APPENDIX II BACKWARD ELIMINATION METHOD The procedure for the backward elimination of variables starts with all the x’s included in the model and deletes one at a time using a partial  or F. At the first step, the partial  for each xi isThe variable with the smallest F or the largest  is deleted. At the second step of backward elimination of variables, a partial  or F is calculated for each q-1 remaining variables and again, the variable which is th ... Continue reading---

         

      TABLE OF CONTENTS - [ Total Page(s): 1 ]TABLE OF CONTENTSPageTitle PageApproval pageDedicationAcknowledgementAbstractTable of ContentsCHAPTER 1: INTRODUCTION1.1    Discriminant Analysis1.2    Stepwise Discriminant analysis1.3    Steps Involved in discriminant Analysis1.4    Goals for Discriminant Analysis1.5    Examples of Discriminant analysis problems1.6    Aims and Obj ectives1.7    Definition of Terms1.7.1    Discriminant function1.7.2    The eigenvalue1.7.3    Discriminant Score1.7.4    Cut off1.7 ... Continue reading---

         

      CHAPTER ONE - [ Total Page(s): 2 ] DEFINITION OF TERMS Discriminant Function This is a latent variable which is created as a linear combination of discriminating variables, such that Y =      L1X1 + L2X2 +          + Lp Xp where the L’s are the discriminant coefficients, the x’s are the discriminating variables. The eigenvalue: This is the ratio of importance of the dimensions which classifies cases of the dependent variables. There is one eigenvalue for each discriminant functio ... Continue reading---

         

      CHAPTER TWO - [ Total Page(s): 3 ] 5 is called the mahalanobis (squared) distance for known parameters. For unknown parameters, the Mahalanobis (squared) distance is obtained by estimating p1, p2 and S by X1, X2 and S, respectively. Following the same technique the Mahalanobis (Squared) distance, D , for the unknown parameters is D2 = (X- X)+S-1 (X1- X2) . The distribution of D can be used to test if there are significant differences between the two groups.2.4 WELCH’S CRITERION Welch (1939) suggest ... Continue reading---

         

      CHAPTER FOUR - [ Total Page(s): 3 ]CHAPTER FOUR DATA ANALYSISMETHOD OF DATA COLLECTIONThe data employed in this work are as collected by G.R. Bryce andR.M. Barker of Brigham Young University as part of a preliminary study of a possible link between football helmet design and neck injuries.Five head measurements were made on each subject, about 30 subjects per group:Group 1    =    High School Football players Group 2    =    Non-football playersThe five variables areWDIM    =    X1    =    head width at wi ... Continue reading---

         

      CHAPTER FIVE - [ Total Page(s): 1 ]CHAPTER FIVERESULTS, CONCLUSION AND RECOMMENDATIONRESULTSAs can be observed from the results of the analysis, when discriminant analysis was employed, the variable CIRCUM(X2) has the highest Wilks’ lambda of 0.999 followed by FBEYE (X2) (0.959). The variable EYEHD (X4) has the least Wilks’ lambda of 0.517 followed by EARHD (X5) (0.705). Also the least F-value was recorded with the variable CIRCUM (X2) (0.074) followed by the variable FBEYE (X2) (2.474), while the variable EYEHD (X4 ... Continue reading---

         

      REFRENCES - [ Total Page(s): 1 ] REFERENCES Anderson, T.W. (1958). An introduction to multivariate statistical Analysis. John Wiley & Sons Inc., New York. Cohen, J. (1968). Multiple regression as a general data-analytic system. Psychological Bulletin 70, 426-443. Cooley W.W. and Lohnes P.R. (1962). Multivariate procedures for the Behavioural Sciences, New York John Wiley and Sons Inc. Efroymson, M.A. (1960). Multiple regression analysis. In A. Raston & H.S. Wilfs (Eds.) Mathematical methods for ... Continue reading---