<< Chapter < Page Chapter >> Page >

8. The next step in the process is to generate the term we will use in the estimation of the grade regression to account for the potential sample selection bias. To do this we will need to find a reference in the literature that offers a clear description of what we need to do. As it turns out, a reasonable explanation of the appropriate estimation technique is available in Jimenez and Kugler (1987). Since much of what follows comes directly from this article, I highly recommend you read it yourself.

The gist of the method suggests that the potential sample bias is accounted for by an inverse Mills ratio for each of the categories. What we need to do is calculate:

λ ^ i = ϕ ( μ ^ j z ^ i ) ϕ ( μ ^ j + 1 z ^ i ) Φ ( μ ^ j + 1 z ^ i ) Φ ( μ ^ j z ^ i ) MathType@MTEF@5@5@+=feaagyart1ev2aaatCvAUfeBSjuyZL2yd9gzLbvyNv2CaerbuLwBLnhiov2DGi1BTfMBaeXatLxBI9gBaerbd9wDYLwzYbItLDharqqtubsr4rNCHbGeaGqiVu0Je9sqqrpepC0xbbL8F4rqqrFfpeea0xe9Lq=Jc9vqaqpepm0xbba9pwe9Q8fs0=yqaqpepae9pg0FirpepeKkFr0xfr=xfr=xb9adbaqaaeGaciGaaiaabeqaamaabaabaaGcbaGafq4UdWMbaKaadaWgaaWcbaGaamyAaaqabaGccqGH9aqpdaWcaaqaaiabew9aMnaabmaabaGafqiVd0MbaKaadaWgaaWcbaGaamOAaaqabaGccqGHsislceWG6bGbaKaadaqhaaWcbaGaamyAaaqaaiabgEHiQaaaaOGaayjkaiaawMcaaiabgkHiTiabew9aMnaabmaabaGafqiVd0MbaKaadaWgaaWcbaGaamOAaiabgUcaRiaaigdaaeqaaOGaeyOeI0IabmOEayaajaWaa0baaSqaaiaadMgaaeaacqGHxiIkaaaakiaawIcacaGLPaaaaeaacqqHMoGrdaqadaqaaiqbeY7aTzaajaWaaSbaaSqaaiaadQgacqGHRaWkcaaIXaaabeaakiabgkHiTiqadQhagaqcamaaDaaaleaacaWGPbaabaGaey4fIOcaaaGccaGLOaGaayzkaaGaeyOeI0IaeuOPdy0aaeWaaeaacuaH8oqBgaqcamaaBaaaleaacaWGQbaabeaakiabgkHiTiqadQhagaqcamaaDaaaleaacaWGPbaabaGaey4fIOcaaaGccaGLOaGaayzkaaaaaaaa@6799@

for the category that the individual actually is in. What we will do is calculate (6) for all of the categories and then sum the product of this number and a dummy variable indicating if a course is the highest math class completed by an individual. Since the dummy variables will equal 0 for math categories an individual is not in, the resulting sum will preserve the value of (6) that is associated with the category the individual does belong to.

It is clear from (6) that we will need to retain the 6 cutoffs. We can do this with the commands:

. generate cutoff1 = _b[_cut1]

. generate cutoff2 = _b[_cut2]

. generate cutoff3 = _b[_cut3]

. generate cutoff4 = _b[_cut4]

. generate cutoff5 = _b[_cut5]

. generate cutoff6 = _b[_cut6]

Technically, this step is not necessary since the parameter estimates are preserved until the next regression is estimated; I suggest doing this purely as a precaution.

9. Preserve the predicted values of the ordered-probit using the command:

. predict zhat, xb

. predict phat1 phat2 phat3 phat4 phat5 phat6 phat7, p

These two commands will generate for each observation the predicted mean category of math classes and the probability that this individual will fall in each category. To see what is going on we will retrieve some representative values of these variables and then graph them for one individual. Table 7 reports these values for 10 individuals in the sample. Now consider individual 2. Fitting a normal distribution with a mean of 4.25 and using the critical values from our estimation yields the probabilities that the individual is in each of the categories. For example, the probability that individual 1 will have completed no math classes is equal to 0.1223. Figure 5 illustrates the results for individual 1. The dashed vertical lines are the six cutoff values that are the same for each individual. The solid vertical line is the zhat for individual 1. The heavy blue line represents the normal probability density function for this individual. While, there is, of course, a different probability distribution for each individual, the cutoff values are the same for all members of the sample.

Predicted values of the ordered probit regression.
Observation Highest Math Class zhat Pr(0) Pr(1) Pr(2) Pr(3) Pr(4) Pr(5) Pr(6)
1 3 3.9657 0.1890 0.0824 0.0194 0.4467 0.0816 0.0568 0.1241
2 0 4.2507 0.1217 0.0640 0.0158 0.4355 0.0975 0.0731 0.1923
165 0 3.5982 0.3036 0.1011 0.0225 0.4149 0.0575 0.0364 0.0640
166 6 4.6914 0.0540 0.0370 0.0098 0.3633 0.1097 0.0922 0.3340
214 3 3.4533 0.3560 0.1056 0.0229 0.3900 0.0483 0.0294 0.0478
215 3 4.0840 0.1587 0.0749 0.0180 0.4459 0.0887 0.0637 0.1501
225 3 3.5250 0.3296 0.1036 0.0228 0.4031 0.0528 0.0328 0.0553
226 3 3.6990 0.2693 0.0969 0.0219 0.4285 0.0641 0.0417 0.0776
453 3 3.9713 0.1875 0.0820 0.0194 0.4468 0.0819 0.0571 0.1253
454 5 4.1650 0.1399 0.0697 0.0170 0.4422 0.0932 0.0684 0.1697
495 3 4.4168 0.0913 0.0533 0.0135 0.4151 0.1043 0.0816 0.2409
496 0 2.9811 0.5410 0.1055 0.0212 0.2797 0.0236 0.0127 0.0162
526 0 2.9247 0.5633 0.1039 0.0207 0.2653 0.0214 0.0114 0.0141
527 3 3.9757 0.1863 0.0817 0.0193 0.4469 0.0822 0.0574 0.1262

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Econometrics for honors students. OpenStax CNX. Jul 20, 2010 Download for free at http://cnx.org/content/col11208/1.2
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Econometrics for honors students' conversation and receive update notifications?

Ask