A COMPANION TO Theoretical Econometrics
Experimental Design and Results
This section details the design and results of a Monte Carlo experiment that we conducted to shed light on the performance of the Gibbs sampling algorithm discussed in Section 2. We generated data according to equations (22.6)-(22.12) using the true parameter values that are listed in column two of Table 22.3. The table does not list the discount rate and the intercepts in the school and home payoff functions, which were set to 0.95, 11,000, and 17,000 respectively, since these are not identified. In all of our experiments we set the number of people, N, to 2,000.
Data from this model were generated using two different assumptions about the way people formed expectations. First, we assumed that people had rational expectations. This required us to solve the resulting dynamic optimization problem once to generate the optimal decision rules. Since the choice set includes only four discrete alternatives it is feasible to do this. Then, to simulate choice and wage paths requires only that we generate realizations of the appropriate stochastic variables. It is important to note that the polynomial future component used in the estimation procedure does not provide a perfect fit to the rational expectations future component. Hence, analysis of this data sheds light on the effect that misspecification of the future component may have on inference.
Next, we assumed that agents used a future component that was actually a polynomial in the state variables to form decisions. Analysis of this data set sheds light on how the algorithm performs when the model is correctly specified. To ensure close comparability with the rational expectations case, we constructed this polynomial by regressing the rational expectations future components on a fourth-order polynomial in the state variables, constructed as described in the discussion preceding (22.13). We used the point estimates from this regression as the coefficients of our polynomial future component (see Appendix A for the specific form of the polynomial).
We found that a fourth-order polynomial provided a good approximation to the future component in the sense that if agents used the approximate instead of optimal decision rule they suffered rather small lifetime earnings losses. Evidence of this is given in Table 22.1, where we report the results of simulations
Table 22.1 Quality of the polynomial approximation to the true future component
* The mean present value of payoffs is the equally-weighted average discounted sum of ex-post lifetime payoffs over 2,000, 40 period lived agents. The values are dollar equivalents. |
under optimal and suboptimal decision rules. The simulations were conducted as follows. First, for N = 2,000 people we drew five sets of lifetime (T = 40) realizations of the model's stochastic components {ei1t, ei2t, (%)у=1,4}. In Table 22.1 these are referred to as error sets one to five. For each of the five error sets we simulated lifetime choice histories for each of the 2,000 people under the optimal and approximate decision rules. We refer to the 10 data sets constructed in this way as 1-EMAX through 5-EMAX and 1-POLY through 5-POLY, respectively. We then calculated the mean of the present value of lifetime payoffs (pecuniary plus nonpecuniary) for each of the 2,000 people under the optimal and approximate decision rules, respectively, for each of the five error sets. These are reported in the second and third rows of Table 22.1. Holding the error set fixed, the source of any difference in the mean present value of lifetime payoffs lies in the use of different decision rules. The mean present values of dollar equivalent losses from using the suboptimal polynomial rules are small, ranging from 287 to 491. The percentage loss ranges from 8 hundredths of 1 percent to 14 hundredths of 1 percent. These findings are similar to those reported by Geweke and Keane (1999a) and Krusell and Smith (1995).
Table 22.2 reports the mean accepted wages and choice frequencies for the data generated from error-set two. The first set of columns report statistics for data generated according to the polynomial approximation (data set 2-POLY) while the second set of columns report results from the optimal decision rule (data set 2-EMAX). Under our parameterization, occupation one can be thought of as "unskilled" labor, while occupation two can be understood as "skilled" labor. The reason is the mean of the wage offer distribution is lower in occupation two early in life, but it rises more quickly with experience. The choice patterns and mean accepted wages are similar under the two decision rules. School is chosen somewhat more often under the optimal decision rule, which helps to generate slightly higher lifetime earnings. Finally, note that selection effects leave the mean accepted wage in occupation two higher than that in occupation one throughout the lifecycle under both decision rules.
Next, we ran the Gibbs algorithm described in section two for 40,000 cycles on each data set. We achieved about three cycles per minute on a Sun ultra-2 workstation.5 Thus, while time requirements were substantial, they were minor compared to what estimation of such a model using a full solution of the dynamic programming problem would entail. Visual inspection of graphs of the draw sequences, as well as application of the split sequence diagnostic suggested by Gelman (1996) - which compares variability of the draws across subsequences - suggests that the algorithm converged for all 10 artificial data sets. In all cases, the final 15,000 draws from each run were used to simulate the parameters' marginal posterior distributions.
Table 22.3 reports the results of the Gibbs sampling algorithm when applied to the data generated with a polynomial future component. In this case, the econometric model is correctly specified. The first column of Table 22.3 is the parameter label, the second column is the true value, and the remaining columns report the structural parameters' posterior means and standard deviations for each of the five data sets.6 The results are extremely encouraging. Across all runs, there was only one instance in which the posterior mean of a parameter for the first wage equation was more than two posterior standard deviations away from its true value: the intercept in data set one. In data sets four and five, all of the structural parameters' posterior means are within two posterior standard deviations of their true values. In the second data set, only the second wage equation's own experience term is slightly more than two posterior standard deviations from its true value. In the third data set the mean of the wage equation's error correlation is slightly more than two posterior standard deviations from the true value, as are a few of the second wage equation's parameters.
Careful examination of Table 22.3 reveals that the standard deviation of the nonpecuniary payoff was the most difficult parameter to pin down. In particular, the first two moments of the marginal posteriors of these parameters vary considerably across experiments, in relation to the variability of the other structural parameters' marginal posteriors. This result reflects earlier findings reported by Geweke and Keane (1999a). In the earlier work they found that relatively large changes in the value of the nonpecuniary component's standard deviation had only a small effect on choices. It appears that this is the case in the current experiment as well.
It is interesting to note that an OLS regression of accepted (observed) log - wages on the log-wage equation's regressors yields point estimates that differ sharply from the results of the Gibbs sampling algorithm. Table 22.4 contains point estimates and standard errors from such an accepted wage regression. Selection bias is apparent in the estimates of the log-wage equation's parameters in all data sets. This highlights the fact that the Bayesian simulation algorithm is doing an impressive job of implementing the appropriate dynamic selection correction.
Perhaps more interesting is the performance of the algorithm when taken to data that were generated using optimal decision rules. Table 22.5 reports the results of this analysis on data sets 1-EMAX to 5-EMAX. Again, the first column labels the parameter, and the second contains its data generating value. The performance of the algorithm is quite impressive. In almost all cases, the posterior means of the wage function parameters deviate only slightly from the true values in percentage terms. Also, the posterior standard deviations are in most cases quite small, suggesting that the data contain a great deal of information about these structural parameters - even without imposing the assumption that agents form the future component "optimally." Finally, despite the fact that the posterior standard deviations are quite small, the posterior means are rarely more than two posterior deviations away from the true values.7 As with the polynomial data, the standard deviation of the nonpecuniary component seems difficult to pin down. Unlike the polynomial data, the school payoff parameters are not pinned down as well as the wage equation parameters. This is perhaps not surprising since school payoffs are never observed.
Figure 22.1 contains the simulated posterior densities for a subset of the structural parameters based on data set 3-EMAX. Each figure includes three triangles on its horizontal axis. The middle triangle defines the posterior mean, and the two flanking triangles mark the points two posterior standard deviations above and below the mean. A vertical line is positioned at the parameters' data generating (true) value. These distributions emphasize the quality of the algorithm's performance in that the true parameter values are typically close to the posterior means. The figures also make clear that not all the parameters have approximately normal distributions. For instance, the posterior density of the wage equations' error correlation is multi-modal.
The results of Table 22.5 indicate that in a case where agents form the future component optimally, we can still obtain reliable and precise inferences about structural parameters of the current payoff functions using a simplified and misspecified model that says the future component is a simple fourth-order polynomial in the state variables. But we are also interested in how well our method approximates the decision rule used by the agents. In Table 22.6 we consider an experiment in which we use the posterior means for the parameters n that
Period Data Set 2 - POLY Data Set 2 - EMAX
|
17 |
0.75 |
0.17 |
18 |
0.73 |
0.17 |
19 |
0.72 |
0.19 |
20 |
0.74 |
0.19 |
21 |
0.71 |
0.23 |
22 |
0.70 |
0.25 |
23 |
0.67 |
0.29 |
24 |
0.66 |
0.30 |
25 |
0.66 |
0.30 |
26 |
0.62 |
0.34 |
27 |
0.62 |
0.34 |
28 |
0.60 |
0.37 |
29 |
0.57 |
0.40 |
ЗО |
0.55 |
0.42 |
31 |
0.52 |
0.45 |
32 |
0.50 |
0.48 |
33 |
0.46 |
0.50 |
34 |
0.43 |
0.54 |
35 |
0.42 |
0.55 |
36 |
0.39 |
0.58 |
37 |
0.36 |
0.60 |
38 |
0.33 |
0.64 |
39 |
0.28 |
0.67 |
40 |
0.26 |
0.70 |
0.02 |
0.06 |
17,886.20 |
0.02 |
0.07 |
18,408.75 |
0.02 |
0.06 |
19,590.88 |
0.02 |
0.05 |
20,186.07 |
0.01 |
0.05 |
21,113.74 |
0.01 |
0.04 |
22,002.82 |
0.01 |
0.03 |
23,259.72 |
0.00 |
0.03 |
23,119.46 |
0.00 |
0.04 |
24,085.78 |
0.01 |
0.04 |
25,399.34 |
0.00 |
0.04 |
26,971.71 |
0.00 |
0.03 |
27,074.62 |
0.00 |
0.03 |
29,049.11 |
0.00 |
0.04 |
30,492.25 |
0.00 |
0.03 |
30,745.54 |
0.00 |
0.03 |
32,078.16 |
0.00 |
0.04 |
34,202.82 |
0.00 |
0.02 |
34,578.60 |
0.00 |
0.04 |
37,084.91 |
0.00 |
0.03 |
37,580.47 |
0.00 |
0.04 |
40,129.34 |
0.00 |
0.03 |
40,101.57 |
0.00 |
0.05 |
43,282.44 |
0.00 |
0.03 |
44,462.69 |
Parameter |
True |
Data Set 1 |
-POLY |
Data Set 2 |
-POLY |
Data Set 3 |
-POLY |
Data Set 4 |
-POLY |
Data Set 5 |
-POLY |
Mean |
SD |
Mean |
SD |
Mean |
SD |
Mean |
SD |
Mean |
SD |
||
Occ. 1 intercept |
9.00000 |
9.01300 |
0.00643 |
9.00125 |
0.00797 |
9.00845 |
0.00600 |
9.00256 |
0.00661 |
9.00309 |
0.00598 |
Occ. 1 own experience |
0.05500 |
0.05440 |
0.00080 |
0.05540 |
0.00086 |
0.05429 |
0.00076 |
0.05462 |
0.00080 |
0.05496 |
0.00060 |
Occ. 2 experience |
0.00000 |
0.00093 |
0.00095 |
-0.00121 |
0.00103 |
-0.00084 |
0.00092 |
0.00114 |
0.00112 |
-0.00129 |
0.00115 |
Education |
0.05000 |
0.04806 |
0.00130 |
0.04881 |
0.00132 |
0.04850 |
0.00137 |
0.04938 |
0.00140 |
0.04924 |
0.00139 |
Occ. 1 exp. squared |
-0.00025 |
-0.00024 |
0.00003 |
-0.00026 |
0.00003 |
-0.00025 |
0.00003 |
-0.00024 |
0.00003 |
-0.00026 |
0.00002 |
Occ. 1 error SD |
0.40000 |
0.398 |
0.002 |
0.402 |
0.002 |
0.401 |
0.002 |
0.400 |
0.002 |
0.400 |
0.002 |
Occ. 2 intercept |
8.95000 |
8.91712 |
0.01625 |
8.97693 |
0.01582 |
8.91699 |
0.01583 |
8.96316 |
0.01551 |
8.92501 |
0.01590 |
Occ. 2 own experience |
0.04000 |
0.04049 |
0.00039 |
0.03918 |
0.00034 |
0.03946 |
0.00037 |
0.03968 |
0.00037 |
0.04055 |
0.00042 |
Occ. 1 experience |
0.06000 |
0.06103 |
0.00175 |
0.06016 |
0.00178 |
0.06461 |
0.00173 |
0.05946 |
0.00180 |
0.06009 |
0.00178 |
Education |
0.07500 |
0.07730 |
0.00187 |
0.07245 |
0.00161 |
0.07782 |
0.00178 |
0.07619 |
0.00167 |
0.07650 |
0.00171 |
Occ. 2 exp. squared |
-0.00090 |
-0.00088 |
0.00008 |
-0.00093 |
0.00008 |
-0.00109 |
0.00008 |
-0.00092 |
0.00008 |
-0.00087 |
0.00008 |
Occ. 2 error SD |
0.40000 |
0.409 |
0.003 |
0.399 |
0.003 |
0.407 |
0.003 |
0.397 |
0.003 |
0.399 |
0.003 |
Error correlation |
0.50000 |
0.481 |
0.031 |
0.512 |
0.025 |
0.420 |
0.033 |
0.528 |
0.037 |
0.438 |
0.042 |
Undergraduate tuition |
-5,000 |
-4,629 |
363 |
-5,212 |
464 |
-5,514 |
404 |
-4,908 |
464 |
-4,512 |
399 |
Graduate tuition |
-15,000 |
-18,006 |
2,085 |
-16,711 |
1,829 |
-16,973 |
1,610 |
-16,817 |
1,692 |
-15,091 |
1,972 |
Return cost |
-15,000 |
-14,063 |
531 |
-16,235 |
894 |
-15,809 |
554 |
-14,895 |
822 |
-15,448 |
679 |
Preference shock SD |
|||||||||||
Occ. 1 |
9,082.95 |
10,121.05 |
255.49 |
9,397.38 |
577.26 |
10,578.31 |
537.01 |
9,253.22 |
615.27 |
9,494.74 |
354.07 |
Occ. 2 |
9,082.95 |
8,686.12 |
456.77 |
11,613.38 |
346.03 |
10,807.31 |
519.13 |
9,610.23 |
457.45 |
9,158.09 |
228.32 |
Occ. 3 |
11,821.59 |
11,569.93 |
281.18 |
12,683.67 |
803.09 |
13,418.79 |
358.98 |
12,019.38 |
417.75 |
12,247.80 |
401.15 |
Preference shock Corr. Occ. 1 with Occ. 2 |
0.89 |
0.90 |
0.01 |
0.96 |
0.01 |
0.93 |
0.01 |
0.91 |
0.01 |
0.90 |
0.02 |
Occ. 1 with Occ. 3 |
0.88 |
0.89 |
0.01 |
0.88 |
0.01 |
0.90 |
0.01 |
0.88 |
0.01 |
0.88 |
0.01 |
Occ. 2 with Occ. 3 |
0.88 |
0.89 |
0.01 |
0.89 |
0.01 |
0.89 |
0.01 |
0.89 |
0.01 |
0.89 |
0.01 |
Occupation |
One |
Occupation |
Two |
Wage Error SDs |
||||||||
Data Set |
Intercept |
Occ. 1 Experience |
Occ. 2 Experience |
Education |
Occ. 1 Exp. Squared |
Intercept |
Occ. 1 Experience |
Occ. 2 Experience |
Education |
Occ. 2 Exp. Squared |
Occ. 1 |
Occ. 2 |
TRUE |
9.00000 |
0.05500 |
0.00000 |
0.05000 |
-0.00025 |
8.95000 |
0.04000 |
0.06000 |
0.07500 |
-0.00090 |
0.40000 |
0.40000 |
1 - POLY |
9.15261 0.00520 |
0.04236 0.00076 |
0.01708 0.00074 |
0.04247 0.00133 |
0.00012 0.00003 |
9.46735 0.00906 |
0.03516 0.00037 |
0.01953 0.00146 |
0.05589 0.00179 |
0.00038 0.00008 |
0.38845 |
0.36574 |
2 - POLY |
9.14715 0.00528 |
0.04320 0.00076 |
0.01586 0.00073 |
0.04309 0.00134 |
0.00010 0.00003 |
9.47924 0.00888 |
0.03446 0.00036 |
0.02261 0.00136 |
0.05311 0.00170 |
0.00017 0.00007 |
0.38940 |
0.36300 |
3 - POLY |
9.14895 0.00528 |
0.04230 0.00076 |
0.01732 0.00072 |
0.04420 0.00135 |
0.00011 0.00003 |
9.45851 0.00914 |
0.03482 0.00036 |
0.02335 0.00140 |
0.05665 0.00177 |
0.00016 0.00007 |
0.38935 |
0.36495 |
4 - POLY |
9.15157 0.00527 |
0.04220 0.00076 |
0.01734 0.00074 |
0.04261 0.00136 |
0.00012 0.00003 |
9.46413 0.00896 |
0.03480 0.00037 |
0.02346 0.00138 |
0.05469 0.00174 |
0.00015 0.00007 |
0.38900 |
0.36217 |
5 - POLY |
9.14838 0.00521 |
0.04274 0.00076 |
0.01695 0.00073 |
0.04408 0.00135 |
0.00011 0.00003 |
9.45131 0.00880 |
0.03570 0.00036 |
0.02021 0.00139 |
0.05671 0.00173 |
0.00035 0.00007 |
0.38772 |
0.35781 |
Standard errors in italics. |
Parameter |
True |
Data Set 1 |
-EMAX |
Data Set 2 |
-EMAX |
Data Set 3 |
-EMAX |
Data Set 4 |
-EMAX |
Data Set 5 |
-EMAX |
Mean |
SD |
Mean |
SD |
Mean |
SD |
Mean |
SD |
Mean |
SD |
||
Occ. 1 intercept |
9.00000 |
9.01342 |
0.00602 |
9.00471 |
0.00527 |
9.01436 |
0.00584 |
9.01028 |
0.00593 |
9.00929 |
0.00550 |
Occ. 1 own experience |
0.05500 |
0.05427 |
0.00073 |
0.05489 |
0.00071 |
0.05384 |
0.00072 |
0.05394 |
0.00072 |
0.05410 |
0.00071 |
Occ. 2 experience |
0.00000 |
0.00111 |
0.00093 |
0.00092 |
0.00114 |
0.00078 |
0.00126 |
0.00107 |
0.00100 |
0.00051 |
0.00093 |
Education |
0.05000 |
0.04881 |
0.00118 |
0.05173 |
0.00126 |
0.04869 |
0.00129 |
0.04961 |
0.00123 |
0.05067 |
0.00124 |
Occ. 1 exp. squared |
-0.00025 |
-0.00023 |
0.00002 |
-0.00025 |
0.00002 |
-0.00023 |
0.00002 |
-0.00022 |
0.00002 |
-0.00023 |
0.00002 |
Occ. 1 error SD |
0.40000 |
0.397 |
0.002 |
0.399 |
0.002 |
0.399 |
0.002 |
0.397 |
0.002 |
0.397 |
0.002 |
Occ. 2 intercept |
8.95000 |
8.90720 |
0.01704 |
8.98989 |
0.01970 |
8.93943 |
0.01850 |
8.93174 |
0.01649 |
8.94097 |
0.01410 |
Occ. 2 own experience |
0.04000 |
0.04093 |
0.00037 |
0.03967 |
0.00037 |
0.03955 |
0.00038 |
0.04001 |
0.00037 |
0.04060 |
0.00039 |
Occ. 1 experience |
0.06000 |
0.06087 |
0.00178 |
0.05716 |
0.00190 |
0.06200 |
0.00201 |
0.06211 |
0.00179 |
0.05880 |
0.00157 |
Education |
0.07500 |
0.07822 |
0.00166 |
0.07338 |
0.00171 |
0.07579 |
0.00165 |
0.07743 |
0.00167 |
0.07613 |
0.00159 |
Occ. 2 exp. squared |
-0.00090 |
-0.00087 |
0.00008 |
-0.00081 |
0.00008 |
-0.00098 |
0.00008 |
-0.00101 |
0.00008 |
-0.00084 |
0.00007 |
Occ. 2 error SD |
0.40000 |
0.409 |
0.003 |
0.397 |
0.003 |
0.404 |
0.003 |
0.402 |
0.003 |
0.397 |
0.003 |
Error correlation |
0.50000 |
0.517 |
0.023 |
0.607 |
0.029 |
0.484 |
0.044 |
0.521 |
0.035 |
0.488 |
0.028 |
Undergraduate tuition |
-5,000 |
-2,261 |
313 |
-2,937 |
358 |
-3,407 |
371 |
-3,851 |
426 |
-3,286 |
448 |
Graduate tuition |
-15,000 |
-10,092 |
1,046 |
-10,788 |
1,141 |
-11,983 |
1,188 |
-10,119 |
1,380 |
-11,958 |
1,823 |
Return cost |
-15,000 |
-14,032 |
482 |
-16,014 |
431 |
-16,577 |
500 |
-16,168 |
662 |
-18,863 |
1,065 |
Preference shock SD |
|||||||||||
Occ. 1 |
9,082.95 |
10,634.90 |
423.85 |
10,177.24 |
165.11 |
11,438.63 |
438.72 |
9,973.32 |
371.64 |
9,071.29 |
509.80 |
Occ. 2 |
9,082.95 |
9,436.10 |
372.86 |
12,741.02 |
405.25 |
11,432.19 |
287.69 |
9,310.37 |
718.15 |
7,770.66 |
555.39 |
Occ. 3 |
11,821.59 |
11,450.65 |
338.28 |
12,470.12 |
259.81 |
13,999.95 |
351.33 |
13,183.33 |
471.47 |
13,897.62 |
533.67 |
Preference shock corr. Occ. 1 with Occ. 2 |
0.89 |
0.93 |
0.01 |
0.98 |
0.00 |
0.94 |
0.01 |
0.91 |
0.02 |
0.86 |
0.03 |
Occ. 1 with Occ. 3 |
0.88 |
0.89 |
0.01 |
0.88 |
0.01 |
0.90 |
0.01 |
0.88 |
0.01 |
0.88 |
0.01 |
Occ. 2 with Occ. 3 |
0.88 |
0.87 |
0.01 |
0.90 |
0.01 |
0.90 |
0.01 |
0.89 |
0.02 |
0.89 |
0.02 |
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|
||
|
|||
|
|||
|
|||
|
|||
|
|||
|
|
||
|
|||
|
characterize how agents form expectations to form an estimate of agents' decision rules. We then simulate five new artificial data sets, using the exact same draws for the current period payoffs as were used to generate the original five artificial data sets. The only difference is that the estimated future component is substituted for the true future component in forming the decision rule. The results in Table 22.6 indicate that the mean wealth losses from using the estimated decision rule range from five-hundredths to three-tenths of 1 percent. The percentage of choices that agree between agents who use the optimal versus the approximate rules ranges from 89.8 to 93.5 percent. These results suggest that our estimated polynomial approximations to the optimal decision rules are reasonably accurate.
Figure 22.2 provides an alternative way to examine the quality of the polynomial approximation to the future component. This figure plots the value of the approximate and the true EMAX future components when evaluated at the mean
Table 22.6 Wealth loss when posterior polynomial approximation is used in place of true future component*
Data set Using Using
True EMAX** Posterior EMAX**
Mean present value of payoffs*** |
Mean present value of payoffs*** |
Mean dollar equivalent loss |
Mean percent loss (%) |
Aggregate choice agreement (%) |
Percent with 0-35 agreement (%) |
Percent with 36-39 agreements (%) |
Percent choosing same path (%) |
|
1-EMAX |
356,796 |
356,134 |
663 |
0.19 |
90.80 |
34.25 |
42.65 |
23.10 |
2-EMAX |
356,327 |
355,836 |
491 |
0.14 |
91.34 |
33.00 |
44.00 |
23.00 |
3-EMAX |
355,797 |
354,746 |
1,051 |
0.30 |
89.79 |
39.00 |
38.95 |
22.05 |
4-EMAX |
355,803 |
355,450 |
353 |
0.10 |
93.48 |
24.60 |
38.45 |
36.95 |
5-EMAX |
355,661 |
355,485 |
176 |
0.05 |
93.18 |
24.95 |
30.50 |
44.55 |
* Polynomial parameter values are set to the mean of their respective empirical posterior distributions. ** Each simulation includes 2,000 agents that live for exactly 40 periods. *** The mean present value of payoffs is the equal-weight sample average of the discounted streams of ex-post lifetime payoffs. |
|
Occupation two's EMAX and polynomial future component
25.000
20.000
15.000
10.000
5,000
0
of each period's state vector.8 Each vertical axis corresponds to the value of the future component, and the horizontal axis is the period. Clearly, the approximation reflects the true EMAX future component's main features. The fit of the polynomial seems relatively strong for each occupational alternative throughout the lifecycle. The fit is good for school in early periods, but begins to deteriorate later. One reason is that school is chosen very infrequently after the first five periods, so there is increasingly less information about its future component. A second reason is that the contemporaneous return begins to dominate the future component in alternative valuations. Consequently, each data point in later periods contains relatively less information about the future component's value.
Overall, however, these figures furnish additional evidence that the polynomial approximation does a reasonable job of capturing the key characteristics of the true future component.
This chapter described how to implement a simulation based method for inference that is applicable to a wide class of dynamic multinomial choice models. The results of a Monte Carlo analysis demonstrated that the method works very well in relatively large state-space models with only partially observed payoffs where very high dimensional integrations are required. Although our discussion focused on models with discrete choices and independent and identically distributed stochastic terms, the method can also be applied to models with mixed continuous/discrete choice sets and serially correlated shocks (see Houser, 1999).