next up previous contents index
Next: 2. Bootstrap and Resampling Up: csahtml Previous: 1.7 Discussion

References

1
Akaike, H. (1970). Statistical predictor identification. Ann. Inst. Statist. Math., 21: 203-217.

2
Akaike, H. (1973). Information theory and the maximum likelihood principle. International Symposium on Information Theory, eds. V. Petrov and F. Csáki, Budapest: Akademiai Kiádo, pp. 267-281.

3
Allen, D.M. (1974). The relationship between variable selection and data augmentation and a method for prediction. Technometrics, 16: 125-127.

4
Bai, Z.D., Rao, C.R. and Wu, Y. (1999). Model selection with data-oriented penalty. Journal of Statistical Planning and Inference, 77: 103-117.

5
Beran, R. (1996). Bootstrap variable selection and confidence sets. In Rieder, H. (ed), Robust Statistics, Data Analysis and Computer Intensive Methods, Springer Lecture Notes in Statistics 109.

6
Berger, J.O. and Pericchi, L.R. (1996). The intrinsic bayes factor for model selection and prediction. Journal of the American Statistical Association, 91: 109-122.

7
Berger, J.O. and Pericchi, L.R. (2001). Objective bayesian methods for model selection: Introduction and comparison. In Lahiri, P. (ed), Model Selection, Institute of Mathematical Statistics Lecture Notes - Monograph Series volume 38, pp.135-207, Beachwood Ohio.

8
Box, G. E.P. (1976). Science and statistics. Journal of the American Statistical Association, 71: 791-799.

9
Box, G. E.P. and Jenkins, G.M. (1976). Time Series Analysis, Holden-Day.

10
Burnham, K.P. and Anderson, D.R. (2002). Model Selection and Multimodel Inference, 2nd ed., Springer, New York.

11
Chatfield, C. (1995). Model uncertainty, data mining and statistical inference (with discussion). Journal of the Royal Statistical Society B, 158: 419-466.

12
Chipman, H., George, E.I. and McCulloch, R.E. (2001). The practical implementation of bayesian model selection. In Lahiri, P. (ed), Model Selection, Institute of Mathematical Statistics Lecture Notes - Monograph Series volume 38, pp.65-134, Beachwood Ohio.

13
Craven, P. and Wahba, G. (1979). Smoothing noisy data with spline functions. Numer. Math., 31: 377-403.

14
Dette, H., Munk, A. and Wagner, T. (1998). Estimating the variance in nonparametric regression - what is a reasonable choice? Journal of the Royal Statistical Society B, 60: 751-764.

15
Donoho, D.L. and Johnston, I.M. (1994). Ideal spatial adaption by wavelet shrinkage. Biometrika, 81: 425-456.

16
Efron, B. (1986). How biased is the apparent error rate of a prediction rule. Journal of the American Statistical Association, 81: 461-470.

17
Gasser, T., Sroka, L. and Jennen-Steinmetz, C. (1986). Residual variance and residual pattern in nonlinear regression. Biometrika, 73: 625-633.

18
Gelman, A., Carlin, J.B., Stern, H.S. and Rubin, D.B. (1995). Bayesian Data Analysis, Chapman & Hall, Boca Raton.

19
George, E.I. (2000). The variable selection problem. Journal of the American Statistical Association, 95: 1304-1308.

20
Golub, G., Heath, M. and Wahba, G. (1979). Generalized cross validation as a method for choosing a good ridge parameter. Technometrics, 21: 215-224.

21
Gu, C. (1998). Model indexing and smoothing parameter selection in nonparametric function estimation (with discussion). Statistica Sinica, 8: 632-638.

22
Gu, C. (2002). Smoothing Spline ANOVA Models, Springer-Verlag, New York.

23
Hall, P., Kay, J.W. and Titterington, D.M. (1990). Asymptotically optimal difference-based estimation of variance in nonparametric regression. Biometrika, 77: 521-528.

24
Hastie, T. and Tibshirani, R. (1990). Generalized Additive Models, Chapman and Hall.

25
Hastie, T., Tibshirani, R. and Friedman, J. (2002). The Elements of Statistical Learning, Springer, New York.

26
Hinkley, D. (2003). Bootstrap methods for variable selection and shrinkage estimatior confidence sets, Personal Communication.

27
Hoeting, J.A., Madigan, D., Raftery, A.E. and Volinsky, C.T. (1999). Bayesian model averaging: a tutorial (with discussion). Statistical Science, 14: 382-417. Corrected version available at
http://www.stat.washington.edu/www/research/online/hoeting1999.pdf.

28
Hurvich, C.M. and Tsai, C.L. (1989). Regression and time series model selection in small samples. Biometrika, 76: 297-207.

29
Jeffreys, H. (1961). Theorey of Probability, Oxford: Clarendon Press.

30
Jeffreys, W. and Berger, J.O. (1992). Ockham's razor and bayesian analysis. American Scientist, 80: 64-72.

31
Kass, R.E. and Raftery, A. (1995). Bayesian factors. Journal of the American Statistical Association, 90: 773-795.

32
Kass, R.E. and Wasserman, L. (1995). A reference bayesian test for nested hypotheses and its relationship to the schwarz criterion. Journal of the American Statistical Association, 90: 982-934.

33
Ke, C. and Wang, Y. (2002). Nonparametric nonlinear regression models, Technical Report # 385, Department of Statistics and Applied Probability, University of California, Santa Barbara.

34
Li, K.C. (1985). From Stein's unbaised risk estimates to the method of generalized cross-validation. Annals of Statistics, 13: 1352-1377.

35
Li, K.C. (1986). Asymptotic optimality of $ {C_L}$ and generalized cross-validation in ridge regression with application to spline smoothing. Annals of Statistics, 14: 1101-1112.

36
Li, K.C. (1987). Asymptotic optimality of $ {C_p}$, $ {C_L}$, cross-validation and generalized cross-validation: Discrete index set. Annals of Statistics, 15: 958-975.

37
Linhart, H. and Zucchini, W. (1986). Model Selection, Wiley, New York.

38
Mallows, C.L. (1973). Some comments on $ {C_p}$. Technometrics, 12: 661-675.

39
Miller, A. (2002). Subset Selection in Regression, 2nd ed., Chapman & Hall, New York.

40
Opsomer, J., Wang, Y. and Yang, Y. (2001). Nonparametric regression with correlated errors. Statistical Science, 16: 134-153.

41
Rao, C.R. and Wu, Y. (1989). A strongly consistent procedure for model selection in a regreesion problem. Biometrika, 76: 369-374.

42
Rao, J.S. (1999). Bootstrap choice of cost complexity for better subset selection. Statistica Sinica, 9: 273-287.

43
Rao, J.S. and Tibshirani, R. (1997). Discussion to ''an asympototic theory for model selection'' by Jun Shao. Statistica Sinica, 7: 249-252.

44
Rice, J.A. (1984). Bandwidth choice for nonparametric regression. Annals of Statistics, 12: 1215-1230.

45
Schwarz, G. (1978). Estimating the dimension of a model. Annals of Statistics, 12: 1215-1231.

46
Shao, J. (1993). Linear model selection by cross-validation. Journal of the American Statistical Association, 88: 486-494.

47
Shao, J. (1997). An asymptotic theory for linear model selection (with discussion), Statistica Sinica, 7: 221-264.

48
Shen, X. and Ye, J. (2002). Adaptive model selection. Journal of the American Statistical Association, 97: 210-221.

49
Stone, M. (1974). Cross-validatory choice and assessment of statistical prediction, Journal of the Royal Statistical Society B, 36: 111-147.

50
Wahba, G. (1990). Spline Models for Observational Data, SIAM, Philadelphia. CBMS-NSF Regional Conference Series in Applied Mathematics, Vol. 59.

51
Wahba, G. and Wang, Y. (1993). Behavior near zero of the distribution of GCV smoothing parameter estimates for splines. Statistics and Probability Letters, 25: 105-111.

52
Wahba, G. and Wold, S. (1975). A completely automatic french curve. Communications in Statistics, 4: 1-17.

53
Wang, Y. (1998). Smoothing spline models with correlated random errors. Journal of the American Statistical Association, 93: 341-348.

54
Wang, Y. and Ke, C. (2002). ASSIST: A suite of s-plus functions implementing spline smoothing techniques, Manual for the ASSIST package. Available at http://www.pstat.ucsb.edu/faculty/yuedong/software.

55
Wang, Y., Guo, W. and Brown, M.B. (2000). Spline smoothing for bivariate data with applications to association between hormones. Statistica Sinica, 10: 377-397.

56
Xiang, D. and Wahba, G. (1996). A genralized approximate cross validation for smoothing splines with non-gaussian data. Statistica Sinica, 6: 675-692.

57
Yang, Y. (1999). Model selection for nonparametric regression. Statistica Sinica, 9: 475-499.

58
Ye, J. (1998). On measuring and correcting the effects of data mining and model selection. Journal of the American Statistical Association, 93: 120-131.

59
Zhang, H., Wahba, G., Lin, Y., Voelker, M., Ferris, M., Klein, R. and Klein, B. (2002). Variable selection and model building via likelihood basis pursuit, Technical Report No. 1059, Department of Statistics, University of Wisconsin.

60
Zhang, P. (1993). Model selection via multifold cross validation. Annals of Statistics, 21: 299-313.

61
Zhou, H. and Huang, J.T. (2004). Shrinkage estimation toward the data chosen reduced model with applications to wavelet analysis and principal component regression analysis. Annals of Statistics, to appear.



Subsections