Model selection, large deviations and consistency of data-driven tests

M. Langovoy

Onderzoeksoutput: Boek/rapportRapportAcademic

46 Downloads (Pure)


We consider three general classes of data-driven statistical tests. Neyman's smooth tests, data-driven score tests and data-driven score tests for statistical inverse problems serve as important special examples for the classes of tests under consideration. Our tests are additionally incorporated with model selection rules. The rules are based on the penalization idea. Most of the optimal penalties, derived in statistical literature, can be used in our tests. We prove general consistency theorems for the tests from those classes. Our proofs make use of large deviations inequalities for deterministic and random quadratic forms. The paper shows that the tests can be applied for simple and composite parametric, semi- and nonparametric hypotheses. Applications to testing in statistical inverse problems and statistics for stochastic processes are also presented.
Originele taal-2Engels
Plaats van productieEindhoven
Aantal pagina's30
StatusGepubliceerd - 2009

Publicatie series

NaamReport Eurandom
ISSN van geprinte versie1389-2355


Duik in de onderzoeksthema's van 'Model selection, large deviations and consistency of data-driven tests'. Samen vormen ze een unieke vingerafdruk.

Citeer dit