Cookies helfen uns bei der Bereitstellung unserer Dienste. Durch die Nutzung unserer Dienste erklären Sie sich damit einverstanden, dass wir Cookies setzen.

Cuvillier Verlag

32 Jahre Kompetenz im wissenschaftlichen Publizieren
Internationaler Fachverlag für Wissenschaft und Wirtschaft

Cuvillier Verlag

De En Es
Statistical Issues in Machine Learning Towards Reliable Split Selection and Variable Importance Measures

EUR 28,00 EUR 26,60

EUR 19,60

Statistical Issues in Machine Learning Towards Reliable Split Selection and Variable Importance Measures

Carolin Strobl (Autor)


Inhaltsverzeichnis, Datei (56 KB)
Leseprobe, Datei (84 KB)

ISBN-13 (Printausgabe) 3867276617
ISBN-13 (Printausgabe) 9783867276610
ISBN-13 (E-Book) 9783736926615
Sprache Englisch
Seitenanzahl 204
Umschlagkaschierung glänzend
Auflage 1 Aufl.
Band 0
Erscheinungsort Göttingen
Promotionsort München
Erscheinungsdatum 30.07.2008
Allgemeine Einordnung Dissertation
Fachbereiche Mathematik
Biochemie, Molekularbiologie, Gentechnologie
Schlagwörter CART, Bagging, Random Forest, Gini Index, Variable Importance

Recursive partitioning methods from machine learning are being widely applied in many scientific fields such as, e.g., genetics and bioinformatics. The present work is concerned with the two main problems that arise in recursive partitioning, instability and biased variable selection, from a statistical point of view. With respect to the first issue, instability, the entire scope of methods from standard classification trees over robustified classification trees and ensemble methods such as TWIX, bagging and random forests is covered in this work.
While ensemble methods prove to be much more stable than single trees, they also loose most of their interpretability. Therefore an adaptive cutpoint selection scheme is suggested with which a TWIX ensemble reduces to a single tree if the partition is sufficiently stable. With respect to the second issue, variable selection bias, the statistical sources of this artifact in single trees and a new form of bias inherent in ensemble methods based on bootstrap samples are investigated. For single trees, one unbiased split selection criterion is evaluated and another one newly introduced here. Based on the results for single trees and further findings on the effects of bootstrap sampling on association measures, it is shown that, in addition to using an unbiased split selection criterion, subsampling instead of bootstrap sampling should be employed in ensemble methods to be able to reliably compare the variable importance scores of predictor variables of different types. The statistical properties and the null hypothesis of a test for the random forest variable importance are critically investigated. Finally, a new, conditional importance measure is suggested that allows for a fair comparison in the case of correlated predictor variables and better reflects the null hypothesis of interest.