Abstract
The proportional odds model (POM) is the most widely used model when the response has ordered categories. In the case of high-dimensional predictor structure the common maximum likelihood approach typically fails when all predictors are included. A boosting technique pomBoost is proposed that fits the model by implicitly selecting the influential predictors. The approach distinguishes between metric and categorical predictors. In the case of categorical predictors, where each predictor relates to a set of parameters, the objective is to select simultaneously all the associated parameters. In addition the approach distinguishes between nominal and ordinal predictors. In the case of ordinal predictors, the proposed technique uses the ordering of the ordinal predictors by penalizing the difference between the parameters of adjacent categories. The technique has also a provision to consider some mandatory predictors (if any) which must be part of the final sparse model. The performance of the proposed boosting algorithm is evaluated in a simulation study and applications with respect to mean squared error and prediction error. Hit rates and false alarm rates are used to judge the performance of pomBoost for selection of the relevant predictors.
Item Type: | Paper |
---|---|
Keywords: | Logistic regression, Proportional odds model, Variable selection, Likelihood-based boosting, Penalization, Hit rate, False alarm rate |
Faculties: | Mathematics, Computer Science and Statistics > Statistics > Technical Reports |
Subjects: | 500 Science > 500 Science |
URN: | urn:nbn:de:bvb:19-epub-12152-8 |
Language: | English |
Item ID: | 12152 |
Date Deposited: | 17. Feb 2011, 12:52 |
Last Modified: | 04. Nov 2020, 12:52 |