On the effect and remedies of shrinkage on classification probability estimation

Shrinkage methods have been shown to be effective for classification problems. As a form of regularization, shrinkage through penalization helps to avoid overfitting and produces accurate classifiers for prediction, especially when the dimension is relatively high. Despite the benefit of shrinkage o...

全面介紹

Saved in:
書目詳細資料
Main Authors: WU, Zhengxiao, LIU, Yufeng
格式: text
語言:English
出版: Institutional Knowledge at Singapore Management University 2013
主題:
在線閱讀:https://ink.library.smu.edu.sg/soe_research_all/12
https://ink.library.smu.edu.sg/cgi/viewcontent.cgi?article=1011&context=soe_research_all
標簽: 添加標簽
沒有標簽, 成為第一個標記此記錄!
機構: Singapore Management University
語言: English
實物特徵
總結:Shrinkage methods have been shown to be effective for classification problems. As a form of regularization, shrinkage through penalization helps to avoid overfitting and produces accurate classifiers for prediction, especially when the dimension is relatively high. Despite the benefit of shrinkage on classification accuracy of resulting classifiers, in this article, we demonstrate that shrinkage creates biases on classification probability estimation. In many cases, this bias can be large and consequently yield poor class probability estimation when the sample size is small or moderate. We offer some theoretical insights into the effect of shrinkage and provide remedies for better class probability estimation. Using penalized logistic regression and proximal support vector machines as examples, we demonstrate that our proposed refit method gives similar classification accuracy and remarkable improvements on probability estimation on several simulated and real data examples.