A Kernel Classifier using Linearised Bregman Iteration

In this paper we introduce a novel kernel classifier based on a iterative shrinkage algorithm developed for compressive sensing. We have adopted Bregman iteration with soft and hard shrinkage functions and generalized hinge loss for solving l1 norm minimization problem for classification. Our experimental results with face recognition and digit classification using SVM as the benchmark have shown that our method has a close error rate compared to SVM but do not perform better than SVM. We have found that the soft shrinkage method give more accuracy and in some situations more sparseness than hard shrinkage methods.




References:
[1] Donoho. D.L.: "Compressed sensing", IEEE Trans. Inform. Theory,
52:1289-1306, (2006)
[2] Daubechies I., Defrise M., De Mol C. "An iterative thresholding algorithm
for linear inverse problems with a sparsity constraint", Comm.
Pure Appl. Math. 57, pp. 14131457 (2004)
[3] Beck A., Teboulle M.,"A Fast Iterative Shrinkage-Thresholding Algorithm
for Linear Inverse Problems", SIAM J. Imaging Sciences,
forthcoming.
[4] YinW., Osher S., Goldfarb D., Darbon J., "Bregman Iterative Algorithms
for l1-Minimization with Applications to Compressed Sensing", SIAM
J. IMAGING SCIENCES Vol. 1, No. 1, pp. 143168 (2008)
[5] Tibshirani R., "Regression selection and shrinkage via the lasso",J. R.
Stat. Soc. Ser. B 58, pp. 267288 (1996)
[6] Zhu J., Rosset S., Hastie T., Tibshirani R., "1-norm Support Vector
Machines", Advances in Neural Information Processing Systems 16,
(2003)
[7] Raina R., Battle A., Lee H., Packer B., Ng. A. Y., "Self-taught learning:
Transfer learning from unlabeled data", ICML 2007
[8] Lee H., Battle A., Raina R., Ng. A. Y., "Efficient sparse coding
algorithms", In Advances in Neural Information Processing Systems
19 (NIPS-06), pages 801808, (2007)
[9] Langford J., Li L. Zhang T., "Sparse Online Learning via Truncated
Gradient", Journal of Machine Learning Research Vol. 10, pp. 777-801
(2009)
[10] Koh K., Kim S., Boyd S., "An Interior-Point Method for Large-
Scale l1-Regularized Logistic Regression", Journal of Machine Learning
Research Vol. 8, pp. 1519-1555, 2007.
[11] Hale E., Yin W., Zhang. Y., "A Fixed-point continuation method
for l1-regularization with application to compressed sensing", CAAM
Technical Report TR07-07, Rice University, Houston, TX, 2007.
[12] Vapnik V. N., Statsitical Leanring Theory, Wiley Interscience, (1998)
[13] Rennie J.D.M., "Smooth Hinge Classfication", Technical report, MIT
(2005)
[14] Bredies K., Lorenz D. A., "Iterated hard shrinkage for minimization
problems with sparsity constraints", SIAM Journal on Scientific Computing,
Vol. 30(2), pp. 657-683, 2008.
[15] Graham D. B., Allinson N. M., "Characterizing Virtual Eigensignatures
for General Purpose Face Recognition", in Face Recognition: From
Theory to Applications , NATO ASI Series F, Computer and Systems
Sciences, Vol. 163. H. Wechsler, P. J. Phillips, V. Bruce, F. Fogelman-
Soulie and T. S. Huang (eds), pp 446-456, 1998.
[16] Campbell C., Cristianini N., "Simple Learning Algorithms for Traning
Support Vector Machines", Technical Report, UNiversity of Bristol,
1998.
[17] Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner. "Gradient-based
learning applied to document recognition", Proceedings of the IEEE,
86(11):2278-2324, November 1998.