From convex to nonconvex: A loss function analysis for binary classification


Zhao L., MƏMMƏDOV M., Yearwood J.

10th IEEE International Conference on Data Mining Workshops, ICDMW 2010, Sydney, Australia, 14 - 17 December 2010, pp.1281-1288, (Full Text) identifier

  • Nəşrin Növü: Conference Paper / Full Text
  • Doi nömrəsi: 10.1109/icdmw.2010.57
  • Çap olunduğu şəhər: Sydney
  • Ölkə: Australia
  • Səhifə sayı: pp.1281-1288
  • Açar sözlər: Classification, Loss function, Non-convex, Optimization, Regularization
  • Açıq Arxiv Kolleksiyası: Konfrans Materialı
  • Adres: Bəli

Qısa məlumat

Problems of data classification can be studied in the framework of regularization theory as ill-posed problems. In this framework, loss functions play an important role in the application of regularization theory to classification. In this paper, we review some important convex loss functions, including hinge loss, square loss, modified square loss, exponential loss, logistic regression loss, as well as some non-convex loss functions, such as sigmoid loss, ø-loss, ramp loss, normalized sigmoid loss, and the loss function of 2 layer neural network. Based on the analysis of these loss functions, we propose a new differentiable non-convex loss function, called smoothed 0-1 loss function, which is a natural approximation of the 0-1 loss function. To compare the performance of different loss functions, we propose two binary classification algorithms for binary classification, one for convex loss functions, the other for non-convex loss functions. A set of experiments are launched on several binary data sets from the UCI repository. The results show that the proposed smoothed 0-1 loss function is robust, especially for those noisy data sets with many outliers. © 2010 IEEE.