Feature Selection Using Smooth Gradient $$L_{1/2}$$ Regularization | SpringerLink
Skip to main content

Feature Selection Using Smooth Gradient \(L_{1/2}\) Regularization

  • Conference paper
  • First Online:
Neural Information Processing (ICONIP 2017)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 10637))

Included in the following conference series:

  • 4510 Accesses

Abstract

In terms of \(L_{1/2}\) regularization, a novel feature selection method for a neural framework model has been developed in this paper. Due to the non-convex, non-smooth and non-Lipschitz characteristics of \(L_{1/2}\) regularizer, it is difficult to directly employ the gradient descent method in training multilayer perceptron neural networks. A smoothing technique has been considered to approximate the original \(L_{1/2}\) regularizer. The proposed method is a two-stage updating approach. First, a multilayer network model with smoothing \(L_{1/2}\) regularizer is trained to eliminate the unimportant features. Second, the compact model without regularization has been simulated until there is no improvements for the performance. The experiments demonstrate that the presented algorithm significantly reduces the redundant features while keeps a considerable model accuracy.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Chapter
JPY 3498
Price includes VAT (Japan)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
JPY 11439
Price includes VAT (Japan)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
JPY 14299
Price includes VAT (Japan)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Jain, A.K., Duin, R.P.W., Mao, J.: Statistical pattern recognition: a review. IEEE Trans. Pattern Anal. Mach. Intell. 22, 4–37 (2000)

    Article  Google Scholar 

  2. Sun, Y., Todorovic, S., Goodison, S.: Local-learning-based feature selection for high-dimensional data analysis. IEEE Trans. Pattern Anal. Mach. Intell. 32, 1610–1626 (2010)

    Article  Google Scholar 

  3. Liu, H., Yu, L.: Toward Integrating Feature Selection Algorithms for Classification and Clustering. IEEE Educational Activities Department (2005)

    Google Scholar 

  4. Bu, H.L., Xia, J., Han, J.B.: A summary of feature selection algorithms and prospect. J. Chaohu Coll. (2008)

    Google Scholar 

  5. Chakraborty, R., Pal, N.R.: Feature selection using a neural framework with controlled redundancy. IEEE Trans. Neural Netw. Learn. Syst. 26, 35–50 (2014)

    Article  MathSciNet  Google Scholar 

  6. He, X., Cai, D., Niyogi, P.: Laplacian score for feature selection. In: International Conference on Neural Information Processing Systems, pp. 507–514. MIT Press (2005)

    Google Scholar 

  7. Gu, Q., Li, Z., Han, J.: Generalized Fisher Score for Feature Selection, pp. 266–273 (2012)

    Google Scholar 

  8. Yao, Y.: Statistical Applications of Linear Programming for Feature Selection via Regularization Methods (2008)

    Google Scholar 

  9. Burden, F., Winkler, D.: Bayesian regularization of neural networks. Methods Mol. Biol. 458, 25 (2008)

    Google Scholar 

  10. Ticknor, J.L.: A Bayesian regularized artificial neural network for stock market forecasting. Expert Syst. Appl. 40, 5501–5506 (2013)

    Article  Google Scholar 

  11. Han, M., Li, D.: An norm 1 regularization term ELM algorithm based on surrogate function and Bayesian framework. Acta Autom. Sinica 37, 1344–1350 (2011)

    Article  MATH  MathSciNet  Google Scholar 

  12. Tibshirani, R.J.: Regression shrinkage and selection via the LASSO. J. Roy. Stat. Soc. Ser. B 58, 267–288 (1996)

    MATH  MathSciNet  Google Scholar 

  13. Xu, Z.B., Guo, H.L., Wang, Y., Zhang, H.: Representative of L1/2 regularization among Lq (0 \(\le \) q \(\le \)1) regularizations: an experimental study based on phase diagram. Acta Autom. Sinica 38, 1225–1228 (2012)

    Google Scholar 

  14. Liu, C., Liang, Y., Luan, X.Z., Leung, K.S., Chan, T.M., Xu, Z.B., Zhang, H.: A improve direct path seeking algorithm for L1/2 regularization, with application to biological feature selection. In: International Conference on Biomedical Engineering and Biotechnology, pp. 8–11. IEEE (2012)

    Google Scholar 

  15. Wu, W., Yang, J.: L1/2 regularization methods for weights sparsication of neural networks (in Chinese). Sci. Sinica 45, 1487–1504 (2015)

    Google Scholar 

  16. Wu, W., Shao, H., Li, Z.: Convergence of batch BP algorithm with penalty for FNN training. In: King, I., Wang, J., Chan, L.-W., Wang, D.L. (eds.) ICONIP 2006. LNCS, vol. 4232, pp. 562–569. Springer, Heidelberg (2006). doi:10.1007/11893028_63

    Chapter  Google Scholar 

  17. Haykin, S.: Neural Networks: A Comprehensive Foundation. Prentice-Hall, Upper Saddle River (1994)

    MATH  Google Scholar 

  18. Hornik, K., Stinchcombe, M., White, H.: Multilayer feedforward networks are universal approximators. Neural Netw. 2, 359–366 (1989)

    Article  Google Scholar 

  19. Kim, Y., Kim, J.: Gradient LASSO for feature selection. In: International Conference on Machine Learning, p. 60. ACM (2004)

    Google Scholar 

  20. Challita, N., Khalil, M., Beauseroy, P.: New feature selection method based on neural network and machine learning. In: Multidisciplinary Conference on Engineering Technology, pp. 81–85. IEEE (2016)

    Google Scholar 

  21. Setiono, R.: A penalty-function approach for pruning feedforward neural networks. Neural Comput. 9, 185 (1997)

    Article  MATH  Google Scholar 

  22. Sun, K., Huang, S.H., Wong, D.S., Jang, S.S.: Design and application of a variable selection method for multilayer perceptron neural network with LASSO. IEEE Trans. Neural Netw. Learn. Syst. 28, 1386–1396 (2017)

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China (Nos. 61305075, 11401185), the China Postdoctoral Science Foundation (No. 2012M520624), the Natural Science Foundation of Shandong Province (Nos. ZR2013FQ004, ZR2013DM015, ZR2015AL014), the Specialized Research Fund for the Doctoral Program of Higher Education of China (No. 20130133120014), the Fundamental Research Funds for the Central Universities (Nos. 14CX05042A, 15CX05053A, 15CX08011A, 15CX02064A) and the University-level Undergraduate Training Program for Innovation and Entrepreneurship (No. 20161349).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shujun Wu .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Gao, H., Yang, Y., Zhang, B., Li, L., Zhang, H., Wu, S. (2017). Feature Selection Using Smooth Gradient \(L_{1/2}\) Regularization. In: Liu, D., Xie, S., Li, Y., Zhao, D., El-Alfy, ES. (eds) Neural Information Processing. ICONIP 2017. Lecture Notes in Computer Science(), vol 10637. Springer, Cham. https://doi.org/10.1007/978-3-319-70093-9_17

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-70093-9_17

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-70092-2

  • Online ISBN: 978-3-319-70093-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics