Learning With Selected Features.

Journal: IEEE transactions on cybernetics
Published Date:

Abstract

The coming big data era brings data of unprecedented size and launches an innovation of learning algorithms in statistical and machine-learning communities. The classical kernel-based regularized least-squares (RLS) algorithm is excluded in the innovation, due to its computational and storage bottlenecks. This article presents a scalable algorithm based on subsampling, called learning with selected features (LSF), to reduce the computational burden of RLS. Almost the optimal learning rate together with a sufficient condition on selecting kernels and centers to guarantee the optimality is derived. Our theoretical assertions are verified by numerical experiments, including toy simulations, UCI standard data experiments, and a real-world massive data application. The studies in this article show that LSF can reduce the computational burden of RLS without sacrificing its generalization ability very much.

Authors

  • Shao-Bo Lin
    College of Mathematics and Information Science, Wenzhou University, Wenzhou, 325035, PR China. Electronic address: sblin1983@gmail.com.
  • Jian Fang
  • Xiangyu Chang
    Industrial and Systems Engineering, University of Washington, Seattle, Washington.