Neural Generalization of Multiple Kernel Learning

Neural Generalization of Multiple Kernel Learning

Neural Generalization of Multiple Kernel Learning Ahamad Navid Ghanizadeha, Kamaledin Ghiasi-Shirazia,∗, Reza Monsefia, Mohammadreza Qaraeib aDepartment of Computer Engineering, Ferdowsi University of Mashhad (FUM), Azadi Sq., Mashhad, Khorasan Razavi, Iran bAalto University, Helsinki, Finland Abstract Multiple Kernel Learning is a conventional way to learn the kernel function in kernel-based methods. MKL algorithms enhance the performance of kernel methods. However, these methods have a lower complexity compared to deep learning models and are inferior to these models in terms of recognition accu- racy. Deep learning models can learn complex functions by applying nonlinear transformations to data through several layers. In this paper, we show that a typical MKL algorithm can be interpreted as a one-layer neural network with linear activation functions. By this interpretation, we propose a Neural General- ization of Multiple Kernel Learning (NGMKL), which extends the conventional multiple kernel learning framework to a multi-layer neural network with non- linear activation functions. Our experiments on several benchmarks show that the proposed method improves the complexity of MKL algorithms and leads to higher recognition accuracy. Keywords: Multiple Kernel Learning; Deep Learning; Kernel Methods; Neural Networks; arXiv:2102.13337v1 [cs.LG] 26 Feb 2021 ∗Corresponding author Email addresses: [email protected] (Ahamad Navid Ghanizadeh), [email protected] (Kamaledin Ghiasi-Shirazi), [email protected] (Reza Monsefi), [email protected] (Mohammadreza Qaraei) Preprint submitted to Elsevier March 5, 2021 1. Introduction Kernel methods have a long history in the field of machine learning [1, 2, 3, 4]. These methods have made significant advancements and have widespread applications in many machine learning problems [5, 6, 7, 8]. However, the performance of kernel methods is highly dependent on the type of the kernel and the parameters of that kernel. This raises the challenge of learning the kernel function. To tackle this challenge, one principled way is to use Multiple Kernel Learning (MKL) algorithms, which learn a combination of a set of kernels [9, 10, 4, 11, 2]. Multiple kernel learning boosts the performance of kernel-based algorithms in machine learning problems [12, 13, 14, 15]. In MKL algorithms, a conven- tional way for kernel learning is to combine the kernel functions linearly. More sophisticated ways for learning the kernel functions is to use a nonlinear combi- nation [16, 17, 18, 19]. However, it is not clear that these methods can obtain better performance compared to a simple linear combination of kernels, and there is still room for improvement. Similar to other kernel-based methods, two well-known properties of typical MKL algorithms are (i) these methods usually converge to a global solution since they can be formulated as a convex optimization problem, and (ii) kernel functions can be used in large-margin classifiers, for example, Support Vector Machines (SVMs), which reduces overfitting and may lead to better generaliza- tion. Although these features have been very appealing in the traditional view of machine learning, their importance have been challenged by remarkable suc- cesses of deep learning methods. Deep learning models pass data through sev- eral layers, which produces a rich representation of the data and provides a way to learn complex functions. There are remarkable differences between the framework of deep models and kernel methods. In contrast to kernel methods, deep learning models are not usually trained by a convex optimization problem. While it might be considered as a drawback, [20] argued that sacrificing con- 2 vexity may be inevitable for learning complex functions. Furthermore, contrary to kernel methods that use the hinge loss, which is margin-based, deep learn- ing models use softmax with cross-entropy in the last layer for classification. However, [21] have shown that softmax with cross-entropy loss and the hinge loss have very similar characteristics, and both entail a loss when the margin is not observed. Therefore, it can be said that deep learning methods that use softmax with cross-entropy are inherently margin-based method. Finally, us- ing stochastic gradient descent with mini-batches for optimizing deep learning models provides a way for highly parallel computations via GPUs and has led to state-of-the-art results in large-scale problems. In recent years, finding the connection between kernel and deep learning methods has become an active research topic. These works have been done mostly with the purpose of boosting kernel methods or a better understanding of deep learning [22, 23, 24, 25, 26]. In this paper, we investigate the connection between deep learning and a well-known problem in kernel methods, namely multiple kernel learning, for improving kernel-based algorithms. We show that the conventional multiple kernel learning problem can be interpreted as a one- layer neural network without nonlinear activation functions in which the input is represented by multiple kernels. From this point of view, we propose a Neu- ral Generalization of Multiple Kernel Learning (NGMKL), which extends the shallow linear neural interpretation of multiple kernel learning to a multilayer neural network with nonlinear activation functions. This practice leads to a MKL model with higher complexity, which improves the capacity of MKL mod- els to learn more complex functions. MKL models with deep structures have been investigated previously through Multilayer Multiple Kernel Learning (MLMKL) methods [27, 28, 29]. However, the framework of these methods still sticks to kernelized SVMs, which makes them significantly different from neural networks. More precisely, firstly, simi- lar to ordinary MKL methods, MLMKL models use SVMs as classifiers, while neural networks utilize softmax for classification. Secondly, these models have some constraints on the parameters, such as the positivity of the combination 3 weights to keep the deep kernel function positive definite. However, we note that the rationality behind imposing positive definiteness over kernel functions is to obtain a convex objective function, which MLMKL will not attain because of their deep structure. As a result, it may not be beneficial to constraint the kernel function to be a Mercer kernel in a deep structure. Finally, in contrast to typical deep learning models, in MLMKL models, the weights are not usually trained simultaneously, and the training algorithm alternates between learning the kernel combination weights and the classifier’s parameters. In contrast to MKL and MLMKL algorithms, our proposed method utilizes a softmax classifier in the last layer. There are no constraints on the parame- ters of NGMKL, such as the positivity of the combination weights, and all the parameters are trained simultaneously. Furthermore, a consequence of neural interpretation of MKLs is that the framework of deep learning algorithms can be leveraged, which, for instance, provides fast parallel computations via GPUs. Moreover, in the context of SGD, minibatching allows high throughput since it can process a large number of input examples in a large number of cores at once in GPU. In the experimental section, we show that NGMKL outperforms conventional MKL algorithms on commonplace datasets in kernel methods. The rest of the paper is organized as follows. In Section 2, we review the related work on relating kernel methods and deep learning. In this section, we also review the multilayer multiple kernel learning models. In Section 3, we describe the conventional formulation of multiple kernel learning problems. In Section 4, we show how the multiple kernel learning can be seen in the neural network framework and propose the NGMKL model. The comparison of NGMKL and ordinary multiple kernel learning algorithms by experiments on several datasets is presented in Section 5. Finally, in section 6, we conclude the paper. 4 2. Related Work There are several works in the literature that investigate the connections between kernel methods and neural networks. In this section, we review three types of these works: (i) the kernel functions which provably imitate the ar- chitecture of multilayer neural networks, (ii) multilayer neural networks con- structed by kernel approximation techniques, and (iii) multiple kernel learning methods with multilayer structures. Some methods link kernel methods with neural networks by proposing ker- nel functions that mimic the computation of neural networks. For example, similar to neural networks, it has been demonstrated that the feature map of arc-cosine kernels consists of a linear transformation using an infinite number of weights followed by a nonlinearity [25]. However, the interpreted weights are fixed and have a Gaussian distribution with zero mean and unit variance. To tackle this drawback, one method learns the covariance of the interpreted weights by stretching a finite number of trained weights to infinite [30], and another method proposed learning the covariance by an iterative approach sim- ilar to restricted boltzmann machines [31]. In arc-cosine kernels, by using two distinct hyperparameters, one can change the nonlinearity of the interpreted neural network and extend the structure to multiple layers. The connections between neural networks with an infinite number of weights and kernel methods are also discussed in the context of Gaussian processes [32, 33, 34, 23], which has attracted attention for a better

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    20 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us