Data Augmentation in HDLSS With

Data Augmentation in HDLSS With

1 Data Augmentation in High Dimensional Low Sample Size Setting Using a Geometry-Based Variational Autoencoder Clement´ Chadebec, Elina Thibeau-Sutre, Ninon Burgos, and Stephanie´ Allassonniere,` for the Alzheimer’s Disease Neuroimaging Initiative, and the Australian Imaging Biomarkers and Lifestyle flagship study of ageing Abstract—In this paper, we propose a new method to perform data augmentation in a reliable way in the High Dimensional Low Sample Size (HDLSS) setting using a geometry-based variational autoencoder. Our approach combines a proper latent space modeling of the VAE seen as a Riemannian manifold with a new generation scheme which produces more meaningful samples especially in the context of small data sets. The proposed method is tested through a wide experimental study where its robustness to data sets, classifiers and training samples size is stressed. It is also validated on a medical imaging classification task on the challenging ADNI database where a small number of 3D brain MRIs are considered and augmented using the proposed VAE framework. In each case, the proposed method allows for a significant and reliable gain in the classification metrics. For instance, balanced accuracy jumps from 66.3% to 74.3% for a state-of-the-art CNN classifier trained with 50 MRIs of cognitively normal (CN) and 50 Alzheimer disease (AD) patients and from 77.7% to 86.3% when trained with 243 CN and 210 AD while improving greatly sensitivity and specificity metrics. F 1 INTRODUCTION VEN though always larger data sets are now available, or to give statistically meaningful results [4]. E the lack of labeled data remains a tremendous issue in A way to address such issues is to perform data aug- many fields of application. Among others, a good example mentation (DA) [5]. In a nutshell, DA is the art of increasing is healthcare where practitioners have to deal most of the the size of a given data set by creating synthetic labeled time with (very) low sample sizes (think of small patient data. For instance, the easiest way to do this on images cohorts) along with very high dimensional data (think of is to apply simple transformations such as the addition neuroimaging data that are 3D volumes with millions of of Gaussian noise, cropping or padding, and assign the voxels). Unfortunately, this leads to a very poor represen- label of the initial image to the created ones. While such tation of a given population and makes classical statistical augmentation techniques have revealed very useful, they analyses unreliable [1], [2]. Meanwhile, the remarkable per- remain strongly data dependent and limited. Some trans- formance of algorithms heavily relying on the deep learning formations may indeed be uninformative or even induce framework [3] has made them extremely attractive and very bias. For instance, think of a digit representing a 6 which popular. However, such results are strongly conditioned by gives a 9 when rotated. While assessing the relevance of aug- the number of training samples since such models usually mented data may be quite straightforward for simple data need to be trained on huge data sets to prevent over-fitting sets, it reveals very challenging for complex data and may require the intervention of an expert assessing the degree of relevance of the proposed transformations. In addition Data used in preparation of this article were obtained from the Alzheimer’s arXiv:2105.00026v1 [stat.ML] 30 Apr 2021 Disease Neuroimaging Initiative (ADNI) database (http://adni.loni.usc.edu). to the lack of data, imbalanced data sets also severely limit As such, the investigators within the ADNI contributed to the design and generalizability since they tend to bias the algorithm toward implementation of ADNI and/or provided data but did not participate in the most represented classes. Oversampling is a method analysis or writing of this report. A complete listing of ADNI investigators can be found at: http://adni.loni.usc.edu/wp-content/uploads/how to apply/ that aims at balancing the number of samples per class by ADNI Acknowledgement List.pdf up-sampling the minority classes. The Synthetic Minority Data used in the preparation of this article was obtained from the Australian Over-sampling TEchnique (SMOTE) was first introduced Imaging Biomarkers and Lifestyle flagship study of ageing (AIBL) funded by in [6] and consists in interpolating data points belonging to the Commonwealth Scientific and Industrial Research Organisation (CSIRO) which was made available at the ADNI database (http://adni.loni.usc.edu). the minority classes in their feature space. This approach The AIBL researchers contributed data but did not participate in analysis or was further extended in other works where the authors writing of this report. AIBL researchers are listed at www.aibl.csiro.au. proposed to over-sample close to the decision boundary • Cl´ementChadebec and St´ephanieAllassonni`ere are with the Universit´e using either the k-Nearest Neighbor (k-NN) algorithm [7] de Paris, Inria, Centre de Recherche des Cordeliers, Inserm, Sorbonne or a support vector machine (SVM) [8] and so insist on sam- Universit´e,Paris, France • Elina Thibeau-Sutre and Ninon Burgos are with Sorbonne Universit´e, ples that are potentially misclassified. Other over-sampling Institut du Cerveau - Paris Brain Institute (ICM), Inserm U 1127, CNRS methods aiming at increasing the number of samples from UMR 7225, AP-HP Hˆopital de la Piti´e Salpˆetri`ere and Inria Aramis the minority classes and taking into account their difficulty project-team, Paris, France to be learned were also proposed [9], [10]. However, these 2 methods hardly scale to high-dimensional data [11], [12]. 2.1 Model Setting The recent rise in performance of generative models Let x be a set of data. A VAE aims at maximizing the 2 X such as generative adversarial networks (GAN) [13] or likelihood of a given parametric model Pθ; θ Θ . It is variational autoencoders (VAE) [14], [15] has made them assumed that there exist latent variables zf living2 in ag lower very attractive models to perform DA. GANs have al- dimensional space , referred to as the latent space, such that ready seen a wide use in many fields of application [16], the marginal distributionZ of the data can be written as: [17], [18], [19], [20], including medicine [21]. For instance, Z GANs were used on magnetic resonance images (MRI) [22], pθ(x) = pθ(x z)q(z)dz ; (1) j [23], computed tomography (CT) [24], [25], X-ray [26], [27], Z [28], positron emission tomography (PET) [29], mass spec- where q is a prior distribution over the latent variables troscopy data [30], dermoscopy [31] or mammography [32], acting as a regulation factor and p (x z) is most of the time [33] and demonstrated promising results. Nonetheless, most θ taken as a simple parametrized distributionj (e.g. Gaussian, of these studies involved either a quite large training set Bernoulli, etc.). Such a distribution is referred to as the (above 1000 training samples) or quite small dimensional decoder, the parameters of which are usually given by neural data, whereas in everyday medical applications it remains networks. Since the integral of Eq. (1) is most of the time very challenging to gather such large cohorts of labeled pa- intractable, so is the posterior distribution: tients. As a consequence, as of today, the case of high dimen- sional data combined with a very low sample size remains pθ(x z)q(z) pθ(z x) = R j : poorly explored. When compared to GANs, VAEs have only j pθ(x z)q(z)dz seen a very marginal interest to perform DA and were Z j mostly used for speech applications [34], [35], [36]. Some This makes direct application of Bayesian inference impos- attempts to use such generative models on medical data sible and so recourse to approximation techniques such as either for classification [37], [38] or segmentation tasks [39], variational inference [42] is needed. Hence, a variational dis- [40], [41] can nonetheless be noted. The main limitation to tribution qφ(z x) is introduced and aims at approximating j a wider use of these models is that they most of the time the true posterior distribution pθ(z x) [14]. This variational produce blurry and fuzzy samples. This undesirable effect distribution is often referred to as thej encoder. In the initial is even more emphasized when they are trained with a small version of the VAE, qφ is taken as a multivariate Gaussian number of samples which makes them very hard to use in whose parameters µφ and Σφ are again given by neural net- practice to perform DA in the high dimensional (very) low works. Importance sampling can then be applied to derive sample size (HDLSS) setting. an unbiased estimate of the marginal distribution pθ(x) we In this paper, we argue that VAEs can actually be used want to maximize in Eq. (1) for data augmentation in a reliable way even in the context pθ(x z)q(z) of HDLSS data, provided that we bring some modeling of p^θ(x) = j and Ez∼qφ p^θ = pθ(x) : qφ(z x) the latent space and amend the way we generate the data. j Hence, in this paper we propose the following contributions: Using Jensen’s inequality allows finding a lower bound on the objective function of Eq. (1) • We propose to combine a proper modeling of the log pθ(x) = log Ez∼qφ p^θ latent space of the VAE, here seen as a Riemannian logp ^ manifold, and a new geometry-aware non-prior-based Ez∼qφ θ ≥ generation procedure which consists in sampling Ez∼qφ log pθ(x; z) log qφ(z x) = ELBO : ≥ − j from the inverse of the Riemannian metric volume el- (2) ement.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    25 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us