Mini-batch learning of exponential family finite mixture models Hien D Nguyen, Florence Forbes, Geoffrey Mclachlan To cite this version: Hien D Nguyen, Florence Forbes, Geoffrey Mclachlan. Mini-batch learning of exponential family finite mixture models. Statistics and Computing, Springer Verlag (Germany), 2020, 30, pp.731-748. 10.1007/s11222-019-09919-4. hal-02415068v2 HAL Id: hal-02415068 https://hal.archives-ouvertes.fr/hal-02415068v2 Submitted on 26 Mar 2020 HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non, lished or not. The documents may come from émanant des établissements d’enseignement et de teaching and research institutions in France or recherche français ou étrangers, des laboratoires abroad, or from public or private research centers. publics ou privés. Mini-batch learning of exponential family finite mixture models Hien D. Nguyen1∗, Florence Forbes2, and Geoffrey J. McLachlan3 September 6, 2019 1Department of Mathematics and Statistics, La Trobe University, Melbourne, Victoria, Australia. 2Univ. Grenoble Alpes, Inria, CNRS, Grenoble INPy, LJK, 38000 Grenoble, France. yInstitute of Engineering Univ. Grenoble Alpes. 3School of Mathematics and Physics, University of Queensland, St. Lucia, Brisbane, Australia. ∗Corresponding author: Hien Nguyen (Email:
[email protected]). Abstract Mini-batch algorithms have become increasingly popular due to the requirement for solving optimization problems, based on large-scale data sets. Using an existing online expectation-- maximization (EM) algorithm framework, we demonstrate how mini-batch (MB) algorithms may be constructed, and propose a scheme for the stochastic stabilization of the constructed mini-batch algorithms.