Deep Transfer Learning with Joint Adaptation Networks

Total Page:16

File Type:pdf, Size:1020Kb

Deep Transfer Learning with Joint Adaptation Networks Deep Transfer Learning with Joint Adaptation Networks Mingsheng Long 1 Han Zhu 1 Jianmin Wang 1 Michael I. Jordan 2 Abstract shallow transfer learning methods bridge the source and tar- get domains by learning invariant feature representations or Deep networks have been successfully applied to estimating instance importance without using target labels learn transferable features for adapting models (Huang et al., 2006; Pan et al., 2011; Gong et al., 2013). Re- from a source domain to a different target domain. cent deep transfer learning methods leverage deep networks In this paper, we present joint adaptation networks to learn more transferable representations by embedding (JAN), which learn a transfer network by aligning domain adaptation in the pipeline of deep learning, which the joint distributions of multiple domain-specific can simultaneously disentangle the explanatory factors of layers across domains based on a joint maximum variations behind data and match the marginal distributions mean discrepancy (JMMD) criterion. Adversarial across domains (Tzeng et al., 2014; 2015; Long et al., 2015; training strategy is adopted to maximize JMMD 2016; Ganin & Lempitsky, 2015; Bousmalis et al., 2016). such that the distributions of the source and target domains are made more distinguishable. Learning Transfer learning becomes more challenging when domains can be performed by stochastic gradient descent may change by the joint distributions of input features and with the gradients computed by back-propagation output labels, which is a common scenario in practical ap- in linear-time. Experiments testify that our model plications. First, deep networks generally learn the complex yields state of the art results on standard datasets. function from input features to output labels via multilayer feature transformation and abstraction. Second, deep fea- tures in standard CNNs eventually transition from general to 1. Introduction specific along the network, and the transferability of features and classifiers decreases when the cross-domain discrepancy Deep networks have significantly improved the state of the increases (Yosinski et al., 2014). Consequently, after feed- arts for diverse machine learning problems and applications. forwarding the source and target domain data through deep Unfortunately, the impressive performance gains come only networks for multilayer feature abstraction, the shifts in the when massive amounts of labeled data are available for joint distributions of input features and output labels still supervised learning. Since manual labeling of sufficient linger in the network activations of multiple domain-specific training data for diverse application domains on-the-fly is higher layers. Thus we can use the joint distributions of the often prohibitive, for a target task short of labeled data, activations in these domain-specific layers to approximately there is strong motivation to build effective learners that can reason about the original joint distributions, which should leverage rich labeled data from a different source domain. be matched across domains to enable domain adaptation. To However, this learning paradigm suffers from the shift in date, this problem has not been addressed in deep networks. data distributions across different domains, which poses a In this paper, we present Joint Adaptation Networks (JAN) arXiv:1605.06636v2 [cs.LG] 17 Aug 2017 major obstacle in adapting predictive models for the target to align the joint distributions of multiple domain-specific task (Quionero-Candela et al., 2009; Pan & Yang, 2010). layers across domains for unsupervised domain adaptation. Learning a discriminative model in the presence of the shift JAN largely extends the ability of deep adaptation networks between training and test distributions is known as transfer (Long et al., 2015) to reason about the joint distributions learning or domain adaptation (Pan & Yang, 2010). Previous as mentioned above, while keeping the training procedure even simpler. Specifically, JAN admits a simple transfer 1 Key Lab for Information System Security, MOE; Tsinghua Na- pipeline, which processes the source and target domain data tional Lab for Information Science and Technology (TNList); NEL- BDS; School of Software, Tsinghua University, Beijing 100084, by convolutional neural networks (CNN) and then aligns China 2University of California, Berkeley, Berkeley 94720. Corre- the joint distributions of activations in multiple task-specific spondence to: Mingsheng Long <[email protected]>. layers. To learn parameters and enable alignment, we derive joint maximum mean discrepancy (JMMD), which measures th Proceedings of the 34 International Conference on Machine the Hilbert-Schmidt norm between kernel mean embedding Learning, Sydney, Australia, PMLR 70, 2017. Copyright 2017 by the author(s). of empirical joint distributions of source and target data. Deep Transfer Learning with Joint Adaptation Networks Thanks to a linear-time unbiased estimate of JMMD, we can the problem tractable (Zhang et al., 2013). As it is not easy easily draw a mini-batch of samples to estimate the JMMD to justify which components of the joint distribution are criterion, and implement it efficiently via back-propagation. changing in practice, our work is transparent to diverse sce- We further maximize JMMD using adversarial training strat- narios by directly manipulating the joint distribution without egy such that the distributions of source and target domains assumptions on the marginal and conditional distributions. are made more distinguishable. Empirical study shows that Furthermore, it remains unclear how to account for the shift our models yield state of the art results on standard datasets. in joint distributions within the regime of deep architectures. 2. Related Work 3. Preliminary Transfer learning (Pan & Yang, 2010) aims to build learning 3.1. Hilbert Space Embedding machines that generalize across different domains following We begin by providing an overview of Hilbert space embed- different probability distributions (Sugiyama et al., 2008; dings of distributions, where each distribution is represented Pan et al., 2011; Duan et al., 2012; Gong et al., 2013; Zhang by an element in a reproducing kernel Hilbert space (RKHS). et al., 2013). Transfer learning finds wide applications in Denote by X a random variable with domain Ω and distribu- computer vision (Saenko et al., 2010; Gopalan et al., 2011; tion P (X), and by x the instantiations of X. A reproducing Gong et al., 2012; Hoffman et al., 2014) and natural lan- kernel Hilbert space (RKHS) H on Ω endowed by a kernel guage processing (Collobert et al., 2011; Glorot et al., 2011). k (x; x0) is a Hilbert space of functions f :Ω 7! R with The main technical problem of transfer learning is how inner product h·; ·iH. Its element k (x; ·) satisfies the repro- to reduce the shifts in data distributions across domains. ducing property: hf (·) ; k (x; ·)iH = f (x). Alternatively, Most existing methods learn a shallow representation model k (x; ·) can be viewed as an (infinite-dimensional) implicit 0 0 by which domain discrepancy is minimized, which cannot feature map φ (x) where k (x; x ) = hφ (x) ; φ (x )iH. Ker- suppress domain-specific exploratory factors of variations. nel functions can be defined on vector space, graphs, time Deep networks learn abstract representations that disentan- series and structured objects to handle diverse applications. gle the explanatory factors of variations behind data (Bengio The kernel embedding represents a probability distribution et al., 2013) and extract transferable factors underlying dif- P by an element in RKHS endowed by a kernel k (Smola ferent populations (Glorot et al., 2011; Oquab et al., 2013), et al., 2007; Sriperumbudur et al., 2010; Gretton et al., 2012) which can only reduce, but not remove, the cross-domain discrepancy (Yosinski et al., 2014). Recent work on deep Z µX (P ) , EX [φ (X)] = φ (x) dP (x); (1) domain adaptation embeds domain-adaptation modules into Ω deep networks to boost transfer performance (Tzeng et al., 2014; 2015; 2017; Ganin & Lempitsky, 2015; Long et al., where the distribution is mapped to the expected feature map, 0 2015; 2016). These methods mainly correct the shifts in i.e. to a point in the RKHS, given that EX [k (x; x )] 6 1. marginal distributions, assuming conditional distributions The mean embedding µX has the property that the expecta- remain unchanged after the marginal distribution adaptation. tion of any RKHS function f can be evaluated as an inner product in H, hµ ; fi [f (X)] ; 8f 2 H. This kind Transfer learning will become more challenging as domains X H , EX of kernel mean embedding provides us a nonparametric per- may change by the joint distributions P (X; Y) of input fea- spective on manipulating distributions by drawing samples tures X and output labels Y. The distribution shifts may from them. We will require a characteristic kernel k such stem from the marginal distributions P (X) (a.k.a. covari- that the kernel embedding µ (P ) is injective, and that the ate shift (Huang et al., 2006; Sugiyama et al., 2008)), the X embedding of distributions into infinite-dimensional feature conditional distributions P (YjX) (a.k.a. conditional shift spaces can preserve all of the statistical features of arbitrary (Zhang et al., 2013)), or both (a.k.a. dataset shift (Quionero- distributions, which removes the necessity of density estima- Candela et al., 2009)). Another line of work (Zhang et al., tion of P . This technique has been widely applied in many 2013; Wang & Schneider, 2014) correct both target and con- tasks, including feature extraction, density estimation and ditional shifts based on the theory of kernel embedding of two-sample test (Smola et al., 2007; Gretton et al., 2012). conditional distributions (Song et al., 2009; 2010; Sriperum- budur et al., 2010). Since the target labels are unavailable, While the true distribution P (X) is rarely accessible, we adaptation is performed by minimizing the discrepancy be- can estimate its embedding using a finite sample (Gretton tween marginal distributions instead of conditional distri- et al., 2012). Given a sample DX = fx1;:::; xng of size n butions.
Recommended publications
  • Transfer Adaptation Learning: a Decade Survey
    JOURNAL OF LATEX CLASS FILES, VOL. -, NO. -, MARCH 2019 1 Transfer Adaptation Learning: A Decade Survey Lei Zhang, Xinbo Gao Abstract—The world we see is ever-changing and it always changes with people, things, and the environment. Domain is referred to as the state of the world at a certain moment. A research problem is characterized as transfer adaptation learning (TAL) when it needs knowledge correspondence between different moments/domains. Conventional machine learning aims to find a model with the minimum expected risk on test data by minimizing the regularized empirical risk on the training data, which, however, supposes that the training and test data share similar joint probability distribution. TAL aims to build models that can perform tasks of target domain by learning knowledge from a semantic related but distribution different source domain. It is an energetic research filed of increasing influence and importance, which is presenting a blowout publication trend. This paper surveys the advances of TAL methodologies in the past decade, and the technical challenges and essential problems of TAL have been observed and discussed with deep insights and new perspectives. Broader solutions of transfer adaptation learning being created by researchers are identified, i.e., instance re-weighting adaptation, feature adaptation, classifier adaptation, deep network adaptation and adversarial adaptation, which are beyond the early semi-supervised and unsupervised split. The survey helps researchers rapidly but comprehensively understand and identify the research foundation, research status, theoretical limitations, future challenges and under-studied issues (universality, interpretability, and credibility) to be broken in the field toward universal representation and safe applications in open-world scenarios.
    [Show full text]
  • Deep Learning for Electromyographic Hand Gesture Signal Classification
    1 Deep Learning for Electromyographic Hand Gesture Signal Classification Using Transfer Learning Ulysse Cotˆ e-Allard,´ Cheikh Latyr Fall, Alexandre Drouin, Alexandre Campeau-Lecours, Clement´ Gosselin, Kyrre Glette, Franc¸ois Laviolettey, and Benoit Gosseliny Abstract—In recent years, deep learning algorithms have widely adopted both in research and clinical settings. The become increasingly more prominent for their unparalleled sEMG signals, which are non-stationary, represent the sum ability to automatically learn discriminant features from large of subcutaneous motor action potentials generated through amounts of data. However, within the field of electromyography- based gesture recognition, deep learning algorithms are seldom muscular contraction [1]. Artificial intelligence can then be employed as they require an unreasonable amount of effort from leveraged as the bridge between sEMG signals and the pros- a single person, to generate tens of thousands of examples. thetic behavior. This work’s hypothesis is that general, informative features can The literature on sEMG-based gesture recognition primarily be learned from the large amounts of data generated by aggre- focuses on feature engineering, with the goal of characterizing gating the signals of multiple users, thus reducing the recording burden while enhancing gesture recognition. Consequently, this sEMG signals in a discriminative way [1], [2], [3]. Recently, paper proposes applying transfer learning on aggregated data researchers have proposed deep learning approaches [4], [5], from multiple users, while leveraging the capacity of deep learn- [6], shifting the paradigm from feature engineering to feature ing algorithms to learn discriminant features from large datasets. learning. Regardless of the method employed, the end-goal Two datasets comprised of 19 and 17 able-bodied participants remains the improvement of the classifier’s robustness.
    [Show full text]
  • A Comprehensive Survey on Transfer Learning
    1 A Comprehensive Survey on Transfer Learning Fuzhen Zhuang, Zhiyuan Qi, Keyu Duan, Dongbo Xi, Yongchun Zhu, Hengshu Zhu, Senior Member, IEEE, Hui Xiong, Fellow, IEEE, and Qing He Abstract—Transfer learning aims at improving the performance of target learners on target domains by transferring the knowledge contained in different but related source domains. In this way, the dependence on a large number of target domain data can be reduced for constructing target learners. Due to the wide application prospects, transfer learning has become a popular and promising area in machine learning. Although there are already some valuable and impressive surveys on transfer learning, these surveys introduce approaches in a relatively isolated way and lack the recent advances in transfer learning. Due to the rapid expansion of the transfer learning area, it is both necessary and challenging to comprehensively review the relevant studies. This survey attempts to connect and systematize the existing transfer learning researches, as well as to summarize and interpret the mechanisms and the strategies of transfer learning in a comprehensive way, which may help readers have a better understanding of the current research status and ideas. Unlike previous surveys, this survey paper reviews more than forty representative transfer learning approaches, especially homogeneous transfer learning approaches, from the perspectives of data and model. The applications of transfer learning are also briefly introduced. In order to show the performance of different transfer learning models, over twenty representative transfer learning models are used for experiments. The models are performed on three different datasets, i.e., Amazon Reviews, Reuters-21578, and Office-31.
    [Show full text]
  • Personalizing EEG-Based Affective Models with Transfer Learning
    Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence (IJCAI-16) Personalizing EEG-Based Affective Models with Transfer Learning 1 1,2,3, Wei-Long Zheng and Bao-Liang Lu ⇤ 1Center for Brain-like Computing and Machine Intelligence Department of Computer Science and Engineering 2Key Laboratory of Shanghai Education Commission for Intelligent Interaction and Cognitive Engineering 3Brain Science and Technology Research Center Shanghai Jiao Tong University, Shanghai, China weilong, bllu @sjtu.edu.cn { } Abstract way to enhance current BCI systems with an increase of infor- mation flow, while at the same time without additional cost. Individual differences across subjects and non- Therefore, aBCIs have attracted increasing interests in both stationary characteristic of electroencephalography research and industry communities, and various studies have (EEG) limit the generalization of affective brain- presented their efficiency and feasibility [Muhl¨ et al., 2014a; computer interfaces in real-world applications. On 2014b; Jenke et al., 2014; Eaton et al., 2015]. Although the the other hand, it is very time consuming and large progress has been obtained about development of aB- expensive to acquire a large number of subject- CIs in recent years, there still exist some challenges such as specific labeled data for learning subject-specific the adaptation to changing environments and individual dif- models. In this paper, we propose to build per- ferences. sonalized EEG-based affective models without la- beled target data using transfer learning techniques. Until now, most of studies emphasized choices of fea- We mainly explore two types of subject-to-subject tures and classifiers [Singh et al., 2007; Jenke et al., 2014; transfer approaches.
    [Show full text]
  • Transfer Learning: Introduction & Application
    Transfer Learning: Introduction & Application Yun Gu Department of Automation Shanghai Jiao Tong University Shanghai, CHINA June 16th, 2014 Overview of Transfer Learning Categorization Applications Conclusion Outline 1 Overview of Transfer Learning 2 Categorization Three Research Issues Different Settings 3 Applications Image Annotation Image Classification Deep Learning 4 Conclusion Y. Gu Transfer Learning: Introduction & Application Overview of Transfer Learning Categorization Applications Conclusion Outline of this section 1 Overview of Transfer Learning 2 Categorization 3 Applications 4 Conclusion Y. Gu Transfer Learning: Introduction & Application Transfer Learning · What to “Transfer”? · Machiine Learniing Scheme.. · How to “Transfer”? · Rellatiionshiip wiith other ML · When to “Transfer”? tech? In Top-Level Conference: 7 papers in CVPR 2014 are related with Transfer Learning; In Top-Level Journal: M.Guilaumin, et.al, "ImageNet Auto-Annotation with Segmentation Propagation", IJCV,2014 Overview of Transfer Learning Categorization Applications Conclusion What is Transfer Learning?[Pan and Yang, 2010] Naive View (Transfer Learning) Transfer Learning (i.e. Knowledge Transfer, Domain Adaption) aims at applying knowledge learned previously to solve new problems faster or with better solutions. Y. Gu Transfer Learning: Introduction & Application Transfer Learning · What to “Transfer”? · Machiine Learniing Scheme.. · How to “Transfer”? · Rellatiionshiip wiith other ML · When to “Transfer”? tech? In Top-Level Conference: 7 papers in CVPR 2014 are related with Transfer Learning; In Top-Level Journal: M.Guilaumin, et.al, "ImageNet Auto-Annotation with Segmentation Propagation", IJCV,2014 Overview of Transfer Learning Categorization Applications Conclusion What is Transfer Learning?[Pan and Yang, 2010] Naive View (Transfer Learning) Transfer Learning (i.e. Knowledge Transfer, Domain Adaption) aims at applying knowledge learned previously to solve new problems faster or with better solutions.
    [Show full text]
  • Transfer Learning for Reinforcement Learning Domains: a Survey
    JournalofMachineLearningResearch10(2009)1633-1685 Submitted 6/08; Revised 5/09; Published 7/09 Transfer Learning for Reinforcement Learning Domains: A Survey Matthew E. Taylor∗ [email protected] Computer Science Department The University of Southern California Los Angeles, CA 90089-0781 Peter Stone [email protected] Department of Computer Sciences The University of Texas at Austin Austin, Texas 78712-1188 Editor: Sridhar Mahadevan Abstract The reinforcement learning paradigm is a popular way to address problems that have only limited environmental feedback, rather than correctly labeled examples, as is common in other machine learning contexts. While significant progress has been made to improve learning in a single task, the idea of transfer learning has only recently been applied to reinforcement learning tasks. The core idea of transfer is that experience gained in learning to perform one task can help improve learning performance in a related, but different, task. In this article we present a framework that classifies transfer learning methods in terms of their capabilities and goals, and then use it to survey the existing literature, as well as to suggest future directions for transfer learning work. Keywords: transfer learning, reinforcement learning, multi-task learning 1. Transfer Learning Objectives In reinforcement learning (RL) (Sutton and Barto, 1998) problems, leaning agents take sequential actions with the goal of maximizing a reward signal, which may be time-delayed. For example, an agent could learn to play a game by being told whether it wins or loses, but is never given the “correct” action at any given point in time. The RL framework has gained popularity as learning methods have been developed that are capable of handling increasingly complex problems.
    [Show full text]
  • Transfer Learning Approach for Occupancy Prediction in Smart Buildings
    Transfer Learning Approach for Occupancy Prediction in Smart Buildings Mohamad Khalil Stephen McGough Zoya Pourmirza School of Engineering School of Computing School of Engineering Newcastle University Newcastle University Newcastle University Newcastle, Uk Newcastle, UK Newcastle, UK [email protected] [email protected] [email protected] Mehdi Pazhoohesh Sara Walker Faculty of Technology School of Engineering De Montfort University Newcastle University Leicester, UK Newcastle, UK [email protected] [email protected] Abstract— Accurate occupancy prediction in smart I. INTRODUCTION buildings is a key element to reduce building energy consumption and control HVAC systems (Heating – Ventilation In buildings the term of occupancy information refers to and– Air Conditioning) efficiently, resulting in an increment of occupants’ presence or absence and their movements. human comfort. This work focuses on the problem of occupancy Accurate real-time occupancy information in smart prediction modelling (occupied / unoccupied) in smart buildings buildings can help HVAC (Heating – Ventilation and– Air using environmental sensor data. A novel transfer learning Conditioning) control systems to optimize their usage and approach was used to enhance occupancy prediction accuracy minimize building energy consumption [1]. In the past years, when the amounts of historical training data are limited. The occupancy prediction modelling in buildings has been studied proposed approach and models are applied to a case study of by many researchers using different types of data such as air three office rooms in an educational building. The data sets used temperature, sound, door status, relative humidity, camera, in this work are actual data collected from the Urban Sciences motion and light.
    [Show full text]
  • Stylegans and Transfer Learning for Generating Synthetic Images in Industrial Applications
    S S symmetry Article StyleGANs and Transfer Learning for Generating Synthetic Images in Industrial Applications Harold Achicanoy 1,2,*,† , Deisy Chaves 2,3,*,† and Maria Trujillo 2,† 1 Alliance of Bioversity International and CIAT, Km 17 Recta Cali-Palmira, Palmira 763537, Colombia 2 School of Computer and Systems Engineering, Universidad del Valle, Cali 760001, Colombia; [email protected] 3 Department of Electrical, Systems and Automation, Universidad de León, 24007 León, Spain * Correspondence: [email protected] (H.A.); [email protected] (D.C.) † These authors contributed equally to this work. Abstract: Deep learning applications on computer vision involve the use of large-volume and repre- sentative data to obtain state-of-the-art results due to the massive number of parameters to optimise in deep models. However, data are limited with asymmetric distributions in industrial applications due to rare cases, legal restrictions, and high image-acquisition costs. Data augmentation based on deep learning generative adversarial networks, such as StyleGAN, has arisen as a way to create train- ing data with symmetric distributions that may improve the generalisation capability of built models. StyleGAN generates highly realistic images in a variety of domains as a data augmentation strategy but requires a large amount of data to build image generators. Thus, transfer learning in conjunction with generative models are used to build models with small datasets. However, there are no reports on the impact of pre-trained generative models, using transfer learning. In this paper, we evaluate a StyleGAN generative model with transfer learning on different application domains—training Citation: Achicanoy, H.; Chaves, D.; with paintings, portraits, Pokémon, bedrooms, and cats—to generate target images with different Trujillo, M.
    [Show full text]
  • Deep Learning Occupancy Activity Detection Approach for Optimising Building Energy Loads
    International Conference on Applied Energy 2020 Dec. 1 - Dec. 10, 2020, Bangkok / Virtual Paper ID: 127 Deep learning occupancy activity detection approach for optimising building energy loads Paige Wenbin Tien*, Shuangyu Wei, John Kaiser Calautit, Jo Darkwa, Christopher Wood Department of Architecture and Built Environment, University of Nottingham, Nottingham, United Kingdom *e-mail: [email protected], [email protected] ABSTRACT Heating, ventilation and air- HVAC The main aim of this paper is to develop a vision- conditioning based deep learning method for real-time occupancy activity detection and recognition to help the operations 1. INTRODUCTION AND LITERATURE REVIEW of building energy systems. A faster region-based The built environment sector is responsible for up to convolutional neural network was developed, trained 35% of the global energy use and energy-related and deployed to an artificial intelligence (AI)-powered emissions [1]. Reducing the energy consumption of camera for the application of real-time occupancy buildings requires innovative methods. Solutions such as activity detection. Initial experimental tests were occupancy-based controls can achieve significant energy performed within an office space of a selected case study savings by eliminating unnecessary energy usage. building. The detection provided correct detections for A significant element affecting the usage of these the majority of the time (97.32%). Average detection energy consumers is the behaviour of the occupants [2]. accuracy of 92.20% was achieved for all activities. For instance, rooms in offices or lecture theatres are not Building energy simulation of the case study building was fully utilised or occupied during the day.
    [Show full text]
  • Use Transfer Learning for Efficient Deep Learning Training on Intel® Xeon® Processors
    WHITE PAPER Artificial Intelligence Transfer Learning Intel AI® Builders Use Transfer Learning For Efficient Deep Learning Training On Intel® Xeon® Processors Authors Introduction Beenish Zia This is an educational white paper on transfer learning, showcasing how existing Intel Corporation deep learning models can be easily and flexibly customized to solve new problems. One of the biggest challenges with deep learning is the large number Ramesh Illikkal of labeled data points that are required to train the deep learning models to Intel Corporation sufficient accuracy. For example, the ImageNet*20 database for image recognition Bob Rogers consists of over 14 million hand-labeled images. While the number of possible Intel Corporation applications of deep learning systems in vision tasks, text processing, speech-to- text translation and many other domains is enormous, very few potential users of deep learning systems have sufficient training data to create models from scratch. A common concern among teams considering the use of deep learning to solve business problems is the need for training data: "Doesn't deep learning need Table of Contents millions of samples and months of training to get good results?" One powerful Introduction . 1 solution is transfer learning, in which part of an existing deep learning model is re-optimized on a small data set to solve a related, but new, problem. In fact, What Is Transfer Learning . .. 2 one of the great attractions of transfer learning is that, unlike most traditional Benefits Using Transfer Learning . 3 approaches to machine learning, we can take models trained on one (perhaps very large) dataset and modify them quickly and easily to work well on a new problem Applications Transfer Learning .
    [Show full text]
  • Deep Learning for EMG-Based Human-Machine Interaction: a Review Dezhen Xiong, Daohui Zhang, Member, IEEE, Xingang Zhao, Member, IEEE, and Yiwen Zhao
    512 IEEE/CAA JOURNAL OF AUTOMATICA SINICA, VOL. 8, NO. 3, MARCH 2021 Deep Learning for EMG-based Human-Machine Interaction: A Review Dezhen Xiong, Daohui Zhang, Member, IEEE, Xingang Zhao, Member, IEEE, and Yiwen Zhao Abstract—Electromyography (EMG) has already been broadly intramuscular EMG (iEMG), according to the electrodes’ used in human-machine interaction (HMI) applications. location. The former is collected from the surface of human Determining how to decode the information inside EMG signals robustly and accurately is a key problem for which we urgently skin, while the latter is collected from needle electrodes need a solution. Recently, many EMG pattern recognition tasks planted inside the human muscle. sEMG has been widely used have been addressed using deep learning methods. In this paper, for hand gesture classification [2], [3], silent speech recogni- we analyze recent papers and present a literature review tion [4], [5], stroke rehabilitation [6], [7], robot control [8], describing the role that deep learning plays in EMG-based HMI. [9], and other applications, mainly because it is cheap and An overview of typical network structures and processing schemes will be provided. Recent progress in typical tasks such as easy to collect and it provides a method for more natural movement classification, joint angle prediction, and force/torque human-machine collaboration. estimation will be introduced. New issues, including multimodal Many approaches, such as video, inertial measurement units sensing, inter-subject/inter-session, and robustness toward (IMU), and EMG, can be used to decode the movement disturbances will be discussed. We attempt to provide a comprehensive analysis of current research by discussing the intention of humans.
    [Show full text]
  • Transfer Learning and Intelligence: an Argument and Approach
    In The First Conference on Artificial General Intelligence (AGI-08), Memphis, Tennessee, March 2008. Transfer Learning and Intelligence: an Argument and Approach Matthew E. TAYLOR, Gregory KUHLMANN, and Peter STONE Department of Computer Sciences The University of Texas at Austin {mtaylor, kuhlmann, pstone}@cs.utexas.edu Abstract. In order to claim fully general intelligence in an autonomous agent, the ability to learn is one of the most central capabilities. Classical machine learn- ing techniques have had many significant empirical successes, but large real-world problems that are of interest to generally intelligent agents require learning much faster (with much less training experience) than is currently possible. This paper presents transfer learning, where knowledge from a learned task can be used to significantly speed up learning in a novel task, as the key to achieving the learning capabilities necessary for general intelligence. In addition to motivating the need for transfer learning in an intelligent agent, we introduce a novel method for select- ing types of tasks to be used for transfer and empirically demonstrate that such a selection can lead to significant increases in training speed in a two-player game. Keywords. Transfer Learning, Game Tree Search, Reinforcement Learning 1. Introduction A generally intelligent agent deployed in any non-trivial environment must be able to learn: no designer can anticipate every possible encountered situation. Specifically, in- telligent agents need to learn how to act, which is the purview of Reinforcement learn- ing [8,15]. Reinforcement learning (RL) problems are defined as those in which learn- ing agents sequentially execute actions with the goal of maximizing a reward signal, which may be time-delayed.
    [Show full text]