Advances in Reinforcement Learning and Their Implications for Intelligent Control

Total Page:16

File Type:pdf, Size:1020Kb

Advances in Reinforcement Learning and Their Implications for Intelligent Control Advances in Reinforcement Learning and Their Implications for Intelligent Control Steven D. Whitehead*, Richard S. Suttoni and Dana H. Ballard* 1 Introduction 0 Robust under Incomplete and Uncertain Domain Knowledge: An intelligent control sys- What is an intelligent control system, and how will we tem must not depend upon a complete and accu- know one when we see it? This question is hard to rate domain model. Even for relatively simple, answer definitively, but intuitively what distinguishes narrow domains, it is extremely difficult to build intelligent control from more mundane control is the models that are complete and accurate [ShaSO]. complexity of the task and the ability of the control If control depends upon a domain model, it must system to deal with uncertain and changing conditions. suffice that the model be incomplete and inaccu- A list of the definitive properties of an intelligent con- rate. If domain knowledge is available, then the trol system might include some of the following: system should be capable of exploiting it, but it should not be a prerequisite for intelligent con- 0 Effective: The first requirement of any control system, intelligent or not, is that it generate con- trol. A system that uses a domain model that is trol actions that adequately control the plant (en- learned incrementally through experience is to be vironment). The adverb adequately is intention- preferred over one that relies upon a complete a ally imprecise since effective control rarely means priori model. optimal, but more often means sufficient for the Perceptual Feasibility: The information pro- purposes of the task. vided directly by a system's sensors is necessar- ily limited, and an intelligent control architecture 0 Reactive: Equally important is that a system's control be timely. Effective control is useless if must take this limitation into account. Instead of it is executed too late. In general, an intelligent assuming that any and all information about the control system cannot require arbitrarily long de- state of the environment is immediately available, lays between control decisions. On the contrary, intelligent system must be designed with limited it must be capable of making decisions on a mo- but flexible sensory-motor systems. Control algo- ments notice-upon demand. If extra time is rithms must consider actions for collecting neces- available that can be exploited to improve the sary state information as well as actions for af- quality of a decision, so much the better, but con- fecting change in the environment. trol decisions must be available any--lime [DB88]. 0 Mathematical Foundations: To facilitate per- formance analysis, it is important that an intelli- 0 Situated: An intelligent control system should be tightly coupled to its environment, and its con- gent control system be based on a framework that trol decisions must be based on the immediate sit- has a solid mathematical foundation. uation. A system must be capable of responding While the above list is not particularly surprising to unexpected contingencies and opportunities as or new, few if any control systems have been built they arise [AC87]. that satisfy all of the requirements. For example, 0 Adaptive An intelligent control system must use the problem-solving architectures that have been the its experience with the environment to improve dominate approach in Artificial Intelligence over the its performance. Adaptability allows a system to past twenty years fall well short of these goals. These refine an initially suboptimal control policy and problem-solving architectures are not particularly re- to maintain effective control in the face of non- active, situated, adaptive, or robust in the face of in- stationary environments. complete and inaccurate domain models. They also tend to make unrealistic assumptions about the ca- Department of Computer Science, University of pabilities of the sensory system. While extensions Rochester, Rochester NY 14627 and modifications to these architectures continue to 'GTE Laboratories Incorporated, Waltham, MA 02254 be popular, other approaches are being considered as 1289 TH0333-5/90/0000/1289$01.OO 0 1990 IEEE Authorized licensed use limited to: IEEE Xplore. Downloaded on January 30, 2009 at 17:01 from IEEE Xplore. Restrictions apply. well. (punishment). This paper focuses on control architectures that are Recent examples of controllers based on reinforce- based on reinforcement learning. In particular, we sur- ment learning include Barto et al.’s pole balancer vey several recent advances in reinforcement learning [BSA83, Sut841, Grefenstette’s simulated flight con- that have substantially increased its viability as a gen- troller [Greg01, Lin’s animats [Lingo], and Franklin’s eral approach to intelligent control. adapt,ive robot controllers [Fra88], among others. We begin by considering the relationship between re- inforcement learning and dynamic programming, both 2.1 Evaluating Reinforcement Learning viewed as methods for solving multi-stage decision problems [Wat89, BSWSO]. It is shown that many re- Reinforcement learning is emerging as an important alternative to classical problem-solving approaches to inforcement learning algorithms can be viewed as a kind of incremental dynamic programming; this pro- intelligent control because it possesses many of the properties for intelligent control that problem-solving vides a mathematical foundation for the study of rein- focrement learning systems. Connecting reinforcement approaches lack. In many respects the two approaches are complimentary and it is likely that eventual intel- learning to dynamic programming has also led to a ligent control architectures will incorporate aspects of strong optimal convergence theorem for one class of both.’ reinforcement learning algorithms and opens the door for similar analyses of other algorithms [Wat89]. Following is a discussion of the degree to which cur- Next we show how reinforcement learning methods rent reinforcement learning systems achieve each of the can be used to go beyond simple trial-and-error learn- properties that we associate with intelligent control. ing. By augmenting them with a predictive domain Effective: Reinforcement learning systems are model and using the model to perform a kind of in- effective in the sense that they eventually learn cremental planning, their learning performance can be effective control strategies. Although a system’s substantially improved. These control architectures initial performance may be poor, with enough in- learn both by performing experiments in the world teraction with the world it will eventually learn and by searching a domain model. ,Because the do- an effective strategy for obtaining reward. For main model need not be complete or accurate, it can the most part, the asymptotic effectiveness of re- be learned incrementally through experience with the inforcement learning systems has been validated world [SutSOa, SutSOb, WB89, Whi89, Lingo]. only empirically, however recent advances in the Finally, we discuss active sensory-motor systems theory of reinforcement learning have yielded for feasible perception and how these systems inter- mathematical results that guarantee optimality in act with reinforcement learning. We find that, with the limit for an important class of reinforcement some modification, many of the ideas from reinforce- learning systems [Wat89]. ment learning can be successfully combined with ac- tive sensory-motor systems. The system then learns Reactive: Decision-making in reinforcement not only an overt control strategy, but also where to learning systems is based on a policy func- focus its at,tention in order to collect necessary sensory tion which maps situations (inputs) directly into information [WBSOb]. actions (outputs) and which can be evaluated The results surveyed in this paper have been re- quickly. Consequently, reinforcement learning ported elsewhere (primarily in the machine learning systems are extremely reactive. literature). The objective here is to summarize them Situated: Reinforcement learning systems are and to consider their implications for the design of in- situated because each action is choser? based on telligent control architectures. the current state of the world. 2 Reinforcement Learning for Adaptive: Reinforcement learning systems are adaptive because they use feedback to improve Intelligent Control their performance. What is reinforcement learning? A reinforcement Incomplete and Uncertain Domain Knowl- learning system is any system that through interac- edge: Reinforcement learning systems do not de- tion with its environment improves its performance by pend upon internal domain models because they receiving feedback in the form of a scalar reward (or learn through trial-and-error experience with the penalty) that is commensurate with the appropriate- world. However, when available, they can ex- ness of the response. By improves its performance, ploit domain knowledge by 1) using prior knowl- we mean that the system uses the feedback to adapt edge about the control task to determine a good its behavior in an effort to maximize some measure of the reward it receives in the future. Intuitively, ‘To some extent this int,egration has already begun to a reinforcement learning system can be viewed as a occur, with the development of reinforcement learning sys- hedonistic automaton whose sole objective is to maxi- tem that learn and use internal domain models to improve mize the positive (reward) and minimize the
Recommended publications
  • Handbook of Intelligent Control
    HANDBOOK OF INTELLIGENT CONTROL NEURAL, FUZZY, AND ADAPTIVE APPROACHES Edited by David A. White Donald A. Sorge \~1l ~ VAN NOSTRAND REINHOLD ~ New York Co,", 10 NEURAL NETWORKS, SYSTEM IDENTIFICATION, AND CONTROL IN THE CHEMICAL PROCESS INDUSTRIES Paull. Werbos NSF Program Director for Neuroengineering Co-Director for Emerging Technologies Initiation Thomas McAvoy and Ted Su Department of Chemical Engineering. University of Maryland 10.1. INTRODUCTION AND SUMMARY This chapter will discuss the application of artificial neural networks (ANNs) in the chemical process industries. with particular emphasis on new designs that could be useful in other applications (neural and nonneuraI) as well: .The use of simple ANNs as the system identification component of model-based predictive control (MPC), a widely used scheme that optimizes over time subject to constraints. .The use of "robust" or "parallel" training of ANNs used in system identification. In tests on real data from a wastewater treatment plant and an oil refinery, this has led to an orders-of -magnitude reduction in prediction error when compared with conventional maximum-likelihood ap- proaches. Earlier tests on nonneural implementations were also promising. There are deep theoretical lessons involved, significant to all forms of ANN adaptation and system identifica- tion. .Techniques for pruning ANNs (a task which has turned out to be crucial to ANN performance). techniques for using ANN-like techniques with other nonlinear specifications (for those who do not want to use ANNs as such) and techniques for using ANN models similar to nonlinear ARMA models or Kalman filtering systems-to be given in later sections. ANNs are already in use in real-world applications by many of the major chemical manufacturers.
    [Show full text]
  • Handbook of Intelligent Control
    HANDBOOK OF INTELLIGENT CONTROL NEURAL, FUZZY, AND ADAPTIVE APPROACHES Edited by David A. White Donald A. Sorge \~1l ~ VAN NOSTRAND REINHOLD ~ New York FOREWORD 1 This book is an outgrowth of discussions that got started in at least three workshops sponsored by the National Science Foundation (NSF): .A workshop on neurocontrol and aerospace applications held in October 1990, under joint sponsorship from McDonnell Douglas and the NSF programs in Dynamic Systems and Control and Neuroengineering .A workshop on intelligent control held in October 1990, under joint sponsorship from NSF and the Electric Power Research Institute, to scope out plans for a major new joint initiative in intelligent control involving a number of NSF programs .A workshop on neural networks in chemical processing, held at NSF in January-February 1991, sponsored by the NSF program in Chemical Reaction Processes The goal of this book is to provide an authoritative source for two kinds of information: (1) fundamental new designs, at the cutting edge of true intelligent control, as well as opportunities for future research to improve on these designs; (2) important real-world applications, including test problems that constitute a challenge to the entire control community. Included in this book are a seriesof realistic test problems, worked out through lengthy discussions between NASA, NetJroDyne, NSF, McDonnell Douglas, and Honeywell, which are more than just benchmarks for evaluating intelligent control designs. Anyone who contributes to solving these problems may well be playing a crucial role in making possible the future development of hypersonic vehicles and subsequently the economic human settlement of outer space.This book also emphasizes chemical process applications (capable of improving the environment as well as increasing profits), the manufacturing of high-qual- ity composite parts, and robotics.
    [Show full text]
  • Complex System and Intelligent Control: Theories and Applications
    Chen et al. / Front Inform Technol Electron Eng 2019 20(1):1-3 1 Frontiers of Information Technology & Electronic Engineering www.jzus.zju.edu.cn; engineering.cae.cn; www.springerlink.com ISSN 2095-9184 (print); ISSN 2095-9230 (online) E-mail: [email protected] Editorial: Complex system and intelligent control: theories and applications Jie CHEN†‡1,2, Ben M. CHEN3,4, Jian SUN1 1School of Automation, Beijing Institute of Technology, Beijing 100081, China 2Tongji University, Shanghai 200092, China 3Department of Mechanical and Automation Engineering, Chinese University of Hong Kong, Hong Kong, China 4Department of Electrical and Computer Engineering, National University of Singapore, Singapore †E-mail: [email protected] https://doi.org/10.1631/FITEE.1910000 Complex systems are the systems that consist of and swarming intelligence. Intelligent control tries to a great many diverse and autonomous but interacting borrow ideas from the sciences of physics, mathe- and interdependent components whose aggregate matics, biology, neuroscience, and others to develop a behaviors are nonlinear. As phased by Aristotle, “the control system to manage a complex process in an whole is more than the sum of its parts;” properties of uncertain environment. Intelligent control is a kind of complex systems are not a simple summation of their interdisciplinary new technology, which integrates individual parts. the knowledges of control theory, computer science, Complex systems are widespread. Typical exam- artificial intelligence, information theory, and other ples of complex systems can be found in the human areas. Intelligent control has been shown to be effec- brain, flocking formation of migrating birds, power tive in controlling complex systems where traditional grid, transportation systems, autonomous vehicles, control methods can hardly perform well.
    [Show full text]
  • The IEEE Control Systems Society (CSS) Technical Committee on Intelligent Control (TCIC) Was Established in 1985 to Foster A
    The IEEE Control Systems Society (CSS) Technical Committee on Intelligent Control (TCIC) was established in 1985 to foster a community and to provide forums for the theoretical and practical considerations of intelligent control techniques and their application to devices and systems. From its inception through the turn of the millennia, rapid advancements established cornerstone techniques for incorporating intelligent control techniques within feedback control approaches. For example, fundamental contributions to our understanding of how to use Lyapunov-based analysis methods to inform the design of gradient and least-squares adaptive update laws and modular and composite adaptive control were established. Neural networks (NN) and fuzzy logic (FL) were transformed from computational tools used for pattern matching based on a finite set of training data to powerful function approximation tools that used backpropagation with adaptation through online state feedback. Iterative (ILC) and repetitive learning control (RLC) methods were developed and matured, and various methods were developed for improved transient performance during adaptation and simultaneous parameter identification. Refinements, generalizations, and the application of these methods are still on-going areas of research. The aforementioned intelligent control techniques have been successfully absorbed into mainstream control systems research. Adoption of intelligent control within control communities is evident from the number of conferences across several societies that
    [Show full text]
  • RECURRENT NEURAL NETWORKS Design and Applications
    RECURRENT NEURAL NETWORKS Design and Applications Edited by L.R. Medsker Departments of Physics and Computer Science and Information Systems American University Washington, D.C. L.C. Jain Knowledge-Based Intelligent Engineering Systems Centre Faculty of Information Technology Director/Founder, KES University of South Australia, Adelaide The Mawson Lakes, SA Australia CRC Press Boca Raton London New York Washington, D.C. PREFACE Recurrent neural networks have been an interesting and important part of neural network research during the 1990's. They have already been applied to a wide variety of problems involving time sequences of events and ordered data such as characters in words. Novel current uses range from motion detection and music synthesis to financial forecasting. This book is a summary of work on recurrent neural networks and is exemplary of current research ideas and challenges in this subfield of artificial neural network research and development. By sharing these perspectives, we hope to illuminate opportunities and encourage further work in this promising area. Two broad areas of importance in recurrent neural network research, the architectures and learning techniques, are addressed in every chapter. Architectures range from fully interconnected to partially connected networks, including recurrent multilayer feedforward. Learning is a critical issue and one of the primary advantages of neural networks. The added complexity of learning in recurrent networks has given rise to a variety of techniques and associated research projects. A goal is to design better algorithms that are both computationally efficient and simple to implement. Another broad division of work in recurrent neural networks, on which this book is structured, is the design perspective and application issues.
    [Show full text]
  • Application of Reinforcement Learning Techniques to the Development of Flight Control Systems for Miniature Uav Rotorcraft
    MACHINE LEARNING FOR INTELLIGENT CONTROL: APPLICATION OF REINFORCEMENT LEARNING TECHNIQUES TO THE DEVELOPMENT OF FLIGHT CONTROL SYSTEMS FOR MINIATURE UAV ROTORCRAFT A thesis submitted in partial fulfilment of the requirements for the Degree of Master of Engineering in Mechanical Engineering in the University of Canterbury by Edwin Hayes University of Canterbury 2013 Abstract This thesis investigates the possibility of using reinforcement learning (RL) techniques to create a flight controller for a quadrotor Micro Aerial Vehicle (MAV). A capable flight control system is a core requirement of any unmanned aerial vehicle. The challenging and diverse applications in which MAVs are destined to be used, mean that consider- able time and effort need to be put into designing and commissioning suitable flight controllers. It is proposed that reinforcement learning, a subset of machine learning, could be used to address some of the practical difficulties. While much research has delved into RL in unmanned aerial vehicle applications, this work has tended to ignore low level motion control, or been concerned only in off-line learning regimes. This thesis addresses an area in which accessible information is scarce: the performance of RL when used for on-policy motion control. Trying out a candidate algorithm on a real MAV is a simple but expensive proposition. In place of such an approach, this research details the development of a suitable simulator environment, in which a prototype controller might be evaluated. Then inquiry then proposes a possible RL-based control system, utilising the Q-learning algorithm, with an adaptive RBF-network providing function approximation. The operation of this prototypical control system is then tested in detail, to determine both the absolute level of performance which can be expected, and the effect which tuning critical parameters of the algorithm has on the functioning of the controller.
    [Show full text]
  • Machine Learning in a Dynamic World" (Invited), M.M.Kokar Organizer, P Roc
    P. J. Antsaklis, "Learning in Control,” in Panel Discussion on "Machine Learning in a Dynamic World" (Invited), M.M.Kokar Organizer, P roc. o f t he 3 rd I EEE I ntern. S ymposium o n I ntelligent C ontro l , pp. 500- 507, Arlington, VA, Aug. 24-26, 1988. MACHINE LEARNING IN A DYNAMIC WORLD Panel Dmussion Miecsyslaw M. Kohr Panel O&m and Editor Northeastern Univdty Boston, Massachwtts Panelists: P. J. Atnsaklis, University of Notre Dame K. A. DeJong, George Mason University A. L. Meyrowitz, Office of Naval Research A. Meystel, Drexel University R. S. Michal~gi,George Mason University R. S. Sutton, GTE Laboratories Machine Learning was established as a research discipline in the lems while ignoring common-senseknowledge. How much of progresd 1970's and experienced a growth expansion in the 1980's. One of in controlling real-world systems can be made without taking into the roots of machine learning research was in Cybernetic Systems account all kinds of imprecise information? and Adaptive Control. Machine Learning has been significantly influenced by Artificial Intelligence, Cognitive Science, Computer In the following we present the views of the panelists. This pre- Science, and other disciplines; Machine Learning has developed its sentation is based on the written material submitted to the panel own research paradigms, methodologies, and a set of research ob- organizers by the panelists. The preeentation has been divided jectives different from those of control systems. In the meantime, a into four sections. The first section is devoted to the trends in the new field of Intelligent Control has emerged.
    [Show full text]
  • Classifying Intelligence in Machines: a Taxonomy of Intelligent Control †
    robotics Article Classifying Intelligence in Machines: A Taxonomy of Intelligent Control † Callum Wilson * , Francesco Marchetti , Marilena Di Carlo , Annalisa Riccardi and Edmondo Minisci Intelligent Computational Engineering Laboratory, Department of Mechanical and Aerospace Engineering, University of Strathclyde, Glasgow G1 1XJ, UK; [email protected] (F.M.); [email protected] (M.D.C.); [email protected] (A.R.); [email protected] (E.M.) * Correspondence: [email protected] † This paper is an extended version of our paper published in Wilson, C.; Marchetti, F.; Di Carlo, M.; Riccardi, A.; Minisci, E. Intelligent Control: A Taxonomy. In Proceedings of the 8th International Conference on Systems and Control (ICSC 2019), Marrakech, Morocco, 23–25 October 2019. Received: 30 June 2020; Accepted: 18 August 2020; Published: 21 August 2020 Abstract: The quest to create machines that can solve problems as humans do leads us to intelligent control. This field encompasses control systems that can adapt to changes and learn to improve their actions—traits typically associated with human intelligence. In this work we seek to determine how intelligent these classes of control systems are by quantifying their level of adaptability and learning. First we describe the stages of development towards intelligent control and present a definition based on literature. Based on the key elements of this definition, we propose a novel taxonomy of intelligent control methods, which assesses the extent to which they handle uncertainties in three areas: the environment, the controller, and the goals. This taxonomy is applicable to a variety of robotic and other autonomous systems, which we demonstrate through several examples of intelligent control methods and their classifications.
    [Show full text]
  • Identification and Predictive Control Using Recurrent Neural Networks
    International Master’s Thesis Identification and Predictive Control Using Recurrent Neural Networks Nima Mohajerin Technology Studies from the Department of Technology at Örebro University örebro 2012 Identification and Predictive Control Using Recurrent Neural Networks Studies from the Department of Technology at Örebro University Nima Mohajerin Identification and Predictive Control Using Recurrent Neural Networks Supervisor: Prof. Ivan Kalaykov Examiners: Dr. Dimitar Dimitrov Dr. Boyko Iliev © Nima Mohajerin, 2012 Title: Identification and Predictive Control Using Recurrent Neural Networks ISSN 1650-8580 Abstract In this thesis, a special class of Recurrent Neural Networks (RNN) is em- ployed for system identification and predictive control of time dependent sys- tems. Fundamental architectures and learning algorithms of RNNs are studied upon which a generalized architecture over a class of state-space represented networks is proposed and formulated. Levenberg-Marquardt (LM) learning algorithm is derived for this architecture and a number of enhancements are introduced. Furthermore, using this recurrent neural network as a system iden- tifier, a Model Predictive Controller (MPC) is established which solves the op- timization problem using an iterative approach based on the LM algorithm. Simulation results show that the new architecture accompanied by LM learn- ing algorithm outperforms some of existing methods. The third approach which utilizes the proposed method in on-line system identification enhances the iden- tification/control process even more. i Acknowledgements I would like to express my sincere gratitude to my supervisor, Professor Ivan Kalaykov, who trusted in me and gave me the opportunity to pursue my pas- sion in studying and conducting my research in Neural Networks. I hope I have fulfilled his expectations.
    [Show full text]
  • The Challenge of Intelligent Systems
    Proceedings of the 15th IEEE International Symposium on Intelligent Control (ISIC 2000) Rio, Patras, GREECE 17-19 July, 2000 THE CHALLENGE OF INTELLIGENT SYSTEMS ALEXANDER MEYSTEL†§, ELENA MESSINA§ †Drexel University, Philadelphia, PA, USA, [email protected] §National Institute of Standards and Technology, Gaithersburg, MD, USA [email protected] Abstract. This paper analyzes the meaning of intelligence within the domain of intelligent control. Because of historical reasons, there are varying and diverse definitions of what constitutes an intelligent control system. Due to the multidisciplinary nature of intelligent control, there are no good formal measures for what is an intelligent controller or for its performance. We argue that there are characteristics which can be used to categorize a controller as being intelligent. Furthermore, the measure of how intelligent a system is should take into account a holistic view of the system’s operations and design requirements. Beginning with the definition of intelligence characteristics and performance measures, we can move towards a science of intelligent systems for control. Key words. Intelligent control, intelligent systems, performance measures In this paper, we analyze the history of intelligent 1. INTRODUCTION control, its multidisciplinarity, and the higher level issues of “meta control” that characterize intelligent “Intelligent Control” is a commonly used term, systems. We argue that there are characteristics that applied to an amorphous set of tools, techniques, and can be found in common among most, if not all, approaches to implementing control systems that systems labeled as intelligent controllers. These have capabilities beyond those attainable through characteristics suggest espousing a broader view of classical control.
    [Show full text]
  • Handbook of Intelligent Control
    HANDBOOK OF INTELLIGENT CONTROL NEURAL, FUZZY, AND ADAPTIVE APPROACHES Edited by David A. White Donald A. Sorge \~1l ~ VAN NOSTRAND REINHOLD ~ New York FOREWORD 1 This book is an outgrowth of discussions that got started in at least three workshops sponsored by the National Science Foundation (NSF): .A workshop on neurocontrol and aerospace applications held in October 1990, under joint sponsorship from McDonnell Douglas and the NSF programs in Dynamic Systems and Control and Neuroengineering .A workshop on intelligent control held in October 1990, under joint sponsorship from NSF and the Electric Power Research Institute, to scope out plans for a major new joint initiative in intelligent control involving a number of NSF programs .A workshop on neural networks in chemical processing, held at NSF in January-February 1991, sponsored by the NSF program in Chemical Reaction Processes The goal of this book is to provide an authoritative source for two kinds of information: (1) fundamental new designs, at the cutting edge of true intelligent control, as well as opportunities for future research to improve on these designs; (2) important real-world applications, including test problems that constitute a challenge to the entire control community. Included in this book are a seriesof realistic test problems, worked out through lengthy discussions between NASA, NetJroDyne, NSF, McDonnell Douglas, and Honeywell, which are more than just benchmarks for evaluating intelligent control designs. Anyone who contributes to solving these problems may well be playing a crucial role in making possible the future development of hypersonic vehicles and subsequently the economic human settlement of outer space.This book also emphasizes chemical process applications (capable of improving the environment as well as increasing profits), the manufacturing of high-qual­ ity composite parts, and robotics.
    [Show full text]
  • Fuzzy Control Systems - Jens Jäkel, Ralf Mikut and Georg Bretthauer
    CONTROL SYSTEMS, ROBOTICS AND AUTOMATION – Vol. XVII - Fuzzy Control Systems - Jens Jäkel, Ralf Mikut and Georg Bretthauer FUZZY CONTROL SYSTEMS Jens Jäkel, Ralf Mikut and Georg Bretthauer Institute of Applied Computer Science, Forschungszentrum Karlsruhe GmbH, Germany Keywords: Adaptation, defuzzification, expert knowledge, fuzzification, fuzzy control, fuzzy controller design, fuzzy control structures, fuzzy logic, fuzzy model, fuzzy operator, fuzzy Petri net, fuzzy set, fuzzy system, fuzzy system analysis, implementation, inference, intelligent control, knowledge-based system, linear system, linguistic term, linguistic variable, Mamdani-type fuzzy system, membership function, nonlinear system, rule base, Takagi-Sugeno-type fuzzy system, supervision Contents 1. Introduction 2. Fuzzy Control – A Simple Example 2.1. Example 2.2. Fuzzy Sets, Linguistic Variables and Fuzzy IF-THEN Rules 2.3. Fuzzification - From Measurements to a Fuzzy Representation of the Input Situation 2.4. Inference - From a Fuzzy Input Representation to a Fuzzy Decision 2.5. Defuzzification - From a Fuzzy Decision to a Real Decision 3. Fuzzy Logic-related Issues in Fuzzy Control 3.1. Fuzzy Sets and Operations 3.2. Types of Rule-based Fuzzy Systems 3.3. Information Processing in Fuzzy Systems 4. Control Issues in Fuzzy Control 4.1. Structures in Fuzzy Control 4.2. Design of Fuzzy Control Systems 4.3. Analysis of Fuzzy Control Systems 4.4. Applications 4.5. Hardware and Software for Fuzzy Control Systems 4.6. Further Concepts Based on Fuzzy Sets Applied to Control Systems 5. Conclusions Glossary BibliographyUNESCO – EOLSS Biographical Sketches Summary SAMPLE CHAPTERS This chapter presents a perspective of fuzzy control systems. Fuzzy control is a form of intelligent control characterized by the use of expert knowledge on the control strategy and/or the behavior of the controlled plant.
    [Show full text]