Deep Learning at the Physical Layer: System Challenges and Applications to 5G and Beyond Francesco Restuccia, Member, IEEE, and Tommaso Melodia, Fellow, IEEE

Total Page:16

File Type:pdf, Size:1020Kb

Deep Learning at the Physical Layer: System Challenges and Applications to 5G and Beyond Francesco Restuccia, Member, IEEE, and Tommaso Melodia, Fellow, IEEE 1 Deep Learning at the Physical Layer: System Challenges and Applications to 5G and Beyond Francesco Restuccia, Member, IEEE, and Tommaso Melodia, Fellow, IEEE Abstract—The unprecedented requirements of the Internet of Things (IoT) have made fine-grained optimization of spectrum resources an urgent necessity. Thus, designing techniques able to extract knowledge from the spectrum in real time and select the optimal spectrum access strategy accordingly has become more important than ever. Moreover, 5G and beyond (5GB) networks will require complex management schemes to deal with problems such as adaptive beam management and rate selection. Although deep learning (DL) has been successful in modeling complex phenomena, commercially-available wireless devices are still very far from actually adopting learning-based techniques to optimize their spectrum usage. In this paper, we first discuss the need for real-time DL at the physical layer, and then summarize the current state of the art and existing limitations. We conclude the paper by discussing an agenda of research challenges and how DL can be applied to address crucial problems in 5GB networks. I. INTRODUCTION The wireless spectrum is undeniably one of nature’s most complex phenomena. This is especially true in the highly- dynamic context of the Internet of Things (IoT), where the Fig. 1: Key issues in today’s wireless optimization approaches. widespread presence of tiny embedded wireless devices seam- lessly connected to people and objects will make spectrum- wireless systems that can learn by themselves the optimal related quantities such as fading, noise, interference, and traffic spectrum actions to take given the current spectrum state. patterns hardly predictable with traditional mathematical mod- Concretely speaking, the big picture is to realize systems able els. Techniques able to perform real-time fine-grained spec- to distinguish on their own different spectrum states (e.g., trum optimization will thus become fundamental to squeeze based on noise, interference, channel occupation, and similar), out any spectrum resource available to wireless devices. and change their hardware and software fabric in real time There are a number of key issues – summarized in Fig. 1 to implement the optimal spectrum action [1, 2]. However, – that make existing wireless optimization approaches not despite the numerous recent advances, so far truly self-adaptive completely suitable to address the spectrum challenges men- and self-resilient cognitive radios have been elusive. On the tioned above. On one hand, model-driven approaches aim at other hand, the success of deep learning at the physical (i) mathematically formalize the entirety of the network, and layer (PHY-DL) in addressing problems such as modulation (ii) optimize an objective function. Although yielding optimal recognition [3], radio fingerprinting [4] and Medium Access solutions, these approaches are usually NP-Hard, and thus, Control (MAC) [5] has taken us many steps in the right unable to be run in real time. Moreover, they rely on a series direction [6]. Thanks to its unique advantages, deep learning (DL) can really be a game-changer, especially when cast in arXiv:2004.10113v2 [cs.NI] 7 Sep 2020 of modeling assumptions (e.g., fading/noise distribution, traffic and mobility patterns, and so on) that may not always be the context of a real-time hardware-based implementation. valid. On the other hand, protocol-driven approaches con- Existing work has mostly focused on generating spectrum sider a specific wireless technology (e.g., WiFi, Bluetooth or data and training models in the cloud. However, a number of Zigbee) and attempt to heuristically change parameters such key system-level issues still remain substantially unexplored. as modulation scheme, coding level, packet size, etc., based To this end, we notice that the most relevant survey work [7, 8] on metrics computed in real time from pilots and/or training introduces research challenges from a learning perspective symbols. Protocol-driven approaches, being heuristic in nature, only. Moreover, [9] and similar survey work focuses on the necessarily yield sub-optimal performance. application of DL to upper layers of the network stack. Since To obtain the best of both worlds, a new approach called DL was not conceived having the constraints and requirements spectrum-driven is being explored. In short, by using real- of wireless communications in mind, it is still unclear what are time machine learning (ML) techniques implemented in the the fundamental limitations of PHY-DL. Moreover, existing hardware portion of the wireless platform, we can design work has still not explored how PHY-DL can be used to address problems in the context of 5G and beyond (5GB) F. Restuccia and T. Melodia are with the Institute for the Wireless Internet of networks. For this reason, the first key contribution of this Things, Northeastern University, Boston, MA USA. Authors emails: ffrestuc, [email protected]. This paper has been accepted for publication in paper is to discuss the research challenges of real-time PHY- IEEE Communications Magazine. DL without considering any particular frequency band or radio 2 technology (Sections III and IV). The second key contribution learn I/Q patterns in the I/Q plane, making them amenable is the introduction of a series of practical problems that may to address different classification problems. Existing work, be addressed by using PHY-DL techniques in the context of indeed, has demonstrated that CNNs can be used for very 5GB (Section IV-C). Notice that 5GB networks are expected different problems, ranging from modulation recognition [3] to to be heavily based on millimeter-wave (mmWave) and ultra- radio fingerprinting [4]. CNNs also keeps latency and energy wideband communications, hence our focus on these issues. consumption constant, as explained in Section II. Fig. 2 shows Since an exhaustive compendium of the existing work in PHY- an example where a learning system is trained to classify DL is outside the scope of this manuscript, we refer the reader modulation, number of carriers and fingerprinting. While DL to [9] for an excellent survey. can concurrently recognize the three parameters, traditional learning requires different feature extraction processes for each II. WHY DEEP LEARNING AT THE PHYSICAL LAYER? of the classification outputs. This, in turn, increases hardware DL is exceptionally suited to address problems where consumption and hinders fine-tuning of the learning model. closed-form mathematical expressions are difficult to obtain Real-Time Fine Tuning. Model-driven optimization of- [10]. For this reason, convolutional neural networks (CNNs) fers predictable performance only when the model actually are now being “borrowed” by wireless researchers to ad- matches the reality of the underlying phenomenon being dress handover and power management in cellular networks, captured. This implies that model-driven systems can yield dynamic spectrum access, resource allocation/slicing/caching, sub-optimal performance when the model assumptions are video streaming, and rate adaptation, just to name a few. Fig. 2 different from what the network is actually experiencing. summarizes why traditional ML may not effectively address For example, a model assuming a Rayleigh fading channel real-time physical-layer problems. Overall, DL relieves from can yield incorrect solutions when placed in a Rician fading the burden of finding the right “features” characterizing a given environment. By using a data-driven approach, PHY-DL may wireless phenomenon. At the physical layer, this key advantage be easily fine-tuned through the usage of fresh spectrum data, comes almost as a necessity for at least three reasons, which which can be used to find in real time a better set of parameters are discussed below. through gradient descent. Hand-tailored model-driven systems Highly-Dimensional Feature Spaces. Classifying waveforms may result to be hard to fine-tune, as they might depend on ultimately boils down to distinguishing small-scale patterns a set of parameters that are not easily adaptable in real time in the in-phase-quadrature (I/Q) plane, which may not be (e.g., channel model). While DL “easily” accomplishes this clearly separable in a low-dimensional feature space. For goal by performing batch gradient descent on fresh input data, example, in radio fingerprinting we want to distinguish among the same is not true for traditional ML, where tuning can be hundreds (potentially thousands) of devices based on the extremely challenging since it would require to completely unique imperfections imposed by the hardware circuitry. While change the circuit itself. legacy low-dimensional techniques can correctly distinguish up to a few dozens of devices [11], DL-based classifiers III. DEEP LEARNING AT THE PHYSICAL LAYER: can scale up to hundreds of devices by learning extremely SYSTEM REQUIREMENTS AND CHALLENGES complex features in the I/Q space [4]. Similarly, O’Shea et al. [3] have demonstrated that on the 24-modulation dataset The target of this section is to discuss existing system- considered, DL models achieve on the average about 20% level challenges in PHY-DL, as well as the state of the art higher classification accuracy than legacy learning models in addressing these issues. For a more detailed compendium under noisy channel conditions. of the state of the art, the interested reader can take a look at the following comprehensive surveys [6–9]. Feature-based Approaches Convolutional Neural Networks To ease the reader into the topic, we summarize at a very high level the main components and operations of a Feature Neural “QPSK” HW Extraction Network (Modulation) Cores learning-based wireless device in Fig. 3. The core feature that distinguishes learning-based devices is that digital signal Feature Neural Convolutional Extraction “#8” Network Neural Weights processing (DSP) decisions are driven by deep neural networks (Fingerprint) Network (DNNs). In particular, in the Receiver (RX) DSP chain the Feature “QPSK” Neural Extraction “64” “64” CPU incoming waveform is first received and placed in an I/Q buffer Network (Carriers) “#8” (step 1).
Recommended publications
  • Logical Link Control and Channel Scheduling for Multichannel Underwater Sensor Networks
    ICST Transactions on Mobile Communications and Applications Research Article Logical Link Control and Channel Scheduling for Multichannel Underwater Sensor Networks Jun Li ∗, Mylene` Toulgoat, Yifeng Zhou, and Louise Lamont Communications Research Centre Canada, 3701 Carling Avenue, Ottawa, ON. K2H 8S2 Canada Abstract With recent developments in terrestrial wireless networks and advances in acoustic communications, multichannel technologies have been proposed to be used in underwater networks to increase data transmission rate over bandwidth-limited underwater channels. Due to high bit error rates in underwater networks, an efficient error control technique is critical in the logical link control (LLC) sublayer to establish reliable data communications over intrinsically unreliable underwater channels. In this paper, we propose a novel protocol stack architecture featuring cross-layer design of LLC sublayer and more efficient packet- to-channel scheduling for multichannel underwater sensor networks. In the proposed stack architecture, a selective-repeat automatic repeat request (SR-ARQ) based error control protocol is combined with a dynamic channel scheduling policy at the LLC sublayer. The dynamic channel scheduling policy uses the channel state information provided via cross-layer design. It is demonstrated that the proposed protocol stack architecture leads to more efficient transmission of multiple packets over parallel channels. Simulation studies are conducted to evaluate the packet delay performance of the proposed cross-layer protocol stack architecture with two different scheduling policies: the proposed dynamic channel scheduling and a static channel scheduling. Simulation results show that the dynamic channel scheduling used in the cross-layer protocol stack outperforms the static channel scheduling. It is observed that, when the dynamic channel scheduling is used, the number of parallel channels has only an insignificant impact on the average packet delay.
    [Show full text]
  • Physical Layer Overview
    ELEC3030 (EL336) Computer Networks S Chen Physical Layer Overview • Physical layer forms the basis of all networks, and we will first revisit some of fundamental limits imposed on communication media by nature Recall a medium or physical channel has finite Spectrum bandwidth and is noisy, and this imposes a limit Channel bandwidth: on information rate over the channel → This H Hz is a fundamental consideration when designing f network speed or data rate 0 H Type of medium determines network technology → compare wireless network with optic network • Transmission media can be guided or unguided, and we will have a brief review of a variety of transmission media • Communication networks can be classified as switched and broadcast networks, and we will discuss a few examples • The term “physical layer protocol” as such is not used, but we will attempt to draw some common design considerations and exams a few “physical layer standards” 13 ELEC3030 (EL336) Computer Networks S Chen Rate Limit • A medium or channel is defined by its bandwidth H (Hz) and noise level which is specified by the signal-to-noise ratio S/N (dB) • Capability of a medium is determined by a physical quantity called channel capacity, defined as C = H log2(1 + S/N) bps • Network speed is usually given as data or information rate in bps, and every one wants a higher speed network: for example, with a 10 Mbps network, you may ask yourself why not 10 Gbps? • Given data rate fd (bps), the actual transmission or baud rate fb (Hz) over the medium is often different to fd • This is for
    [Show full text]
  • Telematics Chapter 3: Physical Layer
    Telematics User Server watching with video Chapter 3: Physical Layer video clip clips Application Layer Application Layer Presentation Layer Presentation Layer Session Layer Session Layer Transport Layer Transport Layer Network Layer Network Layer Network Layer Data Link Layer Data Link Layer Data Link Layer Physical Layer Physical Layer Physical Layer Univ.-Prof. Dr.-Ing. Jochen H. Schiller Computer Systems and Telematics (CST) Institute of Computer Science Freie Universität Berlin http://cst.mi.fu-berlin.de Contents ● Design Issues ● Theoretical Basis for Data Communication ● Analog Data and Digital Signals ● Data Encoding ● Transmission Media ● Guided Transmission Media ● Wireless Transmission (see Mobile Communications) ● The Last Mile Problem ● Multiplexing ● Integrated Services Digital Network (ISDN) ● Digital Subscriber Line (DSL) ● Mobile Telephone System Univ.-Prof. Dr.-Ing. Jochen H. Schiller ▪ cst.mi.fu-berlin.de ▪ Telematics ▪ Chapter 3: Physical Layer 3.2 Design Issues Univ.-Prof. Dr.-Ing. Jochen H. Schiller ▪ cst.mi.fu-berlin.de ▪ Telematics ▪ Chapter 3: Physical Layer 3.3 Design Issues ● Connection parameters ● mechanical OSI Reference Model ● electric and electronic Application Layer ● functional and procedural Presentation Layer ● More detailed ● Physical transmission medium (copper cable, Session Layer optical fiber, radio, ...) ● Pin usage in network connectors Transport Layer ● Representation of raw bits (code, voltage,…) Network Layer ● Data rate ● Control of bit flow: Data Link Layer ● serial or parallel transmission of bits Physical Layer ● synchronous or asynchronous transmission ● simplex, half-duplex, or full-duplex transmission mode Univ.-Prof. Dr.-Ing. Jochen H. Schiller ▪ cst.mi.fu-berlin.de ▪ Telematics ▪ Chapter 3: Physical Layer 3.4 Design Issues Transmitter Receiver Source Transmission System Destination NIC NIC Input Abcdef djasdja dak jd ashda kshd akjsd asdkjhasjd as kdjh askjda Univ.-Prof.
    [Show full text]
  • Face Recognition: a Convolutional Neural-Network Approach
    98 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 8, NO. 1, JANUARY 1997 Face Recognition: A Convolutional Neural-Network Approach Steve Lawrence, Member, IEEE, C. Lee Giles, Senior Member, IEEE, Ah Chung Tsoi, Senior Member, IEEE, and Andrew D. Back, Member, IEEE Abstract— Faces represent complex multidimensional mean- include fingerprints [4], speech [7], signature dynamics [36], ingful visual stimuli and developing a computational model for and face recognition [8]. Sales of identity verification products face recognition is difficult. We present a hybrid neural-network exceed $100 million [29]. Face recognition has the benefit of solution which compares favorably with other methods. The system combines local image sampling, a self-organizing map being a passive, nonintrusive system for verifying personal (SOM) neural network, and a convolutional neural network. identity. The techniques used in the best face recognition The SOM provides a quantization of the image samples into a systems may depend on the application of the system. We topological space where inputs that are nearby in the original can identify at least two broad categories of face recognition space are also nearby in the output space, thereby providing systems. dimensionality reduction and invariance to minor changes in the image sample, and the convolutional neural network provides for 1) We want to find a person within a large database of partial invariance to translation, rotation, scale, and deformation. faces (e.g., in a police database). These systems typically The convolutional network extracts successively larger features return a list of the most likely people in the database in a hierarchical set of layers. We present results using the [34].
    [Show full text]
  • CNN Architectures
    Lecture 9: CNN Architectures Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 1 May 2, 2017 Administrative A2 due Thu May 4 Midterm: In-class Tue May 9. Covers material through Thu May 4 lecture. Poster session: Tue June 6, 12-3pm Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 2 May 2, 2017 Last time: Deep learning frameworks Paddle (Baidu) Caffe Caffe2 (UC Berkeley) (Facebook) CNTK (Microsoft) Torch PyTorch (NYU / Facebook) (Facebook) MXNet (Amazon) Developed by U Washington, CMU, MIT, Hong Kong U, etc but main framework of Theano TensorFlow choice at AWS (U Montreal) (Google) And others... Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 3 May 2, 2017 Last time: Deep learning frameworks (1) Easily build big computational graphs (2) Easily compute gradients in computational graphs (3) Run it all efficiently on GPU (wrap cuDNN, cuBLAS, etc) Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 4 May 2, 2017 Last time: Deep learning frameworks Modularized layers that define forward and backward pass Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 5 May 2, 2017 Last time: Deep learning frameworks Define model architecture as a sequence of layers Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 6 May 2, 2017 Today: CNN Architectures Case Studies - AlexNet - VGG - GoogLeNet - ResNet Also.... - NiN (Network in Network) - DenseNet - Wide ResNet - FractalNet - ResNeXT - SqueezeNet - Stochastic Depth Fei-Fei Li & Justin Johnson & Serena Yeung Lecture 9 - 7 May 2, 2017 Review: LeNet-5 [LeCun et al., 1998] Conv filters were 5x5, applied at stride 1 Subsampling (Pooling) layers were 2x2 applied at stride 2 i.e.
    [Show full text]
  • Medium Access Control Layer
    Telematics Chapter 5: Medium Access Control Sublayer User Server watching with video Beispielbildvideo clip clips Application Layer Application Layer Presentation Layer Presentation Layer Session Layer Session Layer Transport Layer Transport Layer Network Layer Network Layer Network Layer Univ.-Prof. Dr.-Ing. Jochen H. Schiller Data Link Layer Data Link Layer Data Link Layer Computer Systems and Telematics (CST) Physical Layer Physical Layer Physical Layer Institute of Computer Science Freie Universität Berlin http://cst.mi.fu-berlin.de Contents ● Design Issues ● Metropolitan Area Networks ● Network Topologies (MAN) ● The Channel Allocation Problem ● Wide Area Networks (WAN) ● Multiple Access Protocols ● Frame Relay (historical) ● Ethernet ● ATM ● IEEE 802.2 – Logical Link Control ● SDH ● Token Bus (historical) ● Network Infrastructure ● Token Ring (historical) ● Virtual LANs ● Fiber Distributed Data Interface ● Structured Cabling Univ.-Prof. Dr.-Ing. Jochen H. Schiller ▪ cst.mi.fu-berlin.de ▪ Telematics ▪ Chapter 5: Medium Access Control Sublayer 5.2 Design Issues Univ.-Prof. Dr.-Ing. Jochen H. Schiller ▪ cst.mi.fu-berlin.de ▪ Telematics ▪ Chapter 5: Medium Access Control Sublayer 5.3 Design Issues ● Two kinds of connections in networks ● Point-to-point connections OSI Reference Model ● Broadcast (Multi-access channel, Application Layer Random access channel) Presentation Layer ● In a network with broadcast Session Layer connections ● Who gets the channel? Transport Layer Network Layer ● Protocols used to determine who gets next access to the channel Data Link Layer ● Medium Access Control (MAC) sublayer Physical Layer Univ.-Prof. Dr.-Ing. Jochen H. Schiller ▪ cst.mi.fu-berlin.de ▪ Telematics ▪ Chapter 5: Medium Access Control Sublayer 5.4 Network Types for the Local Range ● LLC layer: uniform interface and same frame format to upper layers ● MAC layer: defines medium access ..
    [Show full text]
  • Deep Learning Architectures for Sequence Processing
    Speech and Language Processing. Daniel Jurafsky & James H. Martin. Copyright © 2021. All rights reserved. Draft of September 21, 2021. CHAPTER Deep Learning Architectures 9 for Sequence Processing Time will explain. Jane Austen, Persuasion Language is an inherently temporal phenomenon. Spoken language is a sequence of acoustic events over time, and we comprehend and produce both spoken and written language as a continuous input stream. The temporal nature of language is reflected in the metaphors we use; we talk of the flow of conversations, news feeds, and twitter streams, all of which emphasize that language is a sequence that unfolds in time. This temporal nature is reflected in some of the algorithms we use to process lan- guage. For example, the Viterbi algorithm applied to HMM part-of-speech tagging, proceeds through the input a word at a time, carrying forward information gleaned along the way. Yet other machine learning approaches, like those we’ve studied for sentiment analysis or other text classification tasks don’t have this temporal nature – they assume simultaneous access to all aspects of their input. The feedforward networks of Chapter 7 also assumed simultaneous access, al- though they also had a simple model for time. Recall that we applied feedforward networks to language modeling by having them look only at a fixed-size window of words, and then sliding this window over the input, making independent predictions along the way. Fig. 9.1, reproduced from Chapter 7, shows a neural language model with window size 3 predicting what word follows the input for all the. Subsequent words are predicted by sliding the window forward a word at a time.
    [Show full text]
  • Physical Layer Compliance Testing for 1000BASE-T Ethernet
    Physical Layer Compliance Testing for 1000BASE-T Ethernet –– APPLICATION NOTE Physical Layer Compliance Testing for 1000BASE-T Ethernet APPLICATION NOTE Engineers designing or validating the 1000BASE-T Ethernet 1000BASE-T Physical Layer physical layer on their products need to perform a wide range Compliance Standards of tests, quickly, reliably and efficiently. This application note describes the tests that ensure validation, the challenges To ensure reliable information transmission over a network, faced while testing multi-level signals, and how oscilloscope- industry standards specify requirements for the network’s resident test software enables significant efficiency physical layer. The IEEE 802.3 standard defines an array of improvements with its wide range of tests, including return compliance tests for 1000BASE-T physical layer. These tests loss, fast validation cycles, and high reliability. are performed by placing the device under test in test modes specified in the standard. The Basics of 1000BASE-T Testing While it is recommended to perform as many tests as Popularly known as Gigabit Ethernet, 1000BASE-T has been possible, the following core tests are critical for compliance: experiencing rapid growth. With only minimal changes to IEEE 802.3 Test Mode Test the legacy cable structure, it offers 100 times faster data Reference rates than 10BASE-T Ethernet signals. Gigabit Ethernet, in Peak 40.6.1.2.1 combination with Fast Ethernet and switched Ethernet, offers Test Mode-1 Droof 40.6.1.2.2 Template 40.6.1.2.3 a cost-effective alternative to slow networks. Test Mode-2 Master Jitter 40.6.1.2.5 Test Mode-3 Slave Jitter 1000BASE-T uses four signal pairs for full-duplex Distortion 40.6.1.2.4 transmission and reception over CAT-5 balanced cabling.
    [Show full text]
  • Structure of IEEE 802.11 Packets at Various Physical Layers
    Appendix A: Structure of IEEE 802.11 Packets at Various Physical Layers This appendix gives a detailed description of the structure of packets in IEEE 802.11 for the different physical layers. 1. Packet Format of Frequency Hopping Spread-spectrum Physical Layer (FHSS PHY) The packet is made up of the following elements (Figure A.1): 1.1 Preamble It depends on the physical layer and includes: – Synch: a sequence of 80 bits alterning 0 and 1, used by the physical circuits to select the correct antenna (if more than one are in use), and correct offsets of frequency and synchronization. – SFD: start frame delimiter consists of a pattern of 16 bits: 0000 1100 1011 1101, used to define the beginning of the frame. 1.2 Physical Layer Convergence Protocol Header The Physical Layer Convergence Protocol (PLCP) header is always transmitted at 1 Mbps and carries some logical information used by the physical layer to decode the frame: – Length of word of PLCP PDU (PLW): representing the number of bytes in the packet, useful to the physical layer to detect correctly the end of the packet. – Flag of signalization PLCP (PSF): indicating the supported rate going from 1 to 4.5 Mbps with steps of 0.5 Mbps. Even though the standard gives the combinations of bits for PSF (see Table A.1) to support eight different rates, only the modulations for 1 and 2 Mbps have been defined. – Control error field (HEC): CRC field for error detection of 16 bits (or 32 bits). The polynomial generator used is G(x)=x16 + x12 + x5 +1.
    [Show full text]
  • High-Level Data Link Control
    ELEC3030 (EL336) Computer Networks S Chen High-Level Data Link Control • This class of data link layer protocols includes High-level Data Link Control (HDLC), Link Access Procedure Balanced (LAPB) for X.25, Link Access Procedure for D-channel (LAPD) for ISDN, and Logic Link Control (LLC) for FDDI • The frame format is: Flag Address Control Data FCS Flag Note that address and control bits 8 8 8 variable 16 8 can be extended to 16 bits, so bit position 12 3 4 5 6 7 8 N(S)=send sequence number N(R)=receive sequence number that sequence number is 7-bit Information: 0 N(S) P/F N(R) S=supervisory function bits P/F • Frame flag: 01111110, so bit Supervisory: 1 0 S N(R) M=unumbered function bits stuffing is used Unumbered: 1 1 M P/F M P/F=poll/final bit • Address: For multipoint operation, it is used to identify the terminal that transmits or receives the frame and, in point-to-point link, it is used to distinguish Commands from Responses (2nd bit for C/R: 0/1). The address is now extended to 16 bits (1st bit indicates long/short 16/8 bits) • Checksum: FCS contains the remainder of a 16-bit CRC calculation of the frame. It may be extended to 32 bits, using a 32-bit CRC • Control: Three types of frames, I, S and U frames. The old protocol uses a sliding window with 3-bit sequence number and the maximum window size is N = 7. The control field is now extended to 16 bits with 7-bit sequence number 60 ELEC3030 (EL336) Computer Networks S Chen HDLC (continue) • I-frames: carry user data.
    [Show full text]
  • Introduction-To-Deep-Learning.Pdf
    Introduction to Deep Learning Demystifying Neural Networks Agenda Introduction to deep learning: • What is deep learning? • Speaking deep learning: network types, development frameworks and network models • Deep learning development flow • Application spaces Deep learning introduction ARTIFICIAL INTELLIGENCE Broad area which enables computers to mimic human behavior MACHINE LEARNING Usage of statistical tools enables machines to learn from experience (data) – need to be told DEEP LEARNING Learn from its own method of computing - its own brain Why is deep learning useful? Good at classification, clustering and predictive analysis What is deep learning? Deep learning is way of classifying, clustering, and predicting things by using a neural network that has been trained on vast amounts of data. Picture of deep learning demo done by TI’s vehicles road signs person background automotive driver assistance systems (ADAS) team. What is deep learning? Deep learning is way of classifying, clustering, and predicting things by using a neural network that has been trained on vast amounts of data. Machine Music Time of Flight Data Pictures/Video …any type of data Speech you want to classify, Radar cluster or predict What is deep learning? • Deep learning has its roots in neural networks. • Neural networks are sets of algorithms, modeled loosely after the human brain, that are designed to recognize patterns. Biologocal Artificial Biological neuron neuron neuron dendrites inputs synapses weight axon output axon summation and cell body threshold dendrites synapses cell body Node = neuron Inputs Weights Summation W1 x1 & threshold W x 2 2 Output Inputs W3 Output x3 y layer = stack of ∑ ƒ(x) neurons Wn xn ∑=sum(w*x) Artificial neuron 6 What is deep learning? Deep learning creates many layers of neurons, attempting to learn structured representation of big data, layer by layer.
    [Show full text]
  • Face Recognition Using Popular Deep Net Architectures: a Brief Comparative Study
    future internet Article Face Recognition Using Popular Deep Net Architectures: A Brief Comparative Study Tony Gwyn 1,* , Kaushik Roy 1 and Mustafa Atay 2 1 Department of Computer Science, North Carolina A&T State University, Greensboro, NC 27411, USA; [email protected] 2 Department of Computer Science, Winston-Salem State University, Winston-Salem, NC 27110, USA; [email protected] * Correspondence: [email protected] Abstract: In the realm of computer security, the username/password standard is becoming increas- ingly antiquated. Usage of the same username and password across various accounts can leave a user open to potential vulnerabilities. Authentication methods of the future need to maintain the ability to provide secure access without a reduction in speed. Facial recognition technologies are quickly becoming integral parts of user security, allowing for a secondary level of user authentication. Augmenting traditional username and password security with facial biometrics has already seen impressive results; however, studying these techniques is necessary to determine how effective these methods are within various parameters. A Convolutional Neural Network (CNN) is a powerful classification approach which is often used for image identification and verification. Quite recently, CNNs have shown great promise in the area of facial image recognition. The comparative study proposed in this paper offers an in-depth analysis of several state-of-the-art deep learning based- facial recognition technologies, to determine via accuracy and other metrics which of those are most effective. In our study, VGG-16 and VGG-19 showed the highest levels of image recognition accuracy, Citation: Gwyn, T.; Roy, K.; Atay, M. as well as F1-Score.
    [Show full text]