electronics
Review Optical Camera Communications: Principles, Modulations, Potential and Challenges
Willy Anugrah Cahyadi 1 , Yeon Ho Chung 2,* , Zabih Ghassemlooy 3 and Navid Bani Hassan 3
1 School of Electrical Engineering (SEE) and The Center for Advanced Intelligent Communications (AICOMS), Telkom University, Bandung 40257, Indonesia; [email protected] 2 Department of Information and Communications Engineering, Pukyong National University, Busan 48513, Korea 3 Mathematics, Physics and Electrical Engineering Department, Northumbria University, Newcastle upon Tyne NE1 8ST, UK; [email protected] (Z.G.); [email protected] (N.B.H.) * Correspondence: [email protected]
Received: 20 July 2020; Accepted: 17 August 2020; Published: 19 August 2020
Abstract: Optical wireless communications (OWC) are emerging as cost-effective and practical solutions to the congested radio frequency-based wireless technologies. As part of OWC, optical camera communications (OCC) have become very attractive, considering recent developments in cameras and the use of fitted cameras in smart devices. OCC together with visible light communications (VLC) is considered within the framework of the IEEE 802.15.7m standardization. OCCs based on both organic and inorganic light sources as well as cameras are being considered for low-rate transmissions and localization in indoor as well as outdoor short-range applications and within the framework of the IEEE 802.15.7m standardization together with VLC. This paper introduces the underlying principles of OCC and gives a comprehensive overview of this emerging technology with recent standardization activities in OCC. It also outlines the key technical issues such as mobility, coverage, interference, performance enhancement, etc. Future research directions and open issues are also presented.
Keywords: optical camera communications; visible light communications; camera receiver; wireless communications; LED-camera
1. Introduction The rapid increase in the data being generated and the growing demand for capacity have pushed existing congested radio frequency (RF)-based wireless communications to their limits. In addition, we are witnessing incommensurate spectral efficiency as well as bandwidth with increasing traffic demands and requirements in the RF domain [1]. From a futuristic perspective, it is foreseen that high-rate and high-capacity RF communications coexist with more affordable, offloading, and unlimited bandwidth wireless communications. As an attractive complementary solution, optical wireless communications (OWC) have been considered for easing the bandwidth usage pressure on RF wireless systems in a harmonious manner. OWC covers the entire optical spectrum of ultraviolet (UV), visible and infrared (IR) [1,2], see Figure1, thus o ffering a massive theoretical bandwidth of 20 THz (IR), 320 THz (visible) and 30 PHz (UV) compared with the bandwidth of 300 GHz in the RF technology [1,3,4]. Within the OWC umbrella, there are a number of transmission schemes for both indoor and outdoor environments including infrared communications (IRC), free space optics (FSO), visible light communications (VLC) and ultraviolet communications (UVC) [1,3–6], as shown in
Electronics 2020, 9, 1339; doi:10.3390/electronics9091339 www.mdpi.com/journal/electronics Electronics 2020, 9, 1339 2 of 44 Electronics 2020, 9, x FOR PEER REVIEW 2 of 45
Figure2. The VLC and IRC are generally considered for short-range indoor applications, while FSO rangeElectronics outdoor 2020, 9 ,environments. x FOR PEER REVIEW Note that FSO can also use the visible band for short to medium 2range of 45 transmissions.(using the IR band) and UVC (using the C band) are predominantly employed for medium-long rangerange outdooroutdoor environments. environments. Note Note that that FSO FSO can canalso also use the use visible the visible band bandfor short for to short medium to medium range rangetransmissions. transmissions.
Figure 1. The electromagnetic spectrum. Figure 1. The electromagnetic spectrum. Figure 1. The electromagnetic spectrum.
Figure 2. Overview of optical wireless communications (OWC) transmission schemes. Figure 2. Overview of optical wireless communications (OWC) transmission schemes. The VLCFigure technology, 2. Overview a subsetof optical of wireless OWC ascommunications defined in the (OWC) IEEE transmission 802.15.7, has schemes. specifically been envisionedThe VLC for technology, high capacity a subset mobile of data OWC networks as defined since in 2000, the toIEEE complement 802.15.7, has the RF-basedspecifically mobile been envisionedcommunicationsThe VLC for hightechnology, [1, 7capacity–10]. The a mobilesubset popularity ofdata OWC networks of VLC as defined development since in2000, the toIEEE is drivencomplement 802.15.7, by its has etheffi ciency specificallyRF-based in utilizing mobile been communicationsexistingenvisioned light-emitting for high [1,7–10]. capacity diodes The mobile popularity (LEDs)-based data ofnetworks VLC lighting development since infrastructure 2000, tois drivencomplement as theby its transmitter theefficiency RF-based (Tx) in utilizing mobile to offer existingsignificantlycommunications light-emitting higher [1,7–10]. luminous diodes The e popularity(LEDfficacy,s)-based data of communications VLC lighting development infrastructure and is indoordriven as the localizationby itstransmitter efficiency [1,9 (Tx)]. in In utilizing contrastto offer significantlytoexisting other lighting light-emitting higher sources, luminous diodes LEDs efficacy, can(LED bes)-based switched data communications lighting at a very infrastructure high speed,and indoor thusas the localization facilitating transmitter data [1,9]. (Tx) transmission In tocontrast offer tothatsignificantly other can belighting picked higher sources, up luminous using LEDs photodiodes efficacy, can databe (PDs) switchedcommunications or image at a sensors very and indoorhigh (ISs). speed, localization thus [1,9].facilitating In contrast data transmissionto otherA VLC lighting system that can sources, employing be picked LEDs an up IS, canusing i.e., be aphotodiodes camera,switched is knownat (PDs) a very asor opticalimagehigh speed,sensors camera thus (ISs). communication facilitating (OCC)data accordingtransmissionA VLC to systemthe that standards, can employing be picked which an up IS, requiresusing i.e., a photodiodes camera, no hardware is known (PDs) modifications as or optical image camerasensors [5,11, 12communication(ISs).]. The IS in a camera, (OCC) accordingwhichA is VLC a to massive systemthe standards, matrixemploying ofwhichµ an-scale IS, requires i.e., PDs, a camera, togetherno hard isware withknown modifications a lens,as optical can becamera [5,11,12]. used communication to provideThe IS in an a imaging camera,(OCC) according to the standards, which requires no hardware modifications [5,11,12]. The IS in a camera, whichmassive is a multiple-input massive matrix multiple-output of μ-scale PDs,(MIMO) together system,with a lens, i.e., enablingcan be used spatial to provide separation an imaging of light which is a massive matrix of μ-scale PDs, together with a lens, can be used to provide an imaging massivesources onmultiple-input the IS. In addition, multiple-output the camera (IS)-based(MIMO) sy receiverstem, i.e., (RX) enabling is capable spatial of capturing separation information of light sourcesmassive on multiple-input the IS. In addition, multiple-output the camera (IS)-based(MIMO) sy receiverstem, i.e., (RX) enabling is capable spatial of capturing separation information of light notsources only onfrom the multiple IS. In addition, TXs (i.e., the LEDs) camera but (IS)-based also from receiver digital (RX) displays is capable such of as capturing liquid crystal information display (LCD)not only and from organic multiple LED (OLED)-basedTXs (i.e., LEDs) screens but also [5,12]. from digital displays such as liquid crystal display (LCD) and organic LED (OLED)-based screens [5,12].
Electronics 2020, 9, 1339 3 of 44 not only from multiple TXs (i.e., LEDs) but also from digital displays such as liquid crystal display (LCD) and organic LED (OLED)-based screens [5,12]. Furthermore, an OCC image capturing increases the communication range in outdoor environments up to a kilometer, due to the use of image processing, long sampling duration and lenses [5]. Nevertheless, most OCC schemes reported in the literature are intended for indoor applications because of smart-devices availability and a lower intensity of ambient light in indoor environment [5]. With regard to the OCC, it is important to note that OCC-based systems are mainly targeted for a low-rate transmission, due to the reception-sampling rate, which is determined by the camera frame rate FR, which is typically in the range of 30 to 60 frames per second (fps) at a common resolution of 1920 1080 pixels. However, with current advances made in cameras, for the same resolution, F of × R new smartphone cameras can reach up to 960 fps [13]. Different from VLC that uses a PD to process the received optical intensity, OCC is based on capturing the source using the IS, i.e., a two-dimension (2D) RX containing millions of pixels, and carrying out a significant amount of image processing to acquire the information. Therefore, the data transmission rate of OCC is mostly limited to either kilobits per second (Kbps) or megabits per second (Mbps) in contrast to multi-gigabit per second (Gbps) transmission rates in PD-based VLC systems [1,5,11,14]. The IEEE 802.15.7 standard was initially formulated in 2011 for VLC without considering the IS-based RX [1,11]. In 2014, however, OCC was included in the revised IEEE 802.15.7r1 standard [11,15], which was managed by the TG7r1 task group and covered wider light spectra such as near infrared (NIR) and near UV [5,11]. OCC is currently organized by the updated task group TG7m that finalized the initial draft in 2018 [14,16]. The task group is a joint consortium of academic researchers from multiple countries including United States, South Korea, Turkey, Taiwan and China [16]. Due to the inclusion of nearly the entire optical spectrum in the standardization, there are a wide range of potential applications for OCC within the context of Internet-of-things (IoT) such as augmented reality, color code, device-to-device communications, digital signage, exhibition/store service, in-flight service, indoor positioning, vehicular communications and underwater/seaside communications [5,17,18]. At a global level, a number of companies including Fraunhofer Heinrich Hertz Institute (Germany), Intel (United States), Panasonic (Japan), Casio (Japan), Huawei Technologies (China), Electronics and Telecommunications Research Institute (Korea), LG Electronics (Korea) and China Telecom (China) have been working on the application of OCC [5]. In recent years, the interest in OCC has been growing, judging from four surveys published in the past four years, mostly driven by the early standardization activity (i.e., IEEE 802.15.7r1) [5,11,17,18]. The first survey focused on key technologies used in OCC [11], while the second survey investigated various implementation challenges associated with OCC with no comprehensive structures [17]. The third survey covered concise performance analysis of OCC intended for inclusion in the fifth generation (5G) wireless networks [5]. Finally, the fourth survey, which was the most recently published work, is a broad survey of OCC covering localization, navigation, motion capture, and some insights into future research direction [18,19]. This survey, however, only introduced several challenges and future research directions in the limited sense. This paper, on the other hand, presents a comprehensive survey on OCC in such a way that it focuses heavily on the principles and modulations of OCC to date, together with potential applications and future research directions in mind. It also outlines key technologies and standardization as well as modulation schemes that have been proposed and investigated. The paper aims to enhance practical implementations of OCC, while disseminating promising OCC technologies and future challenges. The rest of the paper, as illustrated in Figure3, is organized as follows. Section2 describes OCC in detail and Section3 reviews existing modulations of OCC. Section4 presents various performance enhancements in terms of mobility, coverage, and interference. Section5 is dedicated to recent OCC schemes for both indoor and outdoor applications. Section6 discusses future potential developments, Electronics 2020, 9, 1339 4 of 44 enhancements, and challenges in OCC. Finally, Section7 closes the paper with a brief summary. ElectronicsIn addition, 2020, Appendix 9, x FOR PEERA shows REVIEW a list of both abbreviations and acronyms used throughout the paper.4 of 45
Figure 3. The paper organization. 2. Optical Camera Communication 2. Optical Camera Communication 2.1. Principles 2.1. Principles OCC employs either a single or multiple-camera setup as the RX, which can now be found in mostOCC mobile employs devices either such a as single mobile or phones, multiple-camera smart watches, setup laptops,as the RX, tablets, which as can well now as vehiclesbe found and in mostbuildings mobile [2]. devices Therefore, such theas mobile existing phones, LED lighting smart wa fixturestches, andlaptops, widely tablets, used as cameras, well as vehicles i.e., cheaper and buildingsand off-the-shelf [2]. Therefore, devices, the can existing be employed LED lighting in OCC. fixtures While and the majoritywidely used of VLC cameras, systems i.e., are cheaper limited and to off-the-shelfutilizing the visibledevices, light can spectrum, be employed OCC employingin OCC. Wh diilefferent the typesmajority of camera of VLC can systems utilize theare entirelimited light to utilizingspectrum the from visible visible light light, spectrum, NIR to UV-C OCC band.employing In OCC, different the camera types captures of camera images can /utilizevideo streamsthe entire of lightthe intensity spectrum modulated from visible light light, sources, NIR i.e.,to UV-C a single band. LED, In multiple OCC, the LEDs camera and digitalcaptures display images/video screens, streamsand extracts of the the intensity information modulated by means light ofsources, image i. processinge., a single LED, [17]. multiple Thus, camera-based LEDs and digital OCC display can be screens,used in bothand indoorextracts and the outdoorinformation environments by means forof image a range processing of applications [17]. Thus, [11]. Thecamera-based development OCC of cancomplementary be used in metal-oxide both indoor semiconductor and outdoor (CMOS) environments sensor technologies for a range has of created applications a new generation[11]. The developmentof high-pixel-resolution of complementary and high-speed metal-oxide built-in semicond camerasuctor [11 (CMOS),17]. Although sensor technologies most cameras has capturecreated a new generation of high-pixel-resolution and high-speed built-in cameras [11,17]. Although most images at a FR of 30 to 60 fps, the OCC standardization covers high-speed cameras, which can support cameras capture images at a FR of 30 to 60 fps, the OCC standardization covers high-speed cameras, a FR of up to thousands of fps [5,11]. whichSimilar can support to VLC, a F OCCR of up predominantly to thousands reliesof fps on[5,11]. the directional line-of-sight (LOS) transmission mode.Similar In addition, to VLC, OCC OCC conveniently predominantly offers relies a MIMO on the capability directional by extracting line-of-sight the (LOS) data from transmission captured mode. In addition, OCC conveniently offers a MIMO capability by extracting the data from captured multiple light sources. Therefore, for these reasons, OCC can be considered as a subset of VLC. It is interesting to note that the imaging MIMO capability of the camera can also be exploited for other applications, such as range estimation, shape detection, scene detection, etc.
Electronics 2020, 9, 1339 5 of 44 multiple light sources. Therefore, for these reasons, OCC can be considered as a subset of VLC. It is interesting to note that the imaging MIMO capability of the camera can also be exploited for other applications,Electronics 2020, 9 such, x FOR as PEER range REVIEW estimation, shape detection, scene detection, etc. 5 of 45
2.2. Standardization OCC has been in the process of standardization since 2014 by the IEEE 802.15.7r1 task group TG7r1, which is is a a revised revised version version of of IEEE IEEE 802.15.7 802.15.7 standard standard for for VLC VLC [11,20,21]. [11,20,21 ].OCC OCC is considered is considered as asan anextension extension of the of theVLC VLC standard standard in IEEE in IEEE 802.15.7 802.15.7 that was that initially was initially defined defined as a specific as a specific task group task groupto formulate to formulate the standard the standard revision revision for VLC, for VLC,which which is called is called the IEEE the IEEE802.15.7-2011 802.15.7-2011 standard standard [1]. The [1]. Thetask taskgroup group for OCC for OCC was was renamed renamed to IEEE to IEEE 802.15.7m 802.15.7m OWC OWC Task Task Group Group TG7m TG7m in in order order to to develop technical specificationsspecifications for the OCC standardstandard [[14].14]. TG7m covers NIR (700–1000 nm wavelengths), UV (100–400 nm wavelengths) and visible lightlight (400–700(400–700 nmnm wavelengths)wavelengths) bands.bands. The TG7m task group has contributed to various OWC solutions including OCC, LED IdentificationIdentification (LED-ID) and LiFi [[14].14]. OCC OCC is is alternatively known known as IS communications (ISC) or camera-based VLC, since it utilizes an IS as the RX, RX, which which provides provides a supplem supplementaryentary functionality of communications in addition to video streaming andand imageimage capturingcapturing [[22–24].22–24]. Figure 44 illustrates illustrates didifferentfferent examplesexamples ofof OCCOCC consideredconsidered inin TG7m,TG7m, highlightinghighlighting vehicular communications (vehicle-to-vehicle (V2V) and vehicle-to-infrastructure (V2I)) and indoor applications [[14].14]. Additional standardized applications were proposed in late 2018 by TG7m, such as 2D color code, code, TX/RX TX/RX relay, relay, digital digital signage, signage, exhibiti exhibitionon/store/store service, service, in-flight in-flight service, service, underwater underwater (or (orseaside/coastal) seaside/coastal) communications, communications, drone-to-drone, drone-to-drone, smart smart living, living, street-lamp street-lamp access access point, point, smart smart city, city,smart smart factory, factory, and smart and smart house house [14,25,26]. [14,25 These,26]. Theseapplications applications are classified are classified into flicker into and flicker flicker- and flicker-freefree physical physical (PHY) (PHY)layer modes, layer modes, depending depending on the on switching the switching rate rateof the of theTX TXbeing being perceptible perceptible or orimperceptible imperceptible to tohuman human eyes, eyes, respectively. respectively.
Figure 4. Example of promising serv servicesices in IEEE 802.15.7m.
Regarding modulations,modulations, five five additional additional schemes schemes are standardized are standardized to deliver optimumto deliver performance optimum inperformance OCC. That is,in accordingOCC. That to theis, according finalized document to the finalized in 2017, document the TG7m Committeein 2017, the approved TG7m Committee additional setsapproved of OWC additional PHY operating sets of OWC modes, PHY i.e., operating PHY IV to modes PHY, VIII, i.e., PHY as shown IV to inPHY Table VIII,1. PHYas shown IV, V, in and Table VI are1. PHY directly IV, V, related and VI to are OCC directly developments, related to O whileCC developments, PHY VII and while VIII are PHY intended VII and for VIII LiFi. are intended In direct detectionfor LiFi. In VLC direct systems, detection the VLC simple systems, on-o theff keying simple(OOK) on-off keying and variable (OOK) pulseand variable position pulse modulation position (VPPM)modulation schemes (VPPM) were schemes adopted were in the adopted initial version in the in ofitial the version IEEE 802.15.7-2011 of the IEEE standard.802.15.7-2011 Table standard.2 details theTable OCC 2 details modulation the OCC schemes modulation that were schemes introduced that inwe there introduced IEEE 802.15.7m in the standard IEEE 802.15.7m for PHY IV, standard PHY V, andfor PHY PHY IV, VI. PHY The nextV, and revised PHY draftVI. The of thenext standard, revised draft known of asthe IEEE standard, 802.15.7-2018, known as was IEEE finalized 802.15.7- in 20182018, [was27,28 finalized]. More detailed in 2018 OCC[27,28]. technologies More detailed are presentedOCC technologies in Sections are3 andpresented4. in Sections 3 and 4.
Electronics 2020, 9, 1339 6 of 44
Table 1. IEEE 802.15.7m initial operating modes.
PHY Description Modulation Scheme
OOK • I Existing IEEE 802.15.7-2011 VPPM •
OOK • II Existing IEEE 802.15.7-2011 VPPM •
CSK III Existing IEEE 802.15.7-2011 •
UFSOOK • Twinkle VPPM • Image sensor communication S2-PSK IV • modes (added to TG7m) HS-PSK • Offset-VPWM •
PWM/PPM • CM-FSK Image sensor communication • V C-OOK modes (added to TG7m) • RS-FSK •
Table 2. IEEE 802.15.7m: Characteristics and performance estimation of OCC technologies.
Flicker-Free Modulation Flicker Modulation Parameter High Frame Rate RoI Signaling Rolling Shutter Screen/Projector Processing Technique Technique Distance <100 m Hundreds of meters Several meters Several meters
Tens of bps for RoI • signaling stream Tens of kbps for the Data rate Tens of Kbps • <10 Kbps Kbps to Mbps high-data rate stream
Traffic light Car lights/traffic LED panel/arrays • • • Screen TX LEDs-array TX lights/LED of LED-based TX • Typical Intended system High-speed signage TX Rolling shutter • • • cameras RX camera RX RoI-camera RX camera RX •
The integration of a The resolution of • • RoI signaling the LED panel on The LED array • stream and a the captured based on the TX MIMO benefits high-speed data image limits the • allowing higher from the spatial stream is amount of data data transmission resolution of implemented in a per image per frame both TX and RX hybrid system The trade-off Characteristics The detection of • Well suited for • The RoI signaling between distance • light source based • short-range stream allows faster, and data rate on image OWC system reliable detection makes this the processing is utilizing screen and tracking of most suitable the drawback multiple light option for sources simultaneously indoor application Electronics 2020, 9, 1339 7 of 44
2.3. Transceivers The operation and performance of IS-based OCC systems are affected by both the IS characteristics and camera’s image processing [17]. In digital cameras, the IS is typically based on the CMOS architecture, which is a mature, highly developed and less costly technology than charge-coupled device (CCD)-based cameras [29]. Note that the size of CMOS-based ISs is increasingly smaller, while still offering higher resolution images with high pixel density. Smaller sized CMOS ISs are made possible using the advanced electronics fabrication technology with a 7 nm manufacturing scale [30]. As shown in Figure5, a typical camera structure is composed of an imaging lens, a color filter array (CFA), an image sensor (IS) and an image processing unit (IPU) [29]. The lens projects the light illuminated or reflected from a light source or an object on the IS, respectively. The diagonal field of view (FOV) of the camera is determined by the focal length of the camera lens and the sensor size as Electronics 2020, 9, x FOR PEER REVIEW 7 of 45 illustrated in Figure6. In addition, the working distance defines the distance for the camera to capture Electronics 2020, 9, x FOR PEER REVIEW 7 of 45 the image in focus.
Figure 5. A brief schematic of a digital camera. Figure 5. A brief schematic of a digital camera. Figure 5. A brief schematic of a digital camera.
FigureFigure 6.6. AA diagonaldiagonal fieldfield ofof viewview (FOV)(FOV) ofof aa camera.camera. Figure 6. A diagonal field of view (FOV) of a camera. AA CFACFA typicallytypically withwith BayerBayer patternpattern recordsrecords thethe colorcolor informationinformation ofof thethe illuminatedilluminated lightlight inin aa spatial configuration [31]. Bayer pattern is structured based on 2 2 blocks of filters (i.e., one blue, spatialA CFAconfiguration typically [31].with BayerBayer patternpattern isrecords structured the color based information on 2 ×× 2 blocks of the of illuminated filters (i.e., lightone blue, in a one red, and two green), which can have different arrangements of BGGR, RGBG, GRGB or RGGB. spatialone red, configuration and two green), [31]. which Bayer canpattern have is differen structuredt arrangements based on 2 of× 2 BGGR, blocks RGBG,of filters GRGB (i.e., orone RGGB. blue, The reason for more green filters is the human eye, which is more sensitive to the green light; hence, oneThe red,reason and for two more green), green which filters can is thehave human differen eye,t arrangements which is more of sensitive BGGR, RGBG,to the green GRGB light; or RGGB. hence, more information about the green color is required [32]. The image is then processed by an image Themore reason information for more about green the filters green is thecolor human is required eye, which [32]. isThe more image sensitive is then to processed the green bylight; an hence,image processor utilizing a demosaicing algorithm, also known as deBayering, to reconstruct the image colors moreprocessor information utilizing about a demosaicing the green coloralgorithm, is required also known [32]. The as imagedeBayering, is then to processed reconstruct by thean imageimage processorcolors similar utilizing to the a demosaicingreal scene, as algorithm, illustrated also in Figureknown 7as [31,32]. deBayering, The output to reconstruct of the IPU the isimage then colorsprocessed similar to extract to the the real information scene, as illustratedin the image. in Figure 7 [31,32]. The output of the IPU is then processed to extract the information in the image.
Electronics 2020, 9, 1339 8 of 44 similar to the real scene, as illustrated in Figure7[ 31,32]. The output of the IPU is then processed to Electronicsextract the 2020 information, 9, x FOR PEER in REVIEW the image. 8 of 45
Figure 7. An example of the demosaicing process. Figure 7. An example of the demosaicing process. The internal shutter mechanism of cameras determines the exposure of pixels in the IS. Two main typesThe of ISsinternal in terms shutter of the mechanism architecture of are cameras CCD and deter CMOSmines [33 the]. Aexposure CCD sensor of pixels has a largerin the physicalIS. Two maindimension types dueof ISs to in the terms utilization of the ofarchitecture a relatively are larger CCD analog-to-digital and CMOS [33]. converter A CCD sensor (ADC), has compared a larger withphysical a CMOS dimension IS [29 ,34due]. Therefore,to the utilization CCD is of not a generallyrelatively implementedlarger analog-to- in mobiledigital devices,converter especially (ADC), comparedsmartphones. with In a comparisonCMOS IS [29,34]. with CCD,Therefore, CMOS CCD cameras is not provide generally lower implemented power consumption, in mobile devices, smaller especiallyIS, faster readout, smartphones. lower costIn andcomparison more programmability. with CCD, CMOS In addition, cameras since provide pixel scanning lower ispower based consumption,on the X–Y address smaller scheme IS, faster in thereadout, CMOS lower architecture, cost and directmore programmability. access to a desired In part addition, of the since IS is pixel scanning is based on the X–Y address scheme in the CMOS architecture, direct access to a possible, which can increase the FR at the cost of reduced FOV [35,36]. Cameras can have two different desiredexposure part methods, of the IS i.e., is possible, global shutter which (GS)can increase and rolling the shutterFR at the (RS), cost of as reduced illustrated FOV in Figure[35,36].8 a.Cameras In the canglobal-shutter-based have two different ISs, exposure the entire methods, sensor is i.e., exposed global to shutter the light (GS) at the and same rolling time, shutter whereas (RS), in theas illustratedrolling-shutter-based in Figure 8a. cameras, In the global-shutter-based each row is exposed to ISs, the the light entire at a time,sensor thereby is exposed considerably to the light reducing at the samethe power time, consumption.whereas in the CCD rolling-shutter-based sensors inherently came captureras, each images row using is exposed the GS, to while the light the at exposure a time, therebymethod inconsiderably most CMOS reducing ISs is based the onpower RS [29 consum,33]. ption. CCD sensors inherently capture images usingIt the should GS, while be noted the thatexposure in the method global-shutter-based in most CMOS CCD ISs is ISs, based considerably on RS [29,33]. longer frame capture timeIt is should needed be to noted convert that the in the charge global-shutter-ba into a voltagesed signal CCD [29 ISs,]. considerably Figure8b shows longer the frame structures capture of timeCCD is and needed CMOS to convert sensors the [29 charge]. Due into to the a voltage fabrication signal of [29]. much Figure more 8b compact shows the ADC structures for each of pixel,CCD andthe CMOSCMOS sensorsensors is [29]. preferable Due to nowadays the fabrication compared of much with more relatively compact larger ADC CCD for sensors.each pixel, CMOS the CMOS is also sensor is preferable nowadays compared with relatively larger CCD sensors. CMOS is also employed employed for high-speed cameras with FR > 1000 fps, which cannot be accomplished using the CCD forsensors high-speed due to slowercameras ADCs. with F NoteR > 1000 that fps, the CMOSwhich cannot sensor inbe smartphoneaccomplished cameras using isthe capable CCD sensors of reaching due to slower ADCs. Note that the CMOS sensor in smartphone cameras is capable of reaching an FR of an FR of 1000 fps [13,37]. The shutter methods are implemented within the sensor itself; however, 1000the information fps [13,37]. can The be shutter treated methods differently are as implem part ofented image within processing the sensor following itself; image however, capturing, the whichinformation is related can be to thetreated OCC differently modulation as part schemes. of image The processing detailed description following image of modulation capturing, schemes which isbased related on RSto the in OCCOCC ismodulation providedin schemes. Section 3The. detailed description of modulation schemes based on RS inAs OCC mentioned is provided previously, in Section OCC 3. only utilizes CMOS sensors, since it is widely employed in the majorityAs mentioned of mobile devicespreviously, including OCC only smartphones utilizes CMOS [11,17 ].sensors, With regards since it to is image widely capturing employed in in OCC, the majorityCMOS sensors of mobile can devices have either including GS or RSsmartphones mechanism, [11,17]. depending With regards on the image to image processing capturing algorithms in OCC, CMOSand the sensors TX switching can have rate either [5,17 GS,22 or,38 RS,39 mechanis]. For example,m, depending a camera on with the image a full processing high definition algorithms (FHD) andresolution the TX (1920 switching1080 rate pixels) [5,17,22,38,39]. and a frame For rate example, of 60 fps a employingcamera with the a GSfull mechanism high definition can sample(FHD) × resolution60 complete (1920 frames × 1080 every pixels) second. and On a theframe other rate hand, of 60 the fps identical employing camera the with GS mechanism RS processes can individual sample 60rows complete on each frames frame asevery a sample, second. thus On yielding the other a samplinghand, the rate identical of 64,800 camera (60 with1080) RS samples processes per × individualsecond (sps). rows on each frame as a sample, thus yielding a sampling rate of 64,800 (60 × 1080) samples per second (sps).
Electronics 2020, 9, 1339 9 of 44 Electronics 2020, 9, x FOR PEER REVIEW 9 of 45
Figure 8.8. (a) AA comparisoncomparison betweenbetween the GSGS andand RS;RS; ((bb)) aa structuralstructural comparisoncomparison betweenbetween CCDCCD andand CMOS sensors.sensors. The complete frame capture period comprises the exposure and frame-saving periods as shown The complete frame capture period comprises the exposure and frame-saving periods as shown in Figure9a. The exposure time, which can be set by the camera controller, is a period when the IS in Figure 9a. The exposure time, which can be set by the camera controller, is a period when the IS is is exposed to the light. In the RS cameras, the rows of the IS are exposed sequentially. Therefore, exposed to the light. In the RS cameras, the rows of the IS are exposed sequentially. Therefore, each each pixel in a row of frame is treated as an individual sample, which increases the sampling rate by pixel in a row of frame is treated as an individual sample, which increases the sampling rate by thousand times, thus offering data rates much higher than FR. Figure 10 demonstrates the LED status thousand times, thus offering data rates much higher than FR. Figure 10 demonstrates the LED status and the intensity in each row of the IS against the time. It is important to note that the switching rate of and the intensity in each row of the IS against the time. It is important to note that the switching rate the transmitting LED should be set to be lower than the sampling rate of the RS mechanism. of the transmitting LED should be set to be lower than the sampling rate of the RS mechanism. Following the exposure time Texp, the camera’s shutter is closed, followed by the image processing Following the exposure time Texp, the camera’s shutter is closed, followed by the image (including image compression) as well as saving the frame. Customarily, the image processing and processing (including image compression) as well as saving the frame. Customarily, the image frame saving are called the frame saving period. Obviously, the frame saving period is determined processing and frame saving are called the frame saving period. Obviously, the frame saving period by the speed of the IPU and the storage device. In the RS, the frame-saving period is compressed by is determined by the speed of the IPU and the storage device. In the RS, the frame-saving period is pipelining the readout time sequentially on each row as illustrated in Figure9b. compressed by pipelining the readout time sequentially on each row as illustrated in Figure 9b. The RS mechanism in CMOS, however, has an undesirable skew in the captured frame (i.e., RS distortion) during object capturing. This is due to the fact that the object moves at a high speed during the exposure period as illustrated in Figure9b [ 29,39]. In order to avoid this RS distortion for a 10-percent blur threshold, the exposure time must meet the following condition [39]:
∆x Texp (1) ≤ 10v where ∆x and v denote the object’s displacement and velocity, respectively. However, the image distortion is largely discounted as the image-capture functionality of the camera is ignored. The primary concerns of the RS-based OCC are image skew, link range and the limited camera resolution. Meanwhile, in the GS-based OCC with an N M pixels IS with each pixel producing G-level × grayscale signals and FR, a theoretical maximum achievable data rate limit is given by [5,40,41]:
1 r = NM log G F (2) 8 2 R The constant value of 1/8 is a rate reduction factor for three-dimensional signals, which is a downsamplingFigure 9. ( factora) Frame for capture the data period; rate in(b a) RS camera distortion to avoid causing spatial a skewed aliasing, object since in the the capture pixels frame. in an image sensor are densely organized [41]. The rate reduction factor can only be omitted from the equation when each pixel in the image sensor is placed with adequate gap [41]. For example, a camera with a
Electronics 2020, 9, x FOR PEER REVIEW 9 of 45
Figure 8. (a) A comparison between the GS and RS; (b) a structural comparison between CCD and CMOS sensors.
The complete frame capture period comprises the exposure and frame-saving periods as shown in Figure 9a. The exposure time, which can be set by the camera controller, is a period when the IS is exposed to the light. In the RS cameras, the rows of the IS are exposed sequentially. Therefore, each pixel in a row of frame is treated as an individual sample, which increases the sampling rate by thousand times, thus offering data rates much higher than FR. Figure 10 demonstrates the LED status Electronicsand the intensity2020, 9, 1339 in each row of the IS against the time. It is important to note that the switching10 ofrate 44 of the transmitting LED should be set to be lower than the sampling rate of the RS mechanism. Following the exposure time Texp, the camera’s shutter is closed, followed by the image full high definition (FHD) resolution (1920 1080 pixels), 256 levels of grayscale in the IS with a FR processing (including image compression)× as well as saving the frame. Customarily, the image ofprocessing 60 fps yields and aframe theoretical saving maximum are called achievablethe frame sa dataving rate period. of 124.42 Obviously, Mbps [the41]. frame However, saving this period data rateis determined can only beby obtainedthe speed when of the each IPU IS and pixel the is stor illuminatedage device. by In an the independent RS, the frame-saving TX. Unfortunately, period is thiscompressed individual by illuminationpipelining the is readout impractical time due sequentially to the transmission on each row distance as illustrated between in theFigure camera 9b. and the TX.
Electronics 2020, 9, x FOR PEER REVIEW 10 of 45 Figure 9. (a) Frame capture period; (b) RS distortion causing a skewed object in the capture frame. Figure 9. (a) Frame capture period; (b) RS distortion causing a skewed object in the capture frame.
Figure 10. RS mechanism on image sensor for data capture.
ToThe carry RS mechanism out a more in accurate CMOS, measurement however, has of an the undesirable captured image,skew in a minimumthe captured of twoframe pixels (i.e., perRS thedistortion) smallest during feature objectSF is capturing. required. This The is smallest due to the feature fact that indicates the object the sizemoves of aat smallest a high speed identifiable during propertythe exposure in an period image, as i.e., illustrated a clear shapein Figure and 9b color. [29,39]. The In minimum order to ISavoid (camera) this RS resolution distortionS Rforis a then 10- twopercent times blur the threshold, size of FOV the (exposureFC), either time horizontal must meet or vertical the following FOV, divided condition by [39]: the smallest feature SF. Therefore, to capture a clear image, the minimum resolutionΔ of the camera is given as [42]: (1) 10 FC SR = 2 (3) where Δ and denote the object’s displacement andsF velocity, respectively. However, the image distortion is largely discounted as the image-capture functionality of the camera is ignored. The For example, if the FOV covers 100◦ and the smallest feature is 1 mm, the required minimum primary concerns of the RS-based OCC are image skew, link range and the limited camera resolution. resolution is 200 pixels. Thus, a camera with a resolution of 640 480 pixels would be adequate. Meanwhile, in the GS-based OCC with an N × M pixels IS with× each pixel producing G-level This is because 200 pixels are less than 480 pixels, which is the smallest dimension. In other words, grayscale signals and FR, a theoretical maximum achievable data rate limit is given by [5,40,41]: the number of pixels (in IS) define the maximum distance of capturing a clear image, related to the FOV in Equation (3) and its working distance1 illustrated in Figure6. = ( ) (2) In a camera-based OCC, a range of TXs8 can be adopted, e.g., digital display modules and LEDsThe [5,17 constant,22,36,38 value]. Color of 1/8 filter is arrays,a rate reduction wavelength factor division for three-dimensional multiplexing (WDM) signals, and which color shiftis a keyingdownsampling (CSK) based factor on RGBfor the LEDs data or rate displays in a havecamera been to employedavoid spatial in OCC aliasing, [5]. Interestingly, since the pixels the typical in an refreshimage sensor rate of are a digital densely display organized is ~60 [41]. Hz, whichThe rate is inreduction the same factor range can of F onlyR [38 ].be The omitted type offrom digital the displayequation used when as each the TX pixel is either in the LED,image LCD sensor screen, is pl OLED,aced with or e-paperadequate [5 gap,17,38 [41].,43]. For In example, display-to-camera a camera communications,with a full high definition multiple (FHD) colors resolution with multiple (1920 intensity × 1080 pixels), levels are 256 deployed levels of forgrayscale each pixel, in the since IS with the a of 60 fps yields a theoretical maximum achievable data rate of 124.42 Mbps [41]. However, this data rate can only be obtained when each IS pixel is illuminated by an independent TX. Unfortunately, this individual illumination is impractical due to the transmission distance between the camera and the TX. To carry out a more accurate measurement of the captured image, a minimum of two pixels per the smallest feature is required. The smallest feature indicates the size of a smallest identifiable property in an image, i.e., a clear shape and color. The minimum IS (camera) resolution is then two times the size of FOV ( ), either horizontal or vertical FOV, divided by the smallest feature . Therefore, to capture a clear image, the minimum resolution of the camera is given as [42]: