Annual Report Annual
NHK 2019 Science & Technology Research Laboratories 2019 NHK Science & Technology December 2020 Research Laboratories NHK Science & Technology Research Laboratories 2019
Nippon Hoso Kyokai [Japan Broadcasting Corporation] Table of Contents
Greetings ············································· 1 Accomplishments in FY 2019 ···············2
1 Reality Imaging - Spatial imaging ···········4 5 Smart Production - Universal service ···36
1.1 3D imaging technology 4 5.1 Automatic closed-captioning technology 36
1.2 AR (Augmented reality) / VR (Virtual reality) 7 5.2 Audio description technology 36 1.3 3D imaging devices 8 5.3 Machine translation technology 37 5.4 Information presentation technology 38
2 Reality Imaging - 8K Super Hi-Vision ···10 6 2.1 Video systems 10 Devices and materials for next-generation broadcasting ··············41 2.2 Cameras 11 6.1 Imaging technologies 41 2.3 Displays 12 6.2 Recording technologies 43 2.4 Sound systems 13 6.3 Display technologies 44 2.5 Video coding 15 2.6 Satellite broadcasting technology 17 2.7 Terrestrial broadcasting transmission 7 Research-related work ·························46 technology 18 7.1 Joint activities with other organizations 46 2.8 Wireless transmission technology for program contributions 20 7.2 Publication of research results 49 2.9 Wired transmission technology 22 7.3 Applications of research results 52
3 Connected Media ································· 23
3.1 Content provision platform 23
3.2 Content-matching technology for daily life 25 3.3 IP content delivery platform 26 3.4 TV-watching robot 27 3.5 Security 29 3.6 IP-based production platform technology 29
4 Smart Production - Intelligent program production ··········································· 31
4.1 Text big data analysis technology 31
4.2 Image analysis technology 32
4.3 Speech transcription technology 33
4.4 New image representation technique using real-space sensing 34
4.5 Promotion of the use of AI technology 35
NHK Science & Technology Research Laboratories Outline ···················54 Greetings
Kohji MITANI Director of NHK Science & Technology Research Laboratories
HK Science & Technology Research Laboratories (STRL), a part of the public broadcaster NHK and the Nsole research facility in Japan specializing in broadcasting technology, is working to create a rich broadcasting culture through its world-leading R&D on broadcasting technologies.
Fiscal year 2019 marked one year since 4K and 8K satellite broadcasting was launched in December 2018 after many years of R&D. It also saw the start of the test delivery of the simultaneous continuous online streaming and program catch- up service NHK Plus, which was officially launched in April 2020. It was a year that gave us the realization that NHK is making steady progress toward the goal of becoming a “public service media.”
We have been driving our research activities according to the NHK STRL 3-Year R&D Plan (FY 2018–2020), which is aimed at creating new broadcasting technologies and services for the future around 2030 and 2040. We will continue to push ahead with our R&D under the three pillar concepts of “Reality Imaging” technologies to deliver video and audio with a higher sense of presence and reality, “Connected Media” technologies to achieve more convenient broadcasting and services through the use of the internet, and “Smart Production” technologies to utilize artificial intelligence (AI) to support program production and expand universal services.
On June 1, 2020, NHK STRL celebrated the 90th anniversary of its foundation. We are deeply grateful to all those concerned, and the viewers and residents of neighboring communities who have supported us up to today. As in the past, the entire STRL will work as one to engage in R&D to play a leading role in the R&D on broadcasting technologies and services.
This annual report summarizes our research results in FY 2019. It is my hope that this report will serve as an impetus for you to better understand NHK STRL’s R&D activities. I also hope it will help us build collaborative relationships that promote R&D and opportunities for cooperative creation utilizing the results of our efforts.
Finally, I would like to express my sincere gratitude for your support and look forward to your continued cooperation in the future.
90th anniversary
NHK STRL ANNUAL REPORT 2019 | 1 Accomplishments in FY 2019
Reality Imaging - Spatial imaging NHK STRL is researching three-dimensional (3D) imaging technology that offers more natural 3D images without the use of special glasses with the goal of realizing a future 3D image broadcasting service. In FY 2019, we made progress in our research on a method for improving the resolution of 3D images and studied the color moiré reduction of 3D images for portable terminals. We also investigated a technology that achieves both the reduction of the amount of information of 3D images, which contain a huge amount of depth information, and an effective image representation method based on human vision property. In our work on the application of augmented reality (AR) and virtual reality (VR), we developed prototypes that allow people to try new viewing High-resolution VR images experiences and services and promoted the service concept through exhibitions such as projected on a large cylindrical the NHK STRL Open House 2019. To achieve a device that displays more natural 3D images, we conducted fundamental experiments on spatial light modulators and screen researched the display of 2D images using narrow-pixel-pitch liquid crystal and high- speed optical beam scanning with a narrow-pitch optical phased array. →See p. 4 for details.
Reality Imaging - 8K Super Hi-Vision We are conducting R&D on a production system to enable program production in full- featured 8K, which is the ultimate format of Super Hi-Vision (SHV), and research to identify synergy effects between the parameter sets. For imaging technologies, we investigated an autofocus function suitable for 8K three-chip cameras and a method for suppressing image degradation caused by higher frame rates. We also researched ways to increase the sensitivity of solid-state image sensors. For display technologies, we developed a 4K flexible OLED display formed on a plastic film and researched technologies for increasing the image quality of OLED displays. For audio technologies, we researched next-generation object-based audio and technologies for enhancing the spatial sound 4K flexible OLED display representation as well as technologies for reproducing 22.2 multichannel sound at formed on a plastic film home. In addition, we conducted full-featured 8K live production and transmission experiments using a 21-GHz-band broadcasting satellite as the verification of the full- featured 8K video coding standard and technologies for expanding the transmission capacity of satellite transmission systems. With the goal of realizing terrestrial 4K/8K broadcasting following the 4K/8K UHDTV satellite broadcasting launched in December 2018, we researched functional additions to the transmission system and conducted large-scale field experiments as part of Technical Examination Services of the Ministry of Internal Affairs and Communications. We also researched an IP multicast distribution method that uses commercial closed networks of FTTH owned by cable TV service providers. To achieve a wireless transmitter for producing a wide variety of 4K/8K programs, we worked on the development of ARIB Standard for 1.2-GHz/2.3-GHz-band mobile relay FPUs and researched a low latency wireless camera using a millimeter- wave band. →See p. 10 for details.
Connected Media As a study of technologies for providing content linking broadcasting and broadband networks, we researched Hybridcast Connect and other technologies that connect broadcast content with smartphones and IoT-enabled devices. We released a software development kit (SDK) and other tools for Hybridcast Connect as open source software to support service providers in developing device linkage applications. We also investigated the effectiveness of the utilization of viewers’ viewing history through demonstration experiments linking broadcasting and various services. In our research on IP content delivery platform technology that allows viewers to view content by using the internet, we researched a quick-response delivery technology to stabilize the viewing quality of video delivered on the internet and achieve smooth viewing operation and a technology for delivering the same content to many terminals stably and efficiently in a space crowded with users. We also continued with our research on TV-watching robots, which are expected to become partners to make TV watching more enjoyable. We Viewing experiments using incorporated a function for personal information and privacy protection into a robot that TV-watching robot we developed previously and also conducted viewing experiments using the robot. In our research on cryptography and information security, which are essential to ensure high security and reliability of services in the internet age, we studied cryptography algorithms including one that can be used for post-quantum computer measures and one for integrated broadcast-broadband services. In our research on IP-based production platform technology to realize efficient program production utilizing IP, we prototyped multiformat IP transmission equipment that transmits material in different formats (2K/4K/8K) from a venue to a broadcast station efficiently and developed a system monitoring tool to visualize networks to support the construction and operation of IP networks. →See p. 23 for details.
2 | NHK STRL ANNUAL REPORT 2019 Smart Production - Intelligent program production We are researching intelligent program production technologies to achieve an efficient program production environment using AI technology. We made progress in our research on the use of text big data such as social media by studying a data classification technology for accurate information extraction and a technology for analyzing viewers' opinions about programs after broadcasts. For image analysis technologies, we provided program production support in cooperation with relevant NHK departments. We also researched a technology for automatically recognizing texts in video footage and assigning metadata and a face recognition technology to identify the scenes in which a specific person appears from a massive amount of video. For speech transcription technologies, we investigated ways to expand the target of speech recognition to include telephone interview speech and also supported an effort to install the speech transcription equipment that we developed in each NHK regional key station. For new image representation technologies, we Transcription equipment researched a technology for helping produce more user-friendly video content and a technology installed in regionalkey stations for collecting the 3D information of objects by sensing a space. We also contributed to the development of services and systems that respond to the needs of broadcast producers through activities at Secretariat for AI Promotion, which we established to support the use of NHK STRL’s AI technologies for program production and other applications. →See p. 31 for details. Smart Production - Universal service We are conducting research on universal broadcasting services that allow all viewers to access and enjoy information. We conducted a trial service of an automated closed-captioning technology to automatically recognize program speech and convey it in text to those with hearing difficulties and researched a sign language CG generation technology for sports information to explain the status of sports events in sign language CG. In our work on services for visually impaired people, we studied automated audio description and robot commentary to automatically insert commentary speech for supplementing program information and built a large-scale database to improve the quality and expressive power of synthesized speech. In our work on services for inbound visitors to Japan, whose number is expected to increase, we researched a machine translation technology for news translation from Japanese to English. We also studied an automatic translation technology for Haptic devices for conveying newspaper articles in cooperation with external research institutions. In addition, we are sports experience researching the use of sensation other than sight and hearing for conveying program information. As a haptic presentation technology, we prototyped a haptic device that conveys information such as the strength of impact felt in sports and an editor that can edit and control haptic information. We also continued to study the possibility of conveying information by smell. →See p. 36 for details. Devices and materials for next-generation broadcasting We continued researching fundamental technologies of imaging, recording and display for next-generation broadcasting technology. For imaging technologies, we conducted research on highly integrated 3D imaging devices that could be applied to advanced imaging devices and on RGB-stack-type image sensors for compact and lightweight single-chip color cameras. We also began research on basic technologies for computational photography with the aim of realizing a new capture technology to obtain 3D information. For recording technologies, we conducted research on high-performance holographic memory that has a very large capacity and high transfer rate for the long-term storage of 8K video. We also researched a magnetic recording device utilizing magnetic nano-domains that has no moving parts to achieve a high reliability and conducted fundamental experiments to realize magnetic memory using new materials. For display Solution-processed oxide TFTs technologies, we searched for suitable materials and identified their detailed operation principles prototyped on a film substrate for a flexible OLED display with a longer lifetime and improved the characteristics of thin-film transistors (TFTs). We also researched solution-processed oxide TFTs that can be formed on a large film substrate and quantum dot light-emitting diodes with a high color purity. →See p. 41 for details. Research-related work We promoted our research on 8K SHV and other technologies in various ways, including through the NHK STRL Open House, press releases, various exhibitions, and reports. We also actively collaborated with other organizations and program producers. The theme of the FY 2019 NHK STRL Open House was “Taking media beyond the box.” It featured exhibits on our latest research results such as technologies for providing new viewing services that go beyond the limits of traditional TV and was attended by 21,702 visitors. We also conducted tours of our laboratories for visitors from home and abroad. We published articles describing our research results in conference proceedings and journals within and outside Japan and issued press releases. We also continued to consolidate our intellectual property rights and contributed to the development of technical standards by participating in activities at international and domestic standardization organizations. We STRL Open House 2019 cooperated with outside organizations through collaborative research and commissioned research efforts. We hosted visiting researchers from home and abroad, and dispatched NHK STRL researchers overseas for research activities. Our research outcomes were utilized for the program production of NHK. Our 8K slow-motion system was used in the production of sports programs such as grand sumo tournaments, All-Japan Judo Championships, Japan Championships in Athletics and the Rugby World Cup, and our system for colorizing past monochrome video using AI technology performed well in the production of the historical drama “Idaten.” In recognition of our research achievements, we received external awards including the Maejima Award. →See p. 46 for details.
NHK STRL ANNUAL REPORT 2019 | 3 1 Reality Imaging - Spatial imaging
1.1 3D imaging technology
With the goal of developing a future broadcasting media, demonstrated that generating and interpolating multi-view NHK STRL is researching a three-dimensional (3D) imaging images by referring to the obtained depth maps improved the technology that offers the sense of presence and reality that generation accuracy of edge parts of objects. cannot be expressed by conventional two-dimensional (2D) Since 3D images contain a huge amount of information, it is images. A promising way to achieve more natural and easily necessary to develop a high-efficiency coding technology to viewable 3D images without the use of special glasses is a use them for broadcasting service. In FY 2019, we investigated method that reproduces optical images in the air (hereafter ways to reduce the number of multi-view images to be encoded referred to as spatial imaging method). In FY 2019, we in order to reduce the amount of 3D information. We devised a conducted R&D on capture, display and coding technologies method that reduces multi-view images to those of a smaller for high-resolution 3D images using the spatial imaging number of viewpoints and sets a different viewpoint position to method and on a 3D imaging technology for portable terminals. each frame before coding. The experiment results demonstrated We also worked to identify the characteristics of 3D images that this method can reduce the amount of transmission data suitable for diverse viewing environments and studied a depth- by about 30%. We continued to attend MPEG meetings and compressed expression technology for 3D images with higher submitted the results of coding experiments using provided quality. test sequences as input to promote standardization activities for 3D video coding standards. We also exhibited integral 3D ■ High-resolution 3D imaging technology displays at the MPEG meeting and ITU-R meeting and contributed to the publicizing of our 3D imaging technology With the aim of increasing the resolution of 3D images, we and the promotion of standardization of 3D video coding are developing a system called Aktina Vision(1) that uses multi- methods. view images and a special diffusion screen. To realize 3D images having a high-definition television (HDTV)-equivalent ■ 3D imaging technology for portable terminals resolution, in FY 2019, we developed a time-division light ray multiplexing technique that shifts light rays in a time-sharing We are researching an integral 3D display technology using manner to increase the number of light rays equivalently and eye-tracking system with the aim of realizing portable 3D conducted experiments to verify its effectiveness in improving display. The integral 3D method using a direct-view display the resolution characteristics of displayed images. As the causes color moiré because the display’s subpixel structure of method of time-division light ray multiplexing, we devised a red, green and blue is observed through a cyclical lens array. pixel shift method for improving the resolution of each multi- To reduce the color moiré, in FY 2019, we verified a method for view image and a light ray shift method for increasing the shifting the pixels of elemental images four times using a number of multi-view images (Figure 1-1). We evaluated the double wobbling optical device. In the experiments, we effectiveness of these methods in improving resolution combined display equipment consisting of a liquid crystal characteristics through analyses and experiments. The results display (LCD) having a Bayer-pattern pixel structure and a lens showed that the pixel shift method almost doubled the array with a double wobbling optical device each consisting of resolution of 3D images near the screen surface. The light ray a polarization grating and a polarization control element shift method also improved the resolution of 3D images at a (Figure 1-2). Using this system, we demonstrated that shifting distant depth from the screen surface by about 1.7 times(2). the pixels of elemental images four times according to the As a high-resolution capture technology for 3D images, we Bayer-pattern pixel structure reduced the intensity of color continued to research a technology for the efficient generation moiré to 25%. This method also improved the resolution of 3D of high-resolution 3D images using a camera array. While a images at a distant depth from the lens array surface because system that we developed in FY 2018 produced depth maps at it increases the pixel count by about four times. all camera positions to generate light rays, in FY 2019, we Image display with a wide viewing zone and high quality can devised a method for calculating depth maps only at fewer be realized by detecting the viewer’s eye position with a small representative viewpoint positions and using them to generate camera installed on a display and showing 3D images according and interpolate dense light rays. Experiments using a camera to the eye position. In FY 2019, we used this integral 3D display array demonstrated that this method can reduce the processing with eye-tracking system(4) to achieve the 3D display of moving time to about 1/3 that of the conventional method(3). We also used an ordinary camera array in combination with depth cameras to improve the quality of generated images. We Wobbling optical developed a calibration method for depth cameras and device 1 R G R G Wobbling optical 1 pixel Shift direction device 2 G B G B Shift direction Pixel shift Light ray shift R G R G Improve the resolution of multi-view images Increase the number of light rays R G R G G B G GB R G R G B G B G B G B Shift multi-view images Shift light ray B G B G positions obliquely R G BR GG B G obliquely by half a pixel R G R G G R G R Shift direction G B GG RB G R G B G B λ/4 plate Polarization Polarization B G B G control element grating B G B G Projector R G R G Wobbling optical device G R G R using polarization grating Multiplexed Pixel shift control device Polarization switching (Switching by voltage in a time-sharing (Wobbling device) device for light ray shift Light ray shift optical system Multi-view Display screen applied to polarization manner for display (Polarization grating) Time images control element)
Figure 1-1. 3D image display using time-division light ray multiplexing Figure 1-2. Color moiré reduction using a double wobbling optical device
4 | NHK STRL ANNUAL REPORT 2019 1 Reality Imaging - Spatial imaging
images in addition to still images. We generated high-quality impressions change according to the characteristics of 3D models of live-action video from multi-view video captured displayed images through subjective evaluations. We used the with more than 30 cameras surrounding an object. We then semantic differential (SD) method for the subjective evaluations. arranged these 3D models in a virtual space and generated We asked participants to answer which of each pair of multi-view images using a virtual camera array of 64 cameras. adjectives with opposite meanings is closer to their impressions Parallel processing by the graphics processing unit (GPU) of the evaluated images. We conducted these evaluations for enabled the high-speed rendering of elemental images at a multiple adjective pairs and extracted the participants’ potential speed of about 32 fps (frames per second). This made it possible impressions by analyzing the results by statistical techniques. to display integral 3D images of live-action video with a wide We used 3D-modeled computer graphics (CG) images (Figure viewing zone (81.4° horizontal and 47.6° vertical) and high 1-5) as evaluated images. The use of CG images instead of live- quality (Figure 1-3). action images allows us to set camera positions freely and thus We demonstrated these integral 3D images of live-action control binocular disparity (displacement of an image caused video shown on a 3D display with eye-tracking system at many by the different positions of the left and right eyes) and motion exhibitions and international conferences such as the NHK parallax (displacement of an image caused by the difference in STRL Open House 2019, International Broadcasting Convention the head position), which are visual cues for producing depth (IBC) 2019, Asia-Pacific Broadcasting Union (ABU) General perception. Meeting, World Congress of Science & Factual Producers We analyzed the effects of the size of motion components in (WCSFP) 2019 and NHK Science Stadium 2019. images and the binocular disparity on impressions in a viewing To capture wide-viewing-zone images suitable for 3D environment using the fixed display. The results showed that images on portable terminals with a smaller number of the effect of the binocular disparity on multiple impressions cameras, in FY 2019, we developed a multi-stereo robotic such as the sense of presence and the feeling of fluidity tends cameras system consisting of three stereo robotic cameras to be larger when the image contains less motion(5). systems (Figure 1-4), which expanded the horizontal viewing We conducted similar experiments also in a viewing angle for generated 3D models from the conventional value of environment using the HMD. We examined the effect of 30 degrees to 110 degrees. We confirmed the successful binocular disparity and motion parallax by varying the camera capture and display of wide-viewing-zone 3D images by positions in the virtual 3D space of CG images but found no showing the 3D models generated using this system on eye- noticeable effect on image impressions under any condition. tracking integral 3D displays. HMDs have the capabilities to achieve a wider field of view than that of conventional flat-panel displays and to display ■ 3D image characteristics suitable for the viewing appropriate images according to the direction of the viewer’s environment head. We experimentally restricted these capabilities to make the HMD show images in a wide field of vision but only into a We are engaged in research to identify 3D image certain direction, just like a theater screen. The results characteristics suitable for diverse viewing environments. In demonstrated a tendency of some impressions such as the FY 2019, we conducted evaluations on impressions gained feeling of power to be weakened(6). from watching 3D image content in two different video display environments and studied factors affecting the impressions in ■ Depth-compressed expression technology for viewing effect of 3D display. high-quality 3D images In the experiments, we used a 55-inch diagonal fixed display and head-mounted display (HMD) equipped with the Theoretically, 3D image display using the integral method stereoscopic 3D displaying function to measure how can display high-quality images only in a limited depth range. For 3D displays having this characteristic, we are studying the use of depth-compressed expression that enables high-quality display of scenes with a large depth. Depth compression reduces unintended blurring and displays high-quality 3D images by compressing and deforming the shape of a scene to be displayed in the depth direction to make it fit within a depth reconstruction range with acceptable spatial resolution characteristics. It is important to elaborate a way of deforming scenes to ensure that they appear natural as 3D images even after they are deformed greatly. Depth-compressed 3D images are likely to look unnatural when the viewer views the display obliquely (Figure 1-6). The unnaturalness of 3D images depth-compressed by a conventional method tends to be even more conspicuous when they are viewed on portable displays such as tablets.
Figure 1-3. Display of live-action video on an integral 3D display with eye-tracking system
Stereo robotic cameras system Stereo robotic cameras system
Stereo robotic cameras system
Figure 1-4. Multi-stereo robotic cameras system Figure 1-5. Example of CG images used in experiments
NHK STRL ANNUAL REPORT 2019 | 5 1 Reality Imaging - Spatial imaging
Static Without depth compression With depth compression (Conventional method)
Viewpoint moves Large distortion
Examples of displayed images Viewpoint-tracking (Proposed method)
Origin point of depth Small distortion compression also moves as viewpoint moves Virtual display surface position Scene shapes Figure 1-6. Introduction of viewpoint-tracking depth compression
Depth compression (Geometry transformation)
(A) Static depth compression (B) Viewpoint-tracking (dynamic) depth compression Depth compression function 5
4 Display surface position
3.5 Acceptable level Improvement 3.5 Acceptable level Depth compression 3 Depth after transformation [m] function adjustment UI Original depth [m] Scene 2 Cube (Near) Cube (Mid) Cube (Far) Flower (Near) Classroom (Mid) Urban city (Far)
1 approx. 1.3m approx. 10cm Error bars: Standard error
Evaluation scores of unnaturalness (MOS) Figure 1-9. Prototype depth expression adjustment system 012345012345 Depth range after compression [m] expression adjustment system to adjust 3D image expression according to the producer’s intention of production effects Figure 1-7. Efficient depth compression by the proposed method (Figure 1-9)(8). This system controls depth compression parameters in real time using a user interface with multiple sliders. This enabled production effects such as giving depth to a certain object predominantly to make it appear in three Urban city (Far) Classroom (Mid) Flower (Near) dimensions. With this system, we demonstrated the feasibility of 3D video production reflecting the producer’s intentions better by utilizing the limited depth reconstruction capability of displays.
Cube (Far) Cube (Mid) Cube (Near) [References] (1) H. Watanabe, N. Okaichi, T. Omura, M. Kano, H. Sasaki and M. Kawakita: “Aktina Vision: Full-parallax three-dimensional display with 100 million light rays,” Scientific Reports, Vol.9, 17688 (2019) (2) T. Omura, H. Watanabe, N. Okaichi, H. Sasaki and M. Kawakita: “Three-dimensional Display Method with Time-Division Light-ray Long-distance scene Medium-distance scene Short-distance scene Shifting Method using Polarization Grating,” 3D image conference 54.54 [m] 3.05 [m] 0.218 [m] 2019, 3-3 (2019) (in Japanese) (3) M. Kano and M. Kawakita: “Efficient Multi-View Image Generation Method Considering Light Field Display,” ITE Winter Annual Figure 1-8. Scenes used in experiments Convention 2019, 22B-2 (2019) (4) N. Okaichi, H. Sasaki, H. Watanabe, K. Hisatomi and M. Kawakita: Focusing on the fact that portable displays are used for personal “Integral 3D display with eye-tracking system using 8K display,” ITE viewing, we investigated viewpoint-tracking depth Winter Annual Convention 2018, 23D-3 (2018) (in Japanese) compression that adjusts the depth-compression method (5) M. Tadenuma: “Result of Experimental Estimations to Inspect the according to the viewer’s viewing position(7). Figure 1-7 shows Method Estimating Efficiency of 3D Images,” ITE Technical Report, the results of evaluation of images depth-compressed by the HI2019-65, Vol.43, No.14, pp.97-102 (2019) (in Japanese) conventional (static) method and those by the viewpoint- (6) T. Morita, Y. Sawahata, M. Harasawa and K. Komine: “Analysis of tracking (dynamic) method. The scenes used in the evaluation Impressions of Virtual Reality Images with Head-Mounted Display experiments are shown in Figure 1-8. The results showed that and their factors,” ITE Annual Convention 2019, 22E-3 (2019) (in the unnaturalness of the scenes were acceptable even when Japanese) they were compressed into a depth of 10 cm using the (7) Y. Miyashita, Y. Sawahata, M. Katayama and K. Komine: “Depth viewpoint-tracking method while the conventional method boost: Extended depth reconstruction capability on volumetric required a depth of 1.3 m. These results suggest that the viewer display,” ACM SIGGRAPH 2019 Talks, SIGGRAPH 2019 (2019) feels a larger depth than the physically presented depth and (8) Y. Sawahata, Y. Miyashita and K. Komine: “Development of Depth that the new method is effective in virtually extending the Adjustment System for 3D Video Production,” ITE Annual limited depth reconstruction capability of 3D displays. Convention, 14E-5 (2019) (in Japanese) By applying the depth-compressed expression technology to 3D video production, we prototyped a depth-compressed
6 | NHK STRL ANNUAL REPORT 2019 1 Reality Imaging - Spatial imaging
1.2 AR (Augmented reality) / VR (Virtual reality)
We studied the concept of services that will offer new user of actual viewing conditions regarding viewing positions and experiences to viewers by using augmented reality (AR) and the sense of space sharing to see where virtual persons should virtual reality (VR) technologies from two different approaches: be displayed for effective communication. The results “By AR/VR,” which provides new viewing experiences by suggested that the front-back relationship of viewing positions combining existing technologies, and “For AR/VR,” which may exercise an effect on the perception of viewing images implements technologies that have yet to be introduced and together(2). newly developed technologies. We also developed a delivery technology for volumetric data to realize AR/VR content ■ “For AR/VR” approach delivery using the convergence of broadcasting and telecommunications. We investigated a viewing style that utilizes high-resolution 360 images with the goal of realizing a viewing experience that ■ “By AR/VR” approach allows the viewer to view images of any directions beyond the limits of a TV screen and enjoy an excellent sense of As the concept of a new TV viewing service utilizing AR immersiveness and presence that conventional TVs cannot technology, we investigated a space sharing service that allows provide. At the NHK STRL Open House 2019, we produced 180 the viewer to enjoy watching TV together with family and images with a high resolution of about 12K, exceeding 8K, by friends at distant locations while feeling the presence of TV integrating (“stitching”) images captured with three 8K cameras performers nearby as if they have come out of the TV screen. In arranged radially. Using eight 4K projectors (Figure 1-12), we this service, 3D images of TV performers, family and friends at projected the high-resolution VR images on a large 180-degree distant locations and the viewer himself/herself in the past are cylindrical screen (Figure 1-13), which was enjoyed by many synthesized and displayed in their actual size in a space viewed visitors(3). We also reproduced highly immersive sound suitable through AR glasses or a tablet terminal. We proposed the for 180 images using multiple loudspeakers installed above viewing style in which the viewer watches TV while sharing and below the cylindrical screen for sound reproduction. In the space with TV performers, family and friends beyond space addition, we exhibited mock-ups of viewing styles with a high- and time and demonstrated the service at exhibitions such as resolution HMD and a dome-type display for personal viewing the NHK STRL Open House 2019 and IBC 2019 (Figure 1-10). (Figure 1-14). While the space sharing service offers a viewing style in which TV performers and others come to the viewer’s living ■ Transport technology for volumetric data room, we also proposed another viewing style in which the viewer and family or friends in the same space go to a virtual One example of AR/VR content taking advantage of space and share 360 images. We prototyped a system that integrated broadcast-broadband may be a service that allows allows the viewer to enjoy 360 images while interacting with the viewer to view an object in TV images from a free angle persons and objects nearby (Figure 1-11) using extended VR using AR technology by delivering the volumetric data of the technology(1). This technology displays the persons and objects same object to AR glasses or a tablet terminal through near the viewer by cutting out their images captured with a stereo camera installed on the viewer’s head-mounted display (HMD) while simultaneously displaying 360 images in the remaining area. To realize the space sharing service, we conducted a survey
Figure 1-12. High-resolution VR projection system
Figure 1-10. Space sharing service
360 images
Radius of the real space Video see-through HMD
1~2 m Rendering PC
Reality Virtual
Figure 1-11. 360 image sharing system Figure 1-13. Large cylindrical screen
NHK STRL ANNUAL REPORT 2019 | 7 1 Reality Imaging - Spatial imaging
(a) high-resolution HMD (b) dome-type display Figure 1-14. Mock-ups of (a) high-resolution HMD and (b) dome-type display
Figure 1-16. Program-linked AR content Video and audio via broadcast volumetric data of a TV performer simultaneously and offered Synchronized Add time Transmission reproduction using visitors an opportunity to experience content viewing with information delay difference time information free-viewpoint AR at the NHK STRL Open House 2019. This content presents the AR images of the performer wearing different colorful outfits in various settings in her real size in AR content (volumetric data) front of the TV screen. In addition to the NHK STRL Open via broadband House, we exhibited the content at the NHK Yamagata Station Free-viewpoint AR display on tablet Open House in October, the NHK Showcase in ABU Tokyo 2019 General Assembly & Associated Meetings in November and a program session at ITE Winter Annual Convention in December, Figure 1-15. Concept of real-time transmission of volumetric data and offered many visitors the experience of a new viewing style combining TV and AR technology. broadband networks. To realize such services, we are researching a transport technology for volumetric data. [References] We newly developed a real-time transmission technology for (1) H. Kawakita, K. Yoshino, D. Koide, K. Hisatomi, Y. Kawamura and K. live-action volumetric data of human objects(4)(5) that could be Imamura: “AR/VR FOR VARIOUS VIEWING STYLES IN THE FUTURE used for live sports coverage and other live programs in the OF BROADCASTING,” Proc. of IBC2019 (2019) future. Since this technology can transmit live-action (2) H. Kawakita, K. Yoshino and T. Handa: “Survey on TV Viewing for volumetric data of 30 frames per second at about 100 Mbps in Positioning Virtual Human in Space Sharing,” Proc. of HCG real time to present AR content, the viewer does not need to Symposium2019, HCG2019-I-3-8 (2019) download the volumetric data to his/her viewing terminal in (3) D. Koide, H. Kawakita, K. Yoshino, K. Ono and K. Hisatomi: advance. To prevent the gap in presentation timing between “Development of High-Resolution Virtual Reality System by TV images and AR content due to the transmission latency Projecting to Large Cylindrical Screen,” Proc. of ICCE2020, IEEE, difference between broadcasting and broadband, we employed 1.16.1 (2020) a mechanism for adding an absolute time stamp to each frame (4) Y. Kawamura, Y. Yamakami, H. Nagata and K. Imamura: “Real-time of volumetric data before transmission and referring to the Distribution of Dynamic Volumetric Data for Augmented Reality time stamp at the time of presentation, in a similar way to Synchronized with Broadcasting,” ITE Technical Report, BCT2019- MPEG Media Transport (MMT) used for the advanced satellite 46, Vol.43, No.10, pp.41-44 (2019) (in Japanese) broadcasting for 4K/8K. This enabled a service that ensures (5) Y. Kawamura, Y. Yamakami, H. Nagata and K. Imamura: “Real-Time high-accuracy synchronization between TV images and AR Streaming of Sequential Volumetric Data for Augmented Reality content (Figure 1-15). Synchronized with Broadcast Video,” Proc. of ICCE-Berlin, IEEE, We produced AR content provided by integrated broadcast- pp.280-281 (2019) broadband system (Figure 1-16) by capturing the 4K video and
1.3 3D imaging devices
■ High-density spatial light modulator (SLM) accordance with the magnetization direction of the magnetic materials. We previously prototyped an MO light modulator We are engaged in research on electro-holography to display that can switch the magnetization direction via the motion of natural three-dimensional (3D) motion images. Displaying 3D the magnetic domain wall (the boundary between two motion images in a wide viewing zone of 30 degrees or more magnetic domains with different magnetization directions) requires the development of a high-density spatial light induced by pulse currents injected to the magnetic pixels (an modulator (SLM) having a very small pixel pitch of 1 µm or MO light modulator driven by current-induced domain wall less. We are developing a spin-SLM that uses small pixels with motion) and successfully demonstrated its basic operation on magnetic materials and researching a liquid crystal SLM with a a single-element basis. high density. The spin-SLM comprises magneto-optical (MO) With the aim of building an array of MO light modulators, in light modulators that can modulate light by the MO Kerr effect, FY 2019, we prototyped a high-density one-dimensional (1D) in which the polarization plane of reflected light rotates in array and succeeded in current-driven light modulation
8 | NHK STRL ANNUAL REPORT 2019 1 Reality Imaging - Spatial imaging
operation. The prototype array consists of five MO light of phase variation among channels that is caused by accuracy modulators arranged with a 1 µm pitch. Only two of them, the errors generated during the fabrication of OPAs. This 2nd and 4th pixels, are connected to a pulse current source successfully suppressed the ripple component in optical beams, (Figure 1-17(a)). As Figures 1-17(b) and (c) show, all the pixels enabling optical beam scanning at a high speed of 2 MHz(3) appeared white after current injection from right to left (the (Figure 1-18). We also developed a method for calculating the magnetization of the pixel switched to downward), while the phase of each channel from far-field patterns emitted by 2nd and 4th pixels turned dark after current injection from left shifting the optical phase going through any one of the OPA’s to right (the magnetization of the pixel switched to upward), channels four times (phase compensation by phase-shifting indicating successful light modulation induced by current digital holography). This demonstrated that accurate phase injection(1). This showed that the MO light modulators driven compensation can be performed in a short time even if the by current-induced domain wall motion can be applied to a number of channels increases(4) (Figure 1-19). To further high-density SLM with a 1-µm-pitch array. expand the deflection angle, we studied the use of silicon Increasing the pixel density of liquid crystal SLMs would nitride optical waveguides to enable a narrower pitch of OPA’s increase the leakage of the electric field between adjacent output part. We demonstrated that applying a taper structure pixels and reduce the contrast ratio. To address this issue, we to the serially grafted an EO polymer and a silicon nitride previously prototyped a high-density 1D device of ferroelectric optical waveguide decreases optical loss in the waveguide liquid crystal that is turned on and off by applying a positive or connection, which improves the output efficiency from the negative voltage and confirmed that it achieves a higher conventional value of 65% to 82%. contrast ratio than that of common nematic liquid crystal. In FY 2019, we designed and fabricated a high-pixel-density [References] ferroelectric liquid crystal device with a pixel pitch of 1 µm×1 (1) N. Funabashi, R. Higashida, K. Aoshima and K. Machida: “Fabrication µm and demonstrated the display of two-dimensional (2D) of high density one-dimensional array of domain wall motion type static images(2). spin light modulation device,” 2019 ITE Annual Convention Program, 33D-4 (2019) (in Japanese) ■ Narrow-pitch optical phased array (2) S. Aso, Y. Isomae, J. Shibasaki, K. Aoshima, T. Ishinabe, Y. Shibata, K. Machida, H. Fujikake and H. Kikuchi: “Driving Experiment of 1 µm For a future integral 3D display with much higher performance × 1 µm Pixel Pitch Liquid Crystal Devices Using Two-Layer Structure than current displays, we are conducting research on a new Electrodes,” The 67th JSAP Spring Meeting, 12a-PA1-7 (2020) (in beam-steering device that can control the direction of light Japanese) rays (optical beams) from each pixel at a high speed without (3) M. Miura, Y. Miyamoto, Y. Hirano, Y. Motoyama, K. Machida, E. using a lens array. We previously designed and prototyped an Ueda, C. Yamada, T. Yamada, A. Otomo and H. Kikuchi: “High-speed 8-channel optical phased array (OPA) using an electro-optic optical beam steering of optical phased array using electro-optic (EO) polymer, which can change the refractive index at a high polymer,” ITE Winter Annual Convention, 13A-4 (2019) (in Japanese) speed by applying an external voltage, for its optical waveguides (4) M. Miura, Y. Miyamoto, Y. Hirano, Y. Motoyama, K. Machida, E. (channels) and demonstrated optical beam control with a Ueda, C. Yamada, T. Yamada, A. Otomo and H. Kikuchi: “Phase scanning speed of 200 kHz and a deflection angle of 22.1 compensation method for optical phased array based on phase- degrees. shifting digital holography,” Proc. SPIE, 11284, pp.1128424.1- In FY 2019, we applied a phase compensation technology 1128424.6 (2020) based on optimizing the optical intensity to address the issue
Nano magnet Photo Control voltage (2 MHz) detector 2 Magneto-optical light modulator Light modulation layer driven by current-induced Photo pt al te t domain wall motion Electrode Incident Output detector 1 light light OPA Detector 1
N S S N Voltage control (Sine wave) Detector 2
Connected to power source a pt al ea a (b) Change of detected optical ea e e t te te t e t e
N S N S Electrode Figure 1-18. High-speed optical beam scanning experiment of OPA Unconnected to power source Pulse current source (a) Schematic diagram of high-density 1D array
Electrode Electrode (b) Light modulation element (c) Light modulation element