Line Detection for Moth Flight Simulation

Line Detection for Moth Flight Simulation

The University of Saskatchewan Department of Computer Science Technical Report #2006-05 Line Detection for Moth Flight Simulation Brennan Rusnell Mark Eramian August 18, 2008 Contents 1 Introduction i 2 Theory v 3 Experimental Method vi 3.1 UIRemoval....................................... vii 3.2 EdgeRemoval..................................... ........ vii 3.3 HorizonCalculation .............................. ........... vii 3.4 OcclusionDetection .............................. ........... viii 3.5 VerticalEdgeDetermination . ............. ix 4 Setup x 4.1 Installation .................................... .......... x 4.2 Configuration.................................... x 4.3 Execution ....................................... xi 5 Analysis xii 5.1 HorizonOrientation .............................. ........... xii 5.2 OrthogonalEdges................................. .......... xii 6 Conclusion xvi 7 Acknowledgments xviii Bibliography xix i List of Tables 5.1 This data shows how the proposed methodology’s approximation to the horizon orientation compares to the actual horizon, which was manually computed.................. xiii 5.2 Terms/calculations used in analysis. The numbers generated by these calculations facilitate quantitative analysis of the proposed methodology. .................. xiv 5.3 The data in this table shows the correctness and completeness of the proposed methodology fororthogonaledgedetection. ............ xv 5.4 Overall data from Table 5.3. The values from Table 5.3 were summed and used in the equations from Table 5.2 in order to create an overall value for each field.................. xv ii List of Figures 1.1 Example scene capture. This is a frame from a 2D movie of the flight path of a moth and servesasanexampleoftheanalyzeddata.. ............. ii 1.2 Sample scene illustrating the extracted line parameters. ..................... ii 1.3 Image of the depth buffer before image processing. The imagecontrastislow.. iii 1.4 Image of the depth buffer after histogram equalization. Histogram equalization exaggerates thequantizationofdepthvalues. ............ iv 3.1 Example of demo user interface (autocontrasted for viewingpurposes).. vii 3.2 Example scene demonstrating a border surrounding the image. ................. viii 3.3 Outputimage. .................................... ix 6.1 This rendering demonstrates how the limited depth data in the .obj files can be used to performocclusiontesting. ...........xvii iii Abstract During a Hawkmoth’s (Manduca sexta) flight path, much attention is directed towards the orientation and location of vertical edges. Previous studies have collected moth flight data where a moth controls its own virtual three-dimensional visual environment through an interface with the video game Descent 3. Descent 3 is also responsible for the rendering of the three-dimensional virutal environment. Movies of the three-dimensional scenes visible to the moths during its simulated flight are recorded. This research aims to extract the vertical edges from individual video frames by reconstructing the three-dimensional geometry associated with each frame. Scenes were reconstructed using Descent 3 to render the three-dimensional geometry stored within recorded “demo files”. The geometric information sent to the graphics processing unit from the Open Graphics Library was recorded and analyzed to extract the desired vertical edges. Overall, the proposed vertical edge detection methodology had a specificity of 0.983913 and a sensitivity of 1.000000. Most errors resulting in edges incorrectly marked as vertical can be attributed to a single type of failure case. Chapter 1 Introduction This research analyzed data that represents the basic principles of moth flight. Figure 1.1 is a frame from a two-dimensional (2D) movie of the flight path of a moth in a “virtual world”. This figure serves as an example of the analyzed data. In the example scene, rectangular objects emanate from a flat plane. Furthermore, the background is black while the ground plane and rectangular objects share a black and white checkerboard texture. The objective of this research was to find the following data in image-space (screen coordinates): 1. α: the horizon orientation; 2. β: the angle of each rectangular block’s edge (edge refers to one of the four lines that collectively define a face of a rectangular block) that is visible and orthogonal to the horizon; 3. length: the length of these edges; 4. location: the location of the midpoint of these edges. The information of interest can be seen in Figure 1.2. During a Hawkmoth’s (Manduca sexta) flight path, much attention is directed towards the orientation and location of vertical edges [1, 3, 9, 10, 12]. This data is collected to provide metrics of a visual stimulus that can be related to simultaneously recorded physiological data from the moth’s nervous system. The information extracted from these images is beneficial because, combined with the physiological data and knowledge of behavior, it will provide an understanding of what the moth’s nervous system responds to and how this relates to flight steering [1, 3, 9, 10, 12]. The initial methodology for collecting horizon and edge data was to continue the work of Madsen [6] and use edge detection algorithms from the domain of image processing on individual frames. However, edge detection failed due to the texture present in the scenes (or “virtual world”). As depicted in Figure 1.1, the boundary between white and black coloured texture regions creates an edge. Thus, edge detection finds edges within the faces of the rectangular blocks. Furthermore, because the background is black in colour, it becomes difficult to find corners that contain a black texture region. Edge detection algorithms would need to be complemented with line reconstruction algorithms in order to fill in the missing edge information. Due to the large number of edges that existed after edge detection, most of which were from texture boundaries (not geometry), line reconstruction became a tedious, inaccurate process. Thus, other options were explored. The option considered next was edge detection on the Open Graphics Library (OpenGL) depth buffer. GLIntercept (described in Section 2) is a piece of software which allows one to save the depth buffer on a per frame basis. GLIntercept captures an image of the depth buffer by assigning a different intensity of gray to each depth value such that more distant objects have a greater intensity and closer objects have a lower intensity. However, Figure 1.3 shows that the resulting images were low in contrast. This made it difficult to perform edge detection. In an attempt to aid in edge detection, histogram equalization was performed on the images in Matlab, but this process exaggerated the quantization of depth values, as shown in Figure 1.4. As a result, edges were detected along the visible depth discontinuities in the histogram equalized depth buffer image. Furthermore, GLIntercept created multiple depth buffer images per frame, making it difficult to determine which image contained the final scene data. Therefore, this method was found to be ineffective. i Figure 1.1: Example scene capture. This is a frame from a 2D movie of the flight path of a moth and serves as an example of the analyzed data. β α Figure 1.2: Sample scene illustrating the line parameters to be extracted. The solid line denotes an edge of interest which is visible and orthogonal to the horizon. α is the horizon orientation, β is the angle that the edge of interest makes with the vertical, and and the “X” is the midpoint of the highlighted edge. ii Figure 1.3: Image of the depth buffer before image processing. The image contrast is low. Another option was to determine if the three-dimensional (3D) geometry of the scene could be recon- structed on a per frame basis. This is traditionally a difficult problem but was a viable option due to the setup of the experiments which generated the scenes. In the experiments, which were conducted during a previous study [3], a live moth was fixed in place in a simulator with its wings and abdomen free to move. Descent 3 acted as both the simulator and scene generator. Sensors that monitored the position of the freely moving body parts provided feedback that allowed the moth to control its own visual environment. Descent 3 is a 3D “first-person shooter” (FPS) video game noted for popularizing the use of true 3D rendering technology and providing the player with six full degrees of freedom to move and look around. Two important file types within Descent 3 were used in this research: a demo file and a level file. The demo file records the events that take place: in this case, the movement of the live moth. The level file is used to store the geometry of the game environment in a 3D coordinate system. Given that this information is attainable from the demo and level files, solving the problem by replaying the demos using Descent 3 in order to reconstruct the 3D scene geometry was considered. The following approach to edge and horizon detection was used: Record the geometric information being sent to the graphics processing unit (GPU) from OpenGL. Because Descent 3 uses OpenGL for rendering, it is possible to extract the location of the geometry in image space for each scene. This was achieved through use of the programs GLIntercept and OGLE: The OpenGLExtractor (Described in Chapter 2). iii Figure 1.4: Image of the depth buffer after histogram equalization. Histogram equalization exaggerates the quantization of depth values. iv Chapter 2 Theory The successful completion

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    25 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us