Foreground Segmentation Using Adaptive Mixture Models in Color and Depth

Foreground Segmentation Using Adaptive Mixture Models in Color and Depth

Foreground Segmentation Using Adaptive Mixture Models in Color and Depth Michael Harville Gaile Gordon, John Woodfill Hewlett-Packard Labs Tyzx Inc. 1501 Page Mill Rd., Palo Alto, CA 94304 301 Bryant Street, Palo Alto, CA 94301 [email protected] gaile,[email protected] Abstract tempted to bring computer vision applications to the mar- ketplace. While their methods for background subtraction Segmentation of novel or dynamic objects in a scene, of- may have been adequate for controlled laboratory settings ten referred to as “background subtraction” or “foreground and experimental demonstrations, they often fail catastroph- segmentation”, is a critical early in step in most com- ically when confronted with a variety of real-world phe- puter vision applications in domains such as surveillance nomena that commonly occur in settings such as homes, and human-computer interaction. All previously described, offices, and retail stores. Among the most problematic are: real-time methods fail to handle properly one or more common phenomena, such as global illumination changes, Illumination changes: The background appearance shadows, inter-reflections, similarity of foreground color to changes with the lighting, and such changes can be background, and non-static backgrounds (e.g. active video confused with the introduction of foreground objects. displays or trees waving in the wind). The recent advent of Shadows and inter-reflections: Moving foreground ob- hardware and software for real-time computation of depth jects can create local illumination changes on back- imagery makes better approaches possible. We propose a ground parts of the scene. method for modeling the background that uses per-pixel, Background object changes: When objects are added to time-adaptive, Gaussian mixtures in the combined input or removed from the scene (e.g. when someone picks space of depth and luminance-invariant color. This com- up or leaves a package), or when an object considered bination in itself is novel, but we further improve it by in- background is changed (e.g. when a chair is moved), troducing the ideas of 1) modulating the background model we may want to modify the background model if the learning rate based on scene activity, and 2) making color- change persists for a long enough time. based segmentation criteria dependent on depth observa- Camouflage: Similarity between the appearance of a fore- tions. Our experiments show that the method possesses ground object and the background makes it difficult to much greater robustness to problematic phenomena than distinguish between them. the prior state-of-the-art, without sacrificing real-time per- Non-static background: Objects like changing television formance, making it well-suited for a wide range of practi- displays, foliage in wind, or a fan rotating in front of a cal applications in video event detection and recognition. wall, are not easily modeled by simple pixel statistics. High-traffic areas: If the true background is frequently 1. Introduction obstructed by many different foreground objects, it can be difficult to determine what the true background is Most systems that attempt to detect and recognize events and to segment foreground from it. in live video, and particularly such systems in the areas of surveillance and vision-based human-computer interaction, Some of these problems can be handled by very com- rely heavily on an early step, commonly called “background putationally expensive methods, but in many applications, a subtraction”, that segments the scene into novel (“fore- short processing time is required. For instance, most vision- ground”) and non-novel (“background”) regions. While im- based human-computer interfaces must react in a timely provement of the succeeding analysis steps in these sys- manner to a person’s requests in order to be useful, and tems has been the focus of a great deal of recent work, the most vision-based surveillance systems collect far too much systems’ overall reliability usually depends as much, if not video to analyze off-line at some later time. Background more, on the accuracy and robustness of their background subtraction methods used in such systems, therefore, need subtraction methods. to keep pace with the flow of video data, so that the systems Despite its importance, background subtraction remains maintain some degree of “real-time” responsiveness. a poorly solved problem. This is painfully apparent to the This paper describes an algorithm for foreground seg- increasing number of scientists and engineers who have at- mentation that is the first to exhibit robustness to all of the Copyright 2001 IEEE. Published in the Workshop on Detection and Recognition of Events in Video (EVENT-2001), July 8, 2001, Vancouver, Canada. Personal use of this material is permitted. Permission to reprint/republish for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works, must be obtained from: Manager, Copyrights and Permissions / IEEE Service Center / 445 Hoes Lane / P.O. Box 1331 / Piscataway, NJ 08855-1331, USA. Telephone: + Intl. 732-562-3966. Algorithm Characteristics Example Robustness to Various Phenomena Color Depth Time Luminance Multimodal Low Visual Depth Non-Empty Illumination BG Object Shadows & Color Depth Rotating Active High-Traffic Input Input Adaptive Norm’ed Color BG Stats Texture Discontinuities Scene Init Changes Changes Inter-Reflect Camouflage Camouflage Fan Display Areas √ √√ √ √√ √√√√√ √ √√√ √√√√√√ √ √√ √Pfinder[9] √√√√√ √√ √√ Stauffer et.al.[7] √√√√√ √√ √√ √ Eveland et.al.[1] √√ √√√ √√ √√ √ √ √ Gordon et.al.[3] √√√√√√√√√√√ this method Table 1.√ Summary of robustness of several background subtraction methods to a variety of common problems. Methods given a for a particular problem do not all handle it equally well, but at least contain some feature designed to address it. problems described above without sacrificing real-time per- require an “empty” scene for initialization, and can incorpo- formance. We achieve this in large part by taking advantage rate persistent additions, removals, or alterations of objects of the recent development of real-time depth computation into its background model. Furthermore, if color compar- from stereo cameras [5, 6, 8]. To first-order, our method isons are made in a luminance-normalized space, some ro- can be summarized as an extension of the time-adaptive, bustness to shadows and inter-reflections is obtained. Both per-pixel, Gaussian mixture modeling of color video, as de- of these ideas are incorporated in [9], among other systems. scribed recently by Stauffer and Grimson [7], to the com- By changing the per-pixel models of expected background bined input space of color and depth imagery, as explored features from unimodal probability densities, such as single by Gordon, et. al. [3]. We seek to combine the best as- Gaussians, to more sophisticated representations, such as pects of both methods, and we also introduce interesting Gaussian mixture models, as in [2, 7], one can begin to ad- new concepts to further improve the method’s performance. dress complex, non-static backgrounds. All of these meth- The resulting robustness is significantly greater than that of ods still have great difficulty with color camouflage, active prior algorithms, and yet, provided that real-time depth im- video displays, and high-traffic areas, among other things. agery is available1, our method can still run in real-time on Others have taken advantage of the recent development standard hardware. of real-time depth computation from stereo cameras. Be- cause the shapes of scene objects are not affected by illu- mination, shadows, or inter-reflections, depth information 2. Previous work provides much robustness to such phenomena. Depth data Many approaches to background subtraction have been is unreliable, however, in scene locations with little visual proposed over the last few decades. The simplest class of texture or that are not visible to all cameras, and tends to be methods, often found in experimental computer vision sys- noisy even where it is available. Hence, background sub- tems, use color or intensity as input, and employ a back- traction methods based on depth alone [1, 4] produce un- ground model that represents the expected background fea- reliable answers in substantial portions of the scene, and ture values at each pixel as an independent (in image space), often fail to find foreground objects in close proximity to static (in time), unimodal (i.e. single-peaked, in feature the background, such as hands placed on walls or feet on a space) distribution. For example, the background statistics floor. A method using both depth and color has been pro- are often modeled at each pixel using a single Gaussian, posed recently [3], but it lacks time adaptivity and uses a stored as a color mean and covariance matrix. The model relatively simple statistical model of the background. is built during a “learning” phase while the scene is known By combining the best of all of these ideas, our algorithm to be empty of foreground objects, and is never modified is able to perform reasonably well under all of the condi- thereafter. Typically, foreground is detected on a per-pixel tions listed in section 1. Table 1 summarizes the charac- basis wherever the current input image differs significantly teristics and capabilities of the algorithms described above, from the distribution of expected background values. While and compares them against our method. Note that the union such methods are well-suited for fast implementation and of the rows for methods [7] and [3] produces the row cor- run time, they fail when confronted with any of the com- responding to our method. However, due to further innova- mon, real-world phenomena listed in section 1. tions that we introduce, such as activity-based modulation Many methods have attempted to address some, but not of model adaptation, we believe that our method handles all, of these problems, by improving one or more aspects several problem cases significantly better than [7], [3], and of the basic class of methods.

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    9 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us