Technical Program

IS&T/SPIE 19th Annual Symposium

28 January–1 February 2007 San Jose Marriott and San Jose Convention Center San Jose, California USA

Sponsored by: Technical Program

IS&T/SPIE 19th Annual Symposium

Conferences • Continuing Education • Technical Exhibitions • Demonstrations 28 January–1 February 2007 San Jose Marriott and San Jose Convention Center • San Jose, California USA

Welcome to San Jose!

On behalf of IS&T—The Society for Imaging Science and Technology and SPIE—The International Society for Optical Engineering, welcome to the Electronic Imaging 2007 Symposium. Imaging is pervasive in the human experience, be it prehistoric cave drawings, a wall of hieroglyphics in a pyramid, a painting by Monet, a photograph by Brady, watching Casablanca on a digital HDTV, a broken arm revealed by a conventional X-ray, the activity of the brain displayed by a Functional MRI, the animation of a game on X-Box, the 3D image on a cell phone, scientific visualization of a turbulence model, a live view of the surface of Mars, or a digital photograph of friends and loved ones. Images are as central to our everyday life as they are to exploration, medical advancement, entertainment, science, and national security. EI 2007 is the one international conference where papers on all aspects of electronic imaging are presented, and where you can develop both your career and business opportunities by networking with leading researchers and entrepreneurs in the field. While you are here: • EI 2007 is the best place to share your work in electronic imaging by presenting an oral or poster paper session or by showcasing your technology at a special demonstration session. • Learn about leading edge technology and science across a broad range of imaging disciplines by attending the various conferences. • Gain exposure to the latest “hot topics” in the field. • Develop a deep understanding of electronic imaging technology by taking one or more of the Short Course being in their respective fields. • Network with fellow scientists, engineers, mangers, and entrepreneurs to gain breath of knowledge and advance your career. IS&T/SPIE would like to express • Participate in panel sessions that discuss the current and future states of electronic their deepest appreciation to the imaging technologies and products. symposium chairs, conference • Attend the formal and informal receptions and coffee breaks and develop life-long chairs, and program committees relationships with other experts in industry, government, and academia. who have so generously given of • Gain from the insight of recognized giants in the electronic imaging field by attending their time and advice to make this the plenary sessions. symposium possible. The • Enjoy the inviting business, cultural, and entertainment offerings found in Silicon symposium, like our other Valley. conferences and activities, would We look forward to seeing you and sharing with you the full spectrum of science and not be possible without the technology that will shape our future. dedicated contribution of our participants and members. Symposium Chairs: Michael A. Kriss, Robert A. Sprague, Consultant Consultant Contents Conference Index ...... 3 Conference Daily Schedule ...... 4 Course Daily Schedule ...... 5 Plenaries ...... 6-7 Special Events ...... 8 General Information ...... 9

Technical Conferences: 3D Imaging, Interaction, and Measurement ...... 10-15 Imaging, Visualization, and Perception ...... 16-25 Image Processing ...... 26-35 Digital Imaging, Sensors, and Applications ...... 36-41 Multimedia Processing and Applications ...... 42-50 Visual Communications and Image Processing ...... 51

Technical Abstract Summaries . 56-213

Participants ...... 214-222 Publication Order Form ...... 223 Proceedings ...... 224

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 1 Symposium Chairs Michael A. Kriss, Consultant Technical Organizing Committee Robert A. Sprague, Consultant Jaakko T. Astola, Tampere Univ. of Technology (Finland) J.-Angelo Beraldin, National Research Council Canada Symposium Organizing Committee: Morley Blouke, Ball Aerospace & Technologies Corp. Charles A. Bouman, Purdue Univ. Mark T. Bolas, Fakespace Labs., Inc. Gabriel G. Marcu, Apple Computer, Inc. Katy Börner, Indiana Univ. Charles A. Bouman, Purdue Univ. Short Course Chairs Jianfei Cai, Nanyang Technological Univ. (Singapore) Edward J. Delp III, Purdue Univ. Matthias F. Carlsohn, Computer Vision and Image Michael A. Kriss, Consultant Communication (Germany) Berna Erol, Ricoh Innovations, Inc. Chang Wen Chen, Florida Institute of Technology Reiner Creutzburg, Fachhochschule Brandenburg (Germany) Exhibition Chair Luke C. Cui, Lexmark International, Inc. Scott J. Daly, Sharp Labs. of America, Inc. A. Ufuk Agar, Oger Telecom (Turkey) Edward J. Delp, Purdue Univ. Demonstration Chair Jeffrey M. DiCarlo, Hewlett-Packard Labs. Neil A. Dodgson, Univ. of Cambridge (United Kingdom) Neil A. Dodgson, Univ. of Cambridge (United Edward R. Dougherty, Texas A&M Univ. Kingdom) Karen O. Egiazarian, Tampere Univ. of Technology (Finland) Robert F. Erbacher, Utah State Univ. Reiner Eschbach, Xerox Corp. Carsten Griwodz, Univ. of Oslo (Norway) Matti T. Gröhn, Ctr. for Scientific Computing (Finland) Alan Hanjalic, Technische Univ. Delft (Netherlands) Nasser Kehtarnavaz, The Univ. of Texas at Dallas Longin Jan Latecki, Temple Univ. Xiaofan Lin, Riya, Inc. Jiebo Luo, Eastman Kodak Co. Gabriel G. Marcu, Apple Computer, Inc. Russel A. Martin, Foveon, Inc. Ian E. McDowall, Fakespace Labs., Inc. Fabrice Meriaudeau, Univ. de Bourgogne (France) John O. Merritt, The Merritt Group Eric L. Miller, Northeastern Univ. Yoichi Miyake, Chiba Univ. Japan David M. Mount, Univ. of Maryland/College Park Kurt S. Niel, Fachhochschule Wels (Austria) Thrasyvoulos N. Pappas, Northwestern Univ. Ilya Pollak, Purdue Univ. Fabio Remondino, ETH Zürich (Switzerland) Jonathan C. Roberts, Univ. of Kent (United Kingdom) Bernice E. Rogowitz, IBM Corp. Nitin Sampat, Rochester Institute of Technology Raimondo Schettini, /Univ. degli Studi di Milano- Bicocca (Italy) Dan Schonfeld, Univ. of Illinois at Chicago Nicu Sebe, Univ. van Amsterdam (Netherlands) Mark R. Shortis, RMIT Univ. (Australia) Jarmo Takala, Tampere Univ. of Technology (Finland) Ping Wah Wong, Nethra Imaging Andrew J. Woods, Curtin Univ. of Technology (Australia) Angela Y. Wu, American Univ. Berrin A. Yanikoglu, Sabanci Univ. (Turkey) Roger Zimmermann, Univ. of Southern California

2 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Technical Conference Index

3D Imaging, Interaction, and Multimedia Processing and Applications Measurement Conf. 6504 Multimedia Computing and Networking 2007 (Zimmermann, Griwodz) ...... p. 42 Conf. 6490A Stereoscopic Displays and Applications XVIII (Woods, Dodgson, Merritt) ...... p. 10 Conf. 6505 Security, Steganography, and Watermarking of Multimedia Contents IX (Delp, Wong) ...... p. 44 Conf. 6490B The Engineering Reality of Virtual Reality 2007 (Bolas, McDowall) ...... p. 13 Conf. 6506 Multimedia Content Access: Algorithms and Systems (Hanjalic, Schettini, Sebe) ...... p. 47 Conf. 6491 Videometrics IX (Beraldin, Remondino, Shortis)p. 14 Conf. 6507 Multimedia on Mobile Devices 2007 (Creutzburg, Imaging, Visualization, and Perception Takala, Cai) ...... p. 49 Conf. 6492 Human Vision and Electronic Imaging XII Visual Communications and Image (Rogowitz, Pappas, Daly) ...... p. 16 Processing Conf. 6493 Color Imaging XII: Processing, Hardcopy, and Applications (Eschbach, Marcu) ...... p. 19 Conf. 6508 Visual Communications and Image Processing Conf. 6494 Image Quality and System Performance IV 2007 (Chen, Schonfeld, Luo) ...... p. 51 (Cui, Miyake) ...... p. 22 Conf. 6495 Visualization and Data Analysis 2007 (Erbacher, Roberts, Gröhn, Börner) ...... p. 24 Image Processing Conf. 6496 Real-Time Image Processing 2007 (Kehtarnavaz, Carlsohn) ...... p. 26 Conf. 6497 Image Processing: Algorithms and Systems VI (Astola, Egiazarian, Dougherty) ...... p. 28 Conf. 6498 Computational Imaging V (Bouman, Miller, Pollak) ...... p. 30 Conf. 6499 Vision Geometry XV (Latecki, Mount, Wu) . . . p. 33 Conf. 6500 Document Recognition and Retrieval XIV (Lin, Yanikoglu) ...... p. 34 Digital Imaging Sensors and Applications Conf. 6501 Sensors, Cameras, and Systems for Scientific/ Industrial Applications X (Blouke) ...... p. 32 Conf. 6502 Digital Photography III (Martin, DiCarlo, Sampat) ...... p. 38 Conf. 6503 Machine Vision Applications in Industrial Inspection XV (Meriaudeau, Niel) ...... p. 40

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 3 Conference Daily Schedule

Sunday Monday Tuesday Wednesday Thursday 3D Imaging, Interaction, and Measurement

Conf. 6490A Stereoscopic Displays and Applications XVIII (Woods, Dodgson, Merritt), p. 10 Conf. 6490B The Engineering Reality of Virtual Reality 2007 (Bolas, McDowall), p. 13 Conf. 6491 Videometrics IX (Beraldin, Remondino, Shortis), p.14

Imaging, Visualization, and Perception

Conf. 6492 Human Vision and Electronic Imaging XII (Rogowitz, Pappas, Daly), p. 16 Conf. 6493 Color Imaging XII: Processing, Hardcopy, and Applications (Eschbach, Marcu), p. 19 Conf. 6494 Image Quality and System Performance IV (Cui, Miyake), p. 22

Conf. 6495 Visualization and Data Analysis 2007 (Erbacher, Roberts, Gröhn, Börner), p. 24

Image Processing

Conf. 6496 Real-Time Image Processing IV (Kehtarnavaz, Carlsohn), p. 26 Conf. 6497 Image Processing: Algorithms and Systems VI (Astola, Egiazarian, Dougherty), p. 28 Conf. 6499 Vision Geometry XV (Latecki, Mount, Wu), p. 33 Conf. 6498 Computational Imaging V (Bouman, Miller, Pollak), p. 230 Conf. 6500 Document Recognition and Retrieval XIV (Lin, Yanikoglu), p. 37

Digital Imaging Sensors and Applications

Conf. 6502 Digital Photography III (Martin, DiCarlo, Conf. 6501 Sensors, Cameras, and Systems for Scientific/ Sampat), p. 38 Industrial Applications X (Blouke), p. 36 Conf. 6503 Machine Vision Applications in Industrial Inspection XV (Meriaudeau, Niel), p. 40

Multimedia Processing and Applications

Conf. 6504 Multimedia Computing and Networking 2007 (Zimmermann, Griwodz), p. 42 Conf. 6505 Security, Steganography, and Watermarking of Multimedia Contents IX (Delp, Wong), p. 44 Conf. 6507 Multimedia on Mobile Devices 2007 (Creutzburg, Conf. 6506 Multimedia Content Access: Algorithms and Takala, Cai), p. 49 Systems (Hanjalic, Schettini, Sebe), p. 47

Visual Communications and Image Processing

Conf. 6508 Visual Communications and Image Processing 2007 (Chen, Schonfeld, Luo), p. 51

4 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Course Daily Schedule

Sunday Monday Tuesday Wednesday Thursday Color and Perception

SC516 Color Considerations SC811 Color Appearance for Liquid Crystal Displays Modeling Using CIECAM02 (Marcu) 8:30 am to 12:30 pm, (Moroney) 1:30 to 3:30 pm, $280 / $325 $125 / $175 SC060 Stereoscopic Display Application Issues (Merritt, Woods) 8:30 am to 5:30 pm, $460 / $545 SC075 Effective Color Students: Save 50% on Computing (Marcu) 1:30 to course registrations! 5:30 pm, $280 / $325 SC754 Human Shape Perception (Pizlo, Latecki) 1:30 to 5:30 pm, $280 / $325 SC812 Perceptual Metrics for Image Quality Evaluation (Pappas, Hemami) 1:30 to 5:30 pm, $280 / $325 Digital Imaging

SC504 Introduction to CCD SC807 Digital Camera and SC762 Device Simulation for and CMOS Imaging Sensors Scanner Performance Image Quality Evaluation and Applications (Janesick) Metrology: Science, (Farrell) 8:30 am to 12:30 pm, 8:30 am to 5:30 pm, $535 / Standards, and Software $280 / $325 $620 (Burns, Williams) 8:30 am to 5:30 pm, $460 / $545 SC753 The Image Pipeline and How It Influences SC810 Digital Still Camera Quality Measurements Imaging System: The Based on Existing ISO Underlying Technologies Standards (Wueller, (Kriss) 8:30 am to 5:30 pm, Matherson) 8:30 am to 5:30 $460 / $545 pm, $460 / $545 Electronic Imaging Applications

SC764 Filter Banks and SC808 Document Analysis Wavelets: Design and Use in on Mobile Devices Perceptual Coding (Schuller) (Doermann) 8:30 am to 12:30 8:30 am to 12:30 pm, $280 / pm, $280 / $325 $325 SC767 Practical SC816 Multimedia and Implementations of Machine Security: An Introduction to Vision Systems within Cryptography, Digital Technical Processes (Niel) Watermarking, Media 8:30 am to 12:30 pm, $280 / Forensics, and Biometrics $325 and How Things go Together (Dittmann, Vielhauer) 8:30 am to 5:30 pm, $460 / $545 SC809 Real-Time Image and Video Processing (Kehtarnavaz, Gamadia) 8:30 am to 12:30 pm, $280 / $325 Electronic Imaging Technology

SC468 Advanced Image SC766 Information Processing SC491 Neural Networks Processing (Rabbani) 8:30 am for Video Surveillance (Ebrahimi, Applications in Image to 5:30 pm, $460 / $545 Dufaux) 8:30 am to 5:30 pm, $460 / Processing (Nasrabadi) 8:30 $545 am to 5:30 pm, $460 / $545 SC814 Image Processing, Understanding, and the SC813 MPEG Family of Video Analysis of Master Drawings Compression Standards and Paintings (Stork) 8:30 am (Rabbani) 8:30 am to 5:30 pm, to 12:30 pm, $280 / $325 $460 / $545 SC804 Introduction to SC805 Principles of Multivariate Statistics for Imaging Statistics for Imaging (Bajorski) (Bajorski) 1:30 to 5:30 pm, 8:30 am to 12:30 pm, $280 / $325 $280 / $325 SC806 Advanced Multivariate Statistics for Imaging (Bajorski) 1:30 to 5:30 pm, $280 / $325 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 5 Awards and Plenary Presentation

Marriott San Jose Ballroom 1-6 Tuesday 30 January ...... 8:15 to 9:15 am Development and Plasticity in Visual Cortex

Brian A. Wandell, Stanford Univ. Brian Wandell’s teaching at Stanford reflects his multiple areas of expertise. He has taught courses on behavior, perception, cognitive and Visual cortex has been an excellent model system behavioral neuroscience, image systems and computational for developing a quantitative understanding of brain neuroimaging. He has also led classes on color science and computer function. We understand a great deal about the applications for engineers and managers from more than 200 companies. physical signals that initiate vision, and this In addition to numerous scientific articles, Brian Wandell is the author of knowledge has led to a relatively advanced the vision science textbook Foundations of Vision. He is an associate understanding of the organization of major structures editor of the Journal of Vision, the Journal of Neuroscience and Neural in visual cortex, such as visual field maps. This talk Networks. He has served as a consultant and technical advisor for a will explain several measurements and number of corporations and has patented some of the products of his computational methods that are used to understand human brain work. development and plasticity. In 1986, he won the Troland Research Award from the National First, we have developed functional magnetic resonance imaging (fMRI) Academy of Sciences for his work in color vision. He was made a fellow methods for measuring and quantifying the properties of maps in of the Optical Society of America in 1990 and, in 1997, he became a individual human and macaque brains. To understand the development McKnight Senior Investigator and received the Edridge-Green Medal in and plasticity of these maps, we have made measurements in several Ophthalmology for work in visual neuroscience. In 2000, he was awarded cases of abnormal development as well as in controlled experiments the Macbeth Prize from the Inter-Society Color Council. Wandell was using macaque. elected to the US National Academy of Sciences in 2003. Second, we are combining fMRI with diffusion tensor imaging (DTI), a method that can be used to study the white matter fibers, to understand visual development. Specifically, as children develop and learn to read certain visual recognition skills become highly automatized and the brain develops specialized visual circuitry to support skilled reading. We are measuring how certain parts of these circuits develop, and how the signals from these circuits are transmitted to other cortical systems. Brian Wandell is the first Isaac and Madeline Stein Family Professor at Stanford University. He graduated from the University of Michigan in 1973 with a B.S. in mathematics and psychology. In 1977, he earned a Ph.D. in social science from the University of California at Irvine. After a year as a postdoctoral fellow at the University of Pennsylvania, he joined the faculty of Stanford University in 1979. Professor Wandell was promoted to associate professor with tenure in 1984 and became a full professor in 1988. Brian Wandell’s research includes image system engineering and visual neuroscience. In cooperation with Professor Emeritus Joseph Goodman (a faculty member in Stanford’s School of Engineering), Professor Wandell founded the university’s Image Systems Engineering Program. As part of this research, Wandell and his team study and build devices used for digital imaging, including image sensors, high dynamic range displays, and software simulations of the digital imaging pipeline. Professor Wandell’s work in visual neuroscience uses both functional MRI and behavior testing to understand the action of the visual portions of the brain. His team has developed a set of methods for identifying and measuring the signals in several visual areas within the human brain, including regions that respond powerful to motion and color. Recently, his team measured the reorganization of brain function during human development and following brain injury; they are now actively studying the development of visual signals during the age period in which children are learning to read, as well as changes to cortex caused by retinal dysfunction.

6 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Awards and Plenary Session

Marriott San Jose Ballroom 1-6 In 2004, he led the development of portable 3D color scanners and the scanning in Paris and the 3D modeling of famous paintings by Renoir Wednesday 31 January ...... 8:00 to 9:20 am and the Mona Lisa by Leonardo Da Vinci. In September 2005, he was recognized among the top 10 Canadian Innovators by the Globe and 8:00 to 8:40 am: Mail, one of the most important newspapers in Canada. He also received several awards from NRC, and from the Association de la recherche More than a Poplar Plank: The Shape industrielle du Québec (ADRIQ). and Subtle Colors of the Masterpiece His research interests cover various fields in real-time digital and image processing, optics and electro-optics, control, 3D, sensor and vision ‘Mona Lisa’ by Leonardo systems and their applications. He is very active in the field of 3D and François Blais, John Taylor, Luc Cournoyer, the scientific community, and has given many tutorials and conference Michel Picard, Louis Borgeat, Guy Godin, J.- papers on active 3D vision and applications. He has more than 100 Angelo Beraldin, Marc Rioux, National Research publications and 12 patents, of which several have been licensed to Council Canada; Christian Lahanier, Bruno Mottin, Canadian Industry. Ctr. de Recherche et de Restauration des Musées de France 8:40 to 9:20 am: François Blais, National Research Council of Canada Space Exploration, Imaging Technology,

During the autumn of 2004, a series of scientific and the Development of a Virtual examinations of the Mona Lisa were undertaken by the Centre de Presence in Space Recherche et de Restauration des Musées de France (C2RMF) and the Musée du Louvre to help assess and monitor the condition of the panel Eric M. De Jong, Principal Investigator, Solar painting. Never before a piece of art underwent such a massive scientific System Visualization (SSV) Project examination including grazing light, x-radiography, ultraviolet, infrared, microfluorescence, 2D multispectral imaging and a world first, the Understanding our place in the Universe is one of complete 3D digitization of the painting and the creation of a high mankind’s greatest scientific and technological resolution virtualized 3D model of the Mona Lisa. challenges and achievements. The invention of the For this task, a team of 3D imaging scientists from the National telescope, the Copernican Revolution, the Research Council of Canada (NRC) were invited to Paris from 17 to 22 development of Newtonian mechanics, and the October, 2004 to undertake the 3D scanning of the painting. The objective Space Age exploration of our solar system; provided of this project was to scan the Mona Lisa - obverse and reverse - in us with a deeper understanding of our place in the order to provide high-resolution 3D image data of the complete painting Universe; based on better observations and models. As we approach to (1) record the overall shape of the poplar panel, (2) document surface the end of the first decade of the new millennium, the same quest, to features, wood grain structure, edge features and surface lacunae and understand our place in the Universe, remains a great challenge. New (3) provide high resolution pictorial layer images to assist in studies related technologies will enable us to construct and interact with a “Virtual to the artists’ technique as well as for conservation examination. Universe” based on remote and in situ observations of other worlds. As The structure and technique used to paint the Mona Lisa presents a we continue the exploration that began in the last century, electronic unique research and development challenge for 3D imaging unlike any imaging, information, and display technology will enable us to experience other painting scanned to date. The pronounced convex curvature of a “Virtual Presence in Space.” the overall panel shape due to warping, the 3D surface relief shape detail This talk describes the history of Space Exploration, the development on the pictorial layer of the painting due to the wood grain pattern, of imaging technology, and plans to establish a Virtual Presence in Space features such as the split in the panel from the top edge to the head, (VPS). Deep Space observations and frames from High Definition details of the craquelure patterns and previous lacunae and restoration Television (HDTV) mission animations are used to illustrate mission areas, the extremely thin or flat structure of the paint layer composition challenges and science discoveries. Eric M. De Jong and the application of multiple very thin semi-transparent layers or glazes is a Planetary Scientist working in the Earth and using Leonardo’s sfumato technique, are all documented in the color Space Sciences Division of the NASA Jet Propulsion Laboratory, Chief 3D images. Scientist of the Space Science Data Systems Section, and a Visiting With a lateral resolution of 0.05 mm and a depth precision of 0.01 Associate in Planetary Science at Caltech. For the last three decades mm, the 3D images allow a close virtual examination of the Mona Lisa his research has focused on the scientific visualization of planetary by different specialists and by the non-specialist alike, without any risk surfaces and atmospheres, and the evolution and dynamics of planetary of damage to the painting. This presentation will introduce the 3D systems. As the Principal Investigator and Project Scientist for NASA’s technology and will focus on the results obtained from the virtualized Solar System Visualization (SSV) Project, Eric leads a team of high-resolution 3D color model of Leonardo’s famous work, thousands technologists and scientists responsible for developing new science of kilometers away from Paris. algorithms, visualization technologies, images, mosaics, and animations. François Blais is a Principal Research Officer at the National Research The SSV team has pioneered the use of HDTV and IMAX film technologies Council of Canada. He obtained his B.Sc. and M.Sc. in Electrical for science visualizations. Eric and his team, in collaboration with other Engineering from Laval University, Quebec City. Since 1984, his research science investigators, have created planetary image sequences for the in 3D has resulted in the development of the prototypes for the recognized IMAX films “Blue Planet”, “Journey to the Planets”, “Destiny in Space”, NRC 3D range-sensing systems, an important contribution to NRC’s “The Discoverers”, and “L5: First City in Space”, and “Roving Mars”. having become one of the leading research organizations in 3D. Several of the technologies generated from his research have been licensed to industry and are now being commercialized in various industrial sectors: automotive, electronics, forestry, mining, biometrics, and heritage. In 1997, he joined Vitana Corp. as Directory of R&D, before returning to NRC to continue his research and supervise the technology transfer to Neptec that resulted in the 3D laser scanner technology currently used in space onboard NASA’s Space Shuttle Discovery in 2001 and 2005 and more recently in July 2006.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 7 Special Events

Electronic Imaging Technical Meeting Symposium Demonstration Session Marriott Willow Glen Room San Jose Convention Center, Exhibit Hall 1 Tuesday 30 January ...... 5:30 to 8:30 pm Monday 29 January ...... 7:30 to 9:30 pm Conference Chair: Neil A. Dodgson, Univ. of Cambridge (United Chair: Gabriel Marcu, Apple Computer, Inc. Kingdom) This meeting addresses diverse research, engineering, and specialized applications of electronic imaging devices or systems. The highly-successful, interactive, hands-on demonstration of Because of the diverse topical areas within electronic imaging, the stereoscopic hardware, software, and display—traditionally a component of the Stereoscopic Display and Applications Conference—will again technical meeting covers image processing, image capture, display be expanded to include research demonstrations and products related and hardcopy, system integration and visualization. Application areas to the entire Electronic Imaging Symposium. are just as far-reaching. They include industrial automation, graphic The annual demonstration, which has previously showcased the largest arts, aerospace sensing, remote sensing, document processing, and most diverse collection of stereoscopic research and products in high-resolution television, medical imaging, and all areas of digital one location, is now open to all EI 2007 symposium attendees and image processing, including analysis, compression and restoration. commercial entities working in the realm of electronic imaging! Attendees are strongly encouraged to propose topics of interest for The EI 2007 Demonstration Session represents a unique networking the next meeting and to submit short articles for publication. All opportunity, a time when attendees can see the latest research in action, attendees are invited to attend this informative meeting that provides compare commercial products, ask questions of technically excellent networking opportunities. knowledgeable demonstrators, and even make purchasing decisions about a range of EI products. The Demonstration Session is scheduled to take place over a three hour period in conjunction with the Symposium Symposium Poster Session Poster Session. San Jose Convention Center, Exhibit Hall 1 All-Conference Reception Tuesday 30 January ...... 5:30 to 7:00 pm Marriott, San Jose Ballrooms 4-6 Conference attendees are invited to the poster session. Authors of Wednesday 31 January ...... 7:30 pm to 9:30 pm poster papers will be on hand during this session to answer questions and provide in-depth discussion concerning their papers. Attendees Plan to join us for this great opportunity to get to know your Electronic are requested to wear their conference registration badges to the poster Imaging colleagues. All attendees are invited to relax, and enjoy a session. pleasant evening with friends old and new! Authors can set up posters after 10:00 am on Tuesday. Poster supplies (pushpins) will be available. Other supplies can be obtained from the Speakers’ Audio Visual Desk. Exhibition and Bookfair Posters can be previewed during the day of the event before the San Jose Convention Center, Concourse 1 Lobby formal poster session begins at 5:30 pm. Tuesday 30 January ...... 10:00 am to 5:00 pm Authors must remove their papers at the conclusion of the poster reception. It is the author’s responsibility to remove their posters Wednesday 31 January ...... 10:00 am to 4:00 pm immediately after the session. Papers not removed will be considered An intimate exhibit and bookfair features select Electronic Imaging unwanted and will be discarded. The Societies assume no responsibility companies showcasing their latest products and technologies. for posters left up after the end of the poster reception.

8 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] General Information

Electronic Imaging 2007 Copy & Computer Services San Jose Convention Center Marriott San Jose Hotel - Business Center. Self-service computers/ 408 S. Almaden Boulevard, San Jose, CA 95110 printers, fax, copiers and internet access is available on 24 hr basis. San Jose Marriott Hotel Access is by your sleeping room key. All Marriott guest rooms include 301 S. Market Street, San Jose, CA 95113 T-1 internet connections. San Jose Convention Center - Business Center at the administration office. Open 8:00 am to 5:00 pm daily. Services available include Registration and Information Hours computers/printers, fax, and copiers. San Jose Convention Center, Concourse 1 Lobby Courses Only: SPIE Bookstore and Membership Booth Sunday 28 January ...... 7:00 am to 10:00 am Monday through Thursday, Open during registration hours. Conference Registration: SPIE publishes a variety of technical books designed to meet diverse Sunday 28 January ...... 10:00 am to 4:00 pm research, reference, and educational needs. Proceedings of SPIE technical conferences from this and related meetings may be purchased Monday-Wednesday 29-31 January ...... 7:00 am to 4:00 pm at the bookstore. Also available are related books in the SPIE PRESS Thursday 1 February ...... 7:00 am to Noon Series, including Tutorial Texts, Milestone Series of Selected Reprints, Critical Reviews in Science & Technology, and Monographs & Speakers Audiovisual Desk Hours Handbooks. San Jose Convention Center, Room E IS&T Bookstore and Membership Booth Monday-Thursday 29 January-1 February ...... 7:30 am to 4:30 pm Monday through Thursday, Open during registration hours. Speakers who have requested to use LCD projection from their laptop, 35mm slide projection, a VHS video player, or an overhead projector are IS&T publishes a variety of books to meet your needs. Proceedings of encouraged to preview their materials at the Audiovisual Desk prior to their past IS&T conferences including Digital Printing Technologies, Color presentation. Speakers who have requested special equipment beyond Imaging as well as Recent Progress series books will be available. IS&T the standard equipment noted here are asked to report to the EI ’07 also distributes selected titles from cooperating publishers of science Audiovisual Desk upon arrival at the meeting to confirm equipment requests. and technology books in the imaging field. Information about upcoming Speakers will be responsible for delivering visual materials to the conference IS&T conferences and IS&T membership benefits, sample journals, and room and may retrieve their presentation materials from the room monitor newsletters are available. in the conference room immediately following the session. Cash Cart: Breakfast Breads, Snacks Short Course Notes and Quick Lunch San Jose Convention Center, Concourse 1 Lobby Short courses will take place in various meeting rooms at the San Jose Marriott Hotel and the San Jose Convention Center. Your room Monday-Thursday 29 January-1 February ...... 7:30 am to 2:30 pm assignment will be given to you with a ticket in your registration materials. The Cash Cart will offer breakfast breads, yogurt, fruit, coffee, juice and Registrants for short courses must exchange each course ticket received other beverages each morning of the conference. Luncheon and snack for their course notes in the course meeting room at the beginning of service will include deli-style sandwiches, salads, snacks and pastries, your class. and beverage. Attendees will need to make their own breakfast arrangements for Monday. Video/Digital Recording Policy For copyright reasons, video or digital recording of any conference Child Care session, short course, or poster session is strictly prohibited without written prior consent from each specific presenter to be recorded. A few child sitting services available in San Jose are as follows. Individuals not complying with this policy will be asked to leave a given 1. Bay Area 2nd MOM Inc. session and to surrender their film or disc. It is the responsibility of the Hotel Nanny Service presenter to notify the conference sponsors if such consent is given. Toll Free Phone: 1-888-926-3666, or (650) 858-2469, ext. 109., Monday-Friday 9:00 am to 5:00 pm. At other times phone (650) 858- 4984; Fax: (650) 493-6598; E-mail: [email protected] or Messages for Attendees [email protected]; Website: www.2ndmom.com Messages for attendees at Electronic Imaging 2007 Symposium can be 2. Sitters Unlimited left by calling the IS&T/SPIE Message Center at 408-271-6100. Messages Phone: (408) 452-0225; E-mail: [email protected] will be taken during registration hours Sunday through Thursday. or www.sittersunlimited.com Attendees should check the message boards at the message center Note: IS&T/SPIE does not imply an endorsement or recommendation of daily to receive their messages. these services. They are provided on an “information-only” basis for your further analysis and decision. Other services may be available. Complimentary Wireless High-Speed rail and regular service buses for a one-day period and can be purchased Internet from your bus operator or light rail ticket vending machines. IS&T/SPIE are pleased to provide complimentary wireless access to the Internet for all conference attendees with wireless-enabled laptop Car Rental computers or PDAs. Hertz Car Rental has been selected as the official car rental agency for this Symposium. To reserve a car, identify yourself as an Electronic Imaging Conference attendee using the Hertz Meeting Code CV# 029B0010. Call 1-800-654-2240.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 9 Conference 6490A • Conv. Ctr. Room A8

Monday-Wednesday 29-31 January 2007 • Part of Proceedings of SPIE Vol. 6490 Stereoscopic Displays and Virtual Reality Systems XIV Stereoscopic Displays and Applications XVIII

Conference Chairs: Andrew J. Woods, Curtin Univ. of Technology (Australia); Neil A. Dodgson, Univ. of Cambridge (United Kingdom); John O. Merritt, The Merritt Group Program Committee: Gregg E. Favalora, Actuality Systems, Inc.; Nicolas S. Holliman, Univ. of Durham (United Kingdom); Janusz Konrad, Boston Univ.; Shojiro Nagata, InterVision (Japan); Steven L. Smith, VRex, Inc.; Vivian K. Walworth, Jasper Associates; Michael A. Weissman, TrueVision Systems, Inc.

Cosponsors:

Monday 29 January SESSION 3 Conv. Ctr. Room A8 ...... Mon. 1:30 to 3:10 pm SESSION 1 Stereoscopic Image Processing and Intermediate View Conv. Ctr. Room A8 ...... Mon. 8:30 to 10:10 am Reconstruction Stereoscopic Display Applications Chair: Janusz Konrad, Boston Univ. Chair: Andrew J. Woods, Curtin Univ. of Technology (Australia) 1:30 pm: A novel triangulation method for building parallel-perspective stereo mosaics, P. K. Gurram, E. Saber, H. Rhody, Rochester Institute of 8:30 am: Stereoscopic applications for design visualization, K. J. Gilson, Technology ...... [6490A-10] Parsons Brinckerhoff ...... [6490A-01] 1:50 pm: Surrogate depth maps for stereoscopic imaging: different edge 8:50 am: A novel technique for visualizing high-resolution 3D terrain maps, types, W. J. Tam, C. A. Vázquez, F. Speranza, Communications Research Ctr. J. F. Dammann, Jr., Army Research Lab...... [6490A-68] Canada (Canada) ...... [6490A-11] 9:10 am: Evolution of the Varrier™ autostereoscopic VR display: 2001-2007, 2:10 pm: Object-based disparity adjusting tool for stereo panorama, C. T. Peterka, R. L. Kooima, J. I. Girado, J. Ge, D. J. Sandin, T. A. DeFanti, Univ. of Wang, A. A. Sawchuk, Univ. of Southern California ...... [6490A-12] Illinois/Chicago ...... [6490A-03] 2:30 pm: View generation for 3D-TV using image reconstruction from 9:30 am: Re-inventing the stereoscope for the 21st Century, J. S. Kollin, irregularly spaced samples, C. A. Vazquez, Communications Research Ctr. Univ. of Washington; A. J. Hollander, Imprint Interactive Technology [6490A-04] Canada (Canada) ...... [6490A-13] 9:50 am: A cohesive system for real-time stereoscopic image processing 2:50 pm: Spline-based intermediate view reconstruction, S. Ince, J. Konrad, and evaluation, E. M. Lazarus, Bristlecone Corp.; R. Galli, ANDXOR Boston Univ.; C. Vázquez, Institut National de la Recherche Scientifique Corp...... [6490A-05] (Canada) ...... [6490A-14] Coffee Break ...... 10:10 to 10:40 am Coffee Break ...... 3:10 to 3:40 pm

SESSION 2 SESSION 4 Conv. Ctr. Room A8 ...... Mon. 10:40 am to 12:00 pm Conv. Ctr. Room A8 ...... Mon. 3:40 to 5:00 pm Medical Applications of Stereoscopy Stereoscopic Human Factors Michael A. Weissman Chair: , TrueVision Systems, Inc. Chair: John O. Merritt, The Merritt Group 10:40 am: Stereoscopic medical imaging collaboration system, F. Okuyama, 3:40 pm: Vertical surround correction for stereoscopic projection, L. Lipton, Suzuka Univ. of Medical Science (Japan) ...... [6490A-06] Real D ...... [6490A-15] 11:00 am: Using a high-definition stereoscopic video system to teach 4:00 pm: An application driven comparison of depth perception on desktop microscopic surgery, J. F. R. Ilgner, J. J. H. Park, D. Labbe, Univ. Hospital 3D displays, N. S. Holliman, B. Froner, Univ. of Durham (United Kingdom); S. Aachen (Germany) ...... [6490A-07] Liversedge, Univ. of Southampton (United Kingdom) ...... [6490A-16] 11:20 am: Stereoscopic video system provides real-time registration for 4:20 pm: Visual discomfort in stereoscopic and autostereoscopic displays: endoscopic surgery, A. Yaron, N. Horesh, M. Shechterman, Visionsense Ltd. a review of concepts, measurement methods, and empirical results, M. (Israel) ...... [6490A-08] Lambooij, W. A. IJsselsteijn, Technische Univ. Eindhoven (Netherlands); I. E. J. 11:40 am: A virtual reality oriented clinical experiment on post-stroke Heynderickx, Philips Research Labs. (Netherlands) and Technische Univ. Delft rehabilitation: performance and preference comparison among different (Netherlands) ...... [6490A-17] stereoscopic displays, S. Yeh, Sr., A. S. Rizzo III, A. A. Sawchuk, Univ. of 4:40 pm: MOEVision: simple multiview display with clear floating image, H. Southern California ...... [6490A-09] Kakeya, Univ. of Tsukuba (Japan) ...... [6490A-18] Lunch Break ...... 12:00 to 1:30 pm Session Break ...... 5:00 to 5:20 pm

3D Theatre ...... 5:20 to 7:20 Chairs: Andrew J. Woods, Curtin Univ. of Technology (Australia); Chris Ward, Lightspeed Design, Inc. See large-screen examples of how 3D video is being used and produced around the world.

SD&A Dinner ...... 7:30 pm to late A no-host informal dinner open to all SD&A attendees will be held at a local San Jose restaurant. Details will be available at the conference.

10 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A

Tuesday 30 January SESSION 8 Conv. Ctr. Room A8 ...... Tues. 3:40 to 5:20 pm Awards and Plenary Presentation Stereoscopic Developments Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Chair: Vivian K. Walworth, Jasper Associates Development and Plasticity in Visual Cortex 3:40 pm: The compatibility of consumer DLP projectors with time- Brian A. Wandell, Stanford Univ. sequential stereoscopic 3D visualization, A. J. Woods, T. Rourke, Curtin Univ. of Technology (Australia) ...... [6490A-30] See pg. 6 for details. 4:00 pm: The fabrication of microretarder for in-cell stereoscopic LCD using reactive liquid crystal, C. Tsai, W. Huang, Industrial Technology SESSION 5 Research Institute (Taiwan); G. Tseng, National Taiwan Univ. (Taiwan); K. Lee, K. Conv. Ctr. Room A8 ...... Tues. 9:30 to 10:30 am Huang, Industrial Technology Research Institute (Taiwan); H. Y. Lin, National Stereoscopy on Mobile Devices Taiwan Univ. (Taiwan) ...... [6490A-31] Chair: Julien C. Flack, Dynamic Digital Depth Australia Pty. Ltd. 4:20 pm: Enabling stereoscopic 3D technology, M. G, Robinson, G. D. (Australia) Sharp, ColorLink, Inc...... [6490A-32] 9:30 am: Autostereoscopic display technology for mobile 3DTV 4:40 pm: Realization of electronic 3D display combining multiview and applications, J. Harrold, G. J. Woodgate, Ocuity Ltd. (United volumetric solutions, H. Ebisu, T. Kimura, H. Kakeya, Univ. of Tsukuba Kingdom) ...... [6490A-19] (Japan) ...... [6490A-33] 9:50 am: A 470 x 235 ppi high-resolution 2D and 3D autostereoscopic 5:00 pm: Advanced multiscopic display via temporal multiplexing, V. B . display and its user aptitude investigation results in mobile phone Markov, S. A. Kupiec, MetroLaser, Inc.; A. R. L. Travis, Univ. of Cambridge applications, N. Takanashi, S. Uehara, H. Asada, NEC Corp. (Japan)[6490A-20] (United Kingdom); D. G. Hopper, G. S. Saini, Air Force Research Lab.[6490A-34] 10:10 am: A prototype 3D mobile phone equipped with a next-generation autostereoscopic display, J. C. Flack, Dynamic Digital Depth Australia Pty. Posters and Demonstrations-Tuesday (Australia); J. Harrold, G. J. Woodgate, Ocuity Ltd. (United Kingdom) [6490A-21] Demonstrations ...... 5:30 to 8:30 pm Coffee Break ...... 10:30 to 11:00 am A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide SESSION 6 interactive, hands-on demonstrations of a wide-range of products related to Conv. Ctr. Room A8 ...... Tues. 11:00 am to 12:00 pm Electronic Imaging. Multiview Content Posters ...... 5:30 to 7:00 pm Chair: Gregg E. Favalora, Actuality Systems, Inc. Posters will be placed on display after 10:00 am in Exhibit Hall 1. 11:00 am: Multiple footprint stereo algorithms for 3D display content A poster session, with authors present at their posters, generation, F. Boughorbel, Philips Research Labs. (Netherlands) . . [6490A-22] will be held Tuesday evening, 5:30 to 7:00 pm. 11:20 am: Compression of still multiview images for 3D automultiscopic ✔ Time-multiplexed autostereoscopic display with content-defined 2D-3D spatially multiplexed displays, R. Lau, S. Ince, J. Konrad, Boston mode selection, S. A. Shestak, D. Kim, K. Cha, J. Koo, SAMSUNG Univ...... [6490A-23] Electronics Co., Ltd. (South Korea) ...... [6490A-53] 11:40 am: Prediction of depth images across multiple views, Y. Morvan, ✔ 3D from compressed 2D video, I. A. Ideses, L. P. Yaroslavsky, B. Fishbain, Technische Univ. Eindhoven (Netherlands) ...... [6490A-24] Tel Aviv Univ. (Israel); R. Vistuch, Tel-Aviv Univ. (Israel) ...... [6490A-55] Lunch/Exhibition Break ...... 12:00 to 1:30 pm ✔ Volumetric display by inclined-image scanning and three-dimensional image transmission based on optical shape measurement, D. Miyazaki, SESSION 7 K. Soutsuka, T. Honda, T. Mukai, Osaka City Univ. (Japan) ...... [6490A-56] Conv. Ctr. Room A8 ...... Tues. 1:30 to 3:10 pm ✔ Ray-space acquisition using parabolic mirror, T. Ichiyanagi, T. Yendo, T. Autostereoscopic Displays Fujii, M. Tanimoto, Nagoya Univ. (Japan) ...... [6490A-58] ✔ Chair: Steven L. Smith, VRex, Inc. Free viewpoint image generation using multipass dynamic programming, N. Fukushima, T. Yendo, T. Fujii, M. Tanimoto, Nagoya Univ. 1:30 pm: Application of Pi-cells in time-multiplexed stereoscopic and (Japan) ...... [6490A-59] autostereoscopic displays based on LCD panels, S. A. Shestak, D. Kim, ✔ SAMSUNG Electronics Co., Ltd. (South Korea) ...... [6490A-25] Generation of binocular object movies from monocular object movies, Y. Y. Chen, National Taiwan Univ. (Taiwan); Y. Tsai, National Chiao Tung Univ. 1:50 pm: Switchable lenticular based 2D/3D displays, M. Hiddink, S. T. De (Taiwan); Y. Hung, National Taiwan Univ. (Taiwan) ...... [6490A-60] Zwart, O. Willemsen, Philips Research Labs. (Netherlands) ...... [6490A-26] ✔ Volumetric three-dimensional display using transparent rotating 2:10 pm: Multiview autostereoscopic display of 36 view using ultrahigh Lanthanide(III) complexes sheet, S. Hisatake, S. Suda, J. Takahara, T. resolution LCD, B. Lee, H. Hong, J. Park, H. Park, H. Shin, I. J. Chung, LG Kobayashi, Osaka Univ. (Japan) ...... [6490A-61] Philips LCD (South Korea) ...... [6490A-27] ✔ Generation of the multiview 3D video of real scene for the multiview 3D 2:30 pm: Autostereoscopic display with 60 ray directions using LCD with display using multiview depth estimation and view synthesis, G. Um, optimized color filter layout, T. Koike, Hitachi, Ltd. (Japan) and The Univ. of Electronics and Telecommunications Research Institute (South Tokyo (Japan); M. Oikawa, K. Utsugi, M. Kobayashi, M. Yamasaki, Hitachi, Ltd. Korea) ...... [6490A-62] (Japan) ...... [6490A-28] ✔ Usability in virtual and augmented environments: a qualitative and 2:50 pm: Development of SVGA resolution 128-directional display, K. Kikuta, quantitative study, A. E. Pimentel, P. Dias, B. S. Santos, Univ. de Aveiro Y. Takaki, Tokyo Univ. of Agriculture and Technology (Japan) ...... [6490A-29] (Portugal) ...... [6490A-63] Coffee Break ...... 3:10 to 3:40 pm ✔ Improvement of the voxels density of three-dimensional image display by spatially multiplexing fluxes in light-ray-flux reproduction method, S. Hisatake, T. Kobayashi, Osaka Univ. (Japan) ...... [6490A-64] ✔ Curved integral imaging scheme using an additional large-aperture lens, D. Shin, Kwangwoon Univ. (South Korea); B. Lee, Seoul National Univ. (South Korea); E. S. Kim, Kwangwoon Univ. (South Korea) ...... [6490A-65] ✔ Psychophysical estimation of 3D virtual depth of united, synthesized, and mixed-type stereograms by means of simultaneous observation, M. Iizuka, Tokyo Polytechnic Univ. (Japan) ...... [6490A-66]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 11 Conference 6490A

✔ Recognition of 3D objects by use of computational integral imaging SESSION 10 reconstruction, J. Park, D. Hwang, D. Shin, E. S. Kim, Kwangwoon Univ. (South Korea) ...... [6490A-67] Conv. Ctr. Room A8 ...... Wed. 1:30 to 2:10 pm ✔ A point-based tele-immersion system: from acquisition to stereoscopic Stereoscopic Vision display, D. Ruiz, Univ. Catholique de Louvain (Belgium); X. Desurmont, J. Chair: Nicolas S. Holliman, Univ. of Durham (United Kingdom) Bruyelle, Multitel (Belgium); B. Macq, Univ. Catholique de Louvain 1:30 pm: DV tape drive synchronization for immediate preview of dual (Belgium) ...... [6490A-70] camera recordings, P. Wimmer, 3dtv.at (Austria) ...... [6490A-38] ✔ Real-time photorealistic stereoscopic rendering of fire, B. M. Rose, 1:50 pm: Real-time embedded system for stereo video processing for Blackbaud, Inc.; D. F. McAllister, North Carolina State Univ...... [6490A-71] multiview displays, R. M. Berretty, P. Machado, B. Riemens, Philips Research Labs. (Netherlands) ...... [6490A-39] Wednesday 31 January SESSION 11 Plenary Session Conv. Ctr. Room A8 ...... Wed. 2:20 to 2:40 pm Marriott Ballroom 1-6 ...... 8:00 to 9:20 am Keynote Presentation 2:20 pm: A Look at the Past and Future of Stereoscopic Displays 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Through a “(Liquid) Crystal Ball (Presentation Only), J. L. Fergason, Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Fergason Patent Properties ...... [6490A-72] Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Mottin, C.re de Recherche et de Restauration des Musées de France Coffee Break ...... 3:20 to 3:50 pm

8:40 am: Space Exploration, Imaging Technology, and the Development of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, SESSION 12 Solar System Visualization (SSV) Project Conv. Ctr. Room A8 ...... Wed. 3:50 to 5:30 pm

See pg. 7 for details. Integral 3D Displays Chair: Shojiro Nagata, InterVision (Japan) SESSION 9 3:50 pm: Integral photography suitable for small-lot production using mutually perpendicular lenticular sheets and fractional view, K. Yanaka, Conv. Ctr. Room A8 ...... Wed. 9:30 to 10:30 am Kanagawa Institute of Technology (Japan) ...... [6490A-41] Crosstalk in Stereoscopic and Autostereoscopic Displays 4:10 pm: Scan-type image capturing system using a cylindrical lens for Chair: Neil A. Dodgson, Univ. of Cambridge (United Kingdom) one-dimensional integral imaging, Y. Momonoi, K. Taira, Y. Hirayama, Toshiba Corp. (Japan) ...... [6490A-42] 9:30 am: A simple measurement of extinction ratio, M. A. Weissman, TrueVision Systems, Inc...... [6490A-35] 4:30 pm: Integral imaging system using an array of planar mirrors, J. Arai, M. Kawakita, H. Kawai, F. Okano, NHK Science & Technical Research Labs. 9:50 am: Optical cross-talk and visual comfort of a stereoscopic display (Japan) ...... [6490A-43] used in a real-time application, S. Pala, R. F. Stevens, National Physical Lab. (United Kingdom) ...... [6490A-36] 4:50 pm: Enhanced algorithm of image mapping method for computer- generated integral imaging system, S. Min, Virginia Polytechnic Institute and 10:10 am: Crosstalk in automultiscopic 3D displays: blessing in disguise?, State Univ.; K. S. Park, Information and Communications Univ. (South Korea); T. A. Jain, J. Konrad, Boston Univ...... [6490A-37] Poon, Virginia Polytechnic Institute and State Univ...... [6490A-54] Coffee Break ...... 10:30 to 11:00 am 5:10 pm: Improvement of viewing resolution of integral videography by use of rotated prism sheet, H. Liao, The Univ. of Tokyo (Japan) ...... [6490A-45] Discussion Forum: 3D in the Home: How Close Are We? ...... 11:00 am to 12:00 pm 3D in the home is a dream shared by many. To achieve it we need compelling content, simple distribution, great displays and affordable prices. How close are we to achieving these goals? A panel of industry specialists Visit www.stereoscopic.org for the latest information on the will discuss this topic. Stereoscopic Displays and Applications conference.

Lunch/Exhibition Break ...... 12:00 to 1:30 pm

12 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490B • Conv. Ctr. Room B4

Wednesday 31 January 2007 • Part of Proceedings of SPIE Vol. 6490 Stereoscopic Displays and Virtual Reality Systems XIV The Engineering Reality of Virtual Reality 2007

Conference Chairs: Mark T. Bolas, Univ. of Southern California; Ian E. McDowall, Fakespace Labs., Inc. Cochair: Margaret Dolinsky, Indiana Univ.

Wednesday 31 January

Plenary Session Marriott Ballroom 1-6 ...... 8:00 to 9:20 am 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Mottin, C.re de Recherche et de Restauration des Musées de France

8:40 am: Space Exploration, Imaging Technology, and the Development of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Solar System Visualization (SSV) Project

See pg. 7 for details.

SESSION 13 Conv. Ctr. Room B4 ...... Wed. 9:30 to 11:20 am Virtual Environments: Implementation Chair: Ian E. McDowall, Fakespace Labs., Inc. 9:30 am: Three-dimensional cameras on the International Space Station, M. Sabbatini, G. Visentin, European Space Agency (Netherlands); M. Collon, cosine Research B.V. (Netherlands) ...... [6490B-46] 9:50 am: Latency compensation by horizontal scanline selection for head- mounted displays, J. J. Jerald, A. Fuller, A. A. Lastra, M. Whitton, L. Kohli, F. P. Brooks, The Univ. of North Carolina at Chapel Hill ...... [6490B-47] 10:10 am: A novel wide-field-of-view display method with higher central resolution for hyper-realistic head dome projector, A. Hotta, T. Sasaki, H. Okumura, Toshiba Corp. (Japan) ...... [6490B-48] Coffee Break ...... 10:30 to 11:00 am 11:00 am: Tangible mixed-reality desktop for digital-media management, S. Winkler, S. Zhou, H. Yu, National Univ. of Singapore (Singapore) . . . [6490B-49]

SESSION 14 Conv. Ctr. Room B4 ...... Wed. 11:20 am to 12:20 pm Virtual Environments: The Experience Chair: Mark T. Bolas, Fakespace Labs., Inc. 11:20 am: Passing excellence, D. Tsoupikova, Univ. of Illinois/ Chicago ...... [6490B-50] 11:40 am: CytoViz: an artistic mapping of network measurements as living organisms in a VR application, B. Lopez Silva, L. Renambot, Univ. of Illinois/ Chicago ...... [6490B-51] 12:00 pm: Skin: an interactive hyperstereoscopic electro installation, H. G. Kostis, R. L. Kooima, Univ. of Illinois/Chicago; J. Kannenberg, Stasisfield.com ...... [6490B-52]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 13 Conference 6491 • Conv. Ctr. Room C1

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6491 Videometrics IX

Conference Chairs: J.-Angelo Beraldin, National Research Council Canada (Canada); Fabio Remondino, ETH Zürich (Switzerland); Mark R. Shortis, RMIT Univ. (Australia) Program Committee: Hirofumi Chikatsu, Tokyo Denki Univ. (Japan); Sabry F. El-Hakim, National Research Council Canada (Canada); Alan M. Frank, Lawrence Livermore National Lab.; Dieter Fritsch, Univ. Stuttgart (Germany); Joao G. M. Goncalves, European Commission (Italy); Armin Gruen, ETH Zürich (Switzerland); Gabriele Guidi, Politecnico di Milano (Italy); Graham P. Haddleton, QinetiQ (United Kingdom); Henrik G. A. Haggren, Helsinki Univ. of Technology (Finland); George I. Karras, National Technical Univ. of Athens (Greece); Hans-Gerd Maas, Technische Univ. Dresden (Germany); Lars S. Nyland, Univ. of North Carolina/Chapel Hill; Stuart Robson, Univ. College London (United Kingdom); Luc J. Van Gool, Katholieke Univ. Leuven (Belgium); James S. Walton, 4DVIDEO

Monday 29 January 2:30 pm: Range imaging technology: new developments and applications for people identification and tracking, T. Kahlmann, F. Remondino, ETH Zürich (Switzerland) ...... [6491-12] SESSION 1 2:50 pm: Heterodyne range imaging as an alternative to photogrammetry, Conv. Ctr. Room C1 ...... Mon. 8:30 to 10:00 am A. A. Dorrington, M. J. Cree, Univ. of Waikato (New Zealand); D. A. Carnegie, 3D Sensing: Videometrics Victoria Univ. of Wellington (New Zealand); A. D. Payne, Univ. of Auckland (New Zealand); R. M. Conroy, Univ. of Waikato (New Zealand) ...... [6491-13] Chair: Hans-Gerd Maas, Technische Univ. Dresden (Germany) Coffee Break ...... 3:10 to 3:40 pm 8:30 am: A review of the status and trends in underwater videometric measurement (Invited Paper), M. R. Shortis, Royal Melbourne Institute of Technology (Australia); E. S. Harvey, J. W. Seager, The Univ. of Western Australia SESSION 4 (Australia) ...... [6491-01] Conv. Ctr. Room C1 ...... Mon. 3:40 to 5:20 pm 9:00 am: Three-dimensional smart surveillance system using multiple 3D Processing and Modeling cameras, A. K. Mishra, S. Winkler, A. A. B. M. Kassim, National Univ. of Singapore (Singapore) ...... [6491-02] Chair: Fabio Remondino, ETH Zürich (Switzerland) 3:40 pm: Content-based retrieval aided registration for images and range 9:20 am: On the performance evaluation of tracking systems using multiple images, E. Paquet, J.-A. Beraldin, National Research Council Canada pan-tilt-zoom cameras, X. Desurmont, Multitel (Belgium) ...... [6491-03] (Canada) ...... [6491-15] 9:40 am: Sensor orientation for coverage optimization in surveillance 4:00 pm: A master-slaves volumetric framework for 3D reconstruction from applications, F. Janoos, R. Machiraju, R. Parent, J. W. Davis, A. Murray, The images, D. Ruiz, B. Macq, Univ. Catholique de Louvain (Belgium) . . . [6491-16] Ohio State Univ...... [6491-04] 4:20 pm: View planning for wide-area outdoor modeling based on Coffee Break ...... 10:00 to 10:30 am measurement density of objects, T. Asai, M. Kanbara, N. Yokoya, Nara Institute of Science and Technology (Japan) ...... [6491-17] SESSION 2 4:40 pm: A multinational deployment of 3D laser scanning to study Conv. Ctr. Room C1 ...... Mon. 10:30 am to 12:00 pm craniofacial dysmorphology in fetal alcohol spectrum disorders, J. L. 3D Sensing: Range Sensing I Rogers, E. A. Wernert, Indiana Univ.; E. Moore, St. Vincent Indianapolis Hospital; R. Ward, Indiana Univ. School of Liberal Arts; L. F. Wetherill, Indiana University Chair: J.-Angelo Beraldin, National Research Council Canada School of Medicine; T. Foroud, Indiana Univ. School of Medicine . . . . [6491-18] (Canada) 5:00 pm: Measurement and modeling of 4D live mouse heart volumes from 10:30 am: More than a poplar plank, the shape and subtle colors of the CT time series, A. W. Wetzel, Pittsburgh Supercomputing Ctr.; C. T. Badea, masterpiece Mona Lisa by Leonardo: the 3D technology (Invited Paper), F. Duke Univ.; S. Pomerantz, D. Nave, Pittsburgh Supercomputing Ctr.; N. Mistry, Blais, J. M. Taylor, L. Cournoyer, M. Picard, L. Borgeat, G. Godin, J.-A. Beraldin, A. Johnson, Duke Univ...... [6491-19] M. Rioux, National Research Council Canada (Canada); C. Lahanier, B. Mottin, Ctr. de Recherche et de Restauration des Musées de France (France) [6491-05] 11:00 am: Low-cost characterization of 3D laser scanners, M. Russo, G. Guidi, Politecnico di Milano (Italy) ...... [6491-06] 11:20 am: Metrological verification of 3D scanners: a preliminary approach, A. Pietrosanto, C. Liguori, R. Anchini, Univ. degli Studi di Salerno (Italy); G. Strazzullo, FOX BIT s.r.l. (Italy) ...... [6491-07] 11:40 am: TOF laser scanner characterization for low-range applications, G. Guidi, Politecnico di Milano (Italy); C. Bianchini, La Sapienza Univ. (Italy) ...... [6491-08] Lunch Break ...... 12:00 to 1:50 pm

SESSION 3 Conv. Ctr. Room C1 ...... Mon. 1:50 to 3:30 pm 3D Sensing: Range Sensing II Chair: Francois Blais, National Research Council Canada (Canada) 1:50 pm: Real-time range imaging by phase-stamp method using correlation image sensor, A. Kimachi, Osaka Electro-Communication Univ. (Japan); S. Ando, The Univ. of Tokyo (Japan) ...... [6491-10] 2:10 pm: Traceable 3D imaging metrology, J.-A. Beraldin, M. Rioux, L. Cournoyer, F. Blais, M. Picard, J. R. Pekelsky, National Research Council Canada (Canada) ...... [6491-11]

14 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6491

Tuesday 30 January SESSION 8 Conv. Ctr. Room C1 ...... Tues. 4:20 to 5:40 pm Awards and Plenary Presentation Visualization Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Chair: Nicola D’Apuzzo, Homometrica Consulting (Switzerland) Development and Plasticity in Visual Cortex 4:20 pm: Performance evaluation of a coded structural light system for Brian A. Wandell, Stanford Univ. cultural heritage applications, D. Akca, F. Remondino, D. Novak, T. Hanusch, G. Schrotter, A. Gruen, ETH Zürich (Switzerland) ...... [6491-31] See pg. 6 for details. 4:40 pm: A solid texture analysis based on three-dimensional convolution kernels, M. T. Suzuki, National Institute of Multimedia Education SESSION 5 (Japan) ...... [6491-32] Conv. Ctr. Room C1 ...... Tues. 9:30 to 11:10 am 5:00 pm: Acquisition of three-dimensional coordinates of objects from Axi- Motion Analysis Vision image data and application to stereoscopic display using integral Chair: Fabio Remondino, ETH Zürich (Switzerland) photography system, T. Aida, A. Yoneda, T. Uragaki, Osaka City Univ. (Japan); M. Kawakita, Japan Broadcasting Corp. (Japan) ...... [6491-33] 9:30 am: Concepts of single-camera high-speed 3D measurement systems (Invited Paper), H.-G. Maas, Technische Univ. Dresden (Germany) . . [6491-20] 5:20 pm: Crack propagation imaging by the ISIS camera and a video trigger system, T. Okinaka, P. Karimov, T. G. Etoh, Kinki Univ. (Japan) ...... [6491-34] 10:00 am: Configuration of multimirror systems for single high-speed camera based 3D motion analysis, T. Putze, Technische Univ. Dresden (Germany); K. Raguse, Volkswagen AG (Germany); H.-G. Maas, Technische Posters and Demonstrations-Tuesday Univ. Dresden (Germany) ...... [6491-21] Demonstrations ...... 5:30 to 8:30 pm Coffee Break ...... 10:20 to 10:50 am A symposium-wide demonstration session will be open to attendees 5:30 to 10:50 am: Gait analysis in forensic medicine, P. K. Larsen, N. Lynnerup, E. B. 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide Simonsen, Univ. of Copenhagen (Denmark) ...... [6491-22] interactive, hands-on demonstrations of a wide-range of products related to Electronic Imaging. SESSION 6 Conv. Ctr. Room C1 ...... Tues. 11:10 am to 12:50 pm Posters ...... 5:30 to 7:00 pm Industry Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, Chair: Gabriele Guidi, Politecnico di Milano (Italy) will be held Tuesday evening, 5:30 to 7:00 pm. 11:10 am: Developing measurement tools for the future: commercial ✔ Variability of bodily measures of normally dressed people using innovation in the videometrics industry (Invited Paper), R. Godding, AICON PhotoModeler®, P. K. Larsen, N. Lynnerup, Univ. of Copenhagen (Denmark); 3D Systems GmbH (Germany) ...... [6491-23] L. Hansen, National Institute of Occupational Health (Denmark) . . . [6491-39] 11:40 am: Three-dimensional body scanning technology for fashion and ✔ Subpixel location of discrete target images in close-range camera apparel industry, N. D’Apuzzo, Homometrica Consulting (Switzerland)[6491-24] calibration: a novel approach, R. Anchini, C. Liguori, Univ. degli Studi di 12:00 pm: Single CMOS sensor system for high-resolution double-volume Salerno (Italy); J. Beraldin, National Research Council Canada measurement applied to membrane distillation system, M. G. Lorenz, Univ. (Canada) ...... [6491-40] Carlos III de Madrid (Spain); M. A. Izquierdo-Gil, Univ. Complutense de Madrid ✔ Design of an image sensor for an ultra-high-speed and ultra-high- (Spain); R. Sanchez-Reillo, Univ. Carlos III de Madrid (Spain); C. Fernandez- sensitive video microscope, N. Ohtsuka, Kinki Univ. (Japan) . . . . [6491-42] Pineda, Univ. Complutense de Madrid (Spain) ...... [6491-25] ✔ Performance analysis and parameter optimization for iris recognition 12:20 pm: Beyond inspection: the promise of videometrics for industry and using Log-Gabor wavelet, V. A. Pozdin, Y. Du, Indiana Univ.-Purdue Univ. engineering (Invited Paper), K. Edmundson, Geodetic Systems, Inc. [6491-26] Indianapolis ...... [6491-43] Lunch/Exhibition Break ...... 12:50 to 2:20 pm

SESSION 7 Conv. Ctr. Room C1 ...... Tues. 2:20 to 4:20 pm Cultural Heritage Chair: Mark R. Shortis, Royal Melbourne Institute of Technology (Australia) 2:20 pm: Active and passive sensors for art works analysis and investigations (Invited Paper), A. Pelagotti, Art-Test s.a.s. (Italy); A. Del Mastio, Univ. degli Studi di Firenze (Italy); A. V. Razionale, Consultant (Italy) . . [6491-27] 2:50 pm: The potential of 3D techniques for cultural heritage reconstruction and visualization, G. Bitelli, V. A. Girelli, Univ. degli Studi di Bologna (Italy); F. Remondino, ETH Zürich (Switzerland); L. Vittuari, Univ. degli Studi di Bologna (Italy) ...... [6491-28] 3:10 pm: Wood artworks dimensional monitoring through high-resolution 3D cameras, G. Guidi, Politecnico di Milano (Italy); C. Atzeni, Univ. degli Studi di Firenze (Italy); J.-A. Beraldin, National Research Council Canada (Canada) ...... [6491-29] Coffee Break ...... 3:30 to 4:00 pm 4:00 pm: Digital documentation of complex architectures by integration of multiple techniques: the case study of Valer Castle, F. Voltolini, ITC-IRST (Italy); S. F. El-Hakim, National Research Council Canada (Canada); F. Remondino, ETH Zürich (Switzerland); M. Pontin, S. Girardi, A. Rizzi, L. Gonzo, ITC-IRST (Italy) ...... [6491-30]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 15 Conference 6492 • Conv. Ctr. Room A3

Monday-Thursday 29 January-1 February 2007 • Proceedings of SPIE Vol. 6492 Human Vision and Electronic Imaging XII

Conference Chairs: Bernice E. Rogowitz, IBM Thomas J. Watson Research Ctr.; Thrasyvoulos N. Pappas, Northwestern Univ.; Scott J. Daly, Sharp Labs. of America, Inc. Program Committee: Albert J. Ahumada, Jr., NASA Ames Research Ctr.; Jan P. Allebach, Purdue Univ.; Erhardt Barth, Univ. zu Lübeck (Germany); Walter R. Bender, MIT Media Lab.; Michael H. Brill, Datacolor; John C. Dalton, Synthetik Software; Huib de Ridder, Technische Univ. Delft (Netherlands); Gunilla A. M. Derefeldt, Swedish Defence Research Agency (Sweden); Elena A. Fedorovskaya, Eastman Kodak Co.; Jennifer Gille, Qualcomm Inc.; Sheila S. Hemami, Cornell Univ.; Laurent Itti, Univ. of Southern California; Stanley A. Klein, Univ. of California/Berkeley; Jan J. Koenderink, Univ Utrecht (Netherlands); John J. McCann, McCann Imaging; Jeffrey B. Mulligan, NASA Ames Research Ctr.; Karol Myszkowski, Max-Planck-Institut für Informatik (Germany); Adar Pelah, The Univ. of York (United Kingdom); Hawley K. Rising III, Sony Electronics Inc.; Sabine E. Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland); Christopher W. Tyler, Smith-Kettlewell Institute; Andrew B. Watson, NASA Ames Research Ctr.

Monday 29 January Tuesday 30 January

SESSION 1 Awards and Plenary Presentation Conv. Ctr. Room A3 ...... Mon. 9:50 to 11:20 am Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Keynote Session Development and Plasticity in Visual Cortex Chair: Bernice E. Rogowitz, IBM Thomas J. Watson Research Ctr. Brian A. Wandell, Stanford Univ. 9:50 am: New vistas in image and video quality (Invited Paper), A. C. Bovik, See pg. 6 for details. K. Seshadrinathan, S. Sumohana, The Univ. of Texas/Austin ...... [6492-68] 10:20 am: Painterly rendered portraits from photographs using a SESSION 3 knowledge-based approach (Invited Paper), S. DiPaola, Simon Fraser Univ. Conv. Ctr. Room A3 ...... Tues. 9:30 am to 12:00 pm (Canada) ...... [6492-01] Perceptual Image Quality and Compression 10:50 am: Nonlinear encoding in multilayer LNL systems optimized for the representation of natural images (Invited Paper), C. Zetzsche, Univ. Bremen Chair: Thrasyvoulos N. Pappas, Northwestern Univ. (Germany); U. Nuding, Ludwig-Maximilians-Univ. München (Germany) [6492-02] 9:30 am: The role of spatially adapted versus non-spatially adapted structural distortion in supra-threshold compression, M. D. Gaubatz, S. Lunch Break ...... 11:20 am to 1:30 pm Kwan, S. S. Hemami, Cornell Univ...... [6492-12] 9:50 am: Image compression using sparse color sampling combined with SESSION 2 nonlinear image processing, S. Brooks, Dalhousie Univ. (Canada); I. Saunders, Conv. Ctr. Room A3 ...... Mon. 1:30 to 5:30 pm N. A. Dodgson, Univ. of Cambridge (United Kingdom) ...... [6492-13] Perception of Natural Images 10:10 am: Compression of image clusters using Karhunen Loeve Chair: Erhardt Barth, Univ. zu Lübeck (Germany) transforms, M. Kramm, Technische Univ. München (Germany) . . . . . [6492-14] 1:30 pm: Optimal sensor design for estimating local velocity in natural Coffee Break ...... 10:30 to 11:00 am environments (Invited Paper), T. Tversky, W. Geisler, The Univ. of Texas/ 11:00 am: Visual ergonomic aspects of computer displays: glossy screens Austin ...... [6492-03] and angular dependence, K. E. Brunnström, B. Andren, Z. Konstantinides, L. 2:00 pm: Bilinear models of natural images (Invited Paper), B. A. Olshausen, Nordström, Acreo AB (Sweden) ...... [6492-15] Univ. of California/Berkeley ...... [6492-04] 11:20 am: The blur effect: perception and estimation with a new no- 2:30 pm: Locally adaptive image representation (Invited Paper), E. P. reference perceptual blur metric, F. Crete, STMicroelectronics (France) and Simoncelli, New York Univ...... [6492-05] Lab. des Images et des Signaux (France); M. Nicolas, STMicroelectronics (France); T. Dolmiere, P. Ladret, Lab. des Images et des Signaux Coffee Break ...... 3:00 to 3:30 pm (France) ...... [6492-16] 3:30 pm: Spatiotemporal power spectra of motion parallax: the case of 11:40 am: Perceptual quality evaluation of geometric distortions in images, cluttered 3D scenes, D. Rivait, M. S. Langer, McGill Univ. (Canada) . [6492-06] A. D’Angelo, G. Menegaz, M. Barni, Univ. degli Studi di Siena (Italy) . [6492-17] 3:50 pm: Feature category systems for 2nd order local image structure Lunch/Exhibition Break ...... 12:00 to 1:30 pm induced by natural image statistics and otherwise, L. D. Griffin, M. Lillholm, Univ. College London (United Kingdom) ...... [6492-07] 4:10 pm: The independent components of natural images are perceptually dependent, M. Bethge, T. Wiecki, F. A. Wichmann, Max Planck Institut für biologische Kybernetik (Germany) ...... [6492-08] 4:30 pm: Learning optimal features for visual pattern recognition, K. Labusch, Univ. zu Lübeck (Germany); U. Siewert, PLANET intelligent systems GmbH (Germany); T. Martinetz, E. Barth, Univ. zu Lübeck (Germany) . [6492-09] 4:50 pm: Unsupervised learning of a steerable basis for invariant image representations, M. Bethge, S. Gerwinn, Max Planck Institut für biologische Kybernetik (Germany) ...... [6492-10] 5:10 pm: Analysis of segment statistics for semantic classification of natural images, D. Depalov, T. N. Pappas, Northwestern Univ...... [6492-11]

Human Vision and Electronic Imaging Banquet Monday ...... 7:30 to 10:00 pm Adaptation and Synthesis in the Perception of Reality Michael A. Webster, Univ. of Nevada/Reno

16 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492

SESSION 4 ✔ Third- and first-party ground truth collection for auto key frame extraction from consumer video clips, K. Costello, J. Luo, Eastman Kodak Conv. Ctr. Room A3 ...... Tues. 1:40 to 3:00 pm Co...... [6492-63] Perceptual Video Quality ✔ Quantifying the use of structure in cognitive tasks, D. M. Rouse, S. S. Chair: Thrasyvoulos N. Pappas, Northwestern Univ. Hemami, Cornell Univ...... [6492-64] 1:40 pm: Color preference, color naturalness, and annoyance of ✔ Quality metric for H.264/AVC scalable video coding with full scalability, compressed and color-scaled digital videos, C. C. Koh, J. M. Foley, S. K. C. Kim, D. Suh, T. Bae, Y. Ro, Information and Communications Univ. (South Mitra, Univ. of California/Santa Barbara ...... [6492-18] Korea) ...... [6492-65] 2:00 pm: Relation between DSIS and DSCQS for temporal and spatial video ✔ Temporal relation between bottom-up versus top-down strategies for artifacts in a wireless home environment, N. Van den Ende, Philips Research gaze prediction, S. Krishna, P. Kuchi, V. Balasubramanian, J. A. Black, Jr., N. Labs. (Netherlands) and Technische Univ. Eindhoven (Netherlands); L. Meesters, Chatapuram-Krishnan, S. Panchanathan, Arizona State Univ. . . . . [6492-66] Technische Univ. Eindhoven (Netherlands); R. Haakma, Philips Research Labs. ✔ Eigen local color histograms for object recognition and orientation (Netherlands) ...... [6492-19] estimation, D. Muselet, Univ. des Sciences et Technologies de Lille (France); 2:20 pm: Can you see me now? An objective metric for predicting B. V. Funt, Simon Fraser Univ. (Canada); L. Macaire, Univ. des Sciences et intelligibility of compressed American Sign Language video, F. Ciaramello, Technologies de Lille (France) ...... [6492-67] S. S. Hemami, Cornell Univ...... [6492-21] 2:40 pm: Price-dependent quality: examining the effects of price on Wednesday 31 January multimedia quality requirements, D. S. Hands, R. Jacobs, K. Cheng, British Telecommunications plc (United Kingdom); C. Partridge, Univ. of Essex (United Kingdom) ...... [6492-22] Plenary Session Coffee Break ...... 3:00 to 3:30 pm Marriott Ballroom 1-6 ...... 8:00 to 9:20 am SESSION 5 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Conv. Ctr. Room A3 ...... Tues. 3:30 to 4:30 pm Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Visual Optics Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Chair: Albert J. Ahumada, Jr., NASA Ames Research Ctr. Mottin, C.re de Recherche et de Restauration des Musées de France 3:30 pm: Correcting spurious resolution in defocused images, J. W. Yellott, 8:40 am: Space Exploration, Imaging Technology, and the Development J. I. Yellott, Univ. of California/Irvine ...... [6492-23] of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, 3:50 pm: Do focus measures apply to retinal images?, Y. Tian, C. F. Wildsoet, Solar System Visualization (SSV) Project Univ. of California/Berkeley ...... [6492-24] See pg. 7 for details. 4:10 pm: SAR 2DAS algorithm for modeling human vision, J. Burki, C. F. Barnes, Georgia Institute of Technology ...... [6492-25] SESSION 6 Panel Discussion: Image Quality ...... 4:30 to 5:30 pm Conv. Ctr. Room A3 ...... Wed. 9:30 am to 12:00 pm Characterizing Color in Imaging Systems Panel Moderators: Thrasyvoulos N. Pappas, Northwestern Univ.; Sheila S. Hemami, Cornell Univ. 9:30 am: Aperture and object mode appearances in images, J. J. McCann, McCann Imaging ...... [6492-26] Panel Members: Albert J. Ahumada, Jr., NASA Ames Research Ctr.; Luke C. Cui, Lexmark International, Inc.; Yoichi Miyake, 9:50 am: Visibility improvement based on gray matching experiment between dark and ambient condition in mobile display, I. Kim, H. Ok, D. Chiba Univ. (Japan); Rene Rasmussen, Xerox Corp.; Eero Park, SAMSUNG Advanced Institute of Technology (South Korea) . . . [6492-27] Simoncelli, New York Univ.; Sabine E. Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland) 10:10 am: Multispectral color constancy: real image tests, M. Mosny, B. V. Funt, Simon Fraser Univ. (Canada) ...... [6492-28] Posters and Demonstrations-Tuesday Coffee Break ...... 10:30 to 11:00 am 11:00 am: Color balancing based upon gamut and temporal correlations, S. Demonstrations ...... 5:30 to 8:30 pm Kim, H. Lee, B. Kang, S. Lee, D. Park, C. Kim, SAMSUNG Advanced Institute of A symposium-wide demonstration session will be open to attendees 5:30 to Technology (South Korea) ...... [6492-29] 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide 11:20 am: Visibility of hue, saturation, and luminance nonuniformities in interactive, hands-on demonstrations of a wide-range of products related to LEDs, P. J. H. Seuntiens, I. Vogels, Philips Research Labs. Electronic Imaging. (Netherlands) ...... [6492-30] 11:40 am: Paper whiteness and its effect on the reproduction of colors, O. Posters ...... 5:30 to 7:00 pm Norberg, Mid Sweden Univ. (Sweden) ...... [6492-31] Posters will be placed on display after 10:00 am in Exhibit Hall 1. Lunch/Exhibition Break ...... 12:00 to 2:00 pm A poster session, with authors present at their posters, will be held Tuesday evening, 5:30 to 7:00 pm. SESSION 7 ✔ Contrast sensitivity measurements for quantifying defectivity and grading of image sensors, V. Sukumar, Micron Technology, Inc. . [6492-55] Conv. Ctr. Room A3 ...... Wed. 2:00 to 3:00 pm ✔ Making flat art for both eyes, S. Mason, Yavapai College ...... [6492-56] Cognitive Graphics ✔ A novel Bayer-like WRGB color filter array for CMOS image sensors, H. Chair: John E. Tumblin, Northwestern Univ. Honda, Y. Iida, G. Itoh, Y. Egawa, Toshiba Corp. (Japan) ...... [6492-57] 2:00 pm: Higher-order image representations for hyperresolution image synthesis and capture, B. Watson, North Carolina State Univ...... [6492-33] ✔ Improving video captioning for deaf and hearing-impaired people based on eye movement and attention overload, C. Chapdelaine, M. Beaulieu, L. 2:20 pm: Semantic photosynthesis, M. A. Johnson, Univ. of Cambridge Gagnon, CRIM (Canada) ...... [6492-59] (United Kingdom) ...... [6492-34] ✔ Comparison of methods for the simplification of mesh models using 2:40 pm: Revealing pentimenti: the hidden history in a painting, A. Gooch, quality indices and an observer study, S. Silva, B. Sousa Santos, C. M. Univ. of Victoria (Canada) ...... [6492-35] Ferreira, J. Madeira, Univ. de Aveiro (Portugal) ...... [6492-60] Coffee Break ...... 3:00 to 3:30 pm ✔ Influence of motion on contrast perception: supra-threshold spatio- velocity measurements, S. Tourancheau, P. Le Callet, D. Barba, Univ. de Nantes (France) ...... [6492-61]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 17 Conference 6492

SESSION 8 SESSION 10 Conv. Ctr. Room A3 ...... Wed. 3:30 to 5:10 pm Conv. Ctr. Room A3 ...... Thurs. 1:30 to 4:20 pm Perceptual Issues in High-Dynamic Range Imaging Higher Level Vision and Cognition Chair: Scott J. Daly, Sharp Labs. of America, Inc. Chair: Bernice E. Rogowitz, IBM Thomas J. Watson Research Ctr. 3:30 pm: Self-calibrating wide color gamut high-dynamic-range display, H. 1:30 pm: Motion of specularities on low-relief surfaces: frequency domain Seetzen, Brightside Technologies, Inc. (Canada) ...... [6492-36] analysis, Y. Farasat, M. S. Langer, McGill Univ. (Canada) ...... [6492-48] 3:50 pm: Tone mapping for high-dynamic range displays, L. Meylan, S. E. Fully automatic perceptual modeling of near-regular textures, Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland); S. J. Daly, 1:50 pm: G. Menegaz, A. Franceschetti, A. Mecocci, Univ. degli Studi di Siena Sharp Labs. of America, Inc...... [6492-38] (Italy) ...... [6492-49] 4:10 pm: Tracking gaze positions during evaluating high-dynamic range 2:10 pm: Machine perception using the five human defined forms plus images, J. Liu, W. Cheng, National Chiao Tung Univ. (Taiwan) ...... [6492-39] infrared, P. J. DeRego, Honeywell, Inc...... [6492-50] 4:30 pm: High-dynamic range imaging pipeline: perception-motivated The normalized compression distance and image representation of visual content, R. Mantiuk, G. Krawczyk, Max-Planck- 2:30 pm: distinguishability, N. Tran, Santa Clara Univ...... [6492-51] Institut für Informatik (Germany); R. Mantiuk, Szczecin Univ. of Technology (Poland); H. Seidel, Max-Planck-Institut für Informatik (Germany) . . . . [6492-40] 2:50 pm: Cortical network and dynamics of figure-ground organization, L. T. Likova, C. W. Tyler, The Smith-Kettlewell Eye Research Institute . . . . . [6492-52] 4:50 pm: Veiling glare: the dynamic range limit of HDR images, J. J. McCann, McCann Imaging; A. Rizzi, Univ. degli Studi di Milano (Italy) [6492-41] Coffee Break ...... 3:10 to 3:40 pm 3:40 pm: Comparing realness between real objects and images at various Discussion Session ...... 5:10 to 6:10 pm resolutions, K. Masaoka, M. Emoto, M. Sugawara, Y. Nojiri, NHK Science & Technical Research Labs. (Japan) ...... [6492-53] 4:00 pm: Navigation based on a sensorimotor representation: a virtual reality study, C. Zetzsche, K. Schill, G. Christopher, Univ. Bremen Thursday 1 February (Germany) ...... [6492-54]

SESSION 9 Discussion Session ...... 4:20 to 5:30 pm Conv. Ctr. Room A3 ...... Thurs. 9:00 to 11:30 am Eye Movements and Visual Attention 9:00 am: Hidden Markov model-based face recognition using selective attention, A. A. Salah, Bogaziçi Univ. (Turkey); M. Bicego, Univ. degli Studi di Sassari (Italy); L. Akarun, Bogaziçi Univ. (Turkey); E. Grosso, M. Tistarelli, Univ. degli Studi di Sassari (Italy) ...... [6492-42] 9:20 am: The role of eye movement signals in dorsal and ventral processing, J. A. Black, Jr., N. Chatapuram-Krishnan, S. Panchanathan, Arizona State Univ...... [6492-43] 9:40 am: Variable resolution images and their effects on eye movements during free viewing, M. Nystrom, Lunds Univ...... [6492-44] Coffee Break ...... 10:00 to 10:30 am 10:30 am: Hierarchy visual attention map, K. Yang, P. K. Das, C. C. Guest, Univ. of California/San Diego ...... [6492-45] 10:50 am: Attention trees and semantic paths, C. Giusti, G. G. Pieroni, Univ. degli Studi di Udine (Italy); L. Pieroni, Univ. degli Studi di Roma/La Sapienza (Italy) ...... [6492-46] 11:10 am: Motion integration in visual attention models for predicting simple dynamic scenes, A. Bur, H. Hügli, Univ. de Neuchâtel (Switzerland) ...... [6492-47] Lunch Break ...... 11:30 am to 1:30 pm

18 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493 • Conv. Ctr. Room A2

Tuesday-Thursday 30 January-1 February 2007 • Proceedings of SPIE Vol. 6493 Color Imaging XII: Processing, Hardcopy, and Applications

Conference Chairs: Reiner Eschbach, Xerox Corp.; Gabriel G. Marcu, Apple Computer, Inc. Program Committee: A. U. Agar, Oger Telecom (Turkey); Jan P. Allebach, Purdue Univ.; Jan Bares, NexPress Solutions, LLC; Phil J. Green, London College of Communication (United Kingdom); Roger David Hersch, École Polytechnique Fédérale de Lausanne (Switzerland); Patrick G. Herzog, Color AIXperts GmbH (Germany); Choon-Woo Kim, Inha Univ. (South Korea); Michael A. Kriss, Consultant; Shaun T. Love, Lexmark International, Inc.; Alessandro Rizzi, Univ. degli Studi di Milano (Italy); Shoji Tominaga, Osaka Electro-Communication Univ. (Japan); Chris Tuijn, Agfa-Gevaert Group (Belgium)

Tuesday 30 January SESSION 3 Conv. Ctr. Room A2 ...... Tues. 1:40 to 3:00 pm Awards and Plenary Presentation Displays I Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Chair: Gabriel G. Marcu, Apple Computer, Inc. Development and Plasticity in Visual Cortex 1:40 pm: A study on color appearance for large size display under various Brian A. Wandell, Stanford Univ. illumination conditions, S. Y. Choi, R. M. Luo, Univ. of Leeds (United Kingdom); M. R. Pointer, National Physical Lab. (United Kingdom) ...... [6493-07] See pg. 6 for details. 2:00 pm: An investigation of the effect of image size on the color appearance of softcopy reproductions using a contrast matching SESSION 1 technique, M. Nezamabadi, R. S. Berns, E. D. Montag, Rochester Institute of Conv. Ctr. Room A2 ...... Tues. 9:30 to 10:20 am Technology ...... [6493-08] Color Science 2:20 pm: Modeling for hue shift effect of human visual system on high- Chair: Reiner Eschbach, Xerox Corp. luminance display, T. Lee, M. Lee, K. Park, Y. Ha, Kyungpook National Univ. (South Korea) ...... [6493-09] 9:30 am: On the behavior of spatial models of color (Invited Paper), A. Rizzi, Univ. degli Studi di Milano (Italy); J. J. McCann, McCann Imaging . . . [6493-01] 2:40 pm: Optimizing color edge contrast in presence of noise and amplitude quantization errors, F. Lebowsky, STMicroelectronics (France); Y. 10:00 am: Effect of time spacing on the percieved color, S. Roch, J. Y. Huang, STMicroelectronics (Singapore) ...... [6493-10] Hardeberg, Gjøvik Univ. College (Norway) ...... [6493-02] Coffee Break ...... 3:00 to 3:30 pm Coffee Break ...... 10:20 to 10:50 am SESSION 4 SESSION 2 Conv. Ctr. Room A2 ...... Tues. 3:30 to 5:50 pm Conv. Ctr. Room A2 ...... Tues. 10:50 am to 12:10 pm Displays II Compression and Archiving Chair: Choon-Woo Kim, Inha Univ. (South Korea) Chair: Shoji Tominaga, Osaka Electro-Communication Univ. (Japan) 3:30 pm: Correlating 2D NTSC gamutrRatio to its 3D gamut volume, P. Sun, 10:50 am: Segmentation for MRC compression, E. Haneda, Purdue Univ.; J. Shih Hsin Univ. (Taiwan) ...... [6493-11] Yi, SAMSUNG Electronics Co., Ltd. (South Korea); C. A. Bouman, Purdue Univ. [6493-03] 3:50 pm: Device calibration method for 1D light modulator, Y. Bang, A. Baik, D. Park, SAMSUNG Advanced Institute of Technology (South Korea) . [6493-12] 11:10 am: A new approach to JBIG2 binary image compression, M. Figuera, Purdue Univ.; J. Yi, SAMSUNG Electronics Co., Ltd. (South Korea); C. A. 4:10 pm: A LCD TV comparable to a CRT TV for moving image quality, K. Bouman, Purdue Univ...... [6493-04] Kim, LG. Philips LCD Co., Ltd. (South Korea) ...... [6493-13] 11:30 am: Removal of artifacts from JPEG compressed document images, 4:30 pm: Nonlinearity curve optimization in a digital imaging device, S. B. Oztan, Univ. of Rochester; A. Malik, Z. Z. Fan, R. Eschbach, Xerox Quan, X. Jiang, Qualcomm, Inc...... [6493-14] Corp...... [6493-05] 4:50 pm: Reducing rainbow effect for filed sequential color LCDs, P. Sun, 11:50 am: Digital images for eternity: color microfilm as archival medium, C. Shih Hsin Univ. (Taiwan) ...... [6493-15] M. Normand, R. Gschwind, Univ. of Basel (Switzerland) ...... [6493-06] 5:10 pm: Design and implementation of optimal color analyzers for TFT- Lunch/Exhibition Break ...... 12:10 to 1:40 pm LCDs, G. Chang, C. Liao, Y. Lin, National Taiwan Normal Univ. (Taiwan)[6493-16] 5:30 pm: Third-time lucky: why a series of ISO display standards deserves extensive coverage at EI conferences, F. L. van Nes, Technische Univ. Eindhoven (Netherlands) and ErgoNes (Netherlands) ...... [6493-17]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 19 Conference 6493

Posters and Demonstrations-Tuesday SESSION 7 Demonstrations ...... 5:30 to 8:30 pm Conv. Ctr. Room A2 ...... Wed. 1:10 to 2:40 pm Printing A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide Chair: Giordano B. Beretta, Hewlett-Packard Labs. interactive, hands-on demonstrations of a wide-range of products related to 1:10 pm: Modeling an electro-photographic printer, part I: monochrome Electronic Imaging. systems (Invited Paper), M. A. Kriss, Consultant ...... [6493-23] 1:40 pm: Modeling an electro-photographic printer, part II: color systems, Posters ...... 5:30 to 7:00 pm M. A. Kriss, Consultant ...... [6493-24] Posters will be placed on display after 10:00 am in Exhibit Hall 1. 2:00 pm: Printer color calibration using an embedded sensor, Y. Wu , A poster session, with authors present at their posters, Hewlett-Packard Co...... [6493-25] will be held Tuesday evening, 5:30 to 7:00 pm. 2:20 pm: Color measurements on prints containing fluorescent whitening ✔ Clapper-Yule reflectivity of halftone color fluorescent imaging, Y. Zhang, agents, M. Andersson, O. Norberg, Mid Sweden Univ. (Sweden) . . . . [6493-26] Y. Du, Southern Yangtze Univ. (China) ...... [6493-55] ✔ Implement of FGS video encoding based on H.264, Q. Lin, Huaqiao Univ. SESSION 8 (China) ...... [6493-56] Conv. Ctr. Room A2 ...... Wed. 2:40 to 5:30 pm ✔ Study of digital watermarking for halftoneiImages, X. Wan, D. Xie, Wuhan Univ. (China) ...... [6493-51] Printing and Applications ✔ Compression of color images using a hologram of gray tones, A. Chair: Michael A. Kriss, Consultant Restrepo-Martínez, R. Castañeda, Univ. Nacional de Colombia 2:40 pm: A reference printer and color management architecture, P. J . (Colombia) ...... [6493-52] Green, London College of Communication (United Kingdom) ...... [6493-27] ✔ A GIS-based sensitivity mapping as a baseline for coastal resort 3:00 pm: Efficient color printer characterization based on extended planning, M. N. Hegazy, H. M. A. Effat, National Authority for Remote Neugebauer spectral models, P. Soler, J. Arnabat, Hewlett-Packard Co. Sensing and Space Sciences (Egypt) ...... [6493-53] (Spain) ...... [6493-28] Coffee Break ...... 3:20 to 3:50 pm Wednesday 31 January 3:50 pm: Digital camera characterization for color measurements on printed-paper media, M. Andersson, Mid Sweden Univ. (Sweden) . . [6493-29] Plenary Session 4:10 pm: User preferences in color enhancement for unsupervised printing methods, C. E. Parraman, Univ. of the West of England (United Kingdom); A. Marriott Ballroom 1-6 ...... 8:00 to 9:20 am Rizzi, Univ. degli Studi di Milano (Italy) ...... [6493-30] 4:30 pm: A user-friendly digital image processing methodology, R. Shaw, 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the White Rose Digital ...... [6493-31] Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, 4:50 pm: Production planning and automated imposition, C. Tuijn, Agfa- Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Gevaert Group (Belgium) ...... [6493-32] Mottin, C.re de Recherche et de Restauration des Musées de France 5:10 pm: The perfect photo book: hints for the image selection process, R. 8:40 am: Space Exploration, Imaging Technology, and the Development Fageth, W. Schmidt-Sacht, CeWe Color AG & Co. OHG (Germany) . . [6493-33] of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Solar System Visualization (SSV) Project Thursday 1 February See pg. 7 for details. SESSION 9 SESSION 5 Conv. Ctr. Room A2 ...... Thurs. 9:00 to 10:00 am Conv. Ctr. Room A2 ...... Wed. 9:30 to 10:10 am Image Processing Gamut Mapping Chair: Alessandro Rizzi, Univ. degli Studi di Milano (Italy) Chair: Phil J. Green, London College of Communication (United 9:00 am: Automatic processing of images for visual preference, M. Bressan, Kingdom) D. Arregui, H. Poirier, Xerox Research Ctr. Europe (France); C. R. Dance, Xerox Research Ctr. Europe (United Kingdom) ...... [6493-57] 9:30 am: A novel color mapping method for preferred color reproduction, K. Kim, M. Jang, H. S. Oh, S. H. Kim, D. Choi, SAMSUNG Electronics Co., Ltd. 9:20 am: Color recovery from gray image based on analysis of wavelet (South Korea) ...... [6493-18] packet subbands, K. Ko, O. Kwon, C. Son, E. Kwon, Y. Ha, Kyungpook National Univ. (South Korea) ...... [6493-34] 9:50 am: Gamut mapping method for ICC saturated intent, M. Cho, Y. Bang, Y. Kim, S. Kim, H. Choh, SAMSUNG Advanced Institute of Technology (South 9:40 am: Model-based deduction of CMYK surface coverages from visible Korea) ...... [6493-19] and infrared spectral measurements of halftone prints, T. Bugnon, M. Brichon, R. D. Hersch, École Polytechnique Fédérale de Lausanne Coffee Break ...... 10:10 to 10:40 am (Switzerland) ...... [6493-35] SESSION 6 Coffee Break ...... 10:00 to 10:30 am Conv. Ctr. Room A2 ...... Wed. 10:40 to 11:20 am Imaging Solutions and Applications Chair: Chris Tuijn, Agfa-Gevaert Group (Belgium) 10:40 am: Adaptive color artwork, G. B. Beretta, Hewlett-Packard Labs...... [6493-20] 11:00 am: Peteye detection and correction, J. Yen, Toshiba America; H. Luo, Yahoo! Research Berkeley; D. R. Tretter, Hewlett-Packard Labs. . . . . [6493-21] Lunch/Exhibition Break ...... 11:20 am to 1:10 pm

20 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493

SESSION 10 Conv. Ctr. Room A2 ...... Thurs. 10:30 to 11:50 am Multispectral Imaging Chair: Alessandro Rizzi, Univ. degli Studi di Milano (Italy) 10:30 am: Accurate spectral response measurement system for digital color cameras, G. Chang, C. Liao, National Taiwan Normal Univ. (Taiwan) ...... [6493-36] 10:50 am: Interpolation for nonlinear Retinex-type algorithms, D. Shaked, Hewlett-Packard Labs. (Israel) ...... [6493-37] 11:10 am: Omnidirectional scene illuminant estimation using a multispectral imaging system, S. Tominaga, T. Fukuda, Osaka Electro-Communication Univ. (Japan) ...... [6493-38] 11:30 am: Deducing ink-transmittance spectra from reflectance and transmittance measurement of prints, H. Mathieu, R. D. Hersch, École Polytechnique Fédérale de Lausanne (Switzerland) ...... [6493-39] Lunch Break ...... 11:50 am to 1:20 pm

SESSION 11 Conv. Ctr. Room A2 ...... Thurs. 1:20 to 2:20 pm Halftoning I Chair: Jan P. Allebach, Purdue Univ. 1:20 pm: Halftone independent methods for color correction, V. Monga, R. Bala, S. Wang, Xerox Corp...... [6493-40] 1:40 pm: Controlling the error in spectral error diffusion, J. Gerhardt, Hogskolen I Gjovik (Norway); J. Y. Hardeberg, Gjøvik Univ. College (Norway) ...... [6493-41] 2:00 pm: Holladay halftoning using superresolution encoded templates, J. S. McElvain, C. M. Hains, Xerox Corp...... [6493-42]

SESSION 12 Conv. Ctr. Room A2 ...... Thurs. 2:20 to 5:10 pm Halftoning II Chair: Reiner Eschbach, Xerox Corp. 2:20 pm: The hybrid screen: improving the breed, C. Lee, J. P. Allebach, Purdue Univ...... [6493-44] 2:40 pm: Ranked dither for robust color printing, M. R. Gupta, J. J. Bowen, Univ. of Washington ...... [6493-45] Coffee Break ...... 3:00 to 3:30 pm 3:30 pm: Rank-ordered error diffusion: methods and applications, B. Xu, R. P. Loce, Xerox Corp...... [6493-46] 3:50 pm: AM/FM halftoning: improved cost function and training framework, S. W. Han, Purdue Univ...... [6493-47] 4:10 pm: Resolution enhancement techniques for halftoned images, B. T. Ryu, J. O. Lee, C. Kim, Inha Univ. (South Korea); S. H. Kim, SAMSUNG Electronics Co., Ltd. (South Korea) ...... [6493-48] 4:30 pm: Contribution to quality assessment of digital halftoning algorithms, F. Cittadini, S. J. Berche, Océ Print Logic Technologies (France); M. Ben Chouikha, G. Alquié, Univ. Pierre et Marie Curie (France) ...... [6493-49] 4:50 pm: Uniform rosette for moire-free color halftoning, S. Wang, Xerox Corp...... [6493-50]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 21 Conference 6494 • Conv. Ctr. Room C2

Tuesday-Thursday 30 January-1 February 2007 • Proceedings of SPIE Vol. 6494 Image Quality and System Performance IV

Conference Chairs: Luke C. Cui, Lexmark International, Inc.; Yoichi Miyake, Chiba Univ. (Japan) Program Committee: Peter D. Burns, Eastman Kodak Co.; Mark D. Fairchild, Rochester Institute of Technology; Susan P. Farnand, Eastman Kodak Co.; Frans Gaykema, OCE Technologies BV (Netherlands); Dirk W. Hertel, Cypress Semiconductor; Robin B. Jenkin, Cranfield Univ. (United Kingdom); Lindsay W. MacDonald, London College of Communication (United Kingdom); Nathan Moroney, Hewlett-Packard Co.; Rene Rasmussen, Xerox Corp.; Eric K. Zeise, NexPress Solutions, Inc.

SESSION 1 Tuesday 30 January Conv. Ctr. Room C2 ...... Wed. 9:30 to 11:40 am Panel Discussion: Image Quality System Measurement and Modeling: Subjective Chair: Nathan Moroney, Hewlett-Packard Co. Conv. Ctr. Room A3 ...... 4:40 to 5:30 pm 9:30 am: Measuring user experience: theoretical and methodological Panel Moderators: Thrasyvoulos N. Pappas, Northwestern Univ.; issues, J. M. Takatalo, Univ. of Helsinki (Finland); J. P. Häkkinen, Nokia Sheila S. Hemami, Cornell Univ. Research Ctr. (Finland); H. Särkelä, J. Komulainen, J. Kaistinen, G. S. Nyman, Univ. of Helsinki (Finland) ...... [6494-01] Panel Members: Albert J. Ahumada, Jr., NASA Ames Research 9:55 am: Audio-visual quality estimation of mobile phone video cameras Ctr.; Luke C. Cui, Lexmark International, Inc.; Yoichi Miyake, with interpretation-based quality approach, J. E. Radun, T. Virtanen, G. S. Chiba Univ. (Japan); Rene Rasmussen, Xerox Corp.; Eero Nyman, Univ. of Helsinki (Finland); J. Olives, M. Vaahteranoksa, T. Vuori, Nokia Simoncelli, New York Univ.; Sabine E. Süsstrunk, École Research Ctr. (Finland) ...... [6494-02] Polytechnique Fédérale de Lausanne (Switzerland) Coffee Break ...... 10:20 to 10:50 am Posters and Demonstrations-Tuesday 10:50 am: Modeling color quality difference for mobile displays, J. J. Yoo, Y. J. Kim, R. M. Luo, Univ. of Leeds (United Kingdom); S. Lee, D. Park, C. Kim, Demonstrations ...... 5:30 to 8:30 pm SAMSUNG Advanced Institute of Technology (South Korea) ...... [6494-03] A symposium-wide demonstration session will be open to attendees 5:30 to 11:15 am: Threshold value for acceptable video quality using signal-to- 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide noise ratio, M. Vaahteranoksa, Nokia Corp. (Finland) ...... [6494-30] interactive, hands-on demonstrations of a wide-range of products related to Lunch/Exhibition Break ...... 11:40 am to 1:10 pm Electronic Imaging. SESSION 2 Posters ...... 5:30 to 7:00 pm Conv. Ctr. Room C2 ...... Wed. 1:10 to 2:00 pm Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, System Measurement and Modeling: Web-based will be held Tuesday evening, 5:30 to 7:00 pm. Chair: Peter D. Burns, Eastman Kodak Co. ✔ Accurate and cost-effective MTF measurement system for lens modules 1:10 pm: Color differences without probit analysis, N. Moroney, Hewlett- of digital cameras, G. Chang, S. Pan, C. Su, National Taiwan Normal Univ. Packard Co...... [6494-05] (Taiwan) ...... [6494-28] 1:35 pm: Web-based versus controlled-environment psychophysics ✔ Quality improvement by selective regional slice coding implementation experiments, S. Zuffi, C. Brambilla, Consiglio Nazionale delle Ricerche (Italy); G. in H.264/AVC, H. Ryu, W. Cheong, S. Jeong, K. Moon, Electronics and B. Beretta, Hewlett-Packard Labs.; P. Scala, Consiglio Nazionale delle Ricerche Telecommunications Research Institute (South Korea) ...... [6494-29] (Italy) ...... [6494-06] ✔ Quality evaluation of the halftone by halftoning algorithm-based methods and adaptive method, X. Wan, D. Xie, Wuhan Univ. . . . [6494-32] SESSION 3 Conv. Ctr. Room C2 ...... Wed. 2:00 to 2:50 pm Wednesday 31 January System Measurement and Modeling: Objective I Chair: Peter D. Burns, Eastman Kodak Co. Plenary Session 2:00 pm: Video quality assesment using M-SVD, P. Tao, A. M. Eskicioglu, The City Univ. of New York ...... [6494-07] Marriott Ballroom 1-6 ...... 8:00 to 9:20 am 2:25 pm: Method of estimating perceived video quality for mobile multimedia application based on full reference framework, O. Sugimoto, S. 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Sakazawa, A. Koike, KDDI R&D Labs. (Japan) ...... [6494-08] Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Coffee Break ...... 2:50 to 3:20 pm Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Mottin, C.re de Recherche et de Restauration des Musées de France

8:40 am: Space Exploration, Imaging Technology, and the Development of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Solar System Visualization (SSV) Project

See pg. 7 for details.

22 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6494

SESSION 4 SESSION 7 Conv. Ctr. Room C2 ...... Wed. 3:20 to 4:35 pm Conv. Ctr. Room C2 ...... Thurs. 10:40 to 11:55 am System Measurement and Modeling: Objective II Image Quality Standards II Chair: Robin B. Jenkin, Cranfield Univ. (United Kingdom) Chair: Dirk W. Hertel, Cypress Semiconductor Corp. 3:20 pm: Performance evaluation of digital still camera image processing 10:40 am: W1.1 macro-uniformity and color fidelity (Presentation Only), R. pipelines, D. W. Hertel, E. T. Chang, L. C. Shih, Cypress Semiconductor Corp.; Rasmussen, Xerox Co.; D. Gusev, NexPress Solutions, Inc...... [6494-18] J. Sproul, Boston Scientific Corp...... [6494-09] 11:05 am: Recent progress in the development of INCITS W1.1: 3:45 pm: Image quality and automatic color equalization, M. Chambah, Univ. appearance-based image quality standards for printers, T. Bouk, Eastman de Reims Champagne-Ardenne (France); A. Rizzi, Univ. degli Studi di Milano Kodak Co.; E. N. Dalal, Xerox Corp.; K. D. Donohue, Univ. of Kentucky; S. (Italy); C. Saint-Jean, Univ. de La Rochelle (France) ...... [6494-10] Farnand, Rochester Institute of Technology; F. Gaykema, OCE Technologies BV (Netherlands); D. Gusev, Eastman Kodak Co.; A. Haley, Monotype Imaging; P. L. 4:10 pm: A unified framework for physical print quality, A. Eid, B. E. Cooper, Jeran, Hewlett-Packard Co.; D. Kozak, Lexmark International, Inc., W. C. Kress, Lexmark International, Inc...... [6494-11] Toshiba America Business Solutions; O. Martinez, Hewlett-Packard Co. (Spain); D. Mashtare, Xerox Corp.; A. L. McCarthy, Lexmark International, Inc.; Y. S. Ng, SESSION 5 Eastman Kodak Co.; D. R. Rasmussen, Xerox Corp.; M. Robb, Lexmark Conv. Ctr. Room C2 ...... Wed. 4:35 to 5:25 pm International, Inc.; H. Shin, Xerox Corp.; M. Q. Slickers, Hewlett-Packard Co.; E. H. Barney-Smith, Boise State Univ.; M. Tse, Quality Engineering Associates, System Measurement and Modeling: Newer Technologies Inc.; E. K. Zeise, Eastman Kodak Co.; S. Zoltner, Xerox Corp...... [6494-19] Chair: Robin B. Jenkin, Cranfield Univ. (United Kingdom) 11:30 am: Scanners for analytic metrology: the devil in the details, E. K. 4:35 pm: Measurement-based objective metric for the printer resolution, J. Zeise, NexPress Solutions, Inc.; D. R. Williams, P. D. Burns, Eastman Kodak Co.; Hasegawa, T. Y. Hwang, H. C. Kim, D. W. Kim, M. H. Choi, SAMSUNG W. C. Kress, Toshiba ...... [6494-20] Electronics Co., Ltd. (South Korea) ...... [6494-33] Lunch Break ...... 11:55 am to 1:25 pm 5:00 pm: Information distance-based selective feature for iris image quality measure, C. S. Belcher, Y. Du, Indiana Univ.-Purdue Univ. Indianapolis[6494-13] SESSION 8 Conv. Ctr. Room C2 ...... Thurs. 1:25 to 3:05 pm Thursday 1 February Image Quality Attributes: Measurement and Modeling Chair: Rene Rasmussen, Xerox Corp. SESSION 6 1:25 pm: Paper roughness and the color gamut of printed images, J. S. Conv. Ctr. Room C2 ...... Thurs. 8:30 to 10:10 am Arney, Rochester Institue of Technology ...... [6494-21] Image Quality Standards I 1:50 pm: Investigation of two methods to quantify noise in digital images Chair: Eric K. Zeise, NexPress Solutions, Inc. based on the perception of the human eye, J. Kleinmann, Univ. of Applied Sciences Köln (Germany); D. Wueller, Image Engineering (Germany) . [6494-22] 8:30 am: Driving color management into the office, T. Newman, Canon Development Americas, Inc...... [6494-14] 2:15 pm: The effective pictorial information capacity of JPEG 6b as an image quality metric, R. B. Jenkin, Cranfield Univ. (United Kingdom); S. 8:55 am: Appearance can be deceiving: using appearance models in color Triantaphillidou, Univ. of Westminster (United Kingdom); M. A. Richardson, imaging, G. M. Johnson, Apple Computer, Inc. and Rochester Institute of Cranfield Univ. (United Kingdom) ...... [6494-23] Technology ...... [6494-15] 2:40 pm: Objective video quality assessment method for evaluating effects 9:20 am: Applying and extending ISO/TC42 digital camera performance of freeze distortion in arbitrary video scenes, K. Watanabe, J. Okamoto, T. standards to mobile imaging products, D. R. Williams, P. D. Burns, Eastman Kurita, Nippon Telegraph and Telephone Corp. (Japan) ...... [6494-24] Kodak Co...... [6494-16] Coffee Break ...... 3:05 to 3:35 pm 9:45 am: Differential gloss quality scale experiment update: an appearance- based image quality standard initiative (INCITS W1.1), Y. S. Ng, C. Kuo, Eastman Kodak Co.; E. Maggard, Hewlett Packard Co.; D. Mashtare, Xerox SESSION 9 Corp.; O. Morris, Hewlett Packard Co...... [6494-17] Conv. Ctr. Room C2 ...... Thurs. 3:35 to 4:25 pm Coffee Break ...... 10:10 to 10:40 am Image Quality Attributes: Unique Defects Chair: Susan P. Farnand, Eastman Kodak Co. 3:35 pm: Evaluation of algorithms for visual masking of defective LCD sub- pixels, J. W. Stellbrink, Hewlett-Packard Co...... [6494-26] 4:00 pm: Scanner motion error detection and correction, L. C. Cui, Lexmark International, Inc...... [6494-27]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 23 Conference 6495 • Conv. Ctr. Room B1

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6495 Visualization and Data Analysis 2007

Conference Chairs: Robert F. Erbacher, Utah State Univ.; Jonathan C. Roberts, Univ. of Kent (United Kingdom); Matti T. Gröhn, Ctr. for Scientific Computing (Finland); Katy Börner, Indiana Univ. Cochairs: Ming C. Hao, Hewlett-Packard Labs.; Pak C. Wong, Pacific Northwest National Lab. Program Committee: Uwe Brinkschulte, Univ. Karlsruhe (Germany); Philip C. Chen, Future, Inc.; L. Eric Greenwade, Idaho National Engineering and Environmental Lab.; Hans-Georg Pagendarm, Deutsches Zentrum für Luft- und Raumfahrt e.V. (Germany); Alex T. Pang, Univ. of California/ Santa Cruz; Hanspeter Pfister, Mitsubishi Electric Research Labs.; Christopher D. Shaw, Georgia Institute of Technology; Kalpathi R. Subramanian, Univ. of North Carolina/Charlotte; Yinlong Sun, Purdue Univ.; J. Edward Swan II, Naval Research Lab.; Craig M. Wittenbrink, NVIDIA; Yingcai Xiao, Univ. of Akron; William J. Yurcik, Univ. of Illinois at Urbana-Champaign

Monday 29 January SESSION 4 Conv. Ctr. Room B1 ...... Mon. 3:40 to 4:55 pm SESSION 1 Evaluations Conv. Ctr. Room B1 ...... Mon. 8:30 to 10:10 am Chair: Matti T. Gröhn, Ctr. for Scientific Computing (Finland) Visualization Techniques I 3:40 pm: A study on multiple views for tree visualization, S. T. Teoh, San Chair: Robert F. Erbacher, Utah State Univ. José State Univ...... [6495-09] 8:30 am: A visualization method for ontology based distance measure on 4:05 pm: Task-based evaluation of multirelational 3D and standard 2D relation network, S. Cho, J. Park, Information and Communications Univ. parallel coordinates, C. Forsell, Uppsala Univ. (Sweden); J. Johansson, (South Korea) ...... [6495-01] Linköpings Univ. (Sweden) ...... [6495-10] 8:55 am: Incorporating highlighting animations into static visualizations, J. 4:30 pm: Evaluation of mesh simplification algorithms using PolyMeCo: a Woodring, H. Shen, The Ohio State Univ...... [6495-02] case study, S. Silva, Instituto de Engenharia Electrónica e Telemática de Aveiro 9:20 am: Re-using millions of visualizations, R. D. Rimey, D. S. Bolme, (Portugal); F. Silva, Univ. da Beira Interior (Portugal); J. Madeira, B. Sousa Lockheed Martin Corp...... [6495-03] Santos, Instituto de Engenharia Electrónica e Telemática de Aveiro (Portugal) ...... [6495-11] 9:45 am: A visual analysis of multi-attribute data using pixel matrix displays, M. C. Hao, U. Dayal, Hewlett-Packard Labs.; D. C. Keim, T. Schreck, Univ. Konstanz (Germany) ...... [6495-04] Tuesday 30 January Coffee Break ...... 10:10 to 10:40 am Awards and Plenary Presentation SESSION 2 Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Conv. Ctr. Room B1 ...... Mon. 10:40 am to 12:20 pm Development and Plasticity in Visual Cortex Interaction and Exploration Brian A. Wandell, Stanford Univ. Chair: Ming C. Hao, Hewlett-Packard Labs. See pg. 6 for details. 10:40 am: An interactive visual exploration tool for Northern California’s water-monitoring system, J. Sreevalsan-Nair, Univ. of California/Davis; E. van SESSION 5 Nieuwenhuyse, Bureau of Reclamation; I. Hotz, Univ. of California/Davis; L. Conv. Ctr. Room B1 ...... Tues. 9:30 to 10:20 am Linsen, Ernst Moritz Arndt Univ. Greifswald (Germany); B. Hamann, Univ. of California/Davis ...... [6495-05] Visualization Applications 11:05 am: PerfViz: a visualization tool for analyzing, exploring, and Chair: Matti T. Gröhn, Ctr. for Scientific Computing (Finland) comparing storage controller performance data, A. P. Sawant, North Carolina 9:30 am: Situation analysis at a digital situation table with Fovea-Tablett, E. State Univ.; M. Vanninen, Network Appliance, Inc.; C. G. Healey, North Caroline Peinsipp-Byma, R. Eck, N. Rehfeld, J. Geisler, Fraunhofer Institut für State Univ...... [6495-06] Informations-und Datenverarbeitung (Germany) ...... [6495-12] 11:30 am: Supporting interactive graph exploration using edge plucking, N. 9:55 am: Parallel unstructured volume rendering in ParaView, K. Moreland, Wong, M. S. T. Carpendale, Univ. of Calgary (Canada) ...... [6495-07] Sandia National Labs.; L. Avila, Kitware Inc.; L. A. Fisk, Sandia National 11:55 am: Exploratory simulation for astrophysics, R. Walker, P. Kenny, J. Labs...... [6495-13] Miao, Univ. of Kent at Canterbury (United Kingdom) ...... [6495-08] Coffee Break ...... 10:20 to 10:50 am Lunch Break ...... 12:20 to 2:20 pm SESSION 6 SESSION 3 Conv. Ctr. Room B1 ...... Tues. 10:50 am to 12:05 pm Conv. Ctr. Room B1 ...... Mon. 2:20 to 3:10 pm Techniques for Large-scale Data Invited Paper I Chair: Jonathan C. Roberts, Univ. of Kent (United Kingdom) 2:20 pm: Visual analytics and knowledge visualization (Invited Paper, 10:50 am: Spatial prefetching for out-of-core visualization of Presentation Only), W. Ribarsky, The Univ. of North Carolina at multidimensional data, D. R. Lipsa, Armstrong Atlantic State Univ.; P. J. Charlotte ...... [6495-29] Rhodes, Univ. of Mississippi; R. D. Bergeron, T. M. Sparr, Univ. of New Coffee Break ...... 3:10 to 3:40 pm Hampshire ...... [6495-14] 11:15 am: NeuroVis: combining dimensional stacking, and pixelization to visually explore, analyze, and mine multidimensional multivariate data, J. T. Langton, Brandeis Univ...... [6495-15] 11:40 am: A modular extensible visualization system architecture for culled prioritized data streaming, J. P. Ahrens, N. Desai, P. S. McCormick, Los Alamos National Lab.; K. Martin, Kitware Inc.; J. Woodring, The Ohio State Univ...... [6495-16] Lunch/Exhibition Break ...... 12:05 to 2:05 pm

24 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6495

SESSION 7 Conv. Ctr. Room B1 ...... Tues. 2:05 to 2:55 pm Invited Paper II 2:05 pm: Maps as visualization systems (Invited Paper, Presentation Only), D. Rumsey, Luna Imaging ...... [6495-28] Coffee Break ...... 2:55 to 3:25 pm

SESSION 8 Conv. Ctr. Room B1 ...... Tues. 3:25 to 5:30 pm Visualization Techniques II Chair: Robert F. Erbacher, Utah State Univ. 3:25 pm: Patterns of coordination in improvise visualizations, C. Weaver, The Pennsylvania State Univ...... [6495-17] 3:50 pm: A new metaphor for projection-based visual analysis and data exploration, T. Schreck, Univ. Konstanz (Germany); C. Panse, ETH Zürich (Switzerland) ...... [6495-18] 4:15 pm: Visualization of rule productivity in deriving nonstandard spellings, S. Kempken, T. Pilz, W. Luther, Univ. Duisburg-Essen (Germany) . . . . [6495-19] 4:40 pm: Wavelet-based data-hiding of DEM in the context of real-time 3D visualization, K. Hayat, W. Puech, Lab. d’Informatique de Robotique et de Microelectronique de Montpellier (France); G. Gesquière, Univ. de Provence (France); M. Chaumont, Lab. d’Informatique de Robotique et de Microelectronique de Montpellier (France) ...... [6495-20] 5:05 pm: Analyzing sampled terrain volumetrically with regard to error and geologic variation, T. Butkiewicz, R. Chang, Z. J. Wartell, W. Ribarsky, The Univ. of North Carolina at Charlotte ...... [6495-21]

Posters and Demonstrations-Tuesday Demonstrations ...... 5:30 to 8:30 pm A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide interactive, hands-on demonstrations of a wide-range of products related to Electronic Imaging.

Posters ...... 5:30 to 7:00 pm Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, will be held Tuesday evening, 5:30 to 7:00 pm. ✔ GrSim: a system for network visualization and exploration, Y. Chiricota, Univ. du Quebec a Chicoutimi (Canada) ...... [6495-22] ✔ Extension of star coordinates into three dimensions, N. D. Cooprider, The Univ. of Utah; R. P. Burton, Brigham Young Univ...... [6495-23] ✔ A user-driven interface for exploring visualizations, D. H. Hepting, P. Schmiedge, Univ. of Regina (Canada) ...... [6495-24] ✔ A method for testing graph visualisations using games, J. D. Bovey, P. J. Rodgers, Univ. of Kent at Canterbury (United Kingdom) ...... [6495-25] ✔ Visualizing computer lab usage at Indiana University/Bloomington, K. Hanks, Indiana Univ.; J. E. Judd, Columbia Univ.; M. Henry, K. Brunette, Indiana Univ...... [6495-26] ✔ Multiple oil and gas volumetric data visualization with GPU programming, J. C. Lin, Landmark Graphics Corp...... [6495-27]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 25 Conference 6496 • Conv. Ctr. Room B3

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6496 Real-Time Image Processing 2007

Conference Chairs: Nasser Kehtarnavaz, The Univ. of Texas at Dallas; Matthias F. Carlsohn, Computer Vision and Image Communication (Germany) Program Committee: Mohamed Akil, École Supérieure d’Ingénieurs en Électronique et Électrotechnique (France); Carlos R. Castro-Pareja, Univ. of Maryland; Luciano F. da Fontoura Costa, Univ. de São Paolo (Brazil); Philip P. Dang, STMicroelectronics; Xavier Desurmont, Multitel A.S.B.L. (Belgium); Edward R. Dougherty, Texas A&M Univ.; Pierre Graebling, Univ. Louis Pasteur (France); Phillip A. Laplante, The Pennsylvania State Univ.; Sang-Yong Lee, Texas Instruments; Mehrube Mehrubeoglu, Texas A&M Univ.; Volodymyr I. Ponomaryov, Instituto Politécnico Nacional (Mexico); Fatih M. Porikli, Mitsubishi Electric Research Labs.

Monday 29 January SESSION 3 Conv. Ctr. Room B3 ...... Mon. 1:40 to 3:10 pm SESSION 1 Video and Compression Conv. Ctr. Room B3 ...... Mon. 8:30 to 10:00 am Chair: Volodymyr I. Ponomaryov, Instituto Politécnico Nacional Surveillance (Mexico) Chair: Nasser Kehtarnavaz, Univ. of Texas/Dallas 1:40 pm: A generic software-framework for distributed, high-performance processing of multiview video (Invited Paper), D. Farin, P. H. N. de With, 8:30 am: Two-dimensional statistical linear discriminant analysis for real- Technische Univ. Eindhoven (Netherlands) ...... [6496-10] time robust vehicle-type recognition (Invited Paper), E. A. Edirisinghe, I. Zaffar, Loughborough Univ. (United Kingdom) ...... [6496-01] 2:10 pm: Real-time stabilization of long-range observation system turbulent video, B. Fishbain, L. P. Yaroslavsky, I. A. Ideses, A. Shtern, O. Ben-Zvi, Tel Aviv 9:00 am: Determination of vehicle density from traffic images at day- and Univ. (Israel) ...... [6496-11] nighttime, M. Mehrubeoglu, L. McLauchlan, Texas A&M Univ...... [6496-02] 2:30 pm: Real-time aware rendering of scalable arbitrary shaped MPEG-4 9:20 am: Dual-camera system for acquisition of high-resolution images of decoder for multiprocessor systems, M. Pastrnak, P. H. N. de With, moving targets, J. Papon, R. W. Ives, R. P. Broussard, U.S. Naval LogicaCMG Nederland B.V. (Netherlands) and Technische Univ. Eindhoven Academy ...... [6496-03] (Netherlands); J. van Meerbergen, Philips Research Labs. (Netherlands) and 9:40 am: Camera position estimation method based on matching of top- Technische Univ. Eindhoven (Netherlands) ...... [6496-23] view images for running vehicle, T. Teshima, H. Saito, S. Ozawa, Keio Univ. 2:50 pm: Development of new image compression algorithm (Xena), Y. (Japan); K. Yamamoto, T. Ihara, Mitsubishi Fuso Truck and Bus Corp. Sugita, A. Watanabe, Fuji Photo Film Co., Ltd. (Japan) ...... [6496-13] (Japan) ...... [6496-24] Coffee Break ...... 3:10 to 3:40 pm Coffee Break ...... 10:00 to 10:30 am SESSION 4 SESSION 2 Conv. Ctr. Room B3 ...... Mon. 3:40 to 5:10 pm Conv. Ctr. Room B3 ...... Mon. 10:30 am to 12:10 pm Hardware Algorithms Chair: Fatih M. Porikli, Mitsubishi Electric Research Labs. Chair: Mehrube Mehrubeoglu, Texas A&M Univ. 3:40 pm: Real-time 3D video conference on generic hardware (Invited 10:30 am: Print-from-video: computationally efficient outlier reduction Paper), X. Desurmont, Multitel (Belgium) ...... [6496-14] pattern filtering (Invited Paper), Q. Peng, N. Kehtarnavaz, Univ. of Texas/ Dallas ...... [6496-05] 4:10 pm: Hardware-based JPEG2000 video coding system, A. Schuchter, Univ. Salzburg (Austria) ...... [6496-15] 10:50 am: A fast contour descriptor algorithm for supernova image classification, C. R. Aragon, Lawrence Berkeley National Lab.; D. B. Aragon, 4:30 pm: Three-dimensional color image processing procedures using DSP, DCA ...... [6496-06] V. I. Ponomaryov, Instituto Politécnico Nacional (Mexico) ...... [6496-16] 11:10 am: Computationally scalable fast distance transform computation, F. 4:50 pm: Light-sensitive high-speed line scan camera with digital time M. Porikli, Mitsubishi Electric Research Labs.; T. Kocak, Polytechnic delay integration, E. Bodenstorfer, J. Fürtler, J. Brodersen, K. J. Mayer, Austrian Univ...... [6496-07] Research Ctrs. Seibersdorf Research GmbH (Austria); C. Eckel, K. Gravogl, Oregano Systems Design & Consulting GesmbH (Austria); H. Nachtnebel, 11:30 am: Quantification of IR images for sparking defects in resistance Technische Univ. Wien (Austria); P. Rössler, Fachhochschule Technikum Wien welds through segmentation and marphological processing, N. M. (Austria) ...... [6496-17] Nandhitha, N. Manoharan, B. Sheela Rani, Sathyabama Deemed Univ. (India); B. Venkatraman, P. Kalyanasundaram, B. Raj, Indira Gandhi Ctr. for Atomic Research (India) ...... [6496-08] 11:50 am: Measuring the complexity of design in real-time imaging software, R. S. Sangwan, P. A. Laplante, The Pennsylvania State Univ.[6496-09] Lunch Break ...... 12:10 to 1:40 pm

26 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6496

Tuesday 30 January Awards and Plenary Presentation Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Development and Plasticity in Visual Cortex Brian A. Wandell, Stanford Univ. See pg. 6 for details.

Posters and Demonstrations-Tuesday Demonstrations ...... 5:30 to 8:30 pm A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide interactive, hands-on demonstrations of a wide-range of products related to Electronic Imaging.

Posters ...... 5:30 to 7:00 pm Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, will be held Tuesday evening, 5:30 to 7:00 pm. ✔ Motion object detection and tracking for traffic video analysis, I. Rajapandi, Anna Univ. (India) ...... [6496-19] ✔ Real-time speckle and impulsive noise suppression in 3D imaging based on robust linear combinations of order statistics, J. L. Varela-Benítez, F. J. Gallegos-Funes, V. I. Ponomaryov, Instituto Politécnico Nacional (Mexico) ...... [6496-20] ✔ Real-time quadtree analysis using HistoPyramids, G. Ziegler, R. Dimitrov, C. Theobalt, H. Seidel, Max-Planck-Institut für Informatik (Germany)[6496-21] ✔ Tracking object with radical color changes using modified mean shift, I. Whoang, Chung-Ang Univ. (South Korea); S. H. Chang, Spatial Integrated Systems, Inc.; K. Choi, Chung-Ang Univ. (South Korea) ...... [6496-22] ✔ The FPGA-based real-time image contrast adaptive linear enhancement algorithm, Y. Wang III, D. Wang, Y. Hu, Huazhong Univ. of Science and Technology (China) ...... [6496-25] ✔ On the use of real-time agents in distributed video analysis systems, B. Lienard, X. Desurmont, Multitel (Belgium) ...... [6496-26] ✔ A real-time hierarchical rule-based approach for scale independent human face detection, J. Jun, H. H. S. Ip, City Univ. of Hong Kong (China) ...... [6496-27] ✔ Digital architecture for real-time processing in vision systems for control of traffic lights, J. Garcia-Lamont, Instituto Politécnico Nacional (Mexico) ...... [6496-28]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 27 Conference 6497 • Conv. Ctr. Room C2

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6497 Image Processing: Algorithms and Systems V

Conference Chairs: Jaakko T. Astola, Tampere Univ. of Technology (Finland); Karen O. Egiazarian, Tampere Univ. of Technology (Finland); Edward R. Dougherty, Texas A&M Univ. Program Committee: Til Aach, RWTH Aachen (Germany); Sos S. Agaian, The Univ. of Texas at San Antonio; Junior Barrera, Univ. de São Paulo (Brazil); Reiner Creutzburg, Fachhochschule Brandenburg (Germany); Paul D. Gader, Univ. of Florida; Atanas P. Gotchev, Tampere Univ. of Technology (Finland); John C. Handley, Xerox Corp.; Vladimir V. Lukin, National Aerospace Univ. (Ukraine); Stephen Marshall, Univ. of Strathclyde (United Kingdom); Françoise J. Prêteux, Institut National des Télécommunications (France); Giovanni Ramponi, Univ. degli Studi di Trieste (Italy); Jagath K. Samarabandu, The Univ. of Western Ontario (Canada); Akira Taguchi, Musashi Institute of Technology (Japan)

Monday 29 January SESSION 4 Conv. Ctr. Room C2 ...... Mon. 3:30 to 5:10 pm SESSION 1 Applications and Systems II Conv. Ctr. Room C2 ...... Mon. 8:40 to 10:00 am Chair: Patrizio Campisi, Univ. degli Studi di Roma Tre (Italy) Methods and Algorithms I 3:30 pm: Crop/weed discrimination in simulated agronomic images, G. Jones, C. Gee, Etablissement National d’Enseignement Superieur Agronomique Chair: Karen O. Egiazarian, Tampere Univ. of Technology (Finland) de Dijon (France); F. Truchetet, Univ. de Bourgogne (France) ...... [6497-14] 8:40 am: Multilevel surround inhibition: a biologically inspired contour 3:50 pm: Image-based motion stabilization for maritime surveillance, D. D. detector, G. Papari, Rijksuniv. Groningen (Netherlands); P. Campisi, Univ. degli Morris, Northrop Grumman Corp.; B. R. Colonna, F. D. Snyder, General Studi Roma Tre (Italy); N. Petkov, Rijksuniv. Groningen (Netherlands) . [6497-01] Dynamics Robotic Systems ...... [6497-15] 9:00 am: Stochastic resonance investigation of object detection in images, 4:10 pm: Adaptive multi-histogram equalization using human vision D. W. Repperger, A. R. Pinkus, Air Force Research Lab.; J. A. Skipper, Wright thresholding, E. J. Wharton, Tufts Univ.; S. S. Agaian, The Univ. of Texas at San State Univ...... [6497-02] Antonio; K. A. Panetta, Tufts Univ...... [6497-16] 9:20 am: Combining multiple similarity metrics for corner matching, H. A. 4:30 pm: Using block-based spectral angle measure for iris pattern Khater, Univ. of Kent at Canterbury (United Kingdom) ...... [6497-03] matching, Y. Du, D. D. Isklandar, Indiana Univ.-Purdue Univ. 9:40 am: Detection of junction in images, Y. Li, R. L. Stevenson, J. Gai, Univ. Indianapolis ...... [6497-17] of Notre Dame ...... [6497-04] 4:50 pm: Dynamic content-based vehicle tracking and traffic monitoring Coffee Break ...... 10:00 to 10:30 am system, F. R. Bowen, Y. Du, Indiana Univ.-Purdue Univ. at Indianapolis; S. Li, Indiana Dept. of Transportation; Y. Jiang, Purdue Univ.; T. Nantung, S. Noureldin, Indiana Dept. of Transportation; M. J. Knieser, Indiana Univ.-Purdue Univ. at SESSION 2 Indianapolis ...... [6497-18] Conv. Ctr. Room C2 ...... Mon. 10:30 to 11:50 am Methods and Algorithms II Tuesday 30 January Chair: Sos S. Agaian, The Univ. of Texas at San Antonio Awards and Plenary Presentation 10:30 am: Image mosaicking via multiresolution analysis and cut line definition, L. M. G. Fonseca, V. Blagi, Instituto Nacional de Pesquisas Espaciais Marriott Ballroom 1-6 ...... 8:15 to 9:10 am (Brazil) ...... [6497-06] Development and Plasticity in Visual Cortex 10:50 am: Color filter array interpolation based on spatial adaptivity, D. V. Brian A. Wandell, Stanford Univ. Paliy, Tampere Univ. of Technology (Finland); B. C. Radu, Nokia Research Ctr. (Finland); V. Katkovnik, Tampere Univ. of Technology (Finland); M. Vahvilainen, See pg. 6 for details. Nokia Research Ctr. (Finland) ...... [6497-07] 11:10 am: Hierarchical texture motifs, S. D. Newsam, Lawrence Livermore SESSION 5 National Lab...... [6497-08] Conv. Ctr. Room C2 ...... Tues. 9:30 to 11:40 am 11:30 am: Dust and scratch removal in scanned images, R. Bergman, H. Methods and Algorithms III Nachlieli, G. Ruckenstein, Hewlett-Packard Labs. (Israel); D. Greig, Hewlett- Packard Ltd. (United Kingdom) ...... [6497-09] Chair: Reiner Creutzburg, Fachhochschule Brandenburg (Germany) Lunch Break ...... 11:50 am to 1:40 pm 9:30 am: Edge and corner preserving smoothing for artistic imaging, G. Papari, N. Petkov, Rijksuniv. Groningen (Netherlands); P. Campisi, Univ. degli Studi Roma Tre (Italy) ...... [6497-19] SESSION 3 9:50 am: Demosaicing of noisy data: spatially adaptive approach, D. V. Paliy, Conv. Ctr. Room C2 ...... Mon. 1:40 to 3:00 pm Tampere Univ. of Technology (Finland); M. Trimeche, Nokia Research Ctr. Applications and Systems I (Finland); V. Katkovnik, Tampere Univ. of Technology (Finland); S. Alenius, Nokia Research Ctr. (Finland) ...... [6497-20] Chair: Atanas P. Gotchev, Tampere Univ. of Technology (Finland) Coffee Break ...... 10:10 to 10:40 am 1:40 pm: Cloud classification of satellite image performed in two stages, H. Ikeda, M. Matsumoto, S. Hashimoto, Waseda Univ. (Japan) ...... [6497-10] 10:40 am: Prediction of signs of DCT coefficients in block-based lossy image compression, N. N. Ponomarenko, National Aerospace Univ. (Ukraine); Histogram-based template matching for object detection in 2:00 pm: A. V. Bazhyna, K. O. Egiazarian, Tampere Univ. of Technology (Finland) [6497-21] images with varying contrast, C. D. Schrider, J. A. Skipper, Wright State Univ...... [6497-11] 11:00 am: Nonlinear mapping for dynamic range compression in digital images, G. Guarnieri, S. Carrato, G. Ramponi, Univ. degli Studi di Trieste 2:20 pm: Iris identification using contourlet transform, R. F. Zewail, M. K. (Italy) ...... [6497-22] Mandal, N. G. Durdle, Univ. of Alberta (Canada) ...... [6497-12] 11:20 am: Wavelet-based texture image classification using vector Complex sinusoidally modulated imaging for real-time pixel-wise 2:40 pm: quantization, E. P. Lam, Raytheon Co...... [6497-23] optical flow detection, D. Wei, P. Masurel, T. Kurihara, A. Shigeru, The Univ. of Tokyo (Japan) ...... [6497-13] Lunch/Exhibition Break ...... 11:40 am to 1:30 pm Coffee Break ...... 3:00 to 3:30 pm

28 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6497

SESSION 6 ✔ Algorithms for the resizing of binary and grayscale images using a logical transform, E. E. Danahy, Tufts Univ.; S. S. Agaian, The Univ. of Texas Conv. Ctr. Room C2 ...... Tues. 1:30 to 2:10 pm at San Antonio; K. A. Panetta, Tufts Univ...... [6497-40] Classification, Recognition, and Feature Extraction ✔ Texture classification by ICA, D. Coltuc, Univ. Politehnica Bucuresti Chair: Nasser M. Nasrabadi, Army Research Lab. (Romania) ...... [6497-41] 1:30 pm: Classification SAR targets with support vector machine, L. Cao, ✔ A fuzzy system approach for the classification of underwater AUV color Chinese Leihua Electronic Technology Research Institute (China) . . . . [6497-24] images, J. A. Díaz, R. E. Torres, Univ. de Puerto Rico Mayagüez . . [6497-42] 1:50 pm: Adaptive color space transform using independent component ✔ Object classification, segmentation, and parameter estimation in analysis, E. Vera, Univ. of Concepcion (Chile) ...... [6497-26] multichannel images by classifier learning with clustering of local parameters, V. V. Lukin, N. N. Ponomarenko, A. A. Zelensky, National Aerospace Univ. (Ukraine); J. T. Astola, K. O. Egiazarian, Tampere Univ. of SESSION 7 Technology (Finland) ...... [6497-43] Conv. Ctr. Room C2 ...... Tues. 2:10 to 3:10 pm ✔ Automated image registration and fusion for thermal and visible images, Enhancement, Landmark Recognition, Feature Learning D. R. Deepti, Bharat Electronic Ltd. (India) ...... [6497-44] Chair: Syed A. Rizvi, CUNY/College of Staten Island ✔ Motion deblurring for suppression of breathing distortions caused by 2:10 pm: Neural network-based multiscale image restoration approach, A. camera shake, I. Tsubaki, T. Komatsu, T. Saito, Kanagawa Univ. P. A. d. Castro, J. D. S. Silva, Instituto Nacional de Pesquisas Espaciais (Japan) ...... [6497-45] (Brazil) ...... [6497-27] ✔ Fingerprint image enhancement and denoising, A. R. Upadhyay, Asia 2:30 pm: Landmarks recognition for autonomous aerial navigation by Pacific Institute of Information Technology (India) ...... [6497-49] neural networks and Gabor transform, E. H. Shiguemori, M. P. Martins, M. V. ✔ Rank M-type radial basis functions network for medical image T. Monteiro, F. L. L. Medeiros, M. A. P. Domiciano, Comando-Geral de processing applications, J. A. Moreno-Escobar, F. J. Gallegos-Funes, V. I. Tecnologia Aeroespacial (Brazil) ...... [6497-28] Ponomaryov, Instituto Politécnico Nacional (Mexico) ...... [6497-46] 2:50 pm: Autonomous learning in gesture recognition by using lobe ✔ Scanned image enhancement using multilayer neural networks, M. N. component analysis, J. Lu, National Institute of Occupational Safety and Ahmed, Lexmark International Inc...... [6497-47] Health (Japan) ...... [6497-29] ✔ Identification of vegetable diseases using neural network, J. Zhang, J. Coffee Break ...... 3:10 to 3:40 pm Tang, L. Yao, Beijing Normal Univ. (China) ...... [6497-48]

SESSION 8 Wednesday 31 January Conv. Ctr. Room C2 ...... Tues. 3:40 to 4:20 pm Evolutionary Programming, Filtering, and Enhancement Plenary Session Chair: Nasser M. Nasrabadi, Army Research Lab. 3:40 pm: A new design approach to neural network pattern recognition Marriott Ballroom 1-6 ...... 8:00 to 9:20 am systems, C. J. Hu, Southern Illinois Univ./Carbondale ...... [6497-30] 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the 4:00 pm: Speckle noise filtering with edge preservation of medical Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc ultrasound images, A. R. Upadhyay, R. K. Choudhary, Asia Pacific Institute of Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Information Technology (India); S. N. Talbar, Shri Guru Gobind Singhji Institute of Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Engineering & Technology (India) ...... [6497-32] Mottin, C.re de Recherche et de Restauration des Musées de France 8:40 am: Space Exploration, Imaging Technology, and the Development Posters and Demonstrations-Tuesday of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Demonstrations ...... 5:30 to 8:30 pm Solar System Visualization (SSV) Project A symposium-wide demonstration session will be open to attendees 5:30 to See pg. 7 for details. 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide interactive, hands-on demonstrations of a wide-range of products related to Electronic Imaging.

Posters ...... 5:30 to 7:00 pm Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, will be held Tuesday evening, 5:30 to 7:00 pm. ✔ Green noise halftoning with dot diffusion, S. Lippens, Univ. Gent (Belgium) and Institute for the Promotion of Innovation through Science and Technology in Flanders (Belgium); W. R. Philips, Univ. Gent (Belgium) ...... [6497-33] ✔ Fractal video textures synthesis-by-analysis with directionality estimation, P. Campisi, E. Maiorana, A. Neri, Univ. degli Studi Roma Tre (Italy) ...... [6497-34] ✔ Entropic tracker of color objects, A. Akhriev, Russian Scientific Ctr. of Roentgenology and Radiology (Russia); V. Somikov, ELVEES Research Ctr. (Russia) ...... [6497-37] ✔ Background change detection in video with dynamic foreground activity using robust parameter estimation techniques, C. A. Jenkins, J. B. Jordan, New Mexico State Univ.; J. J. Carlson, Sandia National Labs...... [6497-38] ✔ An elastic graph matching approach for motion estimation on feature points in image sequences, R. Feghali, Communications Research Ctr. Canada (Canada) ...... [6497-39]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 29 Conference 6498 • Conv. Ctr. Room A4

Monday-Wednesday 29-31 January 2007 • Proceedings of SPIE Vol. 6498 Computational Imaging V

Conference Chairs: Charles A. Bouman, Purdue Univ.; Eric L. Miller, Northeastern Univ.; Ilya Pollak, Purdue Univ. Program Committee: Samit K. Basu, GE Global Research; Thomas S. Denney, Jr., Auburn Univ.; Peter C. Doerschuk, Purdue Univ.; Peyman Milanfar, Univ. of California/Santa Cruz; Joseph A. O’Sullivan, Washington Univ. in St. Louis; Zygmunt Pizlo, Purdue Univ.; Stanley J. Reeves, Auburn Univ.; Yongyi Yang, Illinois Institute of Technology; Yibin Zheng

Monday 29 January SESSION 3 Conv. Ctr. Room A4 ...... Mon. 1:40 to 3:20 pm SESSION 1 Segmentation and Shape Analysis Conv. Ctr. Room A4 ...... Mon. 8:30 to 10:20 am Chair: Ilya Pollak, Purdue Univ. Medical Imaging I 1:40 pm: Multiscale multigranular image segmentation, E. D. Kolaczyk, Chair: Eric L. Miller, Northeastern Univ. Boston Univ.; J. Ju, South Dakota State Univ.; S. Gopal, Boston Univ. [6498-09] Keynote 2:00 pm: Riemannian metrics on landmark-based shape spaces: theory and applications, M. Micheli, Brown Univ...... [6498-10] 8:30 am: Progress in electrical impedance tomography, D. Isaacson, J. C. Newell, G. Saunier, Rensselaer Polytechnic Institute ...... [6498-50] 2:20 pm: 4D segmentation of cardiac data using active surfaces with spatiotemporal shape priors, A. Abufadel, A. J. Yezzi, Jr., Georgia Institute of 9:00 am: Analysis of reconstructions in full view fluorescence molecular Technology; R. W. Schafer, Hewlett-Packard Labs...... [6498-11] tomography, D. E. Hyde, Northeastern Univ. and Massachusetts General Hospital; A. Soubret, J. Dunham, T. Lasser, Massachusetts General Hospital; E. 2:40 pm: Automated segmentation of alloy microstructures in serial section L. Miller, D. H. Brooks, Northeastern Univ.; V. Ntziachristos, Massachusetts images, J. Dumke, M. L. Comer, Purdue Univ...... [6498-12] General Hospital ...... [6498-01] 3:00 pm: Fast Mumford-Shah segmentation using image scale space 9:20 am: Image reconstruction for small animal SPECT with two opposing bases, C. V. Alvino, Univ. of Pennsylvania; A. J. Yezzi, Jr., Georgia Institute of half cones, Y. Zheng, Univ. of Virginia; H. Li, The Univ. of Texas M.D. Anderson Technology ...... [6498-13] Cancer Ctr.; J. Wang, A. V. Stolin, J. Pole, M. B. Williams, Univ. of Coffee Break ...... 3:20 to 3:50 pm Virginia ...... [6498-02] 9:40 am: Spatio-temporal reconstruction of dynamic gated cardiac SPECT, SESSION 4 Y. Yang, M. Jin, E. Gravier, Illinois Institute of Technology ...... [6498-03] Conv. Ctr. Room A4 ...... Mon. 3:50 to 6:10 pm 10:00 am: Image reconstruction performance as a function of model complexity using information geometry: application to transmission Compression and Compressed Sensing tomographic imaging, J. A. O’Sullivan, L. Xie, D. G. Politte, B. R. Whiting, Chair: Rebecca M. Willett, Duke Univ. Washington Univ. in St. Louis ...... [6498-04] 3:50 pm: A conformal approach to shape representation, M. Feiszli, Brown Coffee Break ...... 10:20 to 10:50 am Univ...... [6498-14] 4:10 pm: Efficient multispectral imaging using compressive sensing, D. SESSION 2 Takhar, J. N. Laska, K. F. Kelly, R. G. Baraniuk, Rice Univ...... [6498-15] Conv. Ctr. Room A4 ...... Mon. 10:50 am to 11:50 am 4:30 pm: Characterizing compression efficiency of sparse image Mosaicing, Interpolation, Superresolution representation techniques, A. Zakhor, Univ. of California/Berkeley . [6498-16] Chair: Mary L. Comer, Purdue Univ. 4:50 pm: Variational methods for compressive sampling, J. Romberg, Georgia Institute of Technology; F. Park, Univ. of Michigan; E. J. Candes, 10:50 am: Creating panoramas on mobile phones, J. J. Boutellier, O. J. California Institute of Technology ...... [6498-17] Silvén, Oulun Yliopisto (Finland); M. Tico, M. Vehviläinen, Nokia Research Ctr. (Finland) ...... [6498-05] 5:10 pm: Compression of noisy Bayer pattern color filter array images, A. V. Bazhyna, K. O. Egiazarian, Tampere Univ. of Technology (Finland); N. N. 11:10 am: Multidimensional image enhancement from a set of unregistered Ponomarenko, V. V. Lukin, National Aerospace Univ. (Ukraine) ...... [6498-18] differently exposed images, P. Vandewalle, L. Meylan, S. E. Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland) ...... [6498-06] 5:30 pm: Multiscale reconstruction for computational spectral imaging, R. M. Willett, Duke Univ...... [6498-19] 11:30 am: Graph-based multiple panorama extraction from unordered image sets, A. Sibiryakov, M. Bober, Mitsubishi Electric ITE B.V. (United 5:50 pm: Importance prioritized coding of aerial surveillance video using Kingdom) ...... [6498-07] H.264, D. Butler, H. Cheng, J. Cook, Sarnoff Corp...... [6498-52] Lunch Break ...... 11:50 am to 1:40 pm

30 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6498

Tuesday 30 January SESSION 8 Conv. Ctr. Room A4 ...... Tues. 4:50 to 5:30 pm Awards and Plenary Presentation Multigrid Methods Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Chair: Charles A. Bouman, Purdue Univ. Development and Plasticity in Visual Cortex 4:50 pm: Multigrid optimal mass transport for image registration and Brian A. Wandell, Stanford Univ. morphing, T. U. Rehman, A. R. Tannenbaum, Georgia Institute of Technology ...... [6498-32] See pg. 6 for details. 5:10 pm: The effect of intergrid operators on multigrid convergence, P. SESSION 5 Navarrete Michelini, E. J. Coyle, Purdue Univ...... [6498-33] Conv. Ctr. Room A4 ...... Tues. 9:30 to 11:50 am Posters and Demonstrations-Tuesday Hardware Systems for Imaging and Tomography Chair: Samit K. Basu, GE Global Research Demonstrations ...... 5:30 to 8:30 pm Keynote A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide 9:30 am: Why do GPUs work so well for acceleration of CT? (Invited Paper), interactive, hands-on demonstrations of a wide-range of products related to K. D. Mueller, F. Xu, N. Neophytou, Stony Brook Univ...... [6498-20] Electronic Imaging. 10:00 am: Hardware-accelerated deformable image registration for image- guided interventions, R. Shekhar, O. Dandekar, Univ. of Maryland Medical Posters ...... 5:30 to 7:00 pm Ctr...... [6498-21] Posters will be placed on display after 10:00 am in Exhibit Hall 1. Coffee Break ...... 10:20 to 10:50 am A poster session, with authors present at their posters, 10:50 am: Tomographic image reconstruction using the cell broadband will be held Tuesday evening, 5:30 to 7:00 pm. engine (CBE) general purpose hardware, M. Kachelriess, M. Knaup, S. ✔ A computational 3D microscope model for reconstruction of translucent Steckmann, Friedrich-Alexander-Univ. Erlangen-Nürnberg (Germany); O. neural stem cells in bright-field time-lapse microscopy, J. H. Degerman, Bockenbach, Mercury Computer Systems GmbH (Germany) ...... [6498-22] Chalmers Tekniska Högskola (Sweden); E. Winterfors, Univ. Pierre et Marie 11:10 am: Iterative reconstruction of cone-beam CT data on a cluster, T. M. Curie (France); T. Gustavsson, Chalmers Tekniska Högskola Benson, GE Global Research; J. Gregor, The Univ. of Tennessee . . . . [6498-23] (Sweden) ...... [6498-44] 11:30 am: Evolution of computer technology for fast cone-beam ✔ Empirical performance evaluation of dissimilarity measures for use in backprojection, I. Goddard, A. Berman, Mercury Computer Systems, Inc.; O. urban structural damage detection, Z. Chen, T. C. Hutchinson, Univ. of Bockenbach, Mercury Computer Systems GmbH (Germany); F. P. Lauginiger, California/Irvine ...... [6498-45] Mercury Computer Systems, Inc.; S. Schuberth, Mercury Computer Systems ✔ Superresolution single-particle Cryo-EM tomography of icosahedral GmbH (Germany); S. Thieret, Mercury Computer Systems, Inc...... [6498-51] virus using compressed sensing theory, M. Kim, J. Choi, L. Yu, J. Ye, Lunch/Exhibition Break ...... 11:50 am to 1:40 pm Korea Advanced Institute of Science and Technology (South Korea) [6498-46] ✔ GPU-based visualization techniques for 3D microscopic imaging data, SESSION 6 Q. Wang, Y. Sun, J. P. Robinson, Purdue Univ...... [6498-47] Conv. Ctr. Room A4 ...... Tues. 1:40 to 3:20 pm ✔ GICEB: automatic segmentation algorithm for biomedical images, Q. Wang, Y. Sun, Purdue Univ...... [6498-48] Restoration, Filtering, and Reconstruction Chair: Peyman Milanfar, Univ. of California/Santa Cruz SESSION 9 1:40 pm: Higher order bilateral filters and their applications, H. Takeda, S. Conv. Ctr. Room A4 ...... Tues. 5:30 to 5:50 pm Farsiu, P. Milanfar, Univ. of California/Santa Cruz ...... [6498-24] Imaging in Materials Science 2:00 pm: The use of levelable regularization functions for MRF restoration of SAR images while preserving reflectivity, J. Darbon, EPITA (France); M. Chair: Charles A. Bouman, Purdue Univ. Sigelle, F. Tupin, École Nationale Supérieure des Télécommunications 5:30 pm: Using transmission electron microscopy to quantify the spatial (France) ...... [6498-25] distribution of nanoparticles suspended in a film, S. D. Newsam, E. Pernice, 2:20 pm: Nonlocal evolutions for image regularization, G. Gilboa, S. J. Osher, J. Jasinski, V. Leppert, Univ. of California/Merced ...... [6498-34] Univ. of California/Los Angeles ...... [6498-26] 2:40 pm: Parametric point spread function modeling and reduction of stray light effects, B. Bitlis, Purdue Univ.; P. A. Jansson, The Univ. of Arizona; J. P. Allebach, Purdue Univ...... [6498-27] 3:00 pm: An image decomposition model using the total variation and the infinity Laplacian, L. A. Vese, C. Elion, Univ. of California/Los Angeles; J. Morel, École Normale Supérieure de Cachan (France) ...... [6498-28] Coffee Break ...... 3:20 to 3:50 pm

SESSION 7 Conv. Ctr. Room A4 ...... Tues. 3:50 to 4:50 pm Tracking, Recognition, and Scene Analysis Chair: Zygmunt Pizlo, Purdue Univ. 3:50 pm: Object tracking using joint-visible and thermal-infrared video sequences, R. Péteri, O. Siler, E. Bichot, P. Courtellemont, Univ. de La Rochelle (France) ...... [6498-29] 4:10 pm: Symmetry detection in 3D scenes, T. Sawada, K. Sanik, Z. Pizlo, Purdue Univ...... [6498-30] 4:30 pm: Object recognition investigation using information-theoretic measures/metrics, D. W. Repperger, A. R. Pinkus, Air Force Research Lab.; J. A. Skipper, Wright State Univ...... [6498-31]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 31 Conference 6498

Wednesday 31 January SESSION 12 Conv. Ctr. Room A4 ...... Wed. 1:30 to 2:50 pm Plenary Session Medical Imaging III 1:30 pm: High-resolution image reconstruction for PET using detector Marriott Ballroom 1-6 ...... 8:00 to 9:20 am response, M. S. Tohme, J. Qi, Univ. of California/Davis ...... [6498-39] 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the 1:50 pm: Superresolution projection reconstruction MR imaging using Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc FOCUSS (PR-FOCUSS), S. Tak, J. Ye, Korea Advanced Institute of Science & Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Technology (South Korea) ...... [6498-40] Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno 2:10 pm: Fast methods for tomographic reconstruction, J. Thibault, GE Mottin, C.re de Recherche et de Restauration des Musées de France Healthcare; Z. Yu, Purdue Univ.; K. Sauer, Univ. of Notre Dame; C. A. Bouman, 8:40 am: Space Exploration, Imaging Technology, and the Development Purdue Univ.; J. Hsieh, GE Healthcare ...... [6498-41] of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, 2:30 pm: Statistical modeling of the myosin lattice in vertabrate muscle, C. Solar System Visualization (SSV) Project Yoon, R. P. Millane, N. D. Blakeley, A. Goyal, Univ. of Canterbury (New Zealand) ...... [6498-53] See pg. 7 for details. SESSION 13 SESSION 10 Conv. Ctr. Room A4 ...... Wed. 2:50 to 3:10 pm Conv. Ctr. Room A4 ...... Wed. 9:30 to 10:30 am Color Imaging Reconstruction, Inpainting, and Prediction 2:50 pm: Training-based algorithm for moiré suppression in scanned Chair: Minh N. Do, Univ. of Illinois at Urbana-Champaign halftone images, H. Siddiqui, C. A. Bouman, Purdue Univ...... [6498-55] 9:30 am: Image inpainting based on energy minimization, N. Kawai, T. Sato, N. Yokoya, Nara Institute of Science and Technology (Japan) ...... [6498-35] 9:50 am: Signal reconstruction from a periodic nonuniform set of samples using H_infinity optimization, H. T. Nguyen, M. N. Do, Univ. of Illinois at Urbana-Champaign ...... [6498-36] 10:10 am: A multiscale statistical model for time series forecasting, W. Wang, Purdue Univ.; V. Portnoy, Jefferies & Co.; I. Pollak, Purdue Univ. [6498-54] Coffee Break ...... 10:30 to 11:00 am

SESSION 11 Conv. Ctr. Room A4 ...... Wed. 11:00 am to 12:00 pm Medical Imaging II Chair: Thomas Talavage, Purdue Univ. 11:00 am: Simultaneous surface and volume registration using harmonic maps, A. A. Joshi, D. W. Shattuck, Univ. of Southern California; P. Thompson, Univ. of California/Los Angeles; R. M. Leahy, Univ. of Southern California ...... [6498-49] 11:20 am: Regional approach to event-related fMRI data analysis using hemodynamic response modeling, L. Liu, A. Rao, T. Talavage, Purdue Univ...... [6498-37] 11:40 am: Fast joint estimation of magnitude, decay, and frequency from a single-shot MRI signal, W. Tang, S. J. Reeves, Auburn Univ.; D. B. Twieg, The Univ. of Alabama/Birmingham ...... [6498-38] Lunch/Exhibition Break ...... 12:00 to 1:30 pm

32 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6499 • Conv. Ctr. Room A4

Thursday 1 February 2007 • Proceedings of SPIE Vol. 6499 Vision Geometry XV

Conference Chairs: Longin Jan Latecki, Temple Univ.; David M. Mount, Univ. of Maryland/College Park; Angela Y. Wu, American Univ. Program Committee: Gady Agam, Illinois Institute of Technology; Gilles Bertrand, Groupe ESIEE (France); Atsushi Imiya, Chiba Univ. (Japan); Jack Koplowitz, Clarkson Univ.; Nathan S. Netanyahu, Bar Ilan Univ. (Israel); Zygmunt Pizlo, Purdue Univ.; Peter Veelaert, Hogeschool Gent (Belgium); Xiaodong Wu, The Univ. of Iowa

Tuesday 30 January SESSION 2 Conv. Ctr. Room A4 ...... Thurs. 10:20 am to 12:00 pm Posters and Demonstrations-Tuesday Surface Analysis and Reconstruction Demonstrations ...... 5:30 to 8:30 pm Chair: Peter F. Stiller, Texas A&M Univ. 10:20 am: Volumetric scattered data modeling based on modified Shepard’s A symposium-wide demonstration session will be open to attendees 5:30 to method, K. Lee, J. M. Paik, Handong Global Univ. (South Korea) . . . . [6499-05] 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide interactive, hands-on demonstrations of a wide-range of products related to 10:45 am: Incremental adaptive subdivision of mesh surfaces, G. Agam, S. Electronic Imaging. Suresh, Illinois Institute of Technology ...... [6499-06] 11:10 am: Three-dimensional mesh generation of an object from an image Posters ...... 5:30 to 7:00 pm by shape-from-shading and anisotropic Delaunay triangulation, M. Doi, Y. Takabe, Osaka Electro-Communication Univ. (Japan) ...... [6499-07] Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, Lunch Break ...... 12:00 to 1:30 pm will be held Tuesday evening, 5:30 to 7:00 pm. ✔ Perspex Machine IX: transreal analysis, A. A. Adams, J. A. D. W. Anderson, SESSION 3 The Univ. of Reading (United Kingdom); N. Volker, Univ. of Essex (United Conv. Ctr. Room A4 ...... Thurs. 1:30 to 3:35 pm Kingdom) ...... [6499-18] Object Recognition and Reconstruction ✔ Perspex Machine X: software development, S. Noble, B. A. Thomas, J. A. D. W. Anderson, The Univ. of Reading (United Kingdom) ...... [6499-19] Chair: Longin Jan Latecki, Temple Univ. ✔ Aerial lidar data classification using expectation maximization, S. K. 1:30 pm: Shear invariant 3D model retrieval, S. S. Naik, B. Prabhakaran, The Lodha, D. M. Fitzpatrick, D. P. Helmbold, Univ. of California/Santa Univ. of Texas at Dallas ...... [6499-09] Cruz ...... [6499-20] 1:55 pm: Reconstruction of shapes of 3D symmetrical objects by using planarity and compactness constraints, Y. Li, Z. Pizlo, Purdue Univ. [6499-10] Thursday 1 February 2:20 pm: Recognizing point configurations in full perspective, P. F. Stiller, K. Abbott, Texas A&M Univ...... [6499-11] SESSION 1 2:45 pm: Three-dimensional face recognition based on geodesic distances, S. Gupta, M. K. Markey, J. K. Aggarwal, A. C. Bovik, The Univ. of Texas/ Conv. Ctr. Room A4 ...... Thurs. 8:20 to 10:00 am Austin ...... [6499-12] Aspects of Vision Geometry 3:10 pm: Alternative face models for 3D face registration, A. A. Salah, N. Chair: David M. Mount, Univ. of Maryland/College Park Alyüz, L. Akarun, Bogaziçi Univ. (Turkey) ...... [6499-13] 8:20 am: Perspex Machine VIII: axioms of transreal arithmetic, J. A. D. W. Coffee Break ...... 3:35 to 4:00 pm Anderson, The Univ. of Reading (United Kingdom); N. Volker, Univ. of Essex (United Kingdom); A. A. Adams, The Univ. of Reading (United Kingdom) ...... [6499-01] SESSION 4 Conv. Ctr. Room A4 ...... Thurs. 3:50 to 5:30 pm 8:45 am: Fidelity analysis of mechanically aided copying/enlarging of Jan van Eyck’s ‘Portrait of Niccolò Albergati’, D. G. Stork, Ricoh Innovations, Inc.; Digital Geometry and Topology M. F. Duarte, Rice Univ...... [6499-02] Chair: Gady Agam, Illinois Institute of Technology 9:10 am: Analysis of difference in orientations and focal lengths of two 3:50 pm: Arithmetic characterization of polynomial-based discrete curves, arbitrary perspective viewing cameras, A. Ramachandran, Arizona State C. Fiorio, J. Toutant, Univ. Montpellier II (France) ...... [6499-14] Univ.; R. Balasubramanian, Indian Institute of Technology Roorkee (India); K. Swaminathan, S. Das, Indian Institute of Technology (India) ...... [6499-03] 4:15 pm: Topological feature extraction using algebraic topology, S. Derdar, Univ. de Sherbrooke (Canada); M. Allili, Bishop’s Univ. (Canada); D. Ziou, Univ. 9:35 am: Biologically motivated composite image sensor for deep-field de Sherbrooke (Canada) ...... [6499-15] target tracking, P. Melnyk, R. A. Messner, Univ. of New Hampshire . [6499-04] 4:40 pm: Complete boundary detection of textured objects via deformable Coffee Break ...... 10:00 to 10:20 am models, R. Dedic, Univ. de Sherbrooke (Canada); M. Allili, Bishop’s Univ. (Canada) ...... [6499-16] 5:05 pm: Finger-shape extraction and expansion by wavelet transform and hand-shape analysis and recognition, C. Su, Da Yeh Univ. (Taiwan) [6499-17]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 33 Conference 6500 • Conv. Ctr. Room C4

Tuesday-Thursday 30 January-1 February 2007 • Proceedings of SPIE Vol. 6500 Document Recognition and Retrieval XIV

Conference Chairs: Xiaofan Lin, Riya Inc.; Berrin A. Yanikoglu, Sabanci Univ. (Turkey) Program Committee: Jan P. Allebach, Purdue Univ.; Tim L. Andersen, Boise State Univ.; Apostolos Antonacopoulos, Univ. of Salford (United Kingdom); Elisa H. Barney Smith, Boise State Univ.; Kathrin Berkner, Ricoh Innovations, Inc.; Hui Chao, Hewlett-Packard Co.; Brian D. Davison, Lehigh Univ.; Xiaoqing Ding, Tsinghua Univ. (China); David S. Doermann, Univ. of Maryland/College Park; Steven J. Harrington, Xerox Corp.; Jianying Hu, IBM Thomas J. Watson Research Ctr.; Matthew F. Hurst, Intelliseek, Inc.; Hisashi Ikeda, Hitachi, Ltd. (Japan); Tapas Kanungo, IBM Almaden Research Ctr.; Daniel P. Lopresti, Lehigh Univ.; Thomas A. Nartker, Univ. of Nevada/Las Vegas; Sargur N. Srihari, SUNY/Univ. at Buffalo; Kazem Taghva, Univ. of Nevada/Las Vegas; George R. Thoma, National Library of Medicine

Tuesday 30 January Wednesday 31 January Awards and Plenary Presentation Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Best Student Paper Award Document Recognition and Retrieval XIV Best Student Paper Award Development and Plasticity in Visual Cortex Brian A. Wandell, Stanford Univ. An award will be presented for the best paper submitted by a student. The award will be presented Wednesday morning during the Plenary Session. See pg. 6 for details. IS&T and SPIE gratefully acknowledge Nuance Communications, Inc. for generously sponsoring this award. SESSION 1 Room: Conv. Ctr. Room C4 ...... Tues. 1:30 to 2:10 pm Invited Paper I Plenary Session 1:30 pm: Industrial OCR approaches: architecture, algorithms, and Marriott Ballroom 1-6 ...... 8:00 to 9:20 am adaptation techniques (Invited Paper), I. Marosi, Nuance-Recognita, Corp (Hungary) ...... [6500-01] 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, SESSION 2 Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Room: Conv. Ctr. Room C4 ...... Tues. 2:10 to 3:10 pm Mottin, C.re de Recherche et de Restauration des Musées de France

Classifiers 8:40 am: Space Exploration, Imaging Technology, and the Development 2:10 pm: A shape descriptor based on CSS, M. Yang, K. Kpalma, J. Ronsin, of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Institut National des Sciences Appliquées de Rennes (France) ...... [6500-02] Solar System Visualization (SSV) Project 2:30 pm: Frequency coding: an effective method for combining See pg. 7 for details. dichotomizers, S. Andra, G. Nagy, Rensselaer Polytechnic Institute; C. Liu, Institute of Automation (China) ...... [6500-03] 2:50 pm: A multi-evidence multi-engine OCR system, I. Zavorin, E. Borovikov, SESSION 4 A. Borovikov, CACI International Inc.; L. Hernandez, Army Research Lab.; K. M. Room: Conv. Ctr. Room C4 ...... Wed. 9:30 to 10:30 am Summers, M. I. Turner, CACI International Inc...... [6500-04] Image Processing Coffee Break ...... 3:10 to 3:30 pm 9:30 am: Shape from parallel geodesics for distortion correction of digital camera document images, K. Fujimoto, J. Sun, H. Takebe, M. Suwa, S. Naoi, SESSION 3 Fujitsu Labs., Ltd. (Japan) ...... [6500-09] Room: Conv. Ctr. Room C4 ...... Tues. 3:30 to 4:50 pm 9:50 am: Multispectral pattern recognition applied to x-ray fluorescence images of the Archimedes Palimpsest, D. M. Hansen, R. L. Easton, Jr., OCR Rochester Institute of Technology ...... [6500-10] 3:30 pm: Optimal interaction for style-constrained OCR, G. Nagy, Rensselaer 10:10 am: Degraded document image enhancement, G. Agam, G. Bal, Illinois Polytechnic Institute; S. Veeramachaneni, ITC-IRST (Italy) ...... [6500-05] Institute of Technology; G. Frieder, The George Washington Univ.; O. Frieder, 3:50 pm: Reading text in consumer digital photographs, V. Vanhoucke, S. B. Illinois Institute of Technology ...... [6500-11] Gokturk, Riya, Inc...... [6500-06] Coffee Break ...... 10:30 to 10:50 am 4:10 pm: Adding contextual information to improve character recognition on the Archimedes Palimpsest, D. J. Walvoord, R. L. Easton, Jr., R. L. Canosa, Rochester Institute of Technology ...... [6500-07] 4:30 pm: OCR result optimization based on pattern matching, J. Shang, C. Liu, X. Ding, Tsinghua Univ. (China) ...... [6500-08]

34 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6500

SESSION 5 SESSION 9 Room: Conv. Ctr. Room C4 ...... Wed. 10:50 am to 12:10 pm Room: Conv. Ctr. Room C4 ...... Thurs. 10:30 to 11:10 am Handwriting Recognition Invited Paper II 10:50 am: Curvelets based tools for the characterization and the 10:30 am: Document recognition serving people with disabilities (Invited classification of Humanistic and Middle-Ages manuscripts, G. Joutel, V. Paper), J. R. Fruchterman, The Benetech Initiative ...... [6500-26] Eglin, H. Emptoz, Institut National des Sciences Appliquées de Lyon (France) ...... [6500-12] SESSION 10 11:10 am: Interactive training for handwriting recognition in historical Room: Conv. Ctr. Room C4 ...... Thurs. 11:10 to 11:50 am document collections, D. J. Kennard, W. A. Barrett, Brigham Young Univ...... [6500-13] Information Extraction and Retrieval II 11:30 am: Online handwritten mathematical expression recognition, B. A. 11:10 am: Title extraction and generation from OCR’d documents, K. Yanikoglu, H. Buyukbayrak, A. Ercil, Sabanci Univ. (Turkey) ...... [6500-14] Taghva, J. Borsack, Univ. of Nevada/Las Vegas ...... [6500-27] 11:50 am: Recognition of degraded handwritten digits using dynamic 11:30 am: Content-based document image retrieval in complex document Bayesian networks, L. Likforman, M. Sigelle, École Nationale Supérieure des collections, G. Agam, S. Argamon, O. Frieder, D. Grossman, Illinois Institute of Télécommunications (France) ...... [6500-15] Technology; D. Lewis, David D. Lewis Consulting ...... [6500-28] Lunch/Exhibition Break ...... 12:10 to 1:40 pm Lunch Break ...... 11:50 am to 1:30 pm

SESSION 6 SESSION 11 Room: Conv. Ctr. Room C4 ...... Wed. 1:40 to 3:00 pm Room: Conv. Ctr. Room C4 ...... Thurs. 1:40 to 3:20 pm Digital Publishing Special Session I Segmentation 1:40 pm: Making the public domain universally accessable, A. Langley, D. S. 1:40 pm: A statistical approach to line segmentation in handwritten Bloomberg, Google Inc...... [6500-16] documents, M. Arivazhagan, H. Srinivasan, S. N. Srihari, Univ. at Buffalo ...... [6500-29] 2:00 pm: Pixel and semantic capabilities from an image object based document representation, M. J. Gormish, K. Berkner, M. P. Boliek, G. Feng, E. 2:00 pm: Segmentation and labeling of documents using conditional L. Schwartz, Ricoh Innovations, Inc...... [6500-17] random fields, S. R. Shetty, H. Srinivasan, M. J. Beal, S. N. Srihari, Univ. at Buffalo ...... [6500-30] 2:20 pm: Presentation of structured documents without a style sheet, S. J. Harrington, E. Wayman, Xerox Corp...... [6500-18] 2:20 pm: Online medical journal article layout analysis, J. Zou, D. X. Le, G. R. Thoma, National Library of Medicine ...... [6500-31] 2:40 pm: Content selection based on compositional image quality, P. Obrador, Hewlett-Packard Co...... [6500-19] 2:40 pm: Transcript mapping for handwritten Arabic documents, L. M. Lorigo, V. Govindaraju, Univ. at Buffalo ...... [6500-32] Coffee Break ...... 3:00 to 3:30 pm 3:00 pm: Document image content inventories, H. S. Baird, M. A. Moll, C. An, Lehigh Univ...... [6500-33] SESSION 7 Room: Conv. Ctr. Room C4 ...... Wed. 3:30 to 4:10 pm Digital Publishing Special Session II 3:30 pm: Generic architecture for professional authoring environments to export XML-based formats, F. Giannetti, Hewlett-Packard Co. (United Kingdom) ...... [6500-20] 3:50 pm: Cost estimating for commercial digital printing, M. G. Keif, California Polytechnic State Univ...... [6500-21]

Thursday 1 February

SESSION 8 Room: Conv. Ctr. Room C4 ...... Thurs. 9:00 to 10:00 am Information Extraction and Retrieval I 9:00 am: A novel approach for nonuniform list fusion, W. Yan, Univ. of California/Irvine ...... [6500-23] 9:20 am: Identification of comment-on sentences in online biomedical documents using support vector machines, I. C. Kim, D. X. Le, G. R. Thoma, National Library of Medicine ...... [6500-24] 9:40 am: Combining text clustering and retrieval for corpus adaptation, F. He, X. Ding, Tsinghua Univ. (China) ...... [6500-25] Coffee Break ...... 10:00 to 10:30 am

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 35 Conference 6501 • Conv. Ctr. Room A6

Wednesday-Thursday 31 January-1 February 2007 • Proceedings of SPIE Vol. 6501 Sensors, Cameras, and Systems for Scientific/ Industrial Applications X

Conference Chair: Morley Blouke, Ball Aerospace & Technologies Corp. Program Committee: Erik Bodegom, Portland State Univ.; Terrence S. Lomheim, The Aerospace Corp.; Kevin J. Matherson, Hewlett-Packard Co.; Gloria G. Putnam, Eastman Kodak Co.; Alice L. Reinheimer, e2v technologies inc.; Nobukazu Teranishi, Matsushita Electric Industrial Co., Ltd. (Japan); Penny G. Warren, Ball Aerospace & Technologies Corp.; Orly Yadid-Pecht, Univ. of Calgary (Canada)

Tuesday 30 January Wednesday 31 January Posters and Demonstrations-Tuesday Demonstrations ...... 5:30 to 8:30 pm Plenary Session A symposium-wide demonstration session will be open to attendees 5:30 to Marriott Ballroom 1-6 ...... 8:00 to 9:20 am 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide interactive, hands-on demonstrations of a wide-range of products related to 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Electronic Imaging. Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Posters ...... 5:30 to 7:00 pm Mottin, C.re de Recherche et de Restauration des Musées de France Posters will be placed on display after 10:00 am in Exhibit Hall 1. 8:40 am: Space Exploration, Imaging Technology, and the Development A poster session, with authors present at their posters, of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, will be held Tuesday evening, 5:30 to 7:00 pm. Solar System Visualization (SSV) Project ✔ Pigment-dispersed colored photo resists for the new generation high- resolution color filters for CMOS image sensors, J. Kim, K. Lee, E. Jeong, See pg. 7 for details. C. Lee, Cheil Industries Inc. (South Korea) ...... [6501-31] ✔ Qualitative measurement of average index profile of symmetric SESSION 1 microstructured optical fiber from microscopic imaging, E. Saucedo, V. P. Conv. Ctr. Room A6 ...... Wed. 9:30 am to 12:00 pm Minkovich, Ctr. de Investigaciones en Óptica, A.C. (Mexico) . . . . . [6501-32] CCD Imagers ✔ A practicable method for color classification using low-cost remission/ reflection sensors, R. Hauser, T. Klinger, C. Madritsch, Fachhochschule Chair: Morley Blouke, Ball Aerospace & Technologies Corp. Technikum Kärnten (Austria); D. Roemhild, D. Starke, CiS Institut für 9:30 am: 100-M pixel high-speed high-resolution CCD, K. L. Boggs, R. A. Mikrosensorik gGmbH (Germany); V. Mannheim, Fachhochschule Technikum Bredthauer, G. R. Bredthauer, Semiconductor Technology Associates Kärnten (Austria) ...... [6501-33] Inc...... [6501-01] ✔ A CMOS vision chip for a contrast-enhanced image using a logarithmic 9:50 am: Computation of dark frames in digital imagers, R. Widenhorn, APS and a switch-selective resistive network, J. Kong, S. Kim, D. Sung, S. Portland State Univ. and Digital Clarity Consultants; A. Rest, Cerro Tololo Inter- Seo, J. Shin, Kyungpook National Univ. (South Korea) ...... [6501-34] American Observatory; M. Blouke, Ball Aerospace & Technologies Corp.; R. L. ✔ Highly sensitive and area-efficient CMOS image sensor using a Berry, Digital Clarity Consultants; E. Bodegom, Portland State Univ. and Digital PMOSFET-type photodetector with a built-in transfer gate, S. Seo, K. Clarity Consultants ...... [6501-02] Kim, J. Kong, J. Shin, P. Choi, Kyungpook National Univ. (South 10:10 am: A novel pixel structure for 2-µm square pixel IT-CCD, G. Sakoda, Korea) ...... [6501-35] M. Sato, D. Sugimoto, K. Maari, S. Miyazawa, T. Takeda, M. Kimura, T. Fujioka, ✔ Microscopic x-ray imaging system for biomedical applications using Sony Corp. (Japan); M. Fujishima, Y. Kani, Sony LSI Design Inc. (Japan); H. synchrotron radiation, K. Umetani, Japan Synchrotron Radiation Research Kanbe, Sony Corp. (Japan) ...... [6501-03] Institute (Japan); M. Kobatake, A. Yamamoto, T. Yamashita, S. Imai, Kawasaki Coffee Break ...... 10:30 to 11:00 am Medical School (Japan) ...... [6501-37] 11:00 am: Smear noise and sensitivity estimation method for CCD image ✔ Imaging based procedures to evaluate harrowing actions on soil vegetal sensors having square cells less than 2.0 um, K. Kikuchi, S. Miyazawa, Y. waste, G. Bonifazi, Univ. degli Studi di Roma/La Sapienza (Italy); P. Uchida, H. Kamata, T. Hirayama, Sony Corp. (Japan) ...... [6501-04] Menesatti, M. Nillozza, Istituto Sperimentale per la Meccanizzazione Agricola (Italy) ...... [6501-38] 11:20 am: Development of fully depleted scientific CCD’s for astronomy, K. L. Boggs, Semiconductor Technology Associates Inc.; M. P. Lesser, The Univ. of ✔ Sensor for moving foreground detection: an address event based Arizona ...... [6501-05] embedded system, N. Milosevic, Austrian Research Ctrs. GmbH (Austria) ...... [6501-39] 11:40 am: Orthogonal transfer arrays for the Pan-STARRS gigapixel camera, B. E. Burke, MIT Lincoln Lab.; J. L. Tonry, Univ. of Hawaii; M. J. Cooper, A. H. Loomis, P. Doherty, D. J. Young, T. A. Lind, MIT Lincoln Lab.; P. M. Onaka, Univ. of Hawaii; D. J. Landers, J. L. Daneu, MIT Lincoln Lab. . [6501-41] Lunch/Exhibition Break ...... 12:00 to 1:20 pm

36 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6501

SESSION 2 10:40 am: A digital high-dynamic-range CMOS image sensor with multi- integration and pixel readout request, A. Guilvard, École Nationale Supérieure Conv. Ctr. Room A6 ...... Wed. 1:20 to 4:30 pm de l’Aéronautique et de l’Espace (France) and STMicroelectronics (France); J. Applications and Systems I Segura, STMicroelectronics (France); P. Magnan, P. Martin-Gonthier, École Nationale Supérieure de l’Aéronautique et de l’Espace (France) . . . . . [6501-20] Chair: Erik Bodegom, Portland State Univ. 11:00 am: Wide dynamic range active pixel sensor arrays for digital x-ray 1:20 pm: Improvement photoelectric conversion efficiency of red light in imaging using a-Si:H technology, J. Lai, A. Nathan, Univ. of Waterloo HARP film, T. Matsubara, Y. Ohkawa, K. Miyakawa, S. Suzuki, M. Kubota, N. (Canada); J. A. Rowlands, Sunnybrook and Women’s Health Sciences Ctr. Egami, K. Tanioka, NHK Science & Technical Research Labs. (Japan); K. Ogusu, (Canada) ...... [6501-21] A. Kobayashi, T. Hirai, T. Kawai, Hamamatsu Photonics K.K. (Japan) . [6501-06] 11:20 am: Demonstration of a low-voltage three-transistor-per-pixel CMOS 1:40 pm: Bifocal liquid lens zoom objective for mobile phone applications, imager based on a pulse-width-modulation readout scheme employed with F. C. Wippermann, P. Schreiber, A. H. Bräuer, Fraunhofer-Institut für Angewandte a one-transistor in-pixel comparator, S. Shishido, I. Nagahata, T. Sasaki, K. Optik und Feinmechanik (Germany); P. Craen, Varioptic SA (France) . [6501-07] Kagawa, M. Nunoshita, J. Ohta, Nara Institute of Science and Technology 2:00 pm: The intermediate size direct detection detector for electron (Japan) ...... [6501-22] microscopy, L. Jin, A. Milazzo, Univ. of California/San Diego; S. Kleinfelder, S. 11:40 am: A 960-fps sub-sampling object extraction CMOS image sensor Li, Univ. of California/Irvine; P. C. Leblanc, F. Duttweiler, J. C. Bouwer, S. T. with 12-bit column parallel ADCs and ALUs, Y. Motohashi, T. Kubo, H. Kanto, Peltier, M. H. Ellisman, N. Xuong, Univ. of California/San Diego . . . . . [6501-08] T. Tate, S. Sugawa, Tohoku Univ. (Japan) ...... [6501-23] 2:20 pm: A novel image-acquisition system using the optical-multiplex 12:00 pm: A temperature-resistant wide dynamic range CMOS image system, T. Narabu, Sony Corp. (Japan) ...... [6501-09] sensor, K. Mizobuchi, S. Adachi, T. Yamashita, S. Okamura, H. Oshikubo, Texas 2:40 pm: A method of automatic focusing based on fuzzy control, D. Shao, Instruments, Inc. (Japan); N. Akahane, S. Sugawa, Tohoku Univ. Harbin Institute of Technology ...... [6501-10] (Japan) ...... [6501-24] Coffee Break ...... 3:00 to 3:30 pm Lunch Break ...... 12:20 to 1:40 pm 3:30 pm: Position detection with hyperacuity using artificial compound eyes, A. Brueckner, J. Duparré, A. H. Bräuer, A. Tuennermann, Fraunhofer- SESSION 4 Institut für Angewandte Optik und Feinmechanik (Germany) ...... [6501-11] Conv. Ctr. Room A6 ...... Thurs. 1:40 to 3:00 pm 3:50 pm: High-dynamic-range video for photometric measurement of Systems and Applications II illumination, J. Unger, S. Gustavson, Linköpings Univ. (Sweden) . . . . [6501-12] Chair: Morley Blouke, Ball Aerospace & Technologies Corp. 4:10 pm: Thermal imaging based procedures for special concretes characterization, G. Bonifazi, M. Animali, L. Damiani, Univ. degli Studi di 1:40 pm: Characterization and system modeling of a 5-Mpixel CMOS array, Roma/La Sapienza (Italy) ...... [6501-14] D. Dorn, K. Cabanas-Holmen, Pelco, Inc...... [6501-42] 2:00 pm: Low-cost optical polarization sensor, P. Baxter, J. M. Raynor, STMicroelectronics (United Kingdom); D. Renshaw, The Univ. of Edinburgh Thursday 1 February (United Kingdom) ...... [6501-26] 2:20 pm: A diffractive multispectral image sensor with on- and off-die SESSION 3 signal processing and on-Die optics in 0.18-micron CMOS, C. J. Thomas, R. Conv. Ctr. Room A6 ...... Thurs. 8:30 am to 12:20 pm I. Hornsey, York Univ. (Canada) ...... [6501-27] CMOS Sensors 2:40 pm: Diffraction grating lens array, T. Korenaga, T. Ando, S. Moriguchi, F. Takami, Y. Takasu, S. Nishiwaki, M. Suzuki, M. Nagashima, Matsushita Electric Chair: Morley Blouke, Ball Aerospace & Technologies Corp. Industrial Co., Ltd. (Japan) ...... [6501-28] 8:30 am: Shared transistor architecture with diagonally connected pixels Coffee Break ...... 3:00 to 3:30 pm for a CMOS image sensor, Y. Kudoh, F. Koga, T. Abe, H. Taniguchi, M. Sato, H. Ishiwata, S. Ooki, R. Suzuki, H. Mori, Sony Corp. (Japan) ...... [6501-15] 8:50 am: Image sensor for objects detection using modulated light, H. SESSION 5 Taruki, A. Ohno, F. Ono, T. Hamamoto, Tokyo Univ. of Science (Japan); T. Sasaki, Conv. Ctr. Room A6 ...... Thurs. 3:30 to 4:30 pm T. Shirai, M. Sakai, The Nippon Signal Co., Ltd. (Japan) ...... [6501-16] Nondestructive Testing 9:10 am: Linear-logarithmic image sensor with low noise and no flickers, S. Chair: Morley Blouke, Ball Aerospace & Technologies Corp. Miyatake, M. Kusuda, T. Iwamoto, K. Kamon, K. Sato, Y. Serita, Konica Minolta Technology Ctr. (Japan) ...... [6501-17] 3:30 pm: Critical assessment of the integrity of thin organic films by shearography, K. J. Habib, Kuwait Institute for Scientific Research (Kuwait) 9:30 am: A wide-dynamic-range CMOS imager by hybrid use of active and [6501-29] passive pixel sensors, K. Kagawa, Y. Adachi, Nara Institute of Science and Technology (Japan); Y. Nose, H. Takashima, K. Tani, A. Wada, SANYO Electric 3:50 pm: Deformation determination of aircraft parts by photogrammetry, Co., Ltd. (Japan); M. Nunoshita, J. Ohta, Nara Institute of Science and M. Varshosaz, K.N. Toosi Univ. of Technology (Iran) ...... [6501-30] Technology (Japan) ...... [6501-18] 4:10 pm: Development of wavelength changeable multiband mixing visual 9:50 am: Small distance detective camera module, I. Oyama, T. Iijima, K. device and its application, F. Ding, Y. Chen, K. Chao, USDA Agricultural Imada, S. Tamaki, T. Suenaga, N. Imamura, T. Ota, M. Suzuki, M. Nagashima, Research Service ...... [6501-40] Matsushita Electric Industrial Co., Ltd. (Japan) ...... [6501-19] Coffee Break ...... 10:10 to 10:40 am

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 37 Conference 6502 • Conv. Ctr. Room A6

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6502 Digital Photography III

Conference Chairs: Russel A. Martin, Foveon, Inc.; Jeffrey M. DiCarlo, Hewlett-Packard Labs.; Nitin Sampat, Rochester Institute of Technology Program Committee: Eiji Atsumi, Nokia Japan Co., Ltd. (Japan); Ted J. Cooper, Foveon, Inc.; Michael A. Kriss, Consultant; Jingqiang Li, Qualcomm, Inc.; Kevin J. Matherson, Hewlett-Packard Co.; Ricardo J. Motta, Pixim, Inc.; Gloria G. Putnam, Eastman Kodak Co.; John R. Reinert-Nash, Lifetouch, Inc.; Brian G. Rodricks, Micron Technology, Inc.; Takahiro Saito, Kanagawa Univ. (Japan); Sabine E. Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland); Dietmar Wueller, Image Engineering (Germany); Feng Xiao, Motorola, Inc.

Monday 29 January SESSION 4 Conv. Ctr. Room A6 ...... Mon. 2:00 to 3:00 pm SESSION 1 Autofocus and White Balance Techniques Conv. Ctr. Room A6 ...... Mon. 8:30 to 9:40 am Chair: Jeffrey M. DiCarlo, Hewlett-Packard Labs. Mobile Imaging 2:00 pm: Autofocus survey: a comparison of algorithms, L. C. Shih, Cypress Chair: Russel A. Martin, Foveon, Inc. Semiconductor Corp...... [6502-10] 8:30 am: Mobile imaging (Invited Paper), K. Johnson, F. Xiao, J. Pincenti, 2:20 pm: Monte Carlo evaluation of ten focus measures, Y. Tian, Univ. of Motorola, Inc...... [6502-01] California/Berkeley ...... [6502-11] 9:00 am: Ultracompact optical zoom lens for mobile phone, K. Matsusaka, S. 2:40 pm: Combining strategies for white balance, F. Gasparini, R. Schettini, Ozawa, R. Yoshida, T. Yuasa, Y. Souma, Konica Minolta Opto, Inc. S. Bianco, Univ. degli Studi di Milano-Bicocca (Italy) ...... [6502-12] (Japan) ...... [6502-02] Coffee Break ...... 3:00 to 3:30 pm 9:20 am: Camera-motion and mobile imaging, F. Xiao, J. Pincenti, G. John, K. Johnson, Motorola, Inc...... [6502-03] SESSION 5 Conv. Ctr. Room A6 ...... Mon. 3:30 to 5:00 pm SESSION 2 Noise Processing in Cameras Conv. Ctr. Room A6 ...... Mon. 9:40 to 11:10 am Chair: Brian G. Rodricks, Micron Technology, Inc. CMOS Sensors 3:30 pm: Removal of signal-dependent noise for a digital camera, T. Saito, Y. Chair: John R. Reinert-Nash, Lifetouch, Inc. Ishii, R. Nosaka, T. Komatsu, Kanagawa Univ. (Japan) ...... [6502-13] 9:40 am: Modeling of pixel reset associated wide dynamic range CMOS log Coffee Break ...... 3:20 to 3:50 pm image sensor, M. Hashimoto, Omron Automotive Electronics Technology (Germany); Y. Tani, Omron Corp. (Japan); J. Sochos, Omron Automotive 3:50 pm: Adaptive color bleeding removal for video and still DCT Electronics Technology (Germany) ...... [6502-04] compressed sequences, A. Castorina, G. Spampinato, A. Bruna, A. Capra, STMicroelectronics (Italy) ...... [6502-14] 10:00 am: Leakage characteristics for the buried photodiode structure on CMOS image sensors, S. Lee, S. Lim, C. Lee, S. Bang, S. Kang, J. Park, J. 4:10 pm: Super-resolution total-variation decoding of JPEG-compressed Park, Dongbu Electronics Co., Ltd. (South Korea); R. A. Martin, S. Bae, Foveon, image data, T. Saito, T. Komatsu, Kanagawa Univ. (Japan) ...... [6502-15] Inc...... [6502-05] Coffee Break ...... 10:20 to 10:50 am SESSION 6 Conv. Ctr. Room A6 ...... Mon. 4:10 to 5:30 pm 10:50 am: CMOS color image sensors with silicon photodiode and overlaid organic photoconductive layer having narrow absorption band, S. Takada, Sensor Processing M. Ihama, Fuji Photo Film Co.,Ltd. (Japan); T. Komatsu, T. Saito, Kanagawa Chair: Ted J. Cooper, Foveon, Inc. Univ. (Japan) ...... [6502-06] 4:10 pm: A bad pixel location algorithm for cell phone cameras, M. Aleksic, S. R. Goma, ATI Technology (Canada) ...... [6502-16] SESSION 3 Conv. Ctr. Room A6 ...... Mon. 11:10 am to 12:10 pm 4:30 pm: An optic model of the appearance of blemishes in digital photographs, A. Zamfir, FotoNation Romania SRL (Romania); A. Drimbarean, Camera Image Processing FotoNation Ireland Ltd. (Ireland); V. V. Buzuloiu, M. Zamfir, Univ. Politehnica Chair: Nitin Sampat, Rochester Institute of Technology Bucuresti (Romania); E. Steinberg, FotoNation Inc.; D. V. Ursu, ProOptica (Romania) ...... [6502-17] 11:10 am: Matrixing, number of color channels, and color gamut of camera devices, F. Gasparini, R. Schettini, S. Bianco, Univ. degli Studi di Milano- 4:50 pm: Kernel-size selection for defect pixel identification and correction, Bicocca (Italy) ...... [6502-07] E. T. Chang, Cypress Semiconductor Corp...... [6502-18] 11:30 am: From spectral sensitivities to noise characteristics, F. Guichard, 5:10 pm: MEMS digital camera, R. C. Gutierrez, T. K. Tang, R. J. Calvet, E. R. DxO Labs. (France); J. Buzzi, École Polytechnique (France); H. Hornung, DxO Fossum, Siimpel Corp...... [6502-36] Labs. (France) ...... [6502-08] 11:50 am: Joint demosaicing and super-resolution imaging from a set of unregistered aliased images, P. Vandewalle, École Polytechnique Fédérale de Lausanne (Switzerland); D. Alleysson, Univ. Pierre-Mendes-France (France); S. E. Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland)[6502-09] Lunch Break ...... 12:10 to 1:40 pm

38 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6502

Tuesday 30 January Posters and Demonstrations-Tuesday Awards and Plenary Presentation Demonstrations ...... 5:30 to 8:30 pm Marriott Ballroom 1-6 ...... 8:15 to 9:10 am A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide Development and Plasticity in Visual Cortex interactive, hands-on demonstrations of a wide-range of products related to Brian A. Wandell, Stanford Univ. Electronic Imaging. See pg. 6 for details. Posters ...... 5:30 to 7:00 pm SESSION 7 Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, Conv. Ctr. Room A6 ...... Tues. 9:30 to 11:20 am will be held Tuesday evening, 5:30 to 7:00 pm. Camera Specifications/Measurements ✔ Fast nonlocal means for image denoising, R. C. Bilcu, M. Vehvilainen, Chair: Sabine E. Süsstrunk, École Polytechnique Fédérale de Nokia Research Ctr. (Finland) ...... [6502-26] Lausanne (Switzerland) ✔ CMOS image sensor noise reduction method for image signal processor 9:30 am: Establishing imaging sensor specifications for digital still in digital cameras and camera phones, Y. Yoo, S. Lee, W. Choe, C. Kim, cameras, M. A. Kriss, Consultant ...... [6502-19] SAMSUNG Advanced Institute of Technology (South Korea) . . . . . [6502-27] 9:50 am: Quality criterion for digital still camera, S. N. Bezryadin, KWE ✔ A novel color filter array with 75% transparent elements, G. Luo, Harvard International, Inc...... [6502-20] Medical School ...... [6502-28] Coffee Break ...... 10:10 to 10:40 am ✔ The usage of digital cameras as luminance meters, D. Wueller, Image Engineering (Germany); H. Gabele, FRAMOS Electronic Vertriebs GmbH 10:40 am: Digital camera resolution measurements by sine modulated (Germany) ...... [6502-29] Siemensstars, C. M. Loebich, Image Engineering (Germany); A. Jaeger, Canon Deutschland GmbH (Germany); B. Klingen, Fachhochschule Köln (Germany); D. ✔ Motion deblurring based on fusing differently exposed images, M. Tico, Wueller, Image Engineering (Germany) ...... [6502-21] M. Vehvilainen, Nokia Research Ctr. (Finland) ...... [6502-30] 11:00 am: Measurement method for image stabilizing systems, B. Golik, D. ✔ Image demosaic via Nevatia-Babu edge detection, C. C. Reinhart, J. Wueller, Image Engineering (Germany) ...... [6502-22] Bodie, California Lutheran Univ...... [6502-31] ✔ Automated digital camera sensor characterization, C. M. Normand, P. R. SESSION 8 Fornaro, R. Gschwind, Univ. of Basel (Switzerland) ...... [6502-32] Conv. Ctr. Room A6 ...... Tues. 11:20 am to 12:20 pm ✔ Identification of in-field defect development in digital image sensors, J. Dudas, L. M. Wu, C. G. Jung, D. C. Chen, G. H. Chapman, Simon Fraser Applications Univ. (Canada); I. Koren, Z. Koren, Univ. of Massachusetts/Amherst [6502-33] Chair: Dietmar Wueller, Image Engineering (Germany) ✔ Adaptive nonlinear imaging characteristic for wide-dynamic-range 11:20 am: New generation of image editing algorithms, S. N. Bezryadin, KWE image sensor using variable integration time, F. Suzuki, T. Hamamoto, International, Inc...... [6502-23] Science Univ. of Tokyo (Japan) ...... [6502-34] 11:40 am: Characteristics of experimental single-chip color HDTV image ✔ Illumination flicker detection and frequency classification methods, T. acquisition system with 8M-pixel image sensor, H. Shimamoto, R. Funatsu, T. Tajbakhsh, R. Grigat, Technische Univ. Hamburg-Harburg Yamashita, K. Mitani, Y. Nojiri, NHK Science & Technical Research Labs. (Germany) ...... [6502-35] (Japan) ...... [6502-24] 12:00 pm: Full spatial resolution RGB digital camera, R. Kumontoy, J. P. Allebach, G. T. C. Chiu, J. Huang, D. Howard, N. Bajaj, Purdue Univ. . [6502-25]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 39 Conference 6503 • Conv. Ctr. Room C3

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6503 Machine Vision Applications in Industrial Inspection XV

Conference Chairs: Fabrice Meriaudeau, Univ. de Bourgogne (France); Kurt S. Niel, Fachhochschule Wels (Austria) Program Committee: Pierrick T. Bourgeat, Commonwealth Scientific and Industrial Research Organisation (Australia); Luciano F. da Fontoura Costa, Univ. de São Paolo (Brazil); Marc M. Ellenrieder, Carl Zeiss AG (Germany); Steven P. Floeder, 3M Co.; David Fofi, Univ. de Bourgogne (France); Ralph M. Ford, The Pennsylvania State Univ.; Edmund Y. Lam, The Univ. of Hong Kong (Hong Kong China); Katia Lebart, Heriot-Watt Univ. (United Kingdom); Dinesh Nair, National Instruments; Paul L. O’Leary, Montan Univ. Leoben (Austria); Jeffery R. Price, Oak Ridge National Lab.; A. Ravishankar Rao, IBM Thomas J. Watson Research Ctr.; Joaquim Salvi, Univ. de Girona (Spain); Hamed Sari-Sarraf, Texas Tech Univ.; Ralph Seulin, Univ. de Bourgogne (France); Christoph Stiller, RWTH Aachen (Germany); Kenneth W. Tobin, Jr., Oak Ridge National Lab.; Yvon Voisin, Univ. de Bourgogne (France)

Monday 29 January SESSION 4 Conv. Ctr. Room C3 ...... Mon. 1:30 to 2:10 pm SESSION 1 3D Applications II Conv. Ctr. Room C3 ...... Mon. 8:30 to 10:10 am Chair: David Fofi, Univ. de Bourgogne (France) Industrial Applications 1:30 pm: Toward an automation of 3D acquisition and post-processing of Chair: Fabrice Meriaudeau, Univ. de Bourgogne (France) cultural heritage artefact, B. Loriot, R. Seulin, P. Gorria, Univ. de Bourgogne (France) ...... [6503-11] 8:30 am: Automatic characterization of cross-sectional coated particle nuclear fuel using greedy coupled Bayesian snakes, J. R. Price, D. Aykac, J. 1:50 pm: Appliance of stereoscopy for bath level detection of steel casting D. Hunn, A. K. Kercher, Oak Ridge National Lab...... [6503-01] pans, C. Rittenschober, K. S. Niel, Fachhochschule Wels (Austria); R. Rössler, voestalpine Stahl AG (Austria) ...... [6503-12] 8:50 am: Basic scanner for parallepipedic manufactured pieces, H. Hrimech, J. Pavéglio, F. S. Marzani, Y. Voisin, Univ. de Bourgogne (France) . . . . [6503-02] SESSION 5 9:10 am: Statistical modeling, detection, and segmentation of stains in digitized fabric images, A. Gururajan, H. Sari-Sarraf, E. F. Hequet, Texas Tech Conv. Ctr. Room C3 ...... Mon. 2:10 to 4:20 pm Univ...... [6503-03] Multiresolution and Mathematical Fitting 9:30 am: Compact multispectral imaging system for contaminant detection Chair: Kurt S. Niel, Fachhochschule Wels (Austria) on poultry carcasses, M. Kise, B. Park, K. C. Lawrence, W. R. Windham, USDA 2:10 pm: Evaluation of two and a half-dimensional surface data with form Agricultural Research Service ...... [6503-04] component and groove bands, B. Xin, Univ. Karlsruhe (Germany) . . [6503-13] 9:50 am: Chinese videotext detection based on directional gradient angular 2:30 pm: Application of edge field analysis to a blurred image, M. information, J. Peng, Xidian Univ. (China) ...... [6503-05] Matsumoto, S. Hashimoto, Waseda Univ. (Japan) ...... [6503-14] Coffee Break ...... 10:10 to 10:30 am 2:50 pm: Camera calibrationless vision calibration for transfer robot system, N. Kimura, T. Moriya, K. Matsumoto, Hitachi, Ltd. (Japan) . . [6503-15] SESSION 2 Coffee Break ...... 3:10 to 3:40 pm Conv. Ctr. Room C3 ...... Mon. 10:30 to 11:10 am 3:40 pm: Subpixel evaluation of distorted circles characterisitcs, F. Multispectral Imaging Mairesse, T. M. Sliwa, S. Binczak, Y. Voisin, Univ. de Bourgogne Chair: Kurt S. Niel, Fachhochschule Wels (Austria) (France) ...... [6503-16] 10:30 am: Spectroscopic imaging from 400 nm to 1800 nm with liquid 4:00 pm: Face recognition by using ring rotation invariant transform and crystal tunable filters, H. Rothe, C. F. Hahlweg, Helmut-Schmidt Univ. coping with image rotation and image shifting problems, C. Su, Da Yeh Univ. (Germany) ...... [6503-07] (Taiwan) ...... [6503-17] 10:50 am: Infrared imaging and machine vision, F. Meriaudeau, Univ. de Bourgogne (France) ...... [6503-08] SESSION 6 Conv. Ctr. Room C3 ...... Mon. 4:20 to 5:20 pm SESSION 3 HW Equipment Conv. Ctr. Room C3 ...... Mon. 11:10 to 11:50 am Chair: Fabrice Meriaudeau, Univ. de Bourgogne (France) 3D Applications I 4:20 pm: Latest developments in microoptical artificial compound eyes: a Chair: David Fofi, Univ. de Bourgogne (France) promising approach for next generation ultracompact machine vision, J. Duparré, P. Dannberg, A. H. Bräuer, Fraunhofer-Institut für Angewandte Optik 11:10 am: Integrating advanced 3D vision technology in manufacturing, A. und Feinmechanik (Germany) ...... [6503-18] Pichler, H. Bauer, C. Eberst, C. Heindl, J. Minichberger, Profactor Produktionsforschungs GmbH (Austria) ...... [6503-09] 4:40 pm: High-performance camera for fast quality inspection in industrial printing application, J. Fürtler, E. Bodenstorfer, K. J. Mayer, J. Brodersen, 11:30 am: High-performance surface inspection method for thin-film Austrian Research Ctrs. Seibersdorf Research GmbH (Austria); C. Eckel, K. sensors, V. V. W. Wieser, S. Larndorfer, B. Moser, F. Kossak, Software Gravogl, Oregano Systems Design & Consulting GmbH (Austria); H. Nachtnebel, Competence Ctr. Hagenberg (Austria) ...... [6503-10] Technische Univ. Wien (Austria); P. Roessler, Fachhochschule Technikum Wien Lunch Break ...... 11:50 am to 1:30 pm (Austria) ...... [6503-19] 5:00 pm: Setting up task-optimal illumination automatically for inspection purposes, J. J. Uusitalo, R. O. Tuokko, Tampere Univ. of Technology (Finland) ...... [6503-20]

40 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6503

Tuesday 30 January Awards and Plenary Presentation Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Development and Plasticity in Visual Cortex Brian A. Wandell, Stanford Univ. See pg. 6 for details.

Posters and Demonstrations-Tuesday Demonstrations ...... 5:30 to 8:30 pm A symposium-wide demonstration session will be open to attendees 5:30 to 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide interactive, hands-on demonstrations of a wide-range of products related to Electronic Imaging.

Posters ...... 5:30 to 7:00 pm Posters will be placed on display after 10:00 am in Exhibit Hall 1. A poster session, with authors present at their posters, will be held Tuesday evening, 5:30 to 7:00 pm. ✔ Design and implementation of confocal imaging systems with a generalized theoretical framework, G. Chang, C. Liao, National Taiwan Normal Univ. (Taiwan) ...... [6503-22] ✔ Real-time vehicle detection and tracking based on traffic scene analysis, Z. Zeng, S. Wang, X. Ding, Tsinghua Univ. (China) ...... [6503-23] ✔ 2DPCA face recognition based on block scan, C. Zhang, Sun Yat-Sen Univ. (China) ...... [6503-24] ✔ Machine vision system for the inspection of relective parts in the automotive industry, F. Meriaudeau, O. Morel, G. Salis, Univ. de Bourgogne (France) ...... [6503-26] ✔ Imaging based logics for ornamental stones quality chart definition, G. Bonifazi, A. Gargiulo, S. Serranti, Univ. degli Studi di Roma/La Sapienza (Italy) ...... [6503-21] ✔ Validation of a site-finding algorithm for NIF optic scattered-light images, M. Reed, Harvey Mudd College; L. M. Kegelmeyer, J. A. Liebman, M. Bolourchi, J. N. McElroy, R. R. Prasad, Lawrence Livermore National Lab...... [6503-27]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 41 Conference 6504 • Conv. Ctr. Room C1

Wednesday-Thursday 31 January-1 February 2007 • Proceedings of SPIE Vol. 6504 Multimedia Computing and Networking 2007

Conference Chairs: Roger Zimmermann, Univ. of Southern California; Carsten Griwodz, Univ. of Oslo (Norway) Program Committee: Tarek F. Abdelzaher, Univ. of Illinois/Urbana-Champaign; Surendar Chandra, Univ. of Notre Dame; David A. Cheok, National Univ. of Singapore (Singapore); Mark Claypool, Worcester Polytechnic Institute; David H. Du, Univ. of Minnesota; Wu-chi Feng, Portland State Univ.; Pascal Frossard, École Polytechnique Fédérale de Lausanne (Switzerland); Ahsan Habib, Siemens TTB Ctr./Berkeley; Pål Halvorsen, Simula Research Lab. (Norway); Martin Karsten, Univ. of Waterloo (Canada); Seon H. Kim, Univ. of Denver; Baochun Li, Univ. of Toronto (Canada); Kang Li, The Univ. of Georgia; Andreas U. Mauthe, Lancaster Univ. (United Kingdom); Ketan D. Mayer-Patel, The Univ. of North Carolina at Chapel Hill; Wei-Tsang Ooi, National Univ. of Singapore (Singapore); Reza Rejaie, Univ. of Oregon; Christoph Rensing, Technische Univ. Darmstadt (Germany); Utz Roedig, Lancaster Univ. (United Kingdom); Karsten Schwan, Georgia Institute of Technology; Nalini Venkatasubramanian, Univ. of California/ Irvine; Michael Zink, Univ. of Massachusetts/Amherst

In cooperation with: ACM SIGMultimedia

Wednesday 31 January SESSION 3 Conv. Ctr. Room C1 ...... Wed. 4:00 to 5:00 pm Plenary Session User Perception 4:00 pm: Olfactory enhanced multimedia applications: perspectives from an Marriott Ballroom 1-6 ...... 8:00 to 9:20 am empirical study, O. Ademoye, G. Ghinea, Brunel Univ. (United Kingdom) ...... [6504-09] 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc 4:30 pm: The effects of resolution on users playing first person shooter Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, games, K. Claypool, Oracle USA; M. Claypool, Worcester Polytechnic Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Institute ...... [6504-10] Mottin, C.re de Recherche et de Restauration des Musées de France Panel Discussion: Getting in Touch with the Real World 8:40 am: Space Exploration, Imaging Technology, and the Development of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Thursday 1 February Solar System Visualization (SSV) Project

See pg. 7 for details. SESSION 4 Conv. Ctr. Room C1 ...... Thurs. 8:30 to 10:00 am SESSION 1 Resource Management Conv. Ctr. Room C1 ...... Wed. 9:30 am to 12:30 pm 8:30 am: Efficient cost-based scheduling for scalable media streaming, N. Video Coding J. Sarhan, B. Qudah, Wayne State Univ...... [6504-11] 9:30 am: Optimal bit allocation for fine-grained scalable video, M. M. 9:00 am: Resource management in a shared infrastructure video CDN, A. J. Hefeeda, C. Hsu, Simon Fraser Univ. (Canada) ...... [6504-01] Cahill, C. J. Sreenan, National Univ. of Ireland/Cork (Ireland) ...... [6504-12] 10:00 am: A framework for MPEG-21 DIA-based adaptation and perceptual 9:30 am: Flash on disk for low-power multimedia computing, L. C. Singleton encryption of H.264 video, R. Iqbal, S. Shirmohammadi, A. El Saddik, Univ. of IV, R. Nathuji, K. Schwan, Georgia Institute of Technology ...... [6504-13] Ottawa (Canada) ...... [6504-02] Coffee Break ...... 10:00 to 10:30 am Coffee Break ...... 10:30 to 11:00 am 11:00 am: FMOE-MR: content-driven multiresolution MPEG-4 fine-grained SESSION 5 scalable layered video encoding, S. Chattopadhyay, X. Luo, S. M. Bhandarkar, Conv. Ctr. Room C1 ...... Thurs. 10:30 to 11:30 am K. Li, The Univ. of Georgia ...... [6504-03] 11:30 am: Impact of reference distance for motion compensation prediction Streaming Scalability on video quality, Y. Wang, EMC Corp.; M. Claypool, R. Kinicki, Worcester 10:30 am: On providing reliability guarantees in live video streaming with Polytechnic Institute ...... [6504-04] collaborative clients, A. Raghuveer, Univ. of Minnesota; Y. Dong, Univ. of Hawaii; D. H. Du, Univ. of Minnesota ...... [6504-14] 12:00 pm: Perceptual multimedia compression based on the predictive Kalman filter eye movement modeling, O. V. Komogortsev, J. I. Khan, Kent 11:00 am: Priority-progress CPU adaptation for elastic real-time State Univ...... [6504-05] applications, C. C. Krasic, The Univ. of British Columbia (Canada) . . [6504-15] Lunch/Exhibition Break ...... 12:30 to 2:00 pm SESSION 6 SESSION 2 Conv. Ctr. Room C1 ...... Thurs. 11:30 am to 12:30 pm Conv. Ctr. Room C1 ...... Wed. 2:00 to 3:30 pm Keynote Presentation Systems and Measurements 11:30 am: Web video search and the emergence of socially immersive media (Presentation Only), R. Sarukkai, Yahoo! Inc...... [6504-16] 2:00 pm: Share with thy neighbors, S. Chandra, X. Yu, Univ. of Notre Dame ...... [6504-06] 2:30 pm: Video to Go: the effects of mobility on streaming media in a CDMA2000 1xEV-DO network, P. Sessini, M. Leventer, A. Mahanti, Univ. of Calgary (Canada) ...... [6504-07] 3:00 pm: iTVP: large-scale content distribution for live and on-demand video services, E. Kusmierek, M. Czyrnek, C. Mazurek, M. Stroinski, Poznan Supercomputing and Networking Ctr. (Poland) ...... [6504-08] Coffee Break ...... 3:30 to 4:00 pm

42 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6504

SESSION 7 Conv. Ctr. Room C1 ...... Thurs. 2:00 to 4:15 pm Short Papers: Multimedia Systems 2:00 pm: CompatPM: enabling energy efficient multimedia workloads for distributed mobile platforms, R. Nathuji, K. J. O’Hara, K. Schwan, T. Balch, Georgia Institute of Technology ...... [6504-17] 2:15 pm: Efficient MPEG-21-based adaptation decision taking for scalable multimedia content, I. Kofler, C. Timmerer, H. Hellwagner, Univ. Klagenfurt (Austria); A. Hutter, F. Sanahuja, Siemens AG (Germany) ...... [6504-18] 2:30 pm: Cross-channel collaborations in peer-to-peer streaming, H. Li, K. Li, The Univ. of Georgia ...... [6504-19] 2:45 pm: Resource estimation methodology for multimedia applications, H. Kalva, R. Shankar, T. Patel, C. Cruz, Florida Atlantic Univ...... [6504-20] Coffee Break ...... 3:00 to 3:30 pm 3:30 pm: Multimodal event streams for virtual reality, J. von Spiczak, Brigham and Women’s Hospital and Univ. Karlsruhe (Germany); E. Samset, Brigham and Women’s Hospital and Univ. of Oslo (Norway) and Harvard Medical School; S. P. DiMaio, Brigham and Women’s Hospital and Harvard Medical School; G. Reitmayr, Univ. of Cambridge (United Kingdom); D. Schmalstieg, Technische Univ. Graz (Austria); C. R. Burghart, Univ. Karlsruhe (Germany); R. Kikinis, Brigham and Women’s Hospital and Harvard Medical School [6504-21] 3:45 pm: A multichannel multi-encoding transmission scheme for wireless video streaming, A. Kolekar, W. Feng, Portland State Univ.; M. Venkatachalam, Intel Corp...... [6504-22] 4:00 pm: Fine granularity adaptive multireceiver video streaming, V. S. W. Eide, Simula Research Lab. (Norway); F. Eliassen, J. A. Michaelsen, F. Jensen, Univ. of Oslo (Norway) ...... [6504-23]

SESSION 8 Conv. Ctr. Room C1 ...... Thurs. 4:15 to 5:45 pm Multimedia Middleware 4:15 pm: CameraCast: flexible access to remote video sensors, J. Kong, I. Ganev, K. Schwan, Georgia Institute of Technology; P. Widener, The Univ. of New Mexico ...... [6504-24] 4:45 pm: Plasma: a scripting language for processing media streams, T. Zhu, P. Korshunov, B. Liu, W. Ooi, National Univ. of Singapore (Singapore) ...... [6504-25] 5:15 pm: SenseTK: a multimodal multimedia sensor networking toolkit, P. Sitbon, Portland State Univ...... [6504-26]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 43 Conference 6505 • Conv. Ctr. Room A5

Monday-Thursday 29 January-1 February 2007 • Proceedings of SPIE Vol. 6505 Security, Steganography, and Watermarking of Multimedia Contents IX

Conference Chairs: Edward J. Delp III, Purdue Univ.; Ping Wah Wong, Nethra Imaging Cochairs: Nasir D. Memon, Polytechnic Univ.; Jana Dittmann, Otto-von-Guericke-Univ. Magdeburg (Germany) Program Committee: Adnan M. Alattar, Digimarc Corp.; Oscar C. L. Au, The Hong Kong Univ. of Science and Technology (Hong Kong China); Mauro Barni, Univ. degli Studi di Siena (Italy); Jeffrey A. Bloom, THOMSON Corporate Research; Gordon W. Braudaway, IBM Corp.; Ee-Chien Chang, National Univ. of Singapore (Singapore); Ingemar J. Cox, Univ. College London (United Kingdom); Ahmet M. Eskicioglu, The City Univ. of New York; Jessica Fridrich, Binghamton Univ.; Teddy Furon, INRIA (France); Ton Kalker, Hewlett-Packard Co.; Reginald L. Lagendijk, Technische Univ. Delft (Netherlands); Benoît Macq, Univ. Catholique de Louvain (Belgium); Bangalore S. Manjunath, Univ. of California/Santa Barbara; Pierre Moulin, Univ. of Illinois at Urbana-Champaign; Fernando Pérez-González, Univ. de Vigo (Spain); Rei Safavi-Naini, Univ. of Wollongong (Australia); Bülent Sankur, Bogaziçi Univ. (Turkey); Gaurav Sharma, Univ. of Rochester; Claus Vielhauer, Otto-von-Guericke-Univ. Magdeburg (Germany); Sviatoslav V. Voloshynovskiy, Univ. de Genève (Switzerland); Min Wu, Univ. of Maryland/College Park

Monday 29 January SESSION 3 Conv. Ctr. Room A5 ...... Mon. 1:30 to 2:50 pm SESSION 1 Attacks Conv. Ctr. Room A5 ...... Mon. 8:30 to 10:30 am Chair: Ton Kalker, Hewlett-Packard Co. Steganalysis 1:30 pm: Efficient noninteractive zero-knowledge watermark detector Chair: Nasir D. Memon, Polytechnic Univ. robust to sensitivity attacks, J. R. Troncoso, F. Pérez-González, Univ. de Vigo (Spain) ...... [6505-10] 8:30 am: Improving steganographic security by minimizing the embedding impact, J. Fridrich, Binghamton Univ...... [6505-01] 1:50 pm: Reverse-engineering a detector with false alarms, S. A. Craver, J. Yu, Binghamton Univ...... [6505-11] 8:50 am: Merging Markov and DCT features for multiclass JPEG steganalysis, T. Pevny, J. Fridrich, Binghamton Univ...... [6505-02] 2:10 pm: On distortion measures associated to random desynchronization attacks, F. Balado, Univ. College Dublin (Ireland); F. Perez-Gonzalez, Univ. de 9:10 am: Batch steganography and the threshold game, A. D. Ker, Univ. of Vigo (Spain) ...... [6505-12] Oxford (United Kingdom) ...... [6505-03] 2:30 pm: Fast and automatic watermark resynchronization based on 9:30 am: Mel-Cepstrum based steganalysis for VoIP steganography, C. Zernike moments, X. Kang, Sun Yat-Sen Univ. (China); W. Zeng, Univ. of Kraetzer, J. Dittmann, Otto-von-Guericke-Univ. Magdeburg (Germany) [6505-04] Missouri/Columbia; J. Huang, Sun Yat-Sen Univ. (China) ...... [6505-13] 9:50 am: Optimally weighted least-squares steganalysis, A. D. Ker, Univ. of Oxford (United Kingdom) ...... [6505-05] SESSION 4 10:10 am: Blind source separation steganalysis, R. Chandramouli, Stevens Conv. Ctr. Room A5 ...... Mon. 2:50 to 5:20 pm Institute of Technology ...... [6505-06] Theoretical Methods Coffee Break ...... 10:30 to 11:00 am Chair: Sviatoslav V. Voloshynovskiy, Univ. de Genève (Switzerland) 2:50 pm: Security of spread-spectrum-based data hiding, L. L. Pérez-Freire, SESSION 2 Univ. de Vigo (Spain); P. Moulin, Univ. of Illinois at Urbana-Champaign; F. Pérez- Conv. Ctr. Room A5 ...... Mon. 11:00 am to 12:00 pm González, Univ. de Vigo (Spain) ...... [6505-14] Audio 3:10 pm: Trellis coded modulation to improve dirty paper trellis Chair: Jeffrey A. Bloom, THOMSON Corporate Research watermarking, C. Wang, G. Doërr, I. J. Cox, Univ. College London (United Kingdom) ...... [6505-16] 11:00 am: Robust message authentication code algorithm for digital audio recordings, S. Zmudzinski, M. Steinebach, Fraunhofer-Institut für Integrierte Coffee Break ...... 3:30 to 4:00 pm Publikations- und Informationssysteme (Germany) ...... [6505-07] 4:00 pm: Cramer-Rao bound on watermark desynchronization parameter 11:20 am: Blind audio watermark synchronization by passive audio estimation accuracy, S. Sadasivam, P. Moulin, Univ. of Illinois at Urbana- fingerprinting, M. Steinebach, Fraunhofer-Institut für Integrierte Publikations- Champaign ...... [6505-17] und Informationssysteme (Germany) ...... [6505-08] 4:20 pm: A novel interpretation of content authentication, P. Comesaña- 11:40 am: Audio watermarking robust to geometrical distortions based on Alfaro, Univ. de Vigo (Spain); F. Balado, Univ. College Dublin (Ireland); F. Pérez- dyadic wavelet transform, Y. Wang, J. Huang, Sun Yat-Sen Univ. González, Univ. de Vigo (Spain) ...... [6505-18] (China) ...... [6505-09] 4:40 pm: Capacity and optimal collusion attack channels for Gaussian Lunch Break ...... 12:00 to 1:30 pm fingerprinting games, Y. Wang, P. Moulin, Univ. of Illinois at Urbana- Champaign ...... [6505-19] 5:00 pm: Sphere packing lower bound on fingerprinting error probability, N. Kiyavash, P. Moulin, Univ. of Illinois at Urbana-Champaign ...... [6505-20]

44 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505

Tuesday 30 January SESSION 8 Conv. Ctr. Room A5 ...... Tues. 3:20 to 5:00 pm Awards and Plenary Presentation Natural Language Watermarking Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Chairs: Mercan Topkara, Purdue Univ.; Edward J. Delp III, Purdue Development and Plasticity in Visual Cortex Univ. Brian A. Wandell, Stanford Univ. 3:20 pm: Encoding with errors: a confusing approach, M. Topkara, U. See pg. 6 for details. Topkara, M. J. Atallah, Purdue Univ...... [6505-31] 3:40 pm: A bibliography of linguistic steganography, R. Bergmair, Univ. of SESSION 5 Cambridge (United Kingdom) ...... [6505-32] Conv. Ctr. Room A5 ...... Tues. 9:30 to 11:20 am 4:00 pm: Syntactic tools for natural language watermarking, H. M. Meral, B. Sankur, S. Ozsoy, Bogaziçi Univ. (Turkey) ...... [6505-33] Benchmarking 4:20 pm: The syntax of concealment: reliable methods for plain text Chair: Jana Dittmann, Otto-von-Guericke-Univ. Magdeburg information hiding, B. Murphy, The Univ. of Dublin, Trinity College (Germany) (Ireland) ...... [6505-34] 9:30 am: Nested object watermarking: comparison of block-luminance and 4:40 pm: Statistically constrained shallow text marking: techniques, blue channel LSB wet paper code image watermarking, C. Vielhauer, J. evaluation paradigm, and results, B. Murphy, C. Vogel, The Univ. of Dublin, Dittmann, Otto-von-Guericke-Univ. Magdeburg (Germany) ...... [6505-21] Trinity College (Ireland) ...... [6505-35] 9:50 am: Deterring watermark collusion attacks using signal processing techniques, A. N. Lemma, M. van der Veen, Philips Research Labs. (Netherlands) ...... [6505-22] Wednesday 31 January 10:10 am: Experimental study on watermark interference in multiple re- watermarking, A. Uhl, Univ. Salzburg (Austria) ...... [6505-23] Plenary Session Coffee Break ...... 10:30 to 11:00 am Marriott Ballroom 1-6 ...... 8:00 to 9:20 am 11:00 am: Evaluation of standard watermarking techniques, E. Marini, F. Autrusseau, P. LeCallet, Univ. de Nantes (France); P. Campisi, Univ. degli Studi 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Roma Tre (Italy) ...... [6505-24] Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno SESSION 6 Mottin, C.re de Recherche et de Restauration des Musées de France Conv. Ctr. Room A5 ...... Tues. 11:20 am to 12:20 pm 8:40 am: Space Exploration, Imaging Technology, and the Development Forensics I of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Chair: Nasir D. Memon, Polytechnic Univ. Solar System Visualization (SSV) Project 11:20 am: Digital imaging sensor identification: further study, M. Goljan, J. See pg. 7 for details. Fridrich, M. Chen, Binghamton Univ...... [6505-25] 11:40 am: Joint forensics and watermarking approach for video SESSION 9 authentication, S. Thiemert, H. Liu, M. Steinebach, L. Croce Ferri, Fraunhofer- Institut für Integrierte Publikations- und Informationssysteme Conv. Ctr. Room A5 ...... Wed. 9:30 am to 12:00 pm (Germany) ...... [6505-26] Biometrics 12:00 pm: Image splicing detection using 2D phase congruency and Chair: Claus Vielhauer, Otto-von-Guericke-Univ. Magdeburg statistical moments of characteristic function, W. Chen, Y. Shi, New Jersey (Germany) Institute of Technology; W. Su, U.S. Army Research Lab...... [6505-27] 9:30 am: Adaptive and distributed cryptography for signature biometrics Lunch/Exhibition Break ...... 12:20 to 2:00 pm protection, P. Campisi, E. Maiorana, M. Gonzalez-Prats, A. Neri, Univ. degli Studi Roma Tre (Italy) ...... [6505-36] SESSION 7 9:50 am: A fingerprint matching algorithm based on curvature feature Conv. Ctr. Room A5 ...... Tues. 2:00 to 3:00 pm alignment, Q. Su, J. Tian, X. Yang, Institute of Automation (China) . . [6505-37] Forensics II 10:10 am: Analyzing a multimodal biometric system using real and virtual users, T. Scheidat, C. Vielhauer, Otto-von-Guericke-Univ. Magdeburg Chair: Jessica Fridrich, Binghamton Univ. (Germany) ...... [6505-38] 2:00 pm: Robust scanner identification based on noise features, H. Gou, A. Coffee Break ...... 10:30 to 11:00 am Swaminathan, M. Wu, Univ. of Maryland/College Park ...... [6505-28] 11:00 am: Digital watermarking of biometric speech references: impact to 2:20 pm: Detection of malevolent changes in digital video for forensic the EER system performance, A. Lang, J. Dittmann, Otto-von-Guericke-Univ. applications, R. Caldelli, N. Mondaini, V. Cappellini, A. Piva, Univ. degli Studi di Magdeburg (Germany) ...... [6505-39] Firenze (Italy); M. Barni, Univ. degli Studi di Siena (Italy) ...... [6505-29] 11:20 am: How to protect biometric templates, Y. Sutcu, Q. Li, N. D. Memon, 2:40 pm: Forensic classification of imaging sensor types, N. Khanna, A. K. Polytechnic Univ...... [6505-40] Mikkilineni, G. T. C. Chiu, J. P. Allebach, E. J. Delp III, Purdue Univ. . . [6505-30] 11:40 am: Error exponent analysis of person identification based on fusion Coffee Break ...... 3:00 to 3:30 pm of independent modalities, O. J. Koval, S. V. Voloshynovskiy, T. Pun, Univ. de Genève (Switzerland) ...... [6505-41] Lunch/Exhibition Break ...... 12:00 to 1:30 pm

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 45 Conference 6505

SESSION 10 SESSION 13 Conv. Ctr. Room A5 ...... Wed. 1:30 to 4:20 pm Conv. Ctr. Room A5 ...... Thurs. 11:00 am to 12:00 pm BOWS Embedding I Chairs: Mauro Barni, Univ. degli Studi di Siena (Italy); Fernando Chair: Benoît Macq, Univ. Catholique de Louvain (Belgium) Pérez-González, Univ. de Vigo (Spain) 11:00 am: A semi-blind image watermarking scheme: embedding a pseudo- 1:30 pm: The first BOWS contest: break our watermarking system, A. Piva, random number sequence in two-frequency bands in the DCT domain, A. Univ. degli Studi di Firenze (Italy); M. Barni, Univ. degli Studi di Siena Koc, A. M. Eskicioglu, The City Univ. of New York ...... [6505-59] (Italy) ...... [6505-42] 11:20 am: Distortion-free robust watermarking: a case study, D. Coltuc, 1:50 pm: The good, the bad, and the ugly: three different approaches to Univ. Valahia din Targoviste (Romania) and Institut National Polytechnique de break their watermarking system, G. Le Guelvouit, Capgemini (France); T. Grenoble (France); J. Chassery, Institut National Polytechnique de Grenoble Furon, IRISA (France); F. Cayre, École Nationale Supérieure d’Ingénieurs (France) ...... [6505-60] Electriciens de Grenoble (France) ...... [6505-43] 11:40 am: Hidden context highlighting with JPEG2000 imagery, R. U. 2:10 pm: Watermarking attack: BOWS contest, J. Bennour, J. E. Dugelay, M. Rosenbaum, Sr., H. Schumann, Univ. Rostock (Germany) ...... [6505-61] Federico, Institut Eurécom (France) ...... [6505-44] Lunch Break ...... 12:00 to 1:30 pm 2:30 pm: Tangential sensitivity analysis with prior information, J. Earl, Univ. of Cambridge (United Kingdom) ...... [6505-45] SESSION 14 2:50 pm: Tackling BOWS with the sensitivity attack, A. Westfeld, Technische Conv. Ctr. Room A5 ...... Thurs. 1:30 to 2:10 pm Univ. Dresden (Germany) ...... [6505-46] Embedding II Coffee Break ...... 3:10 to 3:40 pm Chair: Benoît Macq, Univ. Catholique de Louvain (Belgium) 3:40 pm: Two different approaches for attacking BOWS, P. Comesaña-Alfaro, F. Pérez-González, Univ. de Vigo (Spain) ...... [6505-47] 1:30 pm: Data hiding based wavelet on finite field, A. Benjelloun Touimi, France Telecom R&D (France); L. Godard, École Nationale Supérieure 4:00 pm: How we broke the BOWS watermark, S. A. Craver, I. Atakli, J. Yu, d’Ingénieurs Electriciens de Grenoble (France) ...... [6505-62] Binghamton Univ...... [6505-48] 1:50 pm: Perceptual data hiding in tree structured Haar transform domain, M. Carli, M. Cancellaro, Univ. degli Studi di Roma Tre (Italy); K. O. Egiazarian, J. SESSION 11 T. Astola, Tampere Univ. of Technology (Finland) ...... [6505-63] Conv. Ctr. Room A5 ...... Wed. 4:20 to 5:20 pm Video SESSION 15 Chair: Adnan M. Alattar, Digimarc Corp. Conv. Ctr. Room A5 ...... Thurs. 2:30 to 5:00 pm 4:20 pm: Adaptive MPEG-2 video data hiding scheme, A. Sarkar, U. Madhow, Other Domains S. Chandrasekaran, B. S. Manjunath, Univ. of California/Santa Barbara[6505-49] Chair: Ahmet M. Eskicioglu, The City Univ. of New York 4:40 pm: Digital video watermarking in P-frames, M. Noorkami, R. M. 2:30 pm: A method of text watermarking using presuppositions, B. Macq, O. Mersereau, Georgia Institute of Technology ...... [6505-50] Vybornova, Univ. Catholique de Louvain (Belgium) ...... [6505-65] 5:00 pm: Framework for combined video frame synchronization and 2:50 pm: An evaluation of lightweight JPEG2000 encryption with anisotropic watermark detection, E. Hauer, T. Bölke, M. Steinebach, Fraunhofer-Institut für wavelet packets, D. Engel, A. Uhl, Univ. Salzburg (Austria) ...... [6505-66] Integrierte Publikations- und Informationssysteme (Germany) ...... [6505-51] 3:10 pm: Tamper-proofing of electronic and printed text documents, R. Villan, Sr., S. V. Voloshynovskiy, O. J. Koval, F. Deguillaume, T. Pun, Univ. de Thursday 1 February Genève (Switzerland) ...... [6505-67] Coffee Break ...... 3:30 to 4:00 pm SESSION 12 4:00 pm: Channel model and capacity analysis of printed text documents, Conv. Ctr. Room A5 ...... Thurs. 8:30 to 10:30 am A. K. Mikkilineni, P. Chiang, G. T. C. Chiu, J. P. Allebach, E. J. Delp III, Purdue Univ...... [6505-68] Forensics III 4:20 pm: Watermarking geographic vectordata using a variable stripsize Chair: Min Wu, Univ. of Maryland/College Park scheme, M. Voigt, Technische Univ. Darmstadt (Germany) ...... [6505-69] 8:30 am: Source digital camcorder identification using CCD photo response 4:40 pm: Authentication of H.264 streams by direct watermarking of CAVLC nonuniformity, M. Chen, J. Fridrich, M. Goljan, Binghamton Univ. . . . [6505-53] blocks, B. G. Mobasseri, Y. J. Naik Raikar, Villanova Univ...... [6505-70] 8:50 am: Blind identification of cellular phone cameras, B. Sankur, O. Celiktutan, Bogaziçi Univ. (Turkey); I. Avcibas, Uludag Üniv. (Turkey) . [6505-54] 9:10 am: Forensics for flatbed scanners, T. Gloe, E. Franz, A. Winkler, Technische Univ. Dresden (Germany) ...... [6505-55] 9:30 am: Image authentication via intrinsic fingerprints, A. Swaminathan, M. Wu, K. J. R. Liu, Univ. of Maryland/College Park ...... [6505-56] 9:50 am: Scanner identification using sensor pattern noise, N. Khanna, A. K. Mikkilineni, G. T. C. Chiu, J. P. Allebach, E. J. Delp III, Purdue Univ. . . [6505-57] 10:10 am: A generalized Benford’s law for JPEG coefficients and its applications in image forensics, D. Fu, Y. Shi, New Jersey Institute of Technology; W. Su, U.S. Army Research Lab...... [6505-58] Coffee Break ...... 10:30 to 11:00 am

46 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6506 • Conv. Ctr. Room B1

Wednesday-Thursday 31 January-1 February 2007 • Proceedings of SPIE Vol. 6506 Multimedia Content Access: Algorithms and Systems

Conference Chairs: Alan Hanjalic, Technische Univ. Delft (Netherlands); Raimondo Schettini, DISCo/Univ. degli Studi di Milano-Bicocca (Italy); Nicu Sebe, Univ. van Amsterdam (Netherlands) Cochairs: Edward Y. Chang, Google, Inc.; Simone Santini, Univ. Autónoma de Madrid; Theo Gevers, Univ. van Amsterdam (Netherlands) Program Committee: Kiyoharu Aizawa, The Univ. of Tokyo (Japan); Noboru Babaguchi, Osaka Univ. (Japan); Nozha Boujemaa, INRIA Rocquencourt (France); Augusto Celentano, Univ. Ca’ Foscari di Venezia (Italy); Tsuhan Chen, Carnegie Mellon Univ.; TatSeng Chua, National Univ. of Singapore (Singapore); Rita Cucchiara, Univ. degli Studi di Modena e Reggio Emilia (Italy); Alberto Del Bimbo, Univ. degli Studi di Firenze (Italy); Ajay Divakaran, Mitsubishi Electric Research Labs.; Chitra Dorai, IBM Thomas J. Watson Research Ctr.; Arun Hampapur, IBM Thomas J. Watson Research Ctr.; Alexander G. Hauptmann, Carnegie Mellon Univ.; Roger David Hersch, École Polytechnique Fédérale de Lausanne (Switzerland); Ebroul Izquierdo, Queen Mary Univ. of London (United Kingdom); Alejandro Jaimes, Fuji Xerox Co., Ltd. (Japan); Mohan S. Kankanhalli, National Univ. of Singapore (Singapore); John R. Kender, Columbia Univ.; Anil C. Kokaram, The Univ. of Dublin, Trinity College (Ireland); Clement H. C. Leung, Victoria Univ. of Technology (Australia); Michael S. Lew, Univ. Leiden (Netherlands); Rainer W. Lienhart, Univ. Augsburg (Germany); Kadir A. Peker, Mitsubishi Electric Research Labs.; Silvia Pfeiffer, Commonwealth Scientific and Industrial Research Organisation (Australia); Alan F. Smeaton, Dublin City Univ. (Ireland); John R. Smith, IBM Thomas J. Watson Research Ctr.; Hari Sundaram, Arizona State Univ.; Ahmet M. Tekalp, Univ. of Rochester; Qi Tian, The Univ. of Texas at San Antonio; Alain Trémeau, Univ. Jean Monnet Saint-Etienne (France); Luc J. Van Gool, Katholieke Univ. Leuven (Belgium) and ETH Zürich (Switzerland); Svetha Venkatesh, Curtin Univ. of Technology (Australia); Marcel Worring, Univ. van Amsterdam (Netherlands); Lei Zhang, Microsoft Research Asia (China)

Wednesday 31 January Lunch/Exhibition Break ...... 12:20 to 2:00 pm SESSION 3 Plenary Session Conv. Ctr. Room B1 ...... Wed. 2:00 to 3:00 pm Marriott Ballroom 1-6 ...... 8:00 to 9:20 am Video Analysis and Retrieval I 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Chair: Ahmet Ekin, Philips Research Labs. (Netherlands) Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc 2:00 pm: Efficient re-indexing of automatically annotated image collections Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, using keyword combination, A. G. Yavlinsky, S. Rueger, Imperial College Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno London (United Kingdom) ...... [6506-07] Mottin, C.re de Recherche et de Restauration des Musées de France 2:20 pm: Video to the rescue of audio: shot boundary assisted speaker 8:40 am: Space Exploration, Imaging Technology, and the Development change detection, A. Samour, M. Karaman, L. Goldmann, T. Sikora, Technische of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, Univ. Berlin (Germany) ...... [6506-08] Solar System Visualization (SSV) Project 2:40 pm: A trajectory based video segmentation for surveillance See pg. 7 for details. applications, N. M. Thomas, N. Canagarajah, Univ. of Bristol (United Kingdom) ...... [6506-09] Coffee Break ...... 3:00 to 3:20 pm SESSION 1 Conv. Ctr. Room B1 ...... Wed. 9:30 to 10:30 am SESSION 4 Image Analysis and Retrieval Conv. Ctr. Room B1 ...... Wed. 3:20 to 4:00 pm Chair: Jan Nesvadba, Philips Research Labs. (Netherlands) Applications I 9:30 am: A model-based conceptual clustering of moving objects in video Chair: Cuneyt M. Taskiran, Motorola, Inc. surveillance, J. Lee, Univ. of Bridgeport ...... [6506-01] 3:20 pm: Knowledge discovery for better photographs, J. Yen, Toshiba 9:50 am: Image watermarking based on a color quantization process, A. America Business Solutions; P. Wu, D. R. Tretter, Hewlett-Packard Trémeau, J. Thomas, G. Chareyron, Univ. Jean Monnet Saint-Etienne Labs...... [6506-10] (France) ...... [6506-02] 3:40 pm: Organizing a daily visual diary using multifeature clustering, C. E. 10:10 am: Search and retrieval of medical images for improved diagnosis of Ó Conaire, N. E. O’Connor, A. F. Smeaton, Dublin City Univ. (Ireland) and neurodegenerative diseases, A. Ekin, R. S. Jasinschi, E. Turan, R. Engbers, Adaptive Information Cluster (Ireland); G. Jones, Dublin City Univ. Philips Research Labs. (Netherlands); J. van der Grond, M. van Buchem, Leids (Ireland) ...... [6506-12] Univ. Medisch Ctr. (Netherlands) ...... [6506-03] Coffee Break ...... 10:30 to 10:50 am Panel Discussion ...... 4:00 to 5:00 pm Chair: Ajay Divakaran, Mitsubishi Electric Research Labs. SESSION 2 Conv. Ctr. Room B1 ...... Wed. 10:50 am to 12:20 pm Content Analysis-based Browsing Chair: Ajay Divakaran, Mitsubishi Electric Research Labs. 10:50 am: Assessment of end-user response to sports highlights extraction for personal video recorders (Invited Paper), A. Divakaran, K. Wittenburg, K. A. Peker, Mitsubishi Electric Research Labs.; R. Radhakrishnan, Dolby Labs. Inc...... [6506-04] 11:20 am: Examining user interactions with video retrieval systems (Invited Paper), M. G. Christel, Carnegie Mellon Univ...... [6506-05] 11:50 am: Video summary evaluation and visualization (Invited Paper), C. M. Taskiran, Z. Li, Motorola, Inc...... [6506-06]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 47 Conference 6506

Thursday 1 February

SESSION 5 Conv. Ctr. Room B1 ...... Thurs. 9:00 to 10:00 am Applications II Chair: Michael G. Christel, Carnegie Mellon Univ. 9:00 am: Recognizing persons in images by learning from videos, R. W. Lienhart, E. Hörster, Univ. Augsburg (Germany) ...... [6506-13] 9:20 am: Storage format for personalized broadcasting content consumption, S. Jin, Information and Communications Univ. (South Korea) ...... [6506-14] 9:40 am: A unified and efficient framework for court-net sports video analysis using 3D camera modeling, J. Han, Technische Univ. Eindhoven (Netherlands) ...... [6506-15] Coffee Break ...... 10:00 to 10:20 am

SESSION 6 Conv. Ctr. Room B1 ...... Thurs. 10:20 am to 12:20 pm Bioinformatics Chair: Fons J. Verbeek, Leiden Univ. (Netherlands) 10:20 am: Ontology driven image search engine in multimodal imaging databases (Invited Paper), F. J. Verbeek, Y. Bei, J. Dmitrieva, M. Belmamoune, Leiden Univ. (Netherlands) ...... [6506-16] 10:50 am: Adaptation of video game UVW mapping to 3D visualization of gene expression patterns (Invited Paper), P. D. Vize, V. Gerth, Univ. Calgary (Canada) ...... [6506-17] 11:20 am: Classification of yeast cells from image features to learn pathogen conditions (Invited Paper), F. J. Verbeek, P. van der Putten, Leiden Univ. (Netherlands) ...... [6506-18] Lunch Break ...... 12:20 to 2:00 pm

SESSION 7 Conv. Ctr. Room B1 ...... Thurs. 2:00 to 3:00 pm Video Analysis and Retrieval II Chair: Alan F. Smeaton, Dublin City Univ. (Ireland) 2:00 pm: Analysis of unstructured video based on camera motion, G. Abdollahian, E. J. Delp III, Purdue Univ...... [6506-20] 2:20 pm: Audiovisual shot-reverse-shot analysis, J. Nesvadba, Philips Research (Netherlands); B. Kroon, Technische Univ. Delft (Netherlands)[6506-21] 2:40 pm: Edit while watching: home video editing made easy, M. Campanella, H. Weda, M. Barbieri, Philips Research Labs. (Netherlands) ...... [6506-22] Coffee Break ...... 3:00 to 3:20 pm

SESSION 8 Conv. Ctr. Room B1 ...... Thurs. 3:20 to 5:00 pm Applications III Chair: Raimondo Schettini, Univ. degli Studi di Milano-Bicocca (Italy) 3:20 pm: Multimodule human motion analysis from a monocular video, W. Lao, Technische Univ. Eindhoven (Netherlands) ...... [6506-23] 3:40 pm: A study on video viewing behavior, S. Mongy, C. Djeraba, Univ. des Sciences et Technologies de Lille (France) ...... [6506-24] 4:00 pm: ARGOS: French evaluation campaign for benchmarking of video content analysis methods, P. Joly, Univ. Paul Sabatier (France); J. Benois- Pineau, Univ. Bordeaux I (France); G. Quenot, Univ. Joseph Fourier (France) ...... [6506-25] 4:20 pm: Data mining learning bootstrap through semantic thumbnail analysis, S. Battiato, G. M. Farinella, G. Giuffrida, G. Tribulato, Univ. di Catania (Italy) ...... [6506-26] 4:40 pm: A spatiotemporal decomposition strategy for personal home video management, H. Yi, Univ. of Illinois at Urbana-Champaign; I. V. Kozintsev, M. Polito, Y. Wu, J. Bouguet, A. V. Nefian, C. Dulong, Intel Corp...... [6506-27]

48 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6507 • Conv. Ctr. Room B4

Monday-Tuesday 29-30 January 2007 • Proceedings of SPIE Vol. 6507 Multimedia on Mobile Devices 2007

Conference Chairs: Reiner Creutzburg, Fachhochschule Brandenburg (Germany); Jarmo H. Takala, Tampere Univ. of Technology (Finland); Jianfei Cai, Nanyang Technological Univ. (Singapore) Program Committee: David Akopian, The Univ. of Texas at San Antonio; Alan Chalmers, Univ. of Bristol (United Kingdom); Surendar Chandra, Univ. of Notre Dame; David S. Doermann, Univ. of Maryland/College Park; Uwe Dummann, Siemens AG (Germany); Elizabeth Dykstra-Erickson, Kinoma, Inc.; Lajos Hanzo, Univ. of Southampton (United Kingdom); Zhihai He, Univ. of Missouri/Columbia; Xin Li, West Virginia Univ.; Manzur M. Murshed, Monash Univ. (Australia); Sethuraman Panchanathan, Arizona State Univ.; Kari Pulli, Nokia (Finland); Matthias Rauterberg, Technische Univ. Eindhoven (Netherlands); Phillip A. Regalia, Institut National des Télécommunications (France); Olli J. Silvén, Univ. of Oulu (Finland)

Monday 29 January SESSION 4 Conv. Ctr. Room B4 ...... Mon. 3:30 to 5:10 pm SESSION 1 Media Delivery and Management I Conv. Ctr. Room B4 ...... Mon. 9:30 to 10:30 am Chair: Jarmo H. Takala, Tampere Univ. of Technology (Finland) Video Processing I 3:30 pm: Underflow prevention for AV streaming media under varying Chair: Reiner Creutzburg, Fachhochschule Brandenburg (Germany) channel conditions, S. G. Deshpande, Sharp Labs. of America, Inc. [6507-11] 9:30 am: Color adaptation of videos for mobile devices, S. Kopf, W. 3:50 pm: Seamless content consumption experience across media delivery Effelsberg, Univ. Mannheim (Germany) ...... [6507-01] channels and devices, C. M. Taskiran, B. Gandhi, W. Wang, D. Li, Motorola, Inc...... [6507-12] 9:50 am: Complexity-constrained rated-distortion optimization for Wyner- Ziv video coding, L. Liu, Purdue Univ.; Z. Li, Thomson Corporate Research; E. 4:10 pm: MDC image coding using cascaded correlating transforms, R. J. Delp III, Purdue Univ...... [6507-02] Knoerig, T. Sikora, Technische Univ. Berlin (Germany) ...... [6507-13] 10:10 am: Dynamic full-scalability conversion in scalable video coding, D. S. 4:30 pm: The interactive contents authoring system for terrestrial digital Lee, T. Bae, T. C. Thang, Y. Ro, Information and Communications Univ. (South multimedia broadcasting, W. Cheong, S. Ahn, J. Cha, K. Moon, Electronics Korea) ...... [6507-03] and Telecommunications Research Institute (South Korea) ...... [6507-14] Coffee Break ...... 10:30 to 11:00 am 4:50 pm: Musical slideshow MAF with protection and governance using MPEG-21 IPMP components and REL, M. S. H. Sabirin, H. Tan, M. Kim, Information and Communications Univ. (South Korea) ...... [6507-15] SESSION 2 Conv. Ctr. Room B4 ...... Mon. 11:00 am to 12:10 pm Tuesday 30 January Video Processing II Chair: Reiner Creutzburg, Fachhochschule Brandenburg (Germany) Awards and Plenary Presentation 11:00 am: Displays enabling mobile multimedia (Invited Paper), J. S. Kimmel, Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Nokia Research Ctr. (Finland) ...... [6507-04] Development and Plasticity in Visual Cortex 11:30 am: A neighborhood analysis based technique for real-time error concealment in H.264 intra pictures, S. T. C. Beesley, C. Grecos, A. J. Brian A. Wandell, Stanford Univ. Armstrong, E. A. Edirisinghe, Loughborough Univ. (United Kingdom) . [6507-05] See pg. 6 for details. 11:50 am: Reducing computational complexity of three-dimensional discrete cosine transform in video coding process, J. J. Koivusaari, J. H. SESSION 5 Takala, M. Gabbouj, Tampere Univ. of Technology (Finland) ...... [6507-06] Conv. Ctr. Room B4 ...... Tues. 9:30 to 10:30 am Lunch Break ...... 12:10 to 1:40 pm Multimedia Applications Chair: Jianfei Cai, Nanyang Technological Univ. (Singapore) SESSION 3 9:30 am: The digital aura, a framework for a sensible mobile phone Conv. Ctr. Room B4 ...... Mon. 1:40 to 3:00 pm entertainment experience: a future-oriented evaluation of possibilities of Systems for Multimedia applying sensor technology on mobile phones toward increasing the consumer entertainment experience, A. R. Lugmayr, Jr., Tampere Univ. of Chair: Jianfei Cai, Nanyang Technological Univ. (Singapore) Technology (Finland) ...... [6507-16] 1:40 pm: Restructuring a software based MPEG-4 video decoder for short 9:50 am: Sensometrics: identifying pen digitizers by statistical multimedia latency hardware acceleration, O. J. Silvén, J. J. Boutellier, T. Erdelyi, Univ. of signal processing, A. Oermann, C. Vielhauer, Otto-von-Guericke-Univ. Oulu (Finland) ...... [6507-07] Magdeburg (Germany) ...... [6507-17] 2:00 pm: Hand-held analog television over WiMAX executed in SW, D. S. 10:10 am: Semantic consumption of photos on mobile devices, S. Lee, Iancu, H. Ye, M. Senthilvelan, V. Kotlyar, J. C. Glossner, M. Moudgill, S. Agrawal, Information and Communications Univ. (South Korea) ...... [6507-18] S. Jinturkar, A. Iancu, G. Nacer, S. Stanley, Sandbridge Technologies, Inc.; J. H. Takala, Tampere Univ. of Technology (Finland) ...... [6507-08] Coffee Break ...... 10:30 to 11:00 am 2:20 pm: Efficient implementation of DCT-based video denoising algorithm on OMAP(tm) platform, D. Rusanovskyy, J. J. Koivusaari, K. O. Egiazarian, J. H. Takala, Tampere Univ. of Technology (Finland) ...... [6507-09] 2:40 pm: Multimedia support for J2ME on high-end PDAs, T. Preuss, M. Rambow, M. Lehmann, Brandenburg Univ. of Applied Sciences (Germany) ...... [6507-10] Coffee Break ...... 3:00 to 3:30 pm

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 49 Conference 6507

SESSION 6 Posters and Demonstrations-Tuesday Conv. Ctr. Room B4 ...... Tues. 11:00 am to 12:10 pm Demonstrations ...... 5:30 to 8:30 pm HCI Issues for Multimedia I A symposium-wide demonstration session will be open to attendees 5:30 to Chair: Jianfei Cai, Nanyang Technological Univ. (Singapore) 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide 11:00 am: Wireless video sensor networks: convergence of video, wireless, interactive, hands-on demonstrations of a wide-range of products related to and vision (Invited Paper), Z. He, Univ. of Missouri/Columbia ...... [6507-19] Electronic Imaging. 11:30 am: Smooth transitions for mobile imagery browsing, R. U. Rosenbaum, Sr., H. Schumann, Univ. Rostock (Germany) ...... [6507-20] Posters ...... 5:30 to 7:00 pm 11:50 am: Experienced quality factors: qualitative evaluation approach to Posters will be placed on display after 10:00 am in Exhibit Hall 1. audiovisual quality, S. Jumisko-Pyykkö, Tampere Univ. of Technology (Finland); A poster session, with authors present at their posters, J. P. Häkkinen, Nokia Research Ctr. (Finland) ...... [6507-21] will be held Tuesday evening, 5:30 to 7:00 pm. Lunch/Exhibition Break ...... 12:10 to 1:40 pm ✔ Intuitive user interface for mobile devices based on visual motion detection, S. Winkler, K. Rangaswamy, S. Zhou, National Univ. of Singapore (Singapore) ...... [6507-30] SESSION 7 ✔ Bidirectional traffic status information service based on T-DMB, Y. Jeong, Conv. Ctr. Room B4 ...... Tues. 1:40 to 3:00 pm S. Kim, S. Cho, G. Kim, C. Ahn, S. Lee, Electronics and Telecommunications HCI Issues for Multimedia II Research Institute (South Korea) ...... [6507-33] Chair: Jarmo H. Takala, Tampere Univ. of Technology (Finland) ✔ Codesign toolset for application-specific instruction set processors, P. Jääskeläinen, V. Guzma, A. Cilio, J. H. Takala, T. Pitkänen, Tampere Univ. of 1:40 pm: Tablet PC interaction with digital micromirror device (DMD), H. H. Technology (Finland) ...... [6507-34] Refai, J. J. Sluss, Jr., Univ. of Oklahoma ...... [6507-22] ✔ Dynamic power management for UML modeled applications on 2:00 pm: TiDi browser: a novel photo browsing technique for mobile multiprocessor SoC, P. Kukkala, T. Arpinen, M. Setälä, M. Hännikäinen, T. D. devices, C. Tominski, Univ. Rostock (Germany); G. Bieber, B. Urban, Hämäläinen, Tampere Univ. of Technology (Finland) ...... [6507-35] Fraunhofer-Institut für Graphische Datenverarbeitung (Germany) . . . . [6507-23] ✔ The evolution of ambient learning, R. Creutzburg, Fachhochschule 2:20 pm: Of MOS and men: bridging the gap between objective and Brandenburg (Germany) ...... [6507-36] subjective quality measurements in mobile TV, T. C. M. de Koning, P. Veldhoven, TNO TPD (Netherlands); H. O. Knoche, Univ. College London (United ✔ Ambient learning in a mobile world, R. Creutzburg, Fachhochschule Kingdom); R. E. Kooij, TNO TPD (Netherlands) ...... [6507-24] Brandenburg (Germany) ...... [6507-37] 2:40 pm: Interfaces for mobile image browsing, R. U. Rosenbaum, Sr., H. Schumann, Univ. Rostock (Germany) ...... [6507-25] Coffee Break ...... 3:00 to 3:30 pm

SESSION 8 Conv. Ctr. Room B4 ...... Tues. 3:30 to 4:50 pm Media Delivery and Management II Chair: Jarmo H. Takala, Tampere Univ. of Technology (Finland) 3:30 pm: Novel layered scalable video coding transmission over MIMO wireless systems with partial CSI and adaptive channel selection, D. Song, C. W. Chen, Florida Institute of Technology ...... [6507-26] 3:50 pm: An efficient client-based JPEG2000 image transmission protocol, J. P. García Ortiz, Univ. de Almería (Spain) ...... [6507-27] 4:10 pm: Error resilient image transmission based on virtual SPIHT, R. Liu, J. He, X. Zhang, Beijing Univ. of Aeronautics and Astronautics (China) . . [6507-28] 4:30 pm: Complexity analysis and control in joint channel protection system for wireless video communication, X. Jin, G. Zhu, Huazhong Univ. of Science and Technology (China) ...... [6507-29]

50 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508 • Conv. Ctr. Rooms B2 and B3

Tuesday-Thursday 30 January-1 February 2007 • Proceedings of SPIE Vol. 6508 Visual Communications and Image Processing 2007

Conference Chairs: Chang Wen Chen, Florida Institute of Technology; Dan Schonfeld, Univ. of Illinois at Chicago; Jiebo Luo, Eastman Kodak Co. Program Committee: Kiyoharu Aizawa, The Univ. of Tokyo (Japan); Yucel Altunbasak, Georgia Institute of Technology; Rashid Ansari, Univ. of Illinois at Chicago; John G. Apostolopoulos, Hewlett-Packard Labs.; Mireille Boutin, Purdue Univ.; Alan C. Bovik, The Univ. of Texas at Austin; Charles D. Creusere, New Mexico State Univ.; Gerard de Haan, Philips Research Labs. (Netherlands); Edward J. Delp III, Purdue Univ.; Eric Dubois, Univ. of Ottawa (Canada); Frederic Dufaux, École Polytechnique Fédérale de Lausanne (Switzerland); Touradj Ebrahimi, École Polytechnique Fédérale de Lausanne (Switzerland); Onur G. Guleryuz, DoCoMo Communications Labs. USA, Inc.; Lina J. Karam, Arizona State Univ.; W. C. Karl, Boston Univ.; Janusz Konrad, Boston Univ.; Alex C. Kot, Nanyang Technological Univ. (Singapore); C.-C. J. Kuo, Univ. of Southern California; Reginald L. Lagendijk, Technische Univ. Delft (Netherlands); Shipeng Li, Microsoft Research Asia (China); Bangalore S. Manjunath, Univ. of California/Santa Barbara; Peyman Milanfar, Univ. of California/Santa Cruz; Pierre Moulin, Univ. of Illinois at Urbana-Champaign; Sethuraman Panchanathan, Arizona State Univ.; Thrasyvoulos N. Pappas, Northwestern Univ.; William A. Pearlman, Rensselaer Polytechnic Institute; Fernando Pereira, Instituto Superior Técnico (Portugal); Béatrice Pesquet-Popescu, École Nationale Supérieure des Télécommunications (France); Fatih M. Porikli, Mitsubishi Electric Research Labs.; Majid Rabbani, Eastman Kodak Co.; Kannan Ramchandran, Univ. of California/Berkeley; Kenneth Rose, Univ. of California/ Santa Barbara; Amir Said, Hewlett-Packard Labs.; Paul Salama, Indiana Univ.-Purdue Univ. at Indianapolis; Gaurav Sharma, Univ. of Rochester; Eckehard G. Steinbach, Technische Univ. München; Robert L. Stevenson, Univ. of Notre Dame; Thomas Stockhammer, Nomor Research (Germany); Huifang Sun, Mitsubishi Electric Research Labs.; Ming-Ting Sun, Univ. of Washington; Andrew G. Tescher, AGT Associates; Bhaskaran Vasudev, Epson Palo Alto Lab.; Anthony Vetro, Mitsubishi Electric Research Labs.; John W. Woods, Rensselaer Polytechnic Institute; Ying Wu, Northwestern Univ.; Heather H. Yu, Panasonic Information & Networking Technologies Lab.

VCIP 2007 Best Student Paper Awards VCIP 2007 Best Paper Awards Awards will be presented to two outstanding student papers submitted for Awards will be presented for the two best submitted papers. this year’s symposium. IS&T and SPIE gratefully acknowledge Hewlett Packard Co. and Mitsubishi IS&T and SPIE gratefully acknowledge Eastman Kodak Co. and Motorola Electronic Research Lab. for generously sponsoring these awards. for generously sponsoring these awards.

VCIP 2007 Young Investigator Award The VCIP Awards will be presented during the Plenary Session Tuesday The award will be given for the best paper presented by a young researcher Morning. who has graduated within 5 years from the date of the meeting. IS&T and SPIE gratefully acknowledge Hewlett Packard Co. for generously sponsoring this award. Tuesday 30 January Awards and Plenary Presentation Marriott Ballroom 1-6 ...... 8:15 to 9:10 am Development and Plasticity in Visual Cortex, Brian A. Wandell, Stanford Univ. See pg. 6 for details. Sessions 1 and 2 run concurrently. SESSION 1 SESSION 2 Conv. Ctr. Rooms B2 ...... Tues. 9:30 am to 12:00 pm Conv. Ctr. Room B3 ...... Tues. 9:30 am to 12:00 pm Video Coding I Image and Video Analysis Chairs: John G. Apostolopoulos, Hewlett-Packard Labs.; Shipeng Chairs: Touradj Ebrahimi, École Polytechnique Fédérale de Lausanne Li, Microsoft Research Asia (China) (Switzerland); Philippe Salembier, Univ. Politècnica de Catalunya 9:30 am: Real-time video coding under power constraint based on H.264 (Spain) CODEC, L. Su, Institute of Computing Technology (China); Y. Lu, F. Wu, S. Li, 9:30 am: Feature point tracking combining the interacting multiple model Microsoft Research Asia (China); W. Gao, Institute of Computing Technology filter and an efficient assignment algorithm, D. Marimon, Y. Abdeljaoued, B. (China) ...... [6508-01] Palacios, T. Ebrahimi, École Polytechnique Fédérale de Lausanne 9:50 am: A low bit-rate video coding approach using modified adaptive (Switzerland) ...... [6508-07] warping and long-term spatial memory, Y. Chen, M. J. T. Smith, E. J. Delp III, 9:50 am: Trajectory-based ball detection and tracking with aid of Purdue Univ...... [6508-02] homograph transform in broadcast tennis video, X. Yu, Institute for 10:10 am: Rate-prediction structure complexity analysis for multiview video Infocomm Research (Singapore); N. Jiang, J. Li, National Univ. of Singapore coding using hybrid genetic algorithms, Y. Liu, Q. Dai, Z. You, W. Xu, (Singapore) ...... [6508-08] Tsinghua Univ. (China) ...... [6508-03] 10:10 am: A robustified hidden Markov model for visual tracking with Coffee Break ...... 10:30 to 11:00 am subspace representation, J. Gai, R. L. Stevenson, Univ. of Notre Dame ...... [6508-09] 11:00 am: Comparison of standard-based H.264 error-resilience techniques and multiple-description coding for robust MIMO-enabled video Coffee Break ...... 10:30 to 11:00 am transmission, M. Tesanovic, D. R. Bull, D. Agrafiotis, A. Doufexi, Univ. of Bristol 11:00 am: Optimal multiple sprite generation based on physical camera (United Kingdom) ...... [6508-04] parameter estimation, M. Kunter, A. Krutz, Technische Univ. Berlin (Germany); 11:20 am: Spatial and temporal models for texture-based video coding, F. M. K. Mandal, Univ. of Alberta (Canada); T. Sikora, Technische Univ. Berlin Zhu, Purdue Univ...... [6508-05] (Germany) ...... [6508-10] 11:40 am: Content adaptive motion estimation for efficient video 11:20 am: Geometrical image filtering with connected operators and image compression, L. Liu, Purdue Univ.; Y. Liu, Sun Microsystems, Inc.; E. J. Delp III, inpainting, M. Dimiccoli, P. Salembier, Univ. Politècnica de Catalunya Purdue Univ...... [6508-06] (Spain) ...... [6508-11] 11:40 am: Maximum-entropy expectation-maximization algorithm for image processing, H. Hong, D. Schonfeld, Univ. of Illinois/Chicago ...... [6508-12] Lunch/Exhibition Break ...... 12:00 to 1:30 pm electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 51 Conference 6508

SESSION 3 ✔ Fast luminance and chrominance correction based on motion compensated linear regression for multiview video coding, W. Chen, L. Conv. Ctr. Rooms B2 and B3 ...... Tues. 1:30 to 5:00 pm Ding, L. Chen, National Taiwan Univ. (Taiwan) ...... [6508-74] Special Session: The Mathematics of Imaging ✔ Complexity control of fast motion estimation in H.264/MPEG-4 AVC with Chair: Mireille Boutin, Purdue Univ. rate-distortion-complexity optimization, M. Wu, S. Forchhammer, S. M. Aghito, Danmarks Tekniske Univ. (Denmark) ...... [6508-75] 1:30 pm: Mathematical aspects of shape analysis for object recognition (Invited Paper), P. F. Stiller, Texas A&M Univ...... [6508-13] ✔ Wyner-Ziv residual coding for wireless multiview system, Z. Jin, M. Yu, G. Jiang, Ningbo Univ. (China) ...... [6508-76] 2:00 pm: Challenges in 3DTV (Invited Paper), A. Redert, R. M. Berretty, Philips Research Labs. (Netherlands) ...... [6508-14] ✔ Progressive image transmission with RCPT protection, L. Cao, The Univ. of Mississippi ...... [6508-77] 2:30 pm: A survey of polynomial fitting and segmentation methods for generalized principal component analysis (Invited Paper), S. R. Rao, Univ. of ✔ A sender-driven time-stamp controlling based dynamic light field Illinois at Urbana-Champaign; H. Derksen, Univ. of Michigan; R. M. Fossum, Y. streaming service, Z. Han, Q. Dai, Y. Liu, Tsinghua Univ. (China) . [6508-78] Ma, A. W. Wagner, A. Y. Yang, Univ. of Illinois at Urbana-Champaign . [6508-15] ✔ A fast and quality-preserving method for H.264 encoding using dynamic Coffee Break ...... 3:00 to 3:30 pm SAD maps, D. Wu, Institute for Infocomm Research (Singapore) . . [6508-79] 3:30 pm: A lossy coding heuristic for mixed data segmentation (Invited ✔ Impact of rate control tools on very fast non-embedded wavelet image Paper), H. Derksen, Univ. of Michigan; Y. Ma, W. Hong, J. Wright, Univ. of Illinois encoders, O. M. López Granado, M. O. Martínez Rach, Univ. Miguel at Urbana-Champaign ...... [6508-16] Hernández de Elche (Spain); J. Oliver, Univ. Politècnica de València (Spain); M. J. Perez Malumbres, Univ. Miguel Hernández de Elche (Spain) . [6508-81] 4:00 pm: Integral invariants for curves in 3D: inductive approach (Invited Paper), I. A. Kogan, K. Hamid, S. Feng, North Carolina State Univ. . . [6508-17] ✔ Non-rigid object tracker based on a robust combination of parametric active contour and point distribution model, J. I. Olszewska, Univ. 4:30 pm: Variable separation for 3D from 2D (Invited Paper), J. Zhang, M. Catholique de Louvain (Belgium); T. Mathes, Univ. de Liège (Belgium); C. De Boutin, D. G. Aliaga, Purdue Univ...... [6508-18] Vleeschouwer, Univ. Catholique de Louvain (Belgium); J. Piater, Univ. de Liège (Belgium); B. Macq, Univ. Catholique de Louvain (Belgium) . [6508-82] Posters and Demonstrations-Tuesday ✔ Optimal packet scheduling and rate control for video streaming, E. Demonstrations ...... 5:30 to 8:30 pm Gurses, Norwegian Univ. of Science and Technology (Norway); G. B. Akar, Middle East Technical Univ. (Turkey); N. Akar, Bilkent Univ. (Turkey) [6508-83] A symposium-wide demonstration session will be open to attendees 5:30 to ✔ Evaluation of a combined preprocessing and H.264-compression 8:30 pm Tuesday evening in Exhibit Hall 1. Demonstrators will provide scheme for 3D integral images, R. Olsson, M. Sjöström, Y. Xu, Mid Sweden interactive, hands-on demonstrations of a wide-range of products related to Univ. (Sweden) ...... [6508-84] Electronic Imaging. ✔ Key frame extraction from unstructured consumer video clips, C. Papin, Posters ...... 5:30 to 7:00 pm J. Luo, Eastman Kodak Co...... [6508-85] ✔ Multiresolution mesh segmentation based on surface roughness and Posters will be placed on display after 10:00 am in Exhibit Hall 1. wavelet analysis, C. Roudet, F. Dupont, A. M. Baskurt, Univ. de Claude A poster session, with authors present at their posters, Bernard Lyon I (France) ...... [6508-86] will be held Tuesday evening, 5:30 to 7:00 pm. ✔ Occlusion and split detection and correction for object tracking in ✔ An improved fast interpolation approach for H.264, D. Wu, Institute for surveillance applications, C. A. Vázquez, Communications Research Ctr. Infocomm Research (Singapore) ...... [6508-60] Canada (Canada); M. Ghazal, A. Amer, Concordia Univ. (Canada) . [6508-87] ✔ A comparative study of image compression based on directional ✔ Optimal reverse frame selection for stored video delivery under wavelets, K. Li, W. Xu, Q. Dai, Tsinghua Univ. (China); Y. Wang, Polytechnic constrained resources, D. Tao, J. Cai, Nanyang Technological Univ. Univ...... [6508-61] (Singapore) ...... [6508-88] ✔ Locally adaptive reconstruction of lost low-frequency coefficients in ✔ Post-processing for decoding without update step in motion- wavelet coded images, J. Rombaut, A. Pizurica, W. R. Philips, Univ. Gent compensated lifted wavelet video coding, A. A. Mavlankar, M. H. Flierl, B. (Belgium) ...... [6508-62] Girod, Stanford Univ...... [6508-89] ✔ Pose estimation from video sequences based on Sylvester’s equation, C. ✔ Wavelet-based multiple description coding of 3D geometry, A. Norkin, Chen, J. Yang, D. Schonfeld, Univ. of Illinois/Chicago; M. A. Mohamed, Middle East Technical Univ. (Turkey) and Tampere Univ. of Technology Motorola, Inc...... [6508-63] (Finland); M. O. Bici, G. Bozdagi Akar, Middle East Technical Univ. (Turkey); A. ✔ A fast and efficient method to protect color images, M. Chaumont, W. P. Gotchev, J. T. Astola, Tampere Univ. of Technology (Finland) . . . [6508-90] Puech, Lab. d’Informatique de Robotique et de Microelectronique de ✔ Hysteresis-based selective Gaussian mixture models for real-time Montpellier (France) ...... [6508-64] background maintenance, F. Achkar, A. Amer, Concordia Univ. ✔ 3D face model tracking based on a multiresolution active search, M. (Canada) ...... [6508-91] Chaumont, W. Puech, Lab. d’Informatique de Robotique et de ✔ View interpolation by inverse filtering: generating the center view using Microelectronique de Montpellier (France) ...... [6508-65] multiview images of circular camera array, A. Kubota, Tokyo Institute of ✔ Region-based hidden Markov models for image categorization and Technology (Japan); K. Kodama, National Institute of Informatics (Japan); Y. retrieval, F. Li, Q. Dai, W. Xu, Tsinghua Univ. (China) ...... [6508-66] Hatori, Tokyo Institute of Technology (Japan) ...... [6508-92] ✔ A new support tool for machine learning and pattern recognition using tracking and motion segmentation, E. Bichot, O. Masset, L. Mascarilla, P. Courtellemont, Univ. de La Rochelle (France) ...... [6508-67] ✔ Pet fur color and texture classification, J. Yen, Toshiba America Business Solutions; D. Mukherjee, S. Lim, D. R. Tretter, Hewlett Packard Labs...... [6508-68] ✔ A simple reversed-complexity Wyner-Ziv video coding mode based on a spatial reduction framework, D. Mukherjee, Hewlett-Packard Labs.; B. Macchiavello, R. L. de Queiroz, Univ. de Brasília (Brazil) ...... [6508-69] ✔ Motion refined medium granular scalability, Z. Li, W. Yao, S. Rahardja, Institute for Infocomm Research (Singapore) ...... [6508-71] ✔ An efficient multiframe dynamic search range motion estimation for H.264, Q. Sun, J. Wang, X. Chen, L. Yu, Zhejiang Univ. (China) . . . [6508-72] ✔ An area-efficient VLSI architecture for AVS intraframe encoder, K. Zhang, L. Yu, Zhejiang Univ. (China) ...... [6508-73]

52 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508

✔ A novel framework for improving bandwidth utilization for VBR video Wednesday 31 January delivery over wide-area networks, J. Yuan, S. Roy, Q. Sun, Institute for Infocomm Research (Singapore) ...... [6508-93] ✔ H.264/AVC error resilient video streaming using leaky prediction, Y. Liu, Plenary Session Sun Microsystems Inc.; R. Kurceren, Nokia Americas ...... [6508-94] ✔ A fast and accurate characteristic-based rate-quantization model for Marriott Ballroom 1-6 ...... 8:00 to 9:20 am video transmission, D. Chan, National Chiayi Univ. (Taiwan); W. Chien, C. 8:00 am: More than a Poplar Plank: The Shape and Subtle Colors of the Chang, C. Chou, C. Lin, J. Hu, National Taiwan Univ. (Taiwan) . . . . [6508-95] Masterpiece ‘Mona Lisa’ by Leonardo, Francois Blais, John Taylor, Luc ✔ A double motion-compensated orthogonal transform with energy- Cournoyer, Michel Picard, Louis Borgeat, Guy Godin, J.-Angelo Beraldin, concentration constraint, M. H. Flierl, Stanford Univ...... [6508-96] Marc Rioux, National Research Council Canada; Christian Lahanier, Bruno Mottin, C.re de Recherche et de Restauration des Musées de France ✔ Image compression using constrained relaxation, Z. He, Univ. of Missouri/ Columbia ...... [6508-97] 8:40 am: Space Exploration, Imaging Technology, and the Development of a Virtual Presence in Space, Eric M. DeJong, Principal Investigator, ✔ Rate-distortion optimized color quantization for compound image Solar System Visualization (SSV) Project compression, W. Ding, Univ. of Science and Technology of China (China); Y. Lu, F. Wu, S. Li, Microsoft Research Asia (China) ...... [6508-98] See pg. 7 for details. ✔ Dynamic GOP structure for scalable video coding, H. S. Kim, SungKongHoe Univ. (South Korea); C. W. Chen, Florida Institute of SESSION 4 Technology ...... [6508-99] Conv. Ctr. Rooms B2 ...... Wed. 9:30 am to 12:30 pm ✔ The wavelet-based multiresolution motion estimation using temporal aliasing detection, T. Lee, D. V. Anderson, Georgia Institute of Special Session: Collaborative Object Tracking Technology ...... [6508-100] Chairs: Fatih M. Porikli, Mitsubishi Electric Research Labs.; Ying Wu, ✔ Adaptive P2P video streaming via packet labeling, J. Chakareski, P. Northwestern Univ. Frossard, École Polytechnique Fédérale de Lausanne (Switzerland)[6508-101] 9:30 am: Bayesian distributed articulated object tracking using multiple ✔ Using machine learning for fast intra MB coding in H.264, H. Kalva, L. collaborative trackers (Invited Paper), W. Qu, D. Schonfeld, Univ. of Illinois/ Christodoulou, Florida Atlantic Univ...... [6508-102] Chicago ...... [6508-19] ✔ Subband motion compensation for spatially scalable video coding, M. L. 10:00 am: Tracking people in mixed modality systems (Invited Paper), Y. Comer, R. Zhang, Purdue Univ...... [6508-103] Ivanov, Mitsubishi Electric Research Labs...... [6508-20] ✔ Nonlinear up-sampling for image coding in a spatial pyramid, M. Coffee Break ...... 10:30 to 11:00 am Beermann, Univ. of Ottawa (Canada); J. Ohm, RWTH Aachen 11:00 am: Multiple hypothesis shape tracking using particle filtering and (Germany) ...... [6508-104] Hough-based observation models (Invited Paper), A. Dore, M. Asadi, C. S. ✔ Collusion attack on a scalable AND-ACC fingerprinting scheme, Y. Wu, Z. Regazzoni, Univ. degli Studi di Genova (Italy) ...... [6508-21] Zhao, Institute for Infocomm Research (Singapore) ...... [6508-105] 11:30 am: Collaborative tracking of objects in EPTZ cameras (Invited ✔ Improvements of multiple FGS layers coding for low-delay applications Paper), F. Bashir, F. M. Porikli, Mitsubishi Electric Research Labs. . . . [6508-22] in SVC, Y. Zheng, X. Ji, Institute of Computing Technology (China); F. Wu, 12:00 pm: Particle filter-based camera tracker fusing marker and feature Microsoft Research Asia (China); D. Zhao, Harbin Institute of Technology point-based cues (Invited Paper), D. Marimon, Y. Maret, Y. Abdeljaoued, T. (China); W. Gao, Institute of Computing Technology (China) . . . . . [6508-106] Ebrahimi, École Polytechnique Fédérale de Lausanne (Switzerland) . . [6508-23] ✔ Achieving high compression efficiency with distributed video coding, S. Lunch/Exhibition Break ...... 12:30 to 2:00 pm Milani, Univ. degli Studi di Padova (Italy); J. Wang, K. Ramchandran, Univ. of California/Berkeley ...... [6508-107]

Sessions 5 and 6-7 run concurrently

SESSION 5 SESSION 6 Conv. Ctr. Rooms B2 ...... Wed. 2:00 to 4:30 pm Conv. Ctr. Room B3 ...... Wed. 2:00 to 3:00 pm Distributed Video Coding Wavelet Representation and Coding Chairs: Edward J. Delp III, Purdue Univ.; Kannan Ramchandran, Chair: John W. Woods, Rensselaer Polytechnic Institute Univ. of California/Berkeley 2:00 pm: Scalable direction representation for image compression with 2:00 pm: Robust distributed multiview video compression for wireless direction-adaptive discrete wavelet transform, T. Xu, C. Chang, B. Girod, camera networks, C. Yeo, K. Ramchandran, Univ. of California/ Stanford Univ...... [6508-30] Berkeley ...... [6508-24] 2:20 pm: Video coding with fully separable wavelet and wavelet packet 2:20 pm: Hybrid key/Wyner-Ziv frames with flexible macroblock ordering transforms, M. S. Trocan, B. Pesquet-Popescu, École Nationale Supérieure des for improved low delay distributed video coding, D. Agrafiotis, P. Ferré, D. R. Télécommunications (France) ...... [6508-31] Bull, Univ. of Bristol (United Kingdom) ...... [6508-25] 2:40 pm: Anisotropic multidirectional representation with oriented lifting 2:40 pm: Distributed video coding based on constrained rate adaptive scheme, G. Jeannic, V. Ricordel, D. Barba, Univ. de Nantes (France) . [6508-32] LDPC codes, R. Liu, Beijing Univ. of Aeronautics and Astronautics (China) and Coffee Break ...... 3:00 to 3:30 pm Florida Institute of Technology; G. Hua, C. W. Chen, Florida Institute of Technology ...... [6508-26] SESSION 7 Coffee Break ...... 3:00 to 3:30 pm Conv. Ctr. Room B3 ...... Wed. 3:30 to 4:30 pm 3:30 pm: Unequal error protection using Wyner-Ziv coding, L. Liang, Purdue Univ.; P. Salama, Indiana Univ.-Purdue Univ. at Indianapolis; E. J. Delp III, Image Registration and Recognition Purdue Univ...... [6508-27] Chair: Robert L. Stevenson, Univ. of Notre Dame 3:50 pm: Generalized in-scale motion compensation framework for spatial 3:30 pm: Multimodal image registration based on edges and junctions, Y. Li, scalable video coding, R. Xiong, Institute of Computing Technology (China); J. R. L. Stevenson, Univ. of Notre Dame ...... [6508-33] Xu, F. Wu, S. Li, Microsoft Research Asia (China) ...... [6508-28] 3:50 pm: Automatic target segmentation in color dental images, J. Luo, M. 4:10 pm: Fast prediction model of adaptive GOP structure for SVC, Y. Chen, R. Bolin, Eastman Kodak Co...... [6508-108] C. Lin, C. Chen, L. Chen, National Taiwan Univ. (Taiwan) ...... [6508-29] 4:10 pm: Comparison of compression algorithms’ impact on fingerprint and face recognition accuracy, A. Uhl, Univ. Salzburg (Austria) ...... [6508-35]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 53 Conference 6508

Thursday 1 February

SESSION 8 Conv. Ctr. Rooms B2 ...... Thurs. 8:30 am to 12:00 pm Special Session: Next-Generation Video Coding Technologies Chairs: C.-C. J. Kuo, Univ. of Southern California; Shawmin Lei, Sharp Labs. of America, Inc.; Shijun Sun, Sharp Labs. of America, Inc. 8:30 am: Film grain noise modeling in advanced video coding (Invited Paper), B. T. Oh, Univ. of Southern California; S. Sun, L. Shawmin, Sharp Labs. of America, Inc.; C. J. Kuo, Univ. of Southern California ...... [6508-36] 9:00 am: Advances in hybrid video coding (Invited Paper), T. Wedi, S. Wittmann, T. Palfner, B. Schuur, F. Knicker, Panasonic R&D Ctr. Germany GmbH (Germany) ...... [6508-37] 9:30 am: Next-generation video coding for mobile applications: market trends, requirements, and technologies (Invited Paper), M. Budagavi, M. Zhou, Texas Instruments Inc...... [6508-38] Coffee Break ...... 10:00 to 10:30 am 10:30 am: Adaptive filtering for cross-view prediction in multiview video coding (Invited Paper), P. Lai, Univ. of Southern California; Y. Su, P. Yin, C. Gomila, THOMSON; A. Ortega, Univ. of Southern California ...... [6508-39] 11:00 am: RD-optimized competition scheme for efficient motion prediction (Invited Paper), J. Jung, G. Laroche, France Télécom R&D (France); B. Pesquet-Popescu, École Nationale Supérieure des Télécommunications (France) ...... [6508-40] 11:30 am: High-definition video coding with super-macroblocks (Invited Paper), S. Ma, C. J. Kuo, Univ. of Southern California ...... [6508-41] Lunch Break ...... 12:00 to 1:30 pm

Sessions 9 and 10 run concurrently

SESSION 9 SESSION 10 Conv. Ctr. Rooms B2 ...... Thurs. 1:30 to 3:30 pm Conv. Ctr. Room B3 ...... Thurs. 1:30 to 3:30 pm Media Communication and Networking Imaging Systems Chairs: Zhihai He, Univ. of Missouri/Columbia; Michael W. Marcellin, Chair: Anthony Vetro, Mitsubishi Electric Research Labs. The Univ. of Arizona 1:30 pm: Edge-based automatic white balancing with linear illuminant 1:30 pm: Cross-layer optimization for wireless video communication, D. Wu, constraint, H. H. Chen, C. Shen, National Taiwan Univ. (Taiwan) . . . . [6508-48] Univ. of Florida; Z. He, Univ. of Missouri/Columbia ...... [6508-42] 1:50 pm: Highly automated image recomposition: picture you wish you had 1:50 pm: A more aggressive prefetching scheme for streaming media taken, J. Luo, P. Lei, Eastman Kodak Co...... [6508-49] delivery over the Internet, J. Yuan, Q. Sun, S. Rahardja, Institute for Infocomm 2:10 pm: Symmetric trinocular dense disparity estimation for car Research (Singapore) ...... [6508-43] surrounding camera array, Y. Chang, Y. Tsai, L. Chen, National Taiwan Univ. 2:10 pm: Rate-smoothed encoding for real-time video streaming (Taiwan) ...... [6508-50] applications, J. Wu, J. Cai, Nanyang Technological Univ. (Singapore) [6508-44] 2:30 pm: Surveillance system with mega-pixel scalable transcoder, T. Hata, 2:30 pm: Joint source-channel rate allocation in parallel channels, L. Pu, M. N. Kuwahara, Mitsubishi Electric Corp. (Japan); D. L. Schwenke, A. Vetro, W. Marcellin, I. B. Djordjevic, B. V. Vasic, A. Bilgin, The Univ. of Arizona[6508-45] Mitsubishi Electric Research Labs...... [6508-51] 2:50 pm: Constant quality JPEG2000 rate control for digital cinema, M. 2:50 pm: A 2D gel electrophoresis DNA image analysis algorithm with Smith, Consultant; J. D. Villasenor, Univ. of California/Los Angeles . . [6508-46] automatic thresholding, N. Kaabouch, R. R. Schultz, Univ. of North Dakota ...... [6508-52] 3:10 pm: On preserving robustness-false alarm tradeoff in media hashing, S. Roy, X. Zhu, J. Yuan, Institute for Infocomm Research (Singapore); E. Chang, 3:10 pm: Automatic estimation and compensation of geometric distortions National Univ. of Singapore (Singapore) ...... [6508-47] in copy videos, B. Chupeau, A. Massoudi, F. Lefebvre, THOMSON (France) ...... [6508-53] Coffee Break ...... 3:30 to 4:00 pm

54 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508

Sessions 11 and 12 run concurrently

SESSION 11 SESSION 12 Conv. Ctr. Rooms B2 ...... Thurs. 4:00 to 5:00 pm Conv. Ctr. Room B3 ...... Thurs. 4:00 to 5:00 pm Video Coding II Motion Estimation Chair: Amir Said, Hewlett-Packard Labs. Chair: Antonio Ortega, Univ. of Southern California 4:00 pm: A novel statistical learning-based rate distortion analysis 4:00 pm: Enhanced SAD reuse fast motion estimation, K. L. Tang, K. N. approach for multiscale binary shape coding, Z. Chen, K. N. Ngan, The Ngan, The Chinese Univ. of Hong Kong (Hong Kong China) ...... [6508-57] Chinese Univ. of Hong Kong (Hong Kong China) ...... [6508-54] 4:20 pm: Motion estimation performance models with application to 4:20 pm: Drift-compensated coding optimization for fast bit-rate reduction hardware error tolerance, H. Cheong, A. Ortega, Univ. of Southern transcoding, P. Zhang, X. Ji, W. Gao, Institute of Computing Technology California ...... [6508-58] (China) ...... [6508-55] 4:40 pm: Super-resolution based on region-matching motion estimation, O. 4:40 pm: Fast intermode decision algorithm in H.264 for IPTV broadcasting A. Omer, T. Tanaka, Tokyo Univ. of Agriculture and Technology (Japan) [6508-59] server, G. Y. Kim, B. Y. Yoon, Electronics and Telecommunications Research Institute (South Korea); Y. Ho, Gwangju Institute of Science and Technology (South Korea) ...... [6508-56]

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 55 Technical Abstract Summaries

3D Imaging, Interaction, and Multimedia Processing Measurement and Applications Conf. 6490A Stereoscopic Displays and Applications XVIII Conf. 6504 Multimedia Computing and Networking 2007 (Woods, Dodgson, Merritt) ...... p. 57 (Zimmermann, Griwodz) ...... p. 166 Conf. 6490B The Engineering Reality of Virtual Reality 2007 Conf. 6505 Security, Steganography, and Watermarking (Bolas, McDowall) ...... p. 69 of Multimedia Contents IX (Delp, Wong) . . . p. 171 Conf. 6491 Videometrics IX (Beraldin, Remondino, Conf. 6506 Multimedia Content Access: Algorithms and Shortis) ...... p. 71 Systems (Hanjalic, Schettini, Sebe) ...... p. 182 Conf. 6507 Multimedia on Mobile Devices 2007 Imaging, Visualization, and Perception (Creutzburg, Takala, Cai) ...... p. 184 Conf. 6492 Human Vision and Electronic Imaging XII (Rogowitz, Pappas, Daly) ...... p. 79 Visual Communications and Image Conf. 6493 Color Imaging XII: Processing, Hardcopy, Processing and Applications (Eschbach, Marcu) ...... p. 91 Conf. 6508 Visual Communications and Image Processing Conf. 6494 Image Quality and System Performance IV 2007 (Chen, Schonfeld, Luo) ...... p. 193 (Cui, Miyake) ...... p. 101 Conf. 6495 Visualization and Data Analysis 2007 (Erbacher, Roberts, Gröhn, Börner) ...... p. 107 Image Processing Conf. 6496 Real-Time Image Processing 2007 (Kehtarnavaz, Carlsohn) ...... p. 112 Conf. 6497 Image Processing: Algorithms and Systems VI (Astola, Egiazarian, Dougherty) . p. 117 Conf. 6498 Computational Imaging V (Bouman, Miller, Pollak) ...... p. 126 Conf. 6499 Vision Geometry XV (Latecki, Mount, Wu) . . p. 135 Conf. 6500 Document Recognition and Retrieval XIV (Lin, Yanikoglu) ...... p. 139 Digital Imaging Sensors and Applications Conf. 6501 Sensors, Cameras, and Systems for Scientific/ Industrial Applications X (Blouke) ...... p. 145 Conf. 6502 Digital Photography III (Martin, DiCarlo, Sampat) ...... p. 153 Conf. 6503 Machine Vision Applications in Industrial Inspection XV (Meriaudeau, Niel) ...... p. 160

56 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A: Stereoscopic Displays and Applications XVIII Monday-Wednesday 29-31 January 2007 Part of Proceedings of SPIE Vol. 6490A Stereoscopic Displays and Applications XVIII

6490A-01, Session 1 After outlining the problem and potential of this solution we describe the initial construction of a simple Wheatstone stereoscope using 23" Stereoscopic applications for design visualization LCD displays and then show engineering improvements that increase K. J. Gilson, Parsons Brinckerhoff the FOV and usability of the system. The applicability of a high-immersion, high-resolution display for art, Advances in display technology and 3D design visualization entertainment, and simulation is presented along with a content applications have made real-time stereoscopic visualization of production system that utilizes the capabilities of the system. We then architectural and engineering projects a reality. Parsons Brinkerhoff (PB) discuss the potential use of the system for VR pain control therapy, is a transportation consulting firm that has used digital visualization treatment of post-traumatic stress disorders and other serious games tools from their inception and has helped pioneer the application of applications. those tools to large scale infrastructure projects. PB is now exploring the use of stereoscopic displays and tools to aid in the design and visualization of those projects. PB is one of the first Architecture/ 6490A-05, Session 1 Engineering/Construction (AEC) firms to implement a CAVE- an immersive presentation environment that includes stereoscopic rear- A cohesive system for real-time stereoscopic image projection capability. The firm also employs a portable stereoscopic processing and evaluation front-projection system, and shutter-glass systems for smaller groups. E. M. Lazarus, Bristlecone Corp.; R. Galli, ANDXOR Corp. PB is using commercial real-time 3D applications in combination with traditional 3D modeling programs to visualize and present large AEC In this paper we define an innovative real time modular system to projects to planners, clients and decision makers in stereo. These visualize, capture, manage, securely preserve, store, visualize and presentations create more immersive and spatially realistic playback stereoscopic images. presentations of the proposed designs. This paper will present the The system, called “Solid-Look” together with the stereoscopic basic display tools and applications, and the 3D modeling techniques cameras “StereOpsis” will allow military, EOD specialists, and private PB is using to produce interactive stereoscopic content. The paper will industry operators to literally “see through the robot’s eyes”. discuss several architectural and engineering design visualizations we The stereo cameras will perform zooming and image stabilization for a have produced. controlled and smooth vision. The display will enable high definition stereoscopic vision without the 6490A-03, Session 1 need of glasses. Evolution of the VarrierTM autostereoscopic VR display: The system will enable the operator to control the robot as if his/her head were located on the robot itself, positioning and zooming the 2001-2007 camera to the visual target object using the operator’s eye and head T. Peterka, R. L. Kooima, J. I. Girado, J. Ge, D. J. Sandin, T. A. movement, without any wearable devices and allowing the operator’s DeFanti, Univ. of Illinois/Chicago hands to perform other tasks. Varrier is a barrier strip autostereo virtual reality display system, offering Every image frame will be authenticated, encrypted and timestamped a head-tracked ortho-stereo first person interactive experience, without to allow authenticity and confidentiality during post-capture playback the need for glasses or other gear. A problem-solution timeline of the or to show evidence in court. last six years is presented, including algorithmic and systemic The system will secure the ability to operate it using administrator’s developments such as a fine-resolution barrier, and a GPU-based sub- biometrical authentication. pixel algorithm that improves performance and quality. A camera-based Solid-Look modular design can be used in multiple industries from tracking system developed specifically for Varrier uses multiple Homeland Security to Pharmaceutical including research, forensic and cameras to capture subjects at high frame rates, and a neural network underwater inspections and will certainly provide great benefit to the recognizes and tracks them in 3D space with accuracy comparable to performance, speed and accuracy of the operations. commercial tracking systems. Varrier supports a variety of VR applications, including visualization of polygonal, ray traced, and volume rendered data, autostereo movie playback, and interactive 6490A-68, Session 1 manipulation of 3D models. Local as well as distributed computation is employed in various applications including autostereo teleconferencing. A novel technique for visualizing high-resolution 3D Art, medicine, and science have been exhibited in Varrier, which exists terrain maps in a variety of form factors from large tiled installations to smaller desktops. Newest developments include the use of a dynamic parallax J. F. Dammann, Jr., Army Research Lab. barrier that affords larger depth working range, rapid steering of A new technique is presented for visualizing high-resolution terrain images, and support for two viewers who each receive an independent elevation data. It produces realistic images at small scales and works pair of perspective views. particularly well for natural objects. Better visualization at small scales opens up new applications, like site surveillance, now done in 2-D. Also, Google Earth-type exploration tasks available in 3-D only in 6490A-04, Session 1 restricted, specially-prepared areas can now be extended to much Re-inventing the stereoscope for the 21st Century larger areas requiring only unprepared map data. The traditional technique drapes a continuous surface over the J. S. Kollin, Univ. of Washington; A. J. Hollander, Imprint Interactive regularly-spaced elevation values. This technique works well for large Technology areas, but falls apart at small scales with natural objects like trees. While discussing the current state of stereo head-mounted and 3D The new technique visualizes the terrain as a set of disjoint patches projection displays, the authors came to the realization that flat-panel combined with an algorithm that identifies smooth areas within the LCD displays offer higher resolution than projection for stereo display at scene. Where the terrain is smooth, it warps the patches to create a a low (and continually dropping) cost. More specifically, we can achieve smooth surface. For trees or shrubs, however, the patches are left an angular resolution approaching that of the human eye with a field-of- disjoint making natural terrain look realistic - trees and bushes look view (FOV) greater than 90 degrees for less than $1500. stylized but still look natural. Also, the new technique does not add significant artifacts to the map like extending bridges and overpasses For many immersive applications head tracking is unnecessary and down to the ground. The new technique is illustrated using 1-m sometimes even undesirable, and a low cost/high quality wide FOV resolution 3-D maps from the Rapid Terrain Visualization (RTV) display may significantly increase the application space for 3D display. program and will be demonstrated on a Sharp 3-D laptop. electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 57 Conference 6490A: Stereoscopic Displays and Applications XVIII

6490A-06, Session 2 surface in real-time. It has miniature dimensions, just a few mm in diameter, typically located distally in a scope or on a tool and has the Stereoscopic medical imaging collaboration system potential to be disposable at appropriate volumes. F. Okuyama, Suzuka Univ. of Medical Science (Japan) A single Pentium processor receives the two views of the scene, corrects for spatial distortions and calculates the disparity of each The computerization of the clinical record and the realization of the pixel by matching the left image with the corresponding pixel in the multimedia has brought improvement of the medical service in right image. All algorithms are processed in real time. medical facilities. It is very important for the patients to obtain comprehensible informed consent. The doctor should plainly explain The result of this process creates a disparity map which includes the the purpose and the content of the diagnoses and treatments for the distance in pixels between every two corresponding pixels, and the patient. Therefore there is sees of informed consent and second confidence of this value. The 3D reconstructed surface information opinion which are easy to understand the medical images from the can be used for real-time registration of the patient for navigation patient side. It is useful that a stereoscopic image indicates an systems in non- or minimally-invasive applications. It can be also affected part to understand this medical image easily because the used for registration of various volumetric imaging modalities to the medical image is hard to understand in general. real-time stereoscopic video image, based on anatomical landmarks and/or fiducials. The real-time MIS navigation registration and °Æsee We propose a Stereoscopic Medical Imaging Collaboration System through tissue°Ø fusion solutions enable the development of new MIS that provides a stereoscopic medical image (a diagnosis image such procedures in different surgical segments, such as spine, abdomen, as X-ray CT, and MRI), which is easy for a doctor and a patient to cardio-thoracic and brain. understand, on demand and control of the image in remote. We designed the Stereoscopic Medical Imaging Collaboration System. This system is constructed from the common server computer which 6490A-09, Session 2 stores diagnosis image data, the client computers which the doctors operate and the stereoscopic displays in which the medical image is A virtual reality oriented clinical experiment on post- presented. The high-speed communication networks connect these stroke rehabilitation: performance and preference apparatus. Furthermore, they can operate the stereoscopic image to comparison among different stereoscopic displays observe the affected part easily, for example, the rotation of the S. Yeh, Sr., A. S. Rizzo III, A. A. Sawchuk, Univ. of Southern object, depth operation. Moreover, the same object is observed California simultaneously by different clients and by transmitting an operation parameter to operate image from another client to more than one We have developed a novel VR task: the Dynamic Reaching Test, that client. Thus, medical diagnosis support between the doctors can be measures human forearm movement in 3D space. In this task, three done. different stereoscopic displays: autostereoscopic (AS), shutter glasses (SG) and head mounted display (HMD), are used in tests in which subjects must catch a virtual ball thrown at them. Parameters such as 6490A-07, Session 2 percentage of successful catches, movement efficiency (subject path length compared to minimal path length), and reaction time are Using a high-definition stereoscopic video system to measured to evaluate differences in 3D perception among the three teach microscopic surgery stereoscopic displays. The SG produces the highest percentage of J. F. R. Ilgner, J. J. H. Park, D. Labbe, Univ. Hospital Aachen successful catches, though the difference between the three displays (Germany) is small, implying that users can perform the VR task with any of the displays. The SG and HMD produced the best movement efficiency, Motivation: To evaluate the use of high-definition (HD) stereoscopic while the AS was slightly less efficient. Finally, the AS and HMD video monitoring in microscopic surgery for teaching junior surgeons produced similar reaction times that were slightly higher (by 0.1 s) and medical students. than the SG. We conclude that SG and HMD displays were the most Material and methods: We attached a 1280x1024 HD stereo camera effective, but only slightly better than the AS display. (MicroVision Systems, Santa Barbara, CA) to an operating microscope (Zeiss ProMagis, Zeiss Co., Oberkochen, Germany), attached to a PC workstation. The live image was displayed by two LCD projectors on a 6490A-10, Session 3 rear-projection screen by polarized filters. The junior surgeon A novel triangulation method for building parallel- interacted with all other participants (teaching surgeon, nurse and medical students), sharing the same stereoscopic image. perspective stereo mosaics Results: Four major effects were obtained: A) Stereoscopy facilitated P. K. Gurram, E. Saber, H. Rhody, Rochester Institute of orientation in the operative field for all junior staff. B) By means of the Technology stereoscopic screen image the senior surgeon could instantly In this paper, we propose an improved triangulation method for demonstrate the next surgical steps to his junior colleague. C) The building parallel-perspective stereo mosaics using ray interpolation. theatre nurse anticipated the next instruments which were needed. D) The currently available fast PRISM (Parallel Ray Interpolation for Medical students were immersed in the operative procedure, thus Stereo Mosaicing) uses a constant triangulation technique that does avoiding the need for an extra teaching session. not take advantage of the spatial information. This works well when Conclusion: High definition stereoscopy bears the potential to the inter-frame displacement in the video is small. However, for large compress the learning curve for undergraduate and postgraduate inter-frame displacements, hence large motion parallax, we observe medical professionals. Further studies will focus on the long term visual artifacts in the mosaics when the source triangles in the video effect for operative training and on creating off-line content for frames are warped into destination triangles in the mosaics using interactive teaching media. affine transformations. Our method uses the edge information present in the scene to ensure that triangles do not cross over from one planar facet to the other. The video frames are segmented and the edges are 6490A-08, Session 2 obtained from the boundaries of different segments. The motion over the entire region in each triangle is constant. Hence, this method Stereoscopic video system provides real-time avoids unwanted warping of objects with large motion parallax in the registration for endoscopic surgery scene and reduces the visual artifacts. The performance of our A. Yaron, N. Horesh, M. Shechterman, Visionsense Ltd. (Israel) proposed technique is demonstrated on a series of simple to complex video imagery collected by utilizing cameras and airborne sensors. The Visionsense single stereoscopic vision sensor, matched to a single optical channel with dual apertures, provides simultaneous left and right views of the scene (stereo video) and 3D coordinates of a

58 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A: Stereoscopic Displays and Applications XVIII

6490A-11, Session 3 newly exposed regions using a simple depth-aware inpainting technique. To validate the proposed algorithm we generated new Surrogate depth maps for stereoscopic imaging: different views from a 2D image plus a depth map. The images were viewed on edge types a SynthaGram auto-stereoscopic display (nine views) and on a stereoscopic display (two views) with shutter glasses. Our results W. J. Tam, C. A. Vázquez, F. Speranza, Communications Research show that this technique provides high quality images suitable for Ctr. Canada (Canada) visualization on different display technologies. We have previously demonstrated that surrogate depth maps, consisting of “depth” values at object boundaries in the image of a scene, are effective for stereoscopic depth image based rendering and 6490A-14, Session 3 2D-to-3D conversion. The main goal of the present study is to Spline-based intermediate view reconstruction investigate the use of surrogate depth maps with actual depth edges that are derived from cast shadows (Multiflash method). A secondary S. Ince, J. Konrad, Boston Univ.; C. Vázquez, Institut National de goal was to examine the effect of image content by using two sets of la Recherche Scientifique (Canada) natural imagesone containing mainly curved contours and the other mainly short, straight lines. For stereoscopic display, the original images Intermediate view reconstruction is an essential step in content served as the right eye view. The left eye view was rendered using preparation for multiview 3D displays and free-viewpoint video. surrogate depth maps generated with the Multiflash method and with Although many approaches to view reconstruction have been two previously suggested methods (Sobel and Standard Deviation) for proposed to date, most of them share the need to model and estimate comparison. Both depth quality and visual comfort of the stereoscopic scene depth first, and follow with the estimation of unknown-view images were assessed by a group of viewers. The results indicate that texture using this depth and other views. The approach we present in the stereoscopic images based on the Multiflash method provided an this paper follows this path as well. First, assuming a reliable disparity enhanced sensation of depth compared to a reference monoscopic (depth) map is known between two views, we present a spline-based image, but no advantage of using actual depth edges (Multiflash) over approach to unknown-view texture estimation, and compare its the other methods was found. The results also indicate a slight performance with standard disparity-compensated interpolation. A reduction in visual comfort of the images containing small straight distinguishing feature of the spline-based reconstruction is that all edges. virtual views between the two known views can be reconstructed from a single disparity field, unlike in disparity-compensated interpolation. In the second part of the paper, we concentrate on the recovery of 6490A-12, Session 3 reliable disparities especially at object boundaries. We outline an occlusion-aware disparity estimation method that we recently Object-based disparity adjusting tool for stereo panorama proposed; it jointly computes disparities in visible areas, inpaints disparities in occluded areas and implicitly detects occlusion areas. C. Wang, A. A. Sawchuk, Univ. of Southern California We then show how to combine occlusion-aware disparity estimation We want to create realistic immersive personal virtual environments with spline-based view reconstruction presented earlier, and we using stereo panoramas. We explore methods to adjust the disparity of experimentally demonstrate its benefits compared to occlusion- stereoscopic images to lie within an acceptable range for human unaware disparity-compensated interpolation. viewers and provide a comfortable stereo viewing experience. Peleg et al described a disparity adjusting method which modifies the disparity of selected objects but also the columns including the object. 6490A-15, Session 4 In this paper, we develop a human interactive object-based tool to Vertical surround correction for stereoscopic projection adjust selectively the horizontal disparity in stereo panoramas. It enhances or reduces the stereo visual effect for selected 2D object L. Lipton, Real D regions without changing the disparities of other regions in the image. Two related techniques for displaying projected stereoscopic moving Our interactive object-selecting tool is based on the mean-shift images that are comfortable to view are described. Certain means are segmentation method. The object presented in either a left or right provided for reducing the conflict of cues that occurs at the vertical image is selected by user’s inputting seed points near the desired edges of the screen surround when objects with negative parallax object boundary, and object contours both in left and right image are values are partially occluded. automatically found by our object-selecting algorithm. The complete interactive disparity-adjusting tool allows the user to select the object Perceptual problems arise when objects with such negative parallax either from manual input using a cursor, or by defining an area with a points are occluded by the vertical edges of the screen surround certain distance range, with the ability to observe the results because the mind then perceives a conflict of cues, namely, the immediately on an autostereoscopic display or other stereo display. stereoscopic binocular cue tells the eye/brain that the image is in front of the screen and the cue of interposition, a monocular cue, tells the eye/brain that the image is located behind the plane of the screen. 6490A-13, Session 3 This conflict of cues is disturbing and produces a feeling of discomfort for the viewer and if not addressed reduces the possible depth View generation for 3D-TV using image reconstruction content of the projected image. from irregularly spaced samples This paper describes means for mitigating the vertical surround edge C. A. Vazquez, Communications Research Ctr. Canada (Canada) conflicts. In one case the interaxial separation is reduced at the areas immediately adjacent to the vertical edges of the surround and in 3D-TV receptors will need to render new views from one or a small another case the stereopair is altered to produce the zero parallax number of captured and transmitted views to accommodate different condition. In both cases the transition is proportional to the proximity viewers and display technologies. Depth image-based rendering (DIBR) to the vertical edge. has been identified as a key method to solve this problem. We propose a forward-mapping disparity compensation algorithm for DIBR that does not require any approximation of the disparity map. Our algorithm 6490A-16, Session 4 uses the full precision of disparity vectors and deals with the resulting irregularly sampled image by applying a recently proposed re-sampling An application driven comparison of depth perception on algorithm based on regularization in a spline function space. The desktop 3D displays improved precision in the position of samples implies that geometrical N. S. Holliman, B. Froner, Univ. of Durham (United Kingdom); S. distortions in the rendered images are minimized. We also paid Liversedge, Univ. of Southampton (United Kingdom) attention to the occlusion and disocclusion problems. Our algorithm detects the samples that become occluded in the new view and Increasingly application users are adopting 3D displays for use in removes them from the rendering process. We fill the gaps created by tasks requiring fine depth judgment, for example in medicine and geo-

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 59 Conference 6490A: Stereoscopic Displays and Applications XVIII

science. Since these depth judgments have the potential to affect little aberration even when we use a lens array composed of simple critical decisions we feel it worthwhile investigating how well different convex lenses widely available with low price. classes of 3D display support fine depth judgment, which in this paper we take to mean image disparities of six pixels or less. We first consider in detail how to predict fine depth perception ability 6490A-19, Session 5 for four classes of desktop 3D display (full resolution, row interleaved Autostereoscopic display technology for mobile 3DTV column interleaved and colour-column interleaved). We then present applications the design of a representative task where viewers are required to judge which of two neighboring squares lies in front of the other. We J. Harrold, G. J. Woodgate, Ocuity Ltd. (United Kingdom) report results from a repeated measures within-subjects experiment presenting four levels of disparity between the two squares on seven With Mobile TV now a reality, an opportunity exists for the first mass different desktop 3D displays, both quantitative and subjective results market 3DTV products based on cell phone platforms with switchable are collected. We find differences between our predictions and the 2D/3D autostereoscopic displays. Compared to conventional cell quantitative results and additionally find that participant’s subjective phones, TV phones need to operate for extended periods of time with preferences are not always consistent with task performance results. the display running at full brightness. The efficiency of the 3D optical We discuss reasons for these differences and conclude by system will thus be a key requirement. With extended time of use, the considering which factors are most important in choosing a 3D desire for increased viewing freedom to provide greater viewing display for fine depth judgment tasks. comfort can be met by increasing the number of views presented. A four view lenticular display will have a brightness five times greater than the equivalent parallax barrier display. Therefore, lenticular 6490A-17, Session 4 displays are very strong candidates for cell phone 3DTV. The performance of candidate 3DTV panels will be assessed using Visual discomfort in stereoscopic and autostereoscopic Nyquist boundaries, human visual contrast sensitivity models and displays: a review of concepts, measurement methods, autostereoscopic display optical output simulations. and empirical results Selection of Polarisation Activated MicrolensesTM architectures for M. Lambooij, W. A. IJsselsteijn, Technische Univ. Eindhoven LCD and OLED applications are described together with new (Netherlands); I. E. J. Heynderickx, Philips Research Labs. developments in manufacturing technology for these components. (Netherlands) and Technische Univ. Delft (Netherlands) The technology delivers significant advantages particularly for high pixel density panels and optimising device ruggedness while Visual discomfort has been subject of considerable research in maintaining display brightness. relation to (auto)stereoscopic displays, but remains an ambiguous concept used to denote a variety of subjective symptoms potentially related to different underlying processes. In order to adequately 6490A-20, Session 5 characterise and understand visual discomfort, multiple types of measurements, both objective as subjective, are needed to determine A 470 x 235 ppi high-resolution 2D and 3D the degree of visual discomfort in a sensitive, reliable and valid way. autostereoscopic display and its user aptitude Classical causative factors such as excessive binocular parallax and investigation results in mobile phone applications accommodation-convergence conflict appear to be of minor N. Takanashi, S. Uehara, H. Asada, NEC Corp. (Japan) importance when disparity values do not surpass one degree of visual angle, which still provides sufficient range to allow satisfactory depth Although interest in 3D displays that don°Øt require special glasses is perception in consumer applications. Visual discomfort, however, may growing, developers of 3D displays still face many challenges, such still occur within this limit and we believe the following factors to be as resolution loss in 3D and 3D/2D convertibility problems. We have pertinent in contributing to this: (1) excessive demand of solved these problems with the development of a novel pixel accommodation-convergence linkage, e.g., motion in depth and arrangement, called HDDP (Horizontally Double-Density Pixels). In this spatial direction (combined with large amounts of parallax) and structure, two pictures (one for the left and one for the right eye) on prolonged viewing at short viewing distances, (2) 3D artefacts two adjacent pixels form one square 3D pixel. It doubles the 3D resulting from 2D-to-3D conversion and rendering such as spatial and resolution as high as the 2D display. It shows 3D images anywhere in temporal inconsistencies and unnatural visualizations, and (3) 2D images with the same resolution. The display we have developed unnatural amounts of blur. We will conclude by highlighting a number is a lenticular lens based, is 2.5 diagonal inches in size and has a of specific areas where more research is required to increase our 320x2(RL) x 480x3(RGB) resolution. As a 3D display, the horizontal understanding of visual discomfort. and vertical resolutions are equal (235 ppi each). A user aptitude investigation was conducted in mobile phone applications. Several kinds of 3D pictures were presented to 345 participants, whose ages 6490A-18, Session 4 ranged from 16 to 54, and we asked their preferences. 3D pictures were categorized into three application areas; communication, MOEVision: simple multiview display with clear floating entertainments, and useful tools in the mobile scene. Eighteen image examples of possible 3D applications were listed and put to the vote. The results showed a high acceptance of, and interest in, this mobile H. Kakeya, Univ. of Tsukuba (Japan) 3D display. This paper presents a simple and inexpensive multiview 3D display system composed of a LCD panel, a convex lens array, and a Fresnel lens. The idea of multiview 3D displays composed of multiple 6490A-21, Session 5 projectors and a large convex lens or a concave mirror is old and famous. The conventional methods, however, require diffusers to A prototype 3D mobile phone equipped with a next- show continuous motion parallax, which decays the quality of the generation autostereoscopic display image. To solve this problem we use a convex lens array with no gaps J. C. Flack, Dynamic Digital Depth Australia Pty. (Australia); J. between the lenses, which realizes continuous motion parallax Harrold, G. J. Woodgate, Ocuity Ltd. (United Kingdom) without diffusers. The convex lens array, however, has to produce images without aberration to show the viewer stable 3D images. It is The mobile phone is quickly evolving from a communications device hard and expensive to realize such lens arrays. To produce images to an application platform and in the process has become the focus with little aberration in a simple format, the author proposes an optical for the development of new technologies. The most challenging system where each component lens makes the parallel light rays technical issues for commercializing a 3D phone are a stereoscopic instead of creating an image by keeping the distance between the display technology which is suitable for mobile applications as well as LCD surface and the lens array the same as the focal distance of the a means for driving the display using the limited capabilities of a component lenses. With this configuration we can create images with

60 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A: Stereoscopic Displays and Applications XVIII mobile handset. In this paper we describe a prototype 3D mobile 6490A-24, Session 6 phone which was developed on a commercially available mobile hardware platform. The demonstration handset was retrofitted with a Prediction of depth images across multiple views Polarization Activated MicrolensTM array that is 2D/3D switchable Y. Morvan, Technische Univ. Eindhoven (Netherlands) and provides both class-leading low crosstalk levels, and suitable brightness characteristics and viewing zones for operation without A free viewpoint video system enables the user to select a point of compromising battery running time. This next generation view in the video scene by synthesizing a novel image at the decoder. autostereoscopic display technology, which combines the advantages However, interpolating novel views without geometric description of in brightness of a lenticular 3D display with the 2D/3D switching the video scene is a difficult task. One solution to interpolate novel capability of parallax barrier, is deployed on a 2.2'’ landscape QVGA views at the decoder is to use one depth image per texture image. TFT LCD base panel. The stereoscopic content solution is an However, an independent transmission of depth image data is not essential component of a commercially viable 3D handset. We efficient because it does not exploit the correlation between depth describe how a range of stereoscopic software solutions have been images. In this paper, we focus on the coding of such a 3D data set developed on the phone’s existing application processor without the (N-depth and N-texture). need for custom hardware. We propose to use a reference depth image that enables the prediction of a set of depth images, with the purpose of compressing 6490A-22, Session 6 this set of multiple depth images. The proposed depth-prediction technique works by synthesizing/ Multiple footprint stereo algorithms for 3D display computing the depth of 3D points seen from the predicted view based content generation on the reference depth image. The attractivity of the proposed F. Boughorbel, Philips Research Labs. (Netherlands) algorithm is that the prediction of depth data avoids an independent transmission of depth for each view, while simplifing the view This work focuses on the conversion of stereoscopic video material interpolation by synthesizing depth images for arbitrary view points. into an image + depth format which is suitable for rendering on the We present experimental results for several multiview sequences, that multiview auto-stereoscopic displays of Philips. Content has been one result in a rendering quality of about 30 dB of the predicted depth of the major limiting factors for this new display technology. images. Fortunately, with the recent adoption of digitization and 3D by major studios stereo material is increasingly available, presenting a clear opportunity for the various 3D display technologies. The conversion 6490A-25, Session 7 from the stereo to the input formats of 3D displays is in this regard a key enabling step. While biological 3D perception is yet to be fully Application of Pi-cells in time-multiplexed stereoscopic understood, a large number of computational techniques are and autostereoscopic displays based on LCD panels achieving increasingly better results. In this paper we address the stereo conversion task by using a multiple footprints matching S. A. Shestak, D. Kim, SAMSUNG Electronics Co., Ltd. (South algorithm that generates several depth candidates for each image Korea) pixel. We characterize various footprint shapes and sizes used for We investigated Pi-cell based polarization switches regarding their matching and we devise a robust strategy for extracting high quality applications in both glass type and autostereoscopic 3D displays, estimates from the resulting depth candidates. The proposed which employs LCD as an image panel. Since LCD display belongs to algorithm is based on a robust surface filter that uses all the available the hold type display, Pi-cell should be divided into the number of depth samples in a local neighborhood. Furthermore the filter includes individually addressable segments to be capable of switching in image constraints ensuring that the resulting depth maps can be used synchronism with line-by-line image update. Too course segmentation for high quality 3D rendering. of Pi-cell cause local mismatch of synchronism followed by the increased level of crosstalk between left and right images. We 6490A-23, Session 6 estimated the number of the segments, which is sufficient to reduce the levels of local and average crosstalk to the acceptable level. The Compression of still multiview images for 3D rise/fall time asymmetry inherent in Pi-cells makes the crosstalk in the automultiscopic spatially multiplexed displays left and right eyes unequal thus increasing its peak level. It is found that the crosstalk can be easily symmetrized by slightly improved R. Lau, S. Ince, J. Konrad, Boston Univ. driving method. Automultiscopic (no glasses, multiview) displays are becoming a viable alternative to 3-D displays with glasses. However, since these displays require multiple views the needed transmission bit rate as 6490A-26, Session 7 well as storage space are of concern. In this paper, we describe Switchable lenticular based 2D/3D displays results of our research on the compression of still multiview images for display on lenticular or parallax-barrier screens. In one approach, M. Hiddink, S. T. De Zwart, O. Willemsen, Philips Research Labs. we examine compression of multiplexed images that, unfortunately, (Netherlands) have relatively low spatial correlation and thus are difficult to The use of LCDs equipped with lenticular lenses is an attractive route compress. We also study compression/decompression of individual to achieve autostereoscopic multiview 3D displays without losing views followed by multiplexing at the receiver. However, instead of brightness. However, such displays suffer from a low spatial resolution using full-resolution views, we apply compression to band-limited and since the pixels are divided over various views. To overcome this downsampled views in the so-called N-tile format. Using lower problem we developed switchable displays. By using active resolution images is acceptable since multiplexing at the receiver switchable LC lenticular lenses it is possible to have a high-brightness involves downsampling from full view resolution anyway. We use three 3D display capable to regain the full native 2D resolution of the standard compression techniques: JPEG, JPEG-2000 and H.264. underlying LCD. While both JPEG standards work with still images and can be applied directly to an N-tile image, H.264, a video compression standard, To show the feasibility of the LC-filled switchable lenticulars in TV requires N images of the N-tile format to be treated as a short video applications we made a 9 view 42" 2D/3D display. A dedicated lens sequence. We present numerous experimental results indicating that design was made and special attention was paid to achieve a uniform the H.264 approach achieves significantly better performance than lens array and cell gap. The display is capable of showing high-quality the other three approaches studied. 3D as well as HD resolution 2D image content. For some applications it is advantageous to be able to display 3D and 2D at the same time. We show a prototype having a matrix electrode structure and a novel driving scheme.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 61 Conference 6490A: Stereoscopic Displays and Applications XVIII

A problem with LC-filled lenses is that at oblique angles the 2D state displays. They have the effect of enlarging the depth of field of an eye has a residual lens effect. This effect and possible solutions will be imaging system. However, their resolution was as low as ~QVGA. In discussed as well. the present study, we develop a high-density directional display whose resolution is increased to SVGA to allow investigating the influence of three-dimensional resolution. The newly developed 6490A-27, Session 7 display consists of 128 small projectors having SVGA resolution. Each projector uses one LCOS device and the field sequential technique is Multiview autostereoscopic display of 36 view using used to display a color image. All 128 projectors are aligned in the ultrahigh resolution LCD modified two-dimensional arrangement, i.e., all projectors are aligned B. Lee, H. Hong, J. Park, H. Park, H. Shin, I. J. Chung, LG Philips two-dimensionally and their horizontal positions are made different LCD (South Korea) one another. All images are displayed in the different horizontal directions with the horizontal angle pitch of 0.28º. The horizontal We have developed a 36 view autostereoscopic display using 15.1" viewing angle is 35.7º. The screen size is 12.8 inches. The system is ultra high resolution LCD. The pixel resolution of LCD is 3200x2400 controlled by a PC cluster consisting of 16 PC’s. In order to correct which is QUXGA grade. The type of sub pixel is a RGB stripe type, the image distortion caused by the aberration of imaging systems, which is aligned as vertical lines and the size of sub pixel is images displayed on the LCOS devices are pre-distorted using 0.032x0.096mm. correction tables. We used parallax barriers to split rays that are radiated from sub pixels. We rotated the angle of barriers by 9.46° from the vertical line(arctangent 1/6). And we located the barrier in front of LCD 6490A-30, Session 8 module. One repeated barrier pattern covers approximately 18 LCD The compatibility of consumer DLP projectors with time- sub pixels, and the slanted angle is arctangent 1/6, therefore, the horizontal number of 3D sub pixels is decreased by 6, and the vertical sequential stereoscopic 3D visualization number of 3D sub pixels is decreased by 2. Consequently calculated A. J. Woods, T. Rourke, Curtin Univ. of Technology (Australia) 3D resolution comes down to 533x400. (=3200/6 x 2400/6) However, the perceptional 3D resolution is much higher than 533x400. A range of advertised “Stereo-Ready” DLP projectors are now available in the market which allow high-quality flicker-free Because it’s a barrier type and it has many views the luminance is stereoscopic 3D visualisation using the time-sequential stereoscopic seriously decreased. However, there are lots of merits. 1) Little display method. The ability to use a single projector for stereoscopic eyestrain, 2) Continuous disparity and no Flipping image within the 3D viewing offers a range of advantages including extremely good seeing area, 3) Higher perceptional resolution than calculated stereoscopic alignment and in some cases portability. It has also resolution. We suppose that these three important merits come from recently become known that some consumer DLP projectors can be the fused views which mean overlapping among adjacent views which used for time-sequential stereoscopic visualisation, however it was are not exactly same but very similar. not well understood which projectors are compatible and incompatible, what display modes (frequency and resolution) are compatible, and what stereoscopic display quality attributes are 6490A-28, Session 7 important. We conducted a study to test a wide range of projectors Autostereoscopic display with 60 ray directions using for stereoscopic compatibility. LCD with optimized color filter layout This paper reports on the testing of 44 consumer DLP projectors of widely different specifications (brand, resolution, brightness, etc). The T. Koike, Hitachi, Ltd. (Japan) and The Univ. of Tokyo (Japan); M. projectors were tested for stereoscopic compatibility with various Oikawa, K. Utsugi, M. Kobayashi, M. Yamasaki, Hitachi, Ltd. video formats (PAL, NTSC, 480P, 576P, 720P, and various VGA (Japan) resolutions) and video input connections (composite, SVideo, component, VGA). Fourteen projectors were found to work well at up We developed a mobile-size integral videography (IV) display that to 85Hz stereo in VGA mode. 23 projectors would work at 60Hz stereo reproduces 60 ray directions. IV is an autostereoscopic video image in VGA mode. technique based on integral photography (IP). The IV display consists of a 2D display and a microlens array. The maximal spatial frequency (MSF) and the number of rays appear 6490A-31, Session 8 to be the most important factors in producing realistic autostereoscopic images. Lens pitch usually determines the MSF of The fabrication of microretarder for in-cell stereoscopic the IV displays. The lens pitch and pixel density of the 2D display LCD using reactive liquid crystal determine the number of rays. There is a trade-off between the lens C. Tsai, W. Huang, Industrial Technology Research Institute pitch and the pixel density. The shape of an elemental IV image determines the shape of the viewing area. (Taiwan); G. Tseng, National Taiwan Univ. (Taiwan); K. Lee, K. Huang, Industrial Technology Research Institute (Taiwan); H. Y. We developed an IV display based on the above correlationship. The Lin, National Taiwan Univ. (Taiwan) IV display consists of a 5-inch 900-dpi LCD and a microlens array. The IV display has 60 ray directions, with 4 rays up and down and a A microretarder for stereoscopic display is a film consisting of 0 and maximum of 18 rays from side to side. We optimized the color filter on half-wave optical retardation states on alternate horizontal or vertical the LCD to achieve 60 rays. The resolution of the display is 256Å~192, stripes. An LCD with a microretarder attached on the front side can and the viewing angle is 30 degrees. These parameters are sufficient display good-quality stereoscopic images when viewed with for mobile game use. The IV display can be interactively controlled polarization glasses. It is now the easiest and cheapest way to with a control pad. present stereoscopic images on a flat-panel display. However, the space caused by the glass between the retarder layer and the LC layer of the panel seriously limits the vertical view angle, which, in 6490A-29, Session 7 turn, limits the application of this technology. In this paper, a process for thin-film microretarder is developed using reactive liquid crystal. Development of SVGA resolution 128-directional display The material and the process are essentially compatible with current K. Kikuta, Y. Takaki, Tokyo Univ. of Agriculture and Technology LC panel processes. Since the thin-film microretarder is to be (Japan) fabricated in the cell of an LC panel, the retarder layer and the LC layer can be fabricated as close to each other as possible. A nearly We develop a 128-directional display having SVGA resolution. In order unlimited 3D view angle can be achieved for the display. to explore natural display conditions to solve the visual fatigue problem caused by the accommodation-vergence conflict, we have ever constructed 64-directional, 72-directional and 128-directional

62 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A: Stereoscopic Displays and Applications XVIII

6490A-32, Session 8 set multilayer monochrome TFT liquid crystal panels. On each layer we draw the edges in the depth of that layer, which are extracted from the Enabling stereoscopic 3D technology original image. Here we can use monochrome panels because the M. G. Robinson, G. D. Sharp, ColorLink, Inc. volumetric part is just in charge of expressing edge contrast. This can sufficiently lead proper accommodation since focal accommodation of Stereoscopic 3D has seen a recent resurgence with the introduction our eyes is dependent only on the edge of the image. To connect the of successful high quality 3D cinema productions such as Polar edges of adjacent panels smoothly, we apply DFD approach, where the Express, Chicken Little, Superman, Monster House etc... This point in the middle of two panels is expressed by depiction on both momentum is expected to entice consumers to demand a similar panels. experience in the home. Already, computer games can be played in stereo 3D with minimal modifications to current graphic hardware, if only affordable high quality 3D displays were commercially available. 6490A-34, Session 8 It is ColorLink’s belief that the consumer market for 3D can only flourish if sets can display both high quality 2D and 3D content, with a Advanced multiscopic display via temporal multiplexing modest increase in display cost. We believe that only polarization or V. B. Markov, S. A. Kupiec, MetroLaser, Inc.; A. R. L. Travis, Univ. of wavelength division systems using passive eyewear are feasible. Cambridge (United Kingdom); D. G. Hopper, G. S. Saini, Air Force Polarization encoding favors projection systems, where polarization Research Lab. conservation is possible between projection lens and viewer. Here Recent advances in the characteristics of the Spatial Light Modulators ColorLink is developing an achromatic linear polarization switch (ALPS(tm)) that dynamically switches between orthogonal polarization (SLM) coupled with a corresponding increase in the rendering and data states. When placed in front of a single conventional (though fast transfer capabilities of graphical processing units (GPUs) has lead to frame rate) 2D projector, sequential stereo can be viewed through the greatly increased viability of temporal multiplexing approaches for polarized eyewear. The switch resembles the z-screen implementation 3d display technology. Employing these advances, an initial proof of currently used in cinemas, but differs in performance and flexibility. In concept four-zone multiscopic display has been demonstrated, with a its basic form it enables stereo to be viewed using low cost linear 8-16 zone large format display in development. Both designs employ a polarization eyewear. Adding a simple 45deg quarter wave retarder at high-speed LCD shutter located at the pupil of the optical train the exit of the switch produces ‘head tilt tolerant’ circular polarized synchronized with a high framerate SLM that in our design is made off output similar to that of the z-screen. Preferably, an achromatic deformable mirror device (DMD) projector allowing for the formation of passive quarter wave plate can be used to avoid the necessary color multiple viewing zones via persistence of vision. Present results and balancing and light loss associated with the former chromatic ongoing progress in the design of the optical train, high speed projector approach. A ColorSelect(tm) polarization filter (see and the associated rendering system. In particular we will discuss the www.colorlink.com) can be added to allow modulation between issues of bandwidth management for the display and rendering nodes. colored states, and as such avoids the need for polarization preserving screens. We propose to describe the operation, design and performance of this versatile component and hope to demonstrate it 6490A-53, Poster Session with a Christie Mirage projector at the conference. Time-multiplexed autostereoscopic display with content- TV displays favor a wavelength division based approach, where each defined 2D-3D mode selection eye sees a full color image. That is, the projection system and eyewear make each stereo channel have separate RGB primary S. A. Shestak, D. Kim, K. Cha, J. Koo, SAMSUNG Electronics Co., colors. This demands fine color discriminating eyewear. Here Ltd. (South Korea) ColorLink’s unique ColorSelect(tm) filter technology enables low cost, A time sequential LCD autostereoscopic display with SXGA resolution high performing color filtering ColorComb(tm) eyewear, enabling both is described which does not have any special means for 2D/3D spectral division projection systems and future direct view LED and switching. A 17" prototype of the display is capable of simultaneous laser illuminated systems. Unlike dichroic eyewear, negligible displaying 2D and 3D graphics without deterioration of visual reflections, large viewing angle, large aperture size and stylish performance of 2D images at a frame rate up to 120 Hz. Left and right appearance are all enabled with this technology. Furthermore, in sets of viewing zones are formed by a combination of a patterned future 6 LED illuminated LCD displays, the polarized display output retarder and a lenticular lens array placed behind the LCD panel. Right allows removal of the eyewear’s front polarizer rendering them and left sets of viewing zones are switchable by polarization switch in neutrally colored to all but polarized light. synchronism with displaying left and right images. Application of ColorLink is active in enabling LED and laser displays as a patterned retarder provides smooth light distribution without dark consequence of experience with RPTV’s, having designed JVC’s DILA regions between left and right viewing zones. To reduce dynamic architecture, and currently supplies key components to major CE crosstalk caused by slow switching of LCD cells a combination of Rise companies such as JVC, LG, Sony etc.. ColorLink proposes to enable Time Acceleration (RTA) and scanning backlight have been applied. stereo 3D in future TV multi-primary systems through its unique eyewear. In this presentation, the technology and design of the eyewear will be presented and examples shown. 6490A-55, Poster Session 3D from compressed 2D video 6490A-33, Session 8 I. A. Ideses, L. P. Yaroslavsky, B. Fishbain, Tel Aviv Univ. (Israel); R. Realization of electronic 3D display combining multiview Vistuch, Tel-Aviv Univ. (Israel) and volumetric solutions In this paper we present a method to synthesize 3D video from H. Ebisu, T. Kimura, H. Kakeya, Univ. of Tsukuba (Japan) compressed 2D video. The 2D video is analyzed by computing frame by frame motion maps. Using the computed motion maps, the video The authors propose an electronic 3-D display combining a multiview undergoes analysis and the motion maps are transformed into motion display and a volumetric display. Conventional multiview displays based depth maps. For this computation MPEG motion vectors often give the viewers severe eyestrains because of the contradiction extraction is used. between binocular convergence and focal accommodation of the The frames are then used to synthesize stereo pairs. This is performed eyes. Though volumetric displays are free from the contraction, they by resampling of the image on a grid that is governed by the depth- cannot express occlusion or gloss of the objects. The proposed map. In order to improve the quality of the synthetic video and in order system overcomes these disadvantages at once by displaying colors to enable comfortable 2D as well 3D viewing of the same sequence, by the multiview display part and fine contrast of edges by the several techniques for image enhancement are used. In our test case, volumetric display part. As for the multiview display we use anaglyphs was selected as the 3D visualization method, since this conventional multiview technologies. As for the volumetric part, we

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 63 Conference 6490A: Stereoscopic Displays and Applications XVIII

method is mostly suited to standard displays. The main drawback of high quality imaging, this needs a lot of images or cameras. The this method is ghosting artifacts. In our implementation we minimize reason why is that Ray-Space requires various direction’s and these unwanted artifacts by modifying the computed depth maps using position’s views instead of 3D depth information. non-linear transformations. Defocusing was also used to counter such In this paper, we reduce that flood of information using view-centered artifacts. interpolation. View-centered interpolation means estimating view Our results show that retrieving motion fields from compressed videos dependent depth value at generating view and interpolating that of enable synthesis of high quality 3D videos in real-time minimizing pixel values using multi-view images. The combination of depth unwanted artifacts. estimation and interpolation realizes the rendering photo-realistic images. Unfortunately, however, if depth estimation is week or mistake, a lot of 6490A-56, Poster Session artifacts appears in creating images. Thus powerful depth estimation Volumetric display by inclined-image scanning and three- method is required. When we render the free viewpoint images video, dimensional image transmission based on optical shape we perform the depth estimation at every frame. Thus we want to keep a lid on computing cost. measurement Our depth estimation method is based on dynamic programming (DP). D. Miyazaki, K. Soutsuka, T. Honda, T. Mukai, Osaka City Univ. This method optimizes and solves depth images at the weak (Japan) matching area with high-speed performance. But scan-line noises Volumetric display systems based on three-dimensional (3-D) scanning become appeared because of the limit of DP. So, we perform the DP of an inclined image are reported. An optical image of a two- multi-direction pass and sum-up the result of multi-passed DPs. Our dimensional (2-D) display device, which is placed obliquely in an optical method fulfills the low computation cost and high depth estimation imaging system, is moved laterally by an optical mirror scanner. Inclined performance. cross-sectional images of a 3-D object are displayed on the 2-D display device in accordance with the position of the image plane. A locus of the moving image can be observed as a series of moving afterimages 6490A-60, Poster Session as a result of high-speed scanning. A 3-D real image is thus formed as Generation of binocular object movies from monocular a stack of 2-D cross-sectional images. This 3-D image can satisfy all object movies the criteria for stereoscopic vision. A three-dimensional image consisting of 1024x768x150 pixels was formed by the constructed Y. Y. Chen, National Taiwan Univ. (Taiwan); Y. Tsai, National Chiao system using a digital micro mirror device. In addition, we propose a 3- Tung Univ. (Taiwan); Y. Hung, National Taiwan Univ. (Taiwan) D image transmission system, which can acquire the 3-D information of Object movie (OM) is a well-known media for virtual display of 3D an object and display it as a volumetric image in real time. We constructed a high-speed 3-D measurement system based on the light- objects because of its simplicity in production and its photo-realistic section method using an intelligent vision system, which can capture visual effect in presenting the objects. An OM is constructed form a set of images taken from different perspective around a 3D object. It and process 1000 images per a millisecond. We demonstrate real-time volumetric image generation from the measured profile using the 3-D is quite straightforward to produce a binocular object movie by measurement system and the volumetric display system. capturing the left-eye and right-eye images of the object using two cameras that are horizontally displaced relative to each other. However, the simple approach for creating such a binocular object 6490A-58, Poster Session movie would significantly increase the cost of storage and image processing, e.g., background removal. Furthermore, the stereo Ray-space acquisition using parabolic mirror baseline of the produced binocular OM is fixed. Therefore, this paper presents a framework to generate binocular OMs from monocular T. Ichiyanagi, T. Yendo, T. Fujii, M. Tanimoto, Nagoya Univ. (Japan) ones, which reduces the efforts of image acquisition and image In this paper, we analyzed the distortion of imaged acquired with a Ray- processing, and can arbitrarily produce the binocular OM with the Space acquisition system. desired baseline based on application scenarios. Without loss of generality, we assume the acquired monocular OM is for left eye. In In case an arbitrary viewpoint picture is generated using the Ray-Space the framework, the camera parameters of each acquired image are method, it is necessary to acquire dense ray data. Conventional first estimated fast and then the background of each image is methods for acquiring the Ray-Space data consist of using rotating removed by applying our previous methods. When performing the stages of a camera array. background removal, a 3D model is automatically reconstructed We developed a system consisting of two parabolic mirrors, a simultaneously, and will be used to generate the right-eye images. synchronized galvanometric mirror and a high-speed camera. The Here, a new representation of the 3D model, which combines principle is as follows; if an object is put in the bottom of the parabolic billboard clusters and triangle mesh, is proposed for efficient mirror, the ray which comes out of the object is imaged in the upper generating binocular views. Billboard cluster is a set of multi-layer part, and form a real image. The galvanometer mirror is put on the billboards, as a hybrid of billboard cloud and layer imposters, and is position of a real image, and is made to scan horizontally. The images used to approximate the 3D model. It is an extreme simplification, but of the object of different angles (directions) are then possible to can capture better the shape of the original model, even for curved generate and are captured by the high-speed camera. By capturing surfaces. With billboard cluster technique, the representation can many images each time scanning, Ray-Space is therefore acquirable. achieve a reasonably good rendering quality at high compression However, distortion arises in the real image of the object formed. ratio. In order to obtain better results, a novel approach to extract Consequently, distortion appears in the captured image. Then, it is view-independent texture is also proposed in this work. With view- necessary to correct the captured image to the right image. Here, we independent texture, billboard cluster can be used to perform examine a method to generate corrected images from the acquired relighting so that the binocular OMs can be well integrated into virtual Ray-Space. environments, where the lighting condition differs from the lighting condition of the original object movie. Our work will make great improvement of the application of the stereo object movie with its 6490A-59, Poster Session high compression ratio, arbitrary lighting condition, and arbitrary viewpoint. Free viewpoint image generation using multipass dynamic programming N. Fukushima, T. Yendo, T. Fujii, M. Tanimoto, Nagoya Univ. (Japan) Ray-Space is categorized by Image-Based Rendering, thus generated view become photo-realistic. While this method has the performance of

64 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A: Stereoscopic Displays and Applications XVIII

6490A-61, Poster Session shown in a traditional display without stereo. The second experiment uses an immersive game, consisting of a 3D Volumetric three-dimensional display using transparent maze where the player has to navigate and catch objects in a virtual rotating Lanthanide(III) complexes sheet environment (VR) using the stereoscopic HMD or in a traditional S. Hisatake, S. Suda, J. Takahara, T. Kobayashi, Osaka Univ. desktop display (Virtual Desktop - VD). Besides a qualitative study, (Japan) variables such as time spent, number of collisions, distance covered, and objects caught were recorded for the latter to allow a quantitative We have developed a new type of volumetric three-dimensional (3D) evaluation. The overall conclusion is that VR/AR technologies still display which has a 360-degree, 3D volume group-viewing capability have to overcome hardware limitations. without the use of any special viewing aids or goggles. In our system, However, for short periods of time and specific applications, Virtual a clear thin photopolymer sheet dissolving Lanthanide(III) complexes and Augmented Reality seems to be a promising alternative since is used. The Lanthanide(III) complexes used in our system are interaction is intuitive and natural. Eu(TTA)3 Phen, designed for 615nm luminescence (red) against an exciting light of 395nm. The specific point (x,y,z) of the 3D object is expressed by luminescence of Lanthanide(III) complexes excited by 6490A-64, Poster Session laser diode beam. The direction of the exciting laser beam is controlled by two galvano mirrors so as to illuminate the specific two- Improvement of the voxels density of three-dimensional dimensional point (x,y). The depth direction (z) of the 3D object is image display by spatially multiplexing fluxes in light-ray- expressed by thin sheet rotation synchronized with the galvano flux reproduction method mirrors. Arbitrary luminous point is specified by 3D control of excitation of the Lanthanide(III) complexes, and 3D images are S. Hisatake, T. Kobayashi, Osaka Univ. (Japan) expressed as an aggregate of these luminous point. In the case of “light-ray-flux reproduction method” which is proposed In the experiment, we observed static 3D objects of a triangle and an by our group, a color filter is projected by an array of white-color point octahedron. We also observed 3D animation of a 3D Lissajous curve. light sources to artificially generate a group of light beams Displayed 3D objects integrated with real background because corresponding to light scattered from an object, to thereby generate a rotating thin sheet is transparent (transparent 3D display). 3D image. The voxel density is proportional to the point light source density. The array pitch of the point light sources and distance from the point light sources to the color filter determines the viewing angle. 6490A-62, Poster Session The distance between the point light sources and the color filter determines the distance from the light sources in which 3D images Generation of the multiview 3D video of real scene for can be displayed. Consequently, there is relation of the trade-off the multiview 3D display using multiview depth between the voxel density, display distance and viewing angle. estimation and view synthesis In the developed system, the light-ray-fluxes are spatially multiplexed G. Um, Electronics and Telecommunications Research Institute by rapidly changing the position of the white-color point light sources. (South Korea) A liquid crystal display is used as a color filter. The color filter images are changed in synchronized with the position changes of the white- We present a multi-view 3D video generation system and its color point light sources. In the experiment, position changes of the technique for multi-view 3D display. Usually we should synthesize point light sources are realized by mechanically vibrating periodically virtual images that have smaller separations appropriate for multi-view the slit mask, which covers the point light source array. We confirmed 3D display because it is hard to get small camera separation required that the voxel density was improved to twice without the degradation for the multi-view 3D display using the real multi-view cameras. In the of the viewing angle and the display distance. proposed system, we first capture multi-view video sequences with synchronization using multi-view camera. Then we calibrate multi- view camera and do color and lens distortion corrections. With 6490A-65, Poster Session corrected multi-view video and camera parameters, we then estimate disparity/depth map for each camera-view. For the disparity Curved integral imaging scheme using an additional estimation, we first segment the scene into the foreground and the large-aperture lens background regions with the background image and its color difference. .Matching cost function is newly defined by considering D. Shin, Kwangwoon Univ. (South Korea); B. Lee, Seoul National occlusions, color similarity, and disparity gradient between frames and Univ. (South Korea); E. S. Kim, Kwangwoon Univ. (South Korea) then is minimized using belief propagation. With the estimated In this paper, we propose a curved integral imaging system using an disparity/depth maps and color images from each camera, virtual additional large-aperture lens. The additional large-aperture lens viewpoint images are rendered with the suitable intervals. View provides full-directional curvature effect and improves viewing angle synthesis was done by mapping the depth of each pixel into 3D space compared with conventional system. The proposed system has a and re-projecting it to virtual camera position. Finally, multi-view video simple structure due to the use of well-fabricated flat devices without that contains some of camera viewpoints is played. By doing this, we any modification. A modified computer-generated pickup technique can observe 3D video that have various viewpoints of the scene. based on ABCD matrix is introduced to calculate elemental images of the proposed system. To show the usefulness of the proposed system, optical experiments are carried out and the results are 6490A-63, Poster Session presented. Usability in virtual and augmented environments: a qualitative and quantitative study 6490A-66, Poster Session A. E. Pimentel, P. Dias, B. S. Santos, Univ. de Aveiro (Portugal) Psychophysical estimation of 3D virtual depth of united, A preliminary usability study of a low-cost Virtual and Augmented synthesized, and mixed-type stereograms by means of Reality environment under development at the University of Aveiro is simultaneous observation presented. The environment was evaluated regarding a desktop set- up in two different M. Iizuka, Tokyo Polytechnic Univ. (Japan) scenarios: Augmented Reality (AR) and Virtual Reality (VR). The first There is a lot of information on the internet web site concerning all experiment consisted in showing users data from their own field of aspects of stereogram history, science, social organization, works of expertise both in AR mode, where users see through a stereoscopic art, and free software for generating anaglyph, repeated pattern HMD models aligned with a real marker captured by a camera, and in stereogram (RPS), random dot stereogram (RDS), single image AD (Augmented Desktop) mode, where the same information in stereogram (SIS) etc. A computer-generated stereogram (CGS) is a

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 65 Conference 6490A: Stereoscopic Displays and Applications XVIII two dimensional flat image viewed in such a fashion as to produce a cameras themselves compute the depth maps. However, the disparity three dimensional effect, i.e. fantastic impression of 3D virtual image maps have still to be transformed into clouds of 3D points by with depth. software. The points computed would belong to the model as well as This paper focuses on an influence of the 3D virtual depth perception to the background. and its image quality caused by modifying CGS, i.e. three kinds of The filtering step lies between the acquisition and the conversion of united, synthesized and mixed type stereograms. The united the depth map into a cloud of 3D points. Its first task is to segment stereogram is composed of a set of depth map images (DMIs). On the the images into foreground and background. Then only 3D points other hand, the synthesized stereogram is composed of a corresponding to the foreground of the depth map have to be combination of SISs. The mixed type stereogram is composed of the computed. Its other tasks are to fill holes and remove outliers thanks united stereogram and SIS. An estimation of 3D virtual depth to some quick depth map filtering of the foreground. considering image quality is possible psychophysically owing to the We chose to transmit the cloud of points. The other solution would disparity of two eyes on a computer display screen or hard copy after have been to transmit the left image and the disparity map. The first visual training with a little patience. transmission method does not compress the points but quantize them To summarize, psychophysical problems on 3D virtual depth together and packetize them in a per packet basis. This method tell us the with image quality are discussed for merits and demerits of three main quickest we can do to transmit a given amount of points. The second proposed stereograms from the viewpoint of the simultaneous method performs a Huffman compression of the difference of observation in the case of the parallel view method. successive points. The points are sent either with tcp ip or rtcp. Each packet contains enough information to be decompressed alone. If we lose a packet, we just lose the points of that packet. Packet loss can 6490A-67, Poster Session be recovered thanks to Reed Solomon codes at the cost of extra computation time. Recognition of 3D objects by use of computational integral imaging reconstruction The same computer performs the acquisition, filtering, compression and transmission. The points are sent to a distant computer. At the J. Park, D. Hwang, D. Shin, E. S. Kim, Kwangwoon Univ. (South reception side, points are decompressed and filed to the display unit. Korea) This one renders the points available thanks to a commercial stereoscopic display. Recently, the concept of computational integral imaging reconstruction (CIIR)-based correlation was proposed and showed a good application of CIIR. That method was to recognize partially 6490A-71, Poster Session occluded 3-D objects by using CIIR technique. However their study focused on the object recognition by using the discrimination Real-time photorealistic stereoscopic rendering of fire capability of the optimum filter. B. M. Rose, Blackbaud, Inc.; D. F. McAllister, North Carolina State In this paper, a novel CIIR-based 3-D image correlator to detect Univ. location coordinates of 3-D objects in space. In the proposed method, elemental images of the reference and target objects are imaged We propose a method for realtime photorealistic stereo rendering of through lenslet arrays and captured by CCD cameras, and from which the natural phenomenon of fire. Applications include the use of virtual a set of plane images of the reference and target objects can be reality in fire fighting, military training, and entertainment. reconstructed at each output plane using a CIIR technique. Here, only Rendering fire in realtime presents a challenge because of the the target image reconstructed at the right plane where the target transparency and non-static fluid-like behavior of fire. It is well known object was originally located is clearly focused and the other target that, in general, methods that are effective for monoscopic rendering images reconstructed at the out of this focused planes are out of are not necessarily easily extended to stereo rendering because focused and blurred. Therefore, from the cross-correlations between monoscopic methods often do not provide the depth information the focused reference and object plane images sharp correlation peak necessary to produce the parallax required for binocular disparity in can be expected and from which the location data of the target object stereoscopic rendering. in space can be extracted. To show the feasibility of the proposed scheme, some experiments on recognition of 3D object in a scene We investigate the existing techniques used for monoscopic rendering of fire and discuss their suitability for extension to realtime stereo were carried out and its results are presented as well. rendering. Methods include the use of precomputed textures, dynamic generation of textures, and rendering models resulting from 6490A-70, Poster Session the approximation of solutions of fluid dynamics equations through the use of ray-tracing algorithms. We have found that in order to attain A point-based tele-immersion system: from acquisition realtime frame rates, our method based on billboarding is effective. to stereoscopic display Slicing is used to simulate depth. Texture mapping or 2D images are mapped onto polygons and alpha blending is used to treat D. Ruiz, Univ. Catholique de Louvain (Belgium); X. Desurmont, J. transparency. We can use video recordings or pre-rendered high- Bruyelle, Multitel (Belgium); B. Macq, Univ. Catholique de Louvain quality images of fire as textures to attain photorealistic stereo. (Belgium) In a tele-immersion system, two or more users in different locations 6490A-35, Session 9 collaborate with each other in a shared, simulated environment as if they were in the same physical room. Each user perceives 3D models A simple measurement of extinction ratio of the other users. M. A. Weissman, TrueVision Systems, Inc. Photorealistic models, low latency and real time reconstruction improve immersion. Among the possible reconstruction methods, “Extinction Ratio” is a measure of the crosstalk between stereoscopic point-based models are often considered. They do not require images on a stereoscopic display. It measures to what extent the computing the connectivity, the sheer number of points allowing us to “wrong” image has been extinguished by the system. High values of see the model as a whole, like the 2D pixels of an image. Extinction Ratio are extremely important for comfortable stereo We present a complete point based reconstruction pipeline, from viewing, yet measurements of this parameter are lacking in the 3D acquisition to display through filtering and transmission. Our main display marketplace, perhaps because of its difficulty of goal is to reach real time while reconstructing a photo-realistic model. measurement. I will present a simple method for measuring Extinction Ratio and will encourage all members of the audience - and all At the acquisition side, the clouds of points have to be computed manufacturers - to use it for qualifying their 3D displays. from images. This implies a complex stereo matching step. We decided to use a commercial stereo on chip solution. The stereo

66 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490A: Stereoscopic Displays and Applications XVIII

6490A-36, Session 9 whenever a scene change does not occur at exactly the same time on the left and right tape. In such a case, the first tape is stopped until Optical cross-talk and visual comfort of a stereoscopic the scene change occurs on the second tape as well. Once both display used in a real-time application tapes have reached the scene change, playback on both drives is resumed. Detecting scene changes by means of image analysis is S. Pala, R. F. Stevens, National Physical Lab. (United Kingdom) unreliable and CPU intensive. Consequently, a better method, based Displays that present a three-dimensional image to an observer offer on the recording date and time embedded in the compressed stream, considerable advantages over two-dimensional displays but in has been developed. practice the applications are limited by the level of discomfort To test the suggested solution, the ‘tape synchronization’ feature has experienced by the viewer. Some systems require the observer to been added to the Stereoscopic Multiplexer, a multiplexing solution wear headgear while others, autostereoscopic systems are free of for the Microsoft Windows platform. viewing aids, The latter are generally considered more practical and work by encoding the viewing zone such that the observer sees stereoscopic pairs of images that are combined to give the impression 6490A-39, Session 10 of a 3-D image. A common cause of viewing discomfort is image cross-talk, which is leakage of the optical signal from the channel Real-time embedded system for stereo video processing corresponding to the right eye to the channel corresponding to the left for multiview displays eye or vice versa. In this paper we use two video cameras and two R. M. Berretty, P. Machado, B. Riemens, Philips Research Labs. LCD panels in a Helmholtz set-up, inherently free of cross-talk, to (Netherlands) display a 3D image of two rods. The rods are aligned in depth whilst viewing the 3D image with controlled amounts of cross-talk In video systems, the introduction of 3D video might be the next introduced electronically. Skin conductance is used as a measure of revolution after the introduction of color. Nowadays multiview workload as well as traditional subjective questionnaires. We report autostereoscopic displays are in development. Such displays offer qualitative measurements of user workload. multiple (e.g. nine) views at the same time and what a viewer sees depends upon his position with respect to the screen. His left eye receives a signal that is different from what his right eye gets; this 6490A-37, Session 9 gives Crosstalk in automultiscopic 3D displays: blessing in —provided the signals have been properly processed— the disguise? impression of depth. The various views produced on the display differ with respect to their associated camera positions. A. Jain, J. Konrad, Boston Univ. In this paper, we show how to compute depth from the stereo signal Most of 3-D displays suffer from interocular crosstalk, i.e., the to comply with the multiview display interface format. After initial perception of unintended view in addition to intended one. The algorithm design, resulting “ghosting” at high-contrast object boundaries is the software was mapped and optimised for an embedded media objectionable and interferes with depth perception. In automultiscopic processor (tm3270). This process involved algorithmic adaptations (no glasses, multiview) displays using microlenses or parallax barrier, and extensive performance the effect is compounded since several unintended views may be perceived at once. However, we recently discovered that crosstalk in (cost) versus image quality evaluation. automultiscopic displays can be also beneficial. Since spatial We present the complete system, including novel algorithmic multiplexing of views in order to prepare a composite image for improvements. We illustrate excellent image quality, not even using automultiscopic viewing involves subsampling, prior anti-alias filtering the complete compute capacity of a single processor running at is required. To date, anti-alias filter design has ignored the presence of 240MHz on crosstalk. In this paper, we propose a simple multiplexing model that standard definition input signals (i.e. 2 x 720x480\@30p). Therefore, takes crosstalk into account. Using this model we derive a this system is suitable as real-time, cost-effective realisation for future mathematical expression for the spectrum of single view with products. crosstalk, and show that it leads to reduced spectral aliasing compared to crosstalk-free case. We then propose a new criterion for the characterization of ideal anti-alias pre-filter. In the experimental 6490A-72, Session 11 part, we describe a simple method to measure optical crosstalk between views using digital camera. We use the measured crosstalk A Look at the Past and Future of Stereoscopic Displays parameters to find the ideal frequency response of anti-alias filter and Through a “(Liquid) Crystal Ball we design practical digital filters approximating this response. Having applied the designed filters to a number of multiview images prior to J. L. Fergason, Fergason Patent Properties multiplexing, we conclude that, due to their increased bandwidth, the Many stereoscopic displays use liquid crystals as an important new filters lead to visibly sharper 3-D images without increasing building block. In 1958, image display using liquid crystals began aliasing artifacts. when the chiral nematic liquid crystals (then designated as cholesteric liquid crystals) were used to form images of temperature patterns generated by a wide variety of inputs. This was followed by 6490A-38, Session 10 continuous development going from indirect inputs to the current electric field driven display media. In 1978, the first display fast DV tape drive synchronization for immediate preview of enough for uncompromised viewing was developed using time based dual camera recordings multiplexing. Numerous developments followed involving micro P. Wimmer, 3dtv.at (Austria) display projectors and direct view displays. The stage is set for continuous innovation for the foreseeable future. The presentation will Dual DV or HDV camcorder setups are widely used to record attempt to answer the following questions: What’s old and still a stereoscopic video content. However, even when using a remote source of improved image production? What are some current control during recording (to operate the camcorders synchronously), improvements being made which impact fully using the complete the number of frames per scene on the left and right tape differs vision capabilities of the human eyes while minimizing problems of slightly. This made it impossible to play back the recorded content fatigue and other problem associated with conflicting stereo cues? immediately. Up to now, left and right tape had to be transferred to a What are the areas which will be possible of revolutionary hard disc and edited scene by scene. improvements? Why are there even more LCD’s in your future? This paper presents a software-only solution to synchronize a pair of DV or HDV camcorders: Playback is assumed to be out of sync

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 67 Conference 6490A: Stereoscopic Displays and Applications XVIII

James Fergason published the twisted nematic field effect of liquid a lens array consisting of convex lenses. However, the use of convex crystals in 1969 while at Kent State University. A year later in lens causes problems of geometrical distortion and chromatic September he produced the first fully operational multi-digit TNLCD. aberration. As a result, the reconstructed image is degraded by these Today his inventions form the foundation of the multi-billion dollar problems. The method using an array of planar mirrors has been liquid crystal display industry. In May of 2006, he was named the 2006 reported in order to avoid geometrical distortion and chromatic winner of the $500,000 Lemelson-MIT Prize, the largest cash prize in aberration. In the capturing stage, the virtual images are picked up by the United States for invention, for his work on liquid crystal the camera. In the display stage, these captured images are projected technologies. He received a B.S. degree in physics from the University onto the same mirror array as used in the capturing stage by the of Missouri in 1956, and in 2001 he was awarded an honorary Doctor image projector. The light rays retrace the original routes and then of Science Degree. He holds more than 130 U.S. patents and has 40 converge at the point where the object captured had been, forming a scientific papers. He is currently inventing for Fergason Patent spatial image. However, the capturing direction is opposite to the Properties LLC. viewing direction and this creates a pseudoscopic image with reversed depth. In order to solve the pseudoscopic problem, we propose the use of an afocal system array consisting of gradient- 6490A-41, Session 12 index lenses. Integral photography suitable for small-lot production using mutually perpendicular lenticular sheets and 6490A-45, Session 12 fractional view Improvement of viewing resolution of integral K. Yanaka, Kanagawa Institute of Technology (Japan) videography by use of rotated prism sheet We have developed an integral photography (IP) system that is H. Liao, The Univ. of Tokyo (Japan) suitable for small-lot production and is applicable to animation. IP is an ideal 3D display method because users can see stereoscopic 3D Despite Integral Photography (IP)/ Integral Videography (IV)’s many images from arbitrary directions. However, IP is less popular than advantages, the viewing resolution of its spatial images (image quality) lenticular display with only the horizontal parallax, probably because has thus far been poor. In this study, we propose the use of rotated the initial cost of designing and producing a fly’s eye lens is very high. prism sheet to overcome the upper resolution limit of IP/IV image We used two technologies to solve this problem. First, we used two imposed by the Nyquist sampling theorem. A set of rotated prism mutually perpendicular lenticular sheets instead of a fly’s eye lens sheet is used in stead of the moving lenslet array. We place two prism sheet. A lenticular sheet is much less expensive than a fly’s eye lens sheets with same pitch in front of the IP/IV display and parallel shift because it is easier to produce. Second, we used the “fractional view the entire light ray in the 3D space. When we rotate the prism sheet, method”, in which the ratio of the lens pitch and pixel pitch is not the shifted light ray will be rotated around its original direction within a limited to some specific values, which means that no custom-made shifted area. We can adjust the interval of the prisms set to optimally lenticular lens is necessary. However, the original fractional view shift the light ray with different rotation diameter. The moving speed of method is applicable only to horizontal parallax. We have made it the lenslet array could be control by the rotation speed of prism sheet. applicable to both horizontal and vertical parallaxes by using above- The experimental results show that image quality of the IP/IV could be mentioned two mutually perpendicular lenticular sheets. In addition, improved with the proposed method. We believe that the rotated we developed a simple technique to generate dedicated synthesized prism sheet technique is a simple and correct solution for improving images for the proposed system. Experiments demonstrated the the viewing resolution of IP/IV image without reducing the 3D aspect effectiveness of the 3D display system. of the reconstructed image like the image depth.

6490A-42, Session 12 6490A-54, Session 12 Scan-type image capturing system using a cylindrical Enhanced algorithm of image mapping method for lens for one-dimensional integral imaging computer-generated integral imaging system Y. Momonoi, K. Taira, Y. Hirayama, Toshiba Corp. (Japan) S. Min, Virginia Polytechnic Institute and State Univ.; K. S. Park, Information and Communications Univ. (South Korea); T. Poon, We propose a new capturing system to realize capture of elemental Virginia Polytechnic Institute and State Univ. images for an autostereoscopic display based on a one-dimensional integral imaging (1D-II) method. An enhanced algorithm of image mapping method is proposed for the For a capturing system for an autostereoscopic display based on the computer-generated (CG) integral imaging system. Integral imaging is one-dimensional integral imaging method, two features are required. one of the attractive autostereoscopic three-dimensional (3D) display One feature is a capturing system that can capture elemental images methods since it provides both vertical and horizontal parallaxes with with fine pitch. The other feature is that each elemental image full colors and it needs no special viewing aids on the observer. The containing much horizontal parallax information should be captured CG integral imaging system uses the set of elemental images with wide angle. generated by computer graphics instead of the pickup process using the lens array. The process to obtain the set of elemental images The optical structure consisting of a cylindrical lens, a focusing lens using computer graphics is called image mapping. In this paper, we and an imaging device scanning over the entire object captures propose an enhanced image-mapping algorithm, named viewpoint elemental images having these two features. Furthermore, the system vector rendering, to improve the speed of the elemental-image size can be small by adapting a compact scanning optical system of a generation and to be more robust so that it is less affected by the copy machine. system factors and the object image quality, as compared to the previous methods. Since the proposed algorithm can support the real- time generation of elemental images, it is possible to realize the CG 6490A-43, Session 12 integral imaging system applied for an interactive multimedia system. Integral imaging system using an array of planar mirrors The feasibility and the efficiency of the proposed technique is verified and analyzed by mean of some experimental results. J. Arai, M. Kawakita, H. Kawai, F. Okano, NHK Science & Technical Research Labs. (Japan) Integral photography (IP), as proposed by Lippmann, enables 3-D images corresponding to a viewpoint to be observed under natural light, allowing 3-D images to be viewed under conditions similar to viewing a real object. In a proposed method by Lippmann, it requires

68 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6490B: The Engineering Reality of Virtual Reality 2007 Wednesday 31 January 2007 Part of Proceedings of SPIE Vol. 6490B The Engineering Reality of Virtual Reality 2007

6490B-46, Session 13 6490B-49, Session 13 Three-dimensional cameras on the International Space Tangible mixed-reality desktop for digital-media Station management M. Sabbatini, G. Visentin, European Space Agency (Netherlands); S. Winkler, S. Zhou, H. Yu, National Univ. of Singapore (Singapore) M. Collon, cosine Research B.V. (Netherlands) The design of human-computer interlaces is a challenge. Users are Erasmus Recording Binocular - 3-D video camera for the ISS. required to adapt to the machine and learn the use of the associated The main objectives of the experiment are to test a 3-D video camera interfaces. In fact, input functionality via keyboard and mouse is rather (the Erasmus Recording Binocular) in weightlessness on the ISS, as inefficient. The purpose of our research is to develop a system for well as accurately mapping the interior of the ISS in its current improved human-computer interaction. The specific application of configuration. To achieve this, images from three cameras shall be interest is managing computer files and in particular digital media data used: the ERB 3-D video camera, a Sony PD-150 video camera and a such as photos or video clips by means of a tangible mixed-reality Nikon 3-D still camera. desktop. The sorting and arranging of multimedia information on a screen with a mouse can be very tedious as compared to using photo These images will be used to improve the models available on the prints, for example. Our aim is therefore to facilitate the handling of ground as well as improving the fidelity of the ISS 3-D virtual reality such files, and to provide natural interfaces modalities with real-world simulator at the Erasmus Centre of ESA’s Directorate of Human objects such as the printer or a photo using vision-based Spaceflight, Microgravity and Exploration, located at ESTEC, ESA’s gesture analysis and tracking. In our system, a projector is used to research and technology centre in Noordwijk, the Netherlands. Of project the virtual desktop onto the user’s desk. The interaction is special interest is filming of subjects and/or objects moving to and implemented using stereo cameras for tracking the user’s hand from the camera and filming of objects protruding from a surface such movements and gestures. as cables on experimental racks. 6490B-50, Session 14 6490B-47, Session 13 Passing excellence Latency compensation by horizontal scanline selection for head-mounted displays D. Tsoupikova, Univ. of Illinois/Chicago J. J. Jerald, A. Fuller, A. A. Lastra, M. Whitton, L. Kohli, F. P. This paper describes the research and development of a virtual reality Brooks, The Univ. of North Carolina at Chapel Hill visualization project “Passing excellence” about the world famous architectural ensemble “Kizhi”. The Kizhi Pogost is located on an A fundamental task of a virtual-environment system is to present island in Lake Onega in northern Karelia in Russia. It is an authentic images that change appropriately as the user’s head moves. Latency museum of an ancient wood building tradition which presents a produces registration error causing the scene to appear spatially unique artistic achievement. This ensemble preserves a concentration unstable. To improve the spatial stability of the scene, we built a of masterpieces of the Russian heritage and is included in the List of system that, immediately before scanout to a head-mounted raster Most Endangered Sites of the World Monuments Watch protected by display, selects a portion of each scanline from an image rendered at World Heritage List of UNESCO. The project strives to create a unique a wider display width. The pixel selection corrects for yaw head virtual observation of the dynamics of the architectural changes of the rotations and effectively reduces latency for yaw to the millisecond museum area beginning from the 14th Century up to the 21st Century. range. In informal evaluations, users consistently judged visual scenes The visualization restores the original architecture of Kizhi island more stable and reported no additional visual artifacts with horizontal based on the detailed photographs, architectural and geometric scanline selection than the same system without. Scanline-selection measurements, textural data, video surveys and resources from the hardware can be added to existing virtual-reality systems as an Kizhi State Open-Air Museum archives. The project is being created external device between the graphics card and the raster display. using an application development environment, Electro for the tiled display high-resolution graphics visualization system and can be shown on the virtual reality systems such as the CAVE and the C-Wall 6490B-48, Session 13 environment. A novel wide-field-of-view display method with higher central resolution for hyper-realistic head dome projector 6490B-51, Session 14 A. Hotta, T. Sasaki, H. Okumura, Toshiba Corp. (Japan) CytoViz: an artistic mapping of network measurements We have developed a compact hyperreality display for personal use, as living organisms in a VR application the Head Dome Projector (HDP). It is a head-mountable system and B. Lopez Silva, L. Renambot, Univ. of Illinois/Chicago the image is projected on a dome-shaped screen (40cm diameter) with wide viewing angle of about 150 degrees horizontally by a mobile CytoViz is an artistic, real-time information visualization driven by projector with LED light source. One of the drawbacks of HDP is statistical information gathered during gigabit network data transfers insufficient resolution. Since the image is projected on a dome- to the Scalable Adaptive Graphical Environment at various events. shaped screen with wide viewing angle by a single projector, the Data streams are mapped to cellular organisms defining their image is expanded and the resolution per viewing angle is reduced. structure and behavior as autonomous agents. Network bandwidth Thus, we propose a novel display method to realize a high-resolution drives the growth of each entity and the latency defines its physics- image in a central view field using image processing based on the based independent movements. The collection of entity is bound characteristics of human vision, namely, high central visual acuity and within the 3D representation of the local venue. This visual and low peripheral visual acuity, by using the pixel shift technology, which animated metaphor allows the public to experience the complexity of is one of the resolution-enhancing technologies for projectors. The high-speed network streams that are used in the scientific community. projected image with our method is a fine wide-viewing-angle image with high definition in the central view field. We evaluated the psychological effects of the images projected with our method in terms of sensation of reality. According to the result, we obtained 1.5 times higher resolution in the central vision and a greater sensation of reality by using our method.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 69 Conference 6490B: The Engineering Reality of Virtual Reality 2007

Moreover, CytoViz displays the presence of discoverable Bluetooth devices carried by nearby persons. The concept is to generate an event-specific, real-time visualization that creates informational 3D patterns based on actual local presence. The observed Bluetooth traffic is put in opposition of the wide-area networking traffic by overlaying 2D animations on top of the 3D world. Each device is mapped to an animation fading over time while displaying the name of the detected device and its unique physical address. CytoViz was publicly presented at two major international conferences in 2005 (iGrid2005 in San Diego, CA and SC05 in Seattle, WA).

6490B-52, Session 14 Skin: an interactive hyperstereoscopic electro installation H. G. Kostis, R. L. Kooima, Univ. of Illinois/Chicago; J. Kannenberg, Stasisfield.com It is the uniqueness of Virtual Reality as a medium that calls for the creation of hybrid realities which blur the finite boundaries between physical and digital existence. Virtual Reality’s distinguishing features as an artistic medium embody a distinct form of aesthetics: it is a stereoscopic, immersive, interactive, performative, dynamic, and experiential medium. A Virtual Reality art piece manifests in multiple ways. It can present itself as an interactive virtual archetype, exploring concepts rendered from different perspectives, and as an impetus to challenge the platform’s capabilities, not only theoretically as an artistic practice, but also by calling for the instantiation of authoring tools for the development of virtual reality experiences. The paradigm presented in this paper is a Virtual Reality art piece, called skin, 2006, developed on Electro, which is an open-source cross-platform development environment. skin, 2006, is an interactive hypersteroscopic high-definition audiovisual installation that explores a dialogue between physical and digital senses of “touch”.

70 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6491: Videometrics IX Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6491 Videometrics IX

6491-01, Session 1 Secondly, the Video Content Analysis (VCA) technique to be evaluated must generate results from the test sequences. Thirdly, ground truth A review of the status and trends in underwater needs to be available. Then, the ground truth needs to be compared videometric measurement with the generated results, requiring an unambiguous definition of the metrics. In this paper, we focus on these performance evaluation M. R. Shortis, Royal Melbourne Institute of Technology (Australia); methodologies. E. S. Harvey, J. W. Seager, The Univ. of Western Australia (Australia) Videometric techniques have been used for underwater measurement 6491-04, Session 1 since the 1960s. Commencing with still and movie film cameras, the Sensor orientation for coverage optimization in primary underwater applications were repairs on marine equipment, surveillance applications abundance or habitat surveys of the sea bed, and archaeological site and ship wreck surveys. Synchronised stereo-cameras were first used F. Janoos, R. Machiraju, R. Parent, J. W. Davis, A. Murray, The in the 1980s and provided the significant advantage of dispensing Ohio State Univ. with the reference frame, instead relying on pre-calibration and stable camera geometry. This new versatility enabled the length In any surveillance or remote monitoring system, the location and measurement of mobile species such as epibenthic crustaceans and orientation of pelagic fish, and for the first time the biomass and population (directed) sensors is critical for efficient resource utilization as well as distributions of species could be confidently estimated. The 1990s correct system behavior. Additionally, there may be other saw the introduction of portable camcorders to the underwater requirements like multiple sensor coverage for sensitive areas, fault environment and the first use of digital video image sequences. The tolerance and system reconfiguration to sensor drop out. Any system wider use of stereo systems, plus the substantial improvements in should deal with these requirements in a robust and predictable image resolution and image fidelity is providing new possibilities for fashion. accurate and reliable reconstruction and recognition of 3D surfaces. The sensor placement and orientation problem is not simple and has These advances are leading to significant improvements in the hitherto been addressed either by ad-hoc solutions or, in limited effectiveness of management of marine ecosystems for conservation cases, as a problem in computational geometry. Computational and the estimation of biomass for aquaculture. The paper will review geometers solve the problem as a version of the art-gallery problem the progress and impact of advances in imaging technology on the (AGP), which is known to be NP-hard in 2 and 3 dimensions. use of videometric measurement in the underwater environment. Also, current solutions to the AGP assume unrealistic sensor capabilities, such as unlimited depth of field, unlimited resolution and 6491-02, Session 1 depth of focus, and infinitely expedient reorientation speed. Furthermore, they do not allow for specifying additional coverage Three-dimensional smart surveillance system using requirements like feature importance and multiple-sensor coverage. multiple cameras In this paper we introduce a metric to quantify ground plane coverage A. K. Mishra, S. Winkler, A. A. B. M. Kassim, National Univ. of by a sensor that takes into account the sensor resolution, and Singapore (Singapore) distance of the ground plane from the sensor. Then we determine the orientation (pan/tilt/zoom) of all the sensors that provides optimal In most existing surveillance systems, a single camera is used to coverage. We allow specification of additional coverage requirements observe a given area. However, with 2D techniques it is very hard to like multiple coverage and location importance, by way of a saliency resolve occlusions and track multiple people. We therefore propose a map given by the user. 3D surveillance system using multiple cameras surrounding the Our method provides better modeling of real camera parameters like scene. Our application is concerned with identifying humans in the resolution drop-off with distance , finite depth-of-field and depth-of- scene and then identifying their postures. Such information can help focus, etc. Also, we are able to more accurately estimate the visibility with automatic threat assessment of a scene. viewshed of the sensors, and are able to deal with the complexities of In the current implementation, we have two pairs of stereo cameras. a real scene like occlusions, undulations in the ground plane, etc. The The cameras are fully calibrated and assumed to remain fixed in their algorithm is designed so that camera orientation can be determined positions. After performing correspondence matching between two online in response to changing surveillance requirements and sensor views, the different stereo pairs generate 3D points in the scene. failure. Object detection and interpretation are then performed completely in This algorithm was realized for a portion large urban campus covering 3D space. Using depth information, persons can easily be separated a region of 500 m^2 with multiple buildings, sidewalks, trees and from the background and their posture identified by matching with 3D other occluding objects. We built a 3D CAD model of the environment model templates. and tested the algorithm on it for seven pan-tilt-zoom cameras. The resulting sensor orientations satisify the real-world security and 6491-03, Session 1 surveillance concerns of the environment. On the performance evaluation of tracking systems using 6491-05, Session 2 multiple pan-tilt-zoom cameras X. Desurmont, Multitel (Belgium) More than a poplar plank, the shape and subtle colors of the masterpiece Mona Lisa by Leonardo: the 3D Accurate object tracking in multi-view video sequences from Pan Tilt technology Zoom (PTZ) cameras is an important task in many applications such as video surveillance, traffic monitoring, marketing and sport analysis. F. Blais, J. M. Taylor, L. Cournoyer, M. Picard, L. Borgeat, G. It basically combines in parallel the multi-view matching and fusion of Godin, J. Beraldin, M. Rioux, National Research Council Canada video images using a 3D model of the scene and the positioning of (Canada); C. Lahanier, B. Mottin, Ctr. de Recherche et de mobile objects within this scene using tracking algorithms. Restauration des Musées de France (France) Testing video-analysis algorithms is very important in the academic During the autumn of 2004, a series of scientific examinations of the and industrial communities in order to analyse and enhance these Mona Lisa were undertaken by the C2RMF, the Musée du Louvre and technologies. To enable objective performance evaluation, multiple the NRC, to help assess and monitor the condition of the panel steps are required. Firstly, video sequences must be available.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 71 Conference 6491: Videometrics IX painting. Never before a piece of art underwent such a massive In this paper a new approach, allowing a more reliable scientific examination including grazing light, x-radiography, characterization of 3D scanning systems in the case of applications to ultraviolet, infrared, microfluorescence, 2D multispectral imaging and free-form surfaces, is suggested and described in detail. Problems a world first, the complete 3D digitization of the painting and the that could limit the effectiveness of such a method have been tackled creation of a high resolution virtualized 3D model of the Mona Lisa. and solved. The metrological characterization is performed only with For this task, a team of 3D imaging scientists from the National reference to the coordinate z. Systematic and random effects on the Research Council of Canada were invited to Paris to undertake the 3D values of the coordinates (x,y) give raise to an inaccurate localization scanning of the painting. The objective was to scan the Mona Lisa - of the surface point on the xy plane of the 3D coordinate system obverse and reverse - in order to provide high-resolution 3D image which is considered equivalent to a correct pair of coordinates (x,y) data of the complete painting to: record the shape of the poplar panel, associated with a value of z affected by an error. So the metrological document surface features, wood grain structure, surface lacunae, characterization of the z coordinate itself can be sufficient to describe and provide high resolution pictorial layer images to assist in studies the behaviour that the 3D scanning system exhibits when it localizes related to the artists’ technique as well as for conservation the surface points in the space. examination. This new method uses objects of arbitrary shape which are measured The presentation will highlight the technology and challenges both by the 3D scanner and by a coordinate measuring machine associated with obtaining 3D color images at a resolution of 0.05 mm (CMM) whose accuracy of the measurements uncertainty has been and a depth precision of 0.01 mm. certified. The point coordinates provided by the CMM are considered the conventionally true values and are compared with the corresponding average point coordinates captured in N successive 6491-06, Session 2 scansions by the 3D scanner. Systematic errors can be so corrected, if significant. As final result, a cloud of uncertainty values is provided Low-cost characterization of 3D laser scanners which represent the random variability of the corrected values. M. Russo, G. Guidi, Politecnico di Milano (Italy) The calibration of a range camera greatly influences the whole 3D 6491-08, Session 2 acquisition and modeling process, allowing to minimize the equipment inaccuracy. However, depending on range camera “openness”, we TOF laser scanner characterization for low-range might have systems pre-calibrated only once by the industrial applications manufacturer or systems requiring a regular (and mandatory) end-user G. Guidi, Politecnico di Milano (Italy); C. Bianchini, La Sapienza calibration before any scan session. The metrological system Univ. (Italy) characterization represents a point of paramount importance for making the user aware of the actual performances of his equipment. When TOF laser scanners were introduced on the market, their This permits the choice of appropriate resolution in 3D scan planning, performances were rather poor, having in general a measurement allows to properly interpreting the feedback indices during the uncertainty in the range of centimeters. For this reason it was clear alignment and helps selecting the modification level of geometrical that their application were definitely limited to environment and features in polygonal model editing. These remarks are effective for architecture survey, where the large size of the involved objects both triangulation based instruments, like Minolta Vivid 910, makes acceptable the relative measurement error. But with the ShapeGrabber SG100 and SG1000 evaluated in this paper, or TOF progressive improvement of technology, and the consequent increase based instruments. The proposed experimental method is based on in the measurement precision, the potential range of purposes have post processing of range data produced by acquiring the surface of a been widened. In this paper an application to museum objects have precise test object with 3D laser scanner. In this procedure resolution, been considered. Studying the scanner performance when working at accuracy, and precision parameters are obtained sequentially, through a low range, and using such results to properly interpret the acquired the application of a set of simple geometric processing steps. Such data, it was possible to survey a famous wooden model of S. Peter operating easiness make this approach a possible candidate as a basilica in Rome, re-modeling its shape with a 3D surface modeler. mandatory step in any 3D acquisition and modeling project. Resolution, precision and accuracy have been studied at distances ranging from 1 to 3 meters, in working conditions similar to those imposed by the museum constrains. The results were used to 6491-07, Session 2 properly set-up some post processing steps instrumental characterization as a key step in the 3D modeling process, both for Metrological verification of 3D scanners: a preliminary increasing the geometrical data reliability, and for processing them in approach a way compliant with their metrological characteristics. A. Pietrosanto, C. Liguori, R. Anchini, Univ. degli Studi di Salerno (Italy); G. Strazzullo, FOX BIT s.r.l. (Italy) 6491-10, Session 3 Resolution along the three axes plays an important role among static characteristics of a 3D scanner, since the coordinates x, y, z of each Real-time range imaging by phase-stamp method using point are provided as multiple of the corresponding axes resolutions correlation image sensor Dx, Dy or Dz. However, resolution does not allow the quality of the A. Kimachi, Osaka Electro-Communication Univ. (Japan); S. Ando, measurement results to be quantified. This task can be accomplished The Univ. of Tokyo (Japan) through suitable parameters like probing error, and/or sphere spacing error defined in ISO 10 360. These parameters have been similarly This paper proposes a real-time phase-stamp range finder (PSRF) redefined in VDI/VDE 2634/part 2, with the addition of the flatness using a three-phase correlation image sensor (3PCIS). The system measurement error, thus standardizing a worst-case characterization consists of the same components as in the conventional light-stripe given in terms of accuracy in the measurement of synthetic range finder except the 3PCIS and a three-phase reference signal parameters. To evaluate these error parameters, the 3D scanning generator. During a single scan of the sheet beam, the 3PCIS system must acquire point clouds from the surface of each of “stamps” the phase of the reference signals at the moment the light standard canonical shape objects (planes, spheres, etc..) and process stripe is imaged on each pixel. The three-dimensional shape of the them to measure object parameters (sphere diameter, flatness) whose object is obtained in real time by converting the detected phase to the value is certified. This method has become diffuse and well- angle of the projected sheet beam. Compared to time-stamp VLSI established, but its validity seems to be high only when the objects to range finders, the proposed PSRF is not affected by spatial be reconstructed have canonical shapes. Moreover, this approach nonuniformity in surface reflectance of the object and background does not hold into account the widespread standard of “uncertainty in illumination. It can be realized with a simpler architecture than the measurement”, where a statistical approach is suggested to evaluate modulation-type PSRF previously proposed by the authors. a measurement interval with an estimated confidence level.

72 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6491: Videometrics IX

6491-11, Session 3 6491-15, Session 4 Traceable 3D imaging metrology Content-based retrieval aided registration for images J. Beraldin, M. Rioux, L. Cournoyer, F. Blais, M. Picard, J. R. and range images Pekelsky, National Research Council Canada (Canada) E. Paquet, J. Beraldin, National Research Council Canada This paper summarizes the causes of uncertainty in 3D data (Canada) measurement, some basic theory of 3D imaging that explains the origin The modelling of complex objects and sites involve the acquisition of of some of these causes, and, describes the properties and a large number of texture maps and range images; each one of them performances of a 100 square meter facility to perform research in representing a particular viewing angle. These views must be traceable 3D imaging metrology. Built in 2006, the laboratory space combined and registered in order to create an accurate model of the allows accurate measurements of 3D data from devices operating at original. The complexity of the resultant models, and consequently the standoff distances from a few centimeters up to 10 meters. A laminar number of views required, has increased tremendously over the past flow of 20 degrees C air at 50% humidity level is maintained within ±0.1 decade. Furthermore major projects involve multinational and degree C. The total volume of air in the lab is changed twice a minute multilingual teams, each with different underlying methodologies. In (18 000 cfm). This characteristic combined with the air filtering design such conditions, it is difficult to make sense of the annotation and to allows air cleanliness to be exceptionally good (Class 100). determine which views should be registered together. We propose a new approach in which similar views are found by content-based indexing and retrieval of textures (2D) and range images (3D). The 6491-12, Session 3 views are described automatically, according to their appearance and their 3D geometry. A search engine allows retrieving similar and Range imaging technology: new developments and related views for registration. applications for people identification and tracking T. Kahlmann, F. Remondino, ETH Zürich (Switzerland) 6491-16, Session 4 Range Imaging (RIM) is a new suitable choice for measurement and modeling in many different applications. RIM is a fusion of two different A master-slaves volumetric framework for 3D technologies. According to the terminology, it integrates distance reconstruction from images measurement as well as imaging aspects. The distance measurement principle is dominated by the time-of-flight principle while the imaging D. Ruiz, B. Macq, Univ. Catholique de Louvain (Belgium) array (e.g. CMOS sensor) enables each pixel to store also the distance A system reconstructing arbitrary shapes from images in real time and towards the corresponding object point. Due to the technology’s with enough accuracy would be paramount for a huge number of relatively new appearance on the market, with a few different applications. The difficulty lies in the trade off between accuracy and realizations, the knowledge of its capabilities is very low. computation time. Furthermore, given the image resolution and our In this paper we present our investigations on the range imaging real time needs, only a small number of cameras can be connected to camera SwissRangerTM (realized by the Swiss Center for Electronics a standard computer. The system needs a cluster and a strategy to and Microtechnology - CSEM). Different calibration procedures are share information. We introduce a framework for real time voxel based performed, including a photogrammetric camera calibration and a reconstruction from images on a cluster. From our point of view, the distance system calibration with respect to the reflectivity and the volumetric framework has five major advantages: an equivalent tree distance itself. Furthermore we report about measurement applications representation, an adaptable voxel description, an embedded multi- in the field of surveillance and biometrics. In particular, range imaging resolution capability, an easy fusion of shared information and an easy data of moving people are analyzed, to identify humans, detect their exploitation of inter-frame redundancy; and three minor disadvantage, movements and recover 3D trajectories. its lack of precision with respect to method working at point level, its lack of global constraints on the reconstruction and the need of strongly calibrated cameras. Our goal is to illustrate the advantages 6491-13, Session 3 and disadvantages of the framework in a practical example: the computation of the distributed volumetric inferred visual hull. The Heterodyne range imaging as an alternative to advantages and disadvantages are first discussed in general terms photogrammetry and then illustrated in the case of our concrete example. A. A. Dorrington, M. J. Cree, Univ. of Waikato (New Zealand); D. A. Carnegie, Victoria Univ. of Wellington (New Zealand); A. D. Payne, 6491-17, Session 4 Univ. of Auckland (New Zealand); R. M. Conroy, Univ. of Waikato (New Zealand) View planning for wide-area outdoor modeling based on measurement density of objects Solid-state full-field range imaging technology, capable of determining the distance to objects in a scene simultaneously for every pixel in an T. Asai, M. Kanbara, N. Yokoya, Nara Institute of Science and image, has recently achieved sub-millimeter distance measurement Technology (Japan) precision. With this level of precision, it is becoming practical to use this technology for high precision three-dimensional metrology We propose a support system of range data acquisition by a laser applications. Compared to photogrammetry, range imaging has the rangefinder for wide area outdoor modeling in order to reduce un- advantages of requiring only one viewing angle, a relatively short observed portions in a generated model. The system presents an measurement time, and simplistic fast data processing. In this paper operator a recommendation degree map which illustrates we fist review the range imaging technology, then describe an recommended position for range data acquisition in the objective experiment comparing both photogrammetric and range imaging area. The operator decides the next acquisition position in measurements of a calibration block with attached retro-reflective consideration of movement distance of sensor system and targets. The results show that the range imaging approach exhibits recommendation degrees of the map. The recommendation degree is errors of approximately 0.5 mm in-plane and almost 5 mm out-of- computed based on the measurement density as the index. The plane; however, these errors appear to be mostly systematic. We then recommendation degree at a position is given by the difference proceed to examine the physical nature and characteristics of the between the measurement density acquired by rangefinder and the image ranging technology and discuss the possible causes of these measurement density estimated by the system in reachable area of systematic errors. Also discussed is the potential for further system laser beams. The reachable area of laser beams is estimated by using characterization and calibration to compensate for the range a 3D model generated from the acquired range data. The system determination and other errors, which could possibly lead to three- computes the measurement density considering the reachable area of dimensional measurement precision approaching that of laser beams. The recommendation degrees in the objective area are photogrammetry. computed from the model generated from range data whenever a

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 73 Conference 6491: Videometrics IX range data is acquired. Moreover, the system judges whether 6491-20, Session 5 overlapping portions of the range data can be acquired for the registration by ICP algorithm from a work area where the sensor Concepts of single-camera high-speed 3D measurement system can enter in the objective area. systems H. Maas, Technische Univ. Dresden (Germany) 6491-18, Session 4 The paper gives an overview on single camera 3D measurement system concepts. It presents two concepts of setting up A multinational deployment of 3D laser scanning to study photogrammetric 3D measurement systems based on a single craniofacial dysmorphology in fetal alcohol spectrum highspeed camera in detail, one based on a single camera with a disorders stereo mirror system and one on based on a camera-projector J. L. Rogers, E. A. Wernert, Indiana Univ.; E. Moore, St. Vincent combination: Indianapolis Hospital; R. Ward, Indiana Univ. School of Liberal • A stereo mirror system in front of a camera generates multiple views Arts; L. F. Wetherill, Indiana University School of Medicine; T. on one sensor. These views can be considered virtual cameras of a Foroud, Indiana Univ. School of Medicine multi-camera photogrammetric measurement system. The paper presents a flexible quadruple-mirror system to be used in 3D motion Craniofacial anthropometry (the measurement and analysis of head analysis applications and discusses geometric modeling and system and face dimensions) has been used to assess and describe calibra-tion. abnormal craniofacial variation (dysmorphology) and the facial phenotype in many medical syndromes. Traditionally, anthropometry • Photogrammetric surface measurement systems can be configured measurements have been collected by the direct application of by a camera and a projector device projecting strip or dot patterns calipers and tape measures to the subject’s head and face, and can onto a surface. Using the projector as an active element in the system suffer from inaccuracies due to restless subjects, erroneous landmark design, dense and accurate 3D surface representations may be identification, clinician variability, and other forms of human error. generated using coded light approach or phase shift techniques. In Three-dimensional imaging technologies promise a more effective the geometric modeling and calibration, the projector can be alternative that separates the acquisition and measurement phases to considered an inverse camera. The paper presents results of an reduce these variabilities while also enabling novel measurements and accuracy test obtaining a precision potential in the order of 1/50 pixel longitudinal analysis of subjects. Indiana University (IU) is part of an from a standard consumer-grade beamer. international consortium of researchers studying fetal alcohol Besides the system design and the geometric modeling of the two spectrum disorders (FASD). Fetal alcohol exposure results in concepts, the paper shows results from pilot studies and practical predictable craniofacial dysmorphologies, and anthropometry has applications. been proven to be an effective diagnosis tool for the condition. IU is leading a project to study the use of 3D surface scanning to acquire anthropometry data in order to more accurately diagnose FASD, 6491-21, Session 5 especially in its milder forms. This paper describes our experiences in selecting, verifying, supporting, and coordinating a set of 3D scanning Configuration of multimirror systems for single high- systems for use in collecting facial scans and anthropometric data speed camera based 3D motion analysis from around the world. T. Putze, Technische Univ. Dresden (Germany); K. Raguse, Volkswagen AG (Germany); H. Maas, Technische Univ. Dresden 6491-19, Session 4 (Germany) The use of multi camera systems to determine 3D object coordinates Measurement and modeling of 4D live mouse heart is a common approach in close range photogrammetry. The paper will volumes from CT time series discuss system configurations based on a single camera and mirror A. W. Wetzel, Pittsburgh Supercomputing Ctr.; C. T. Badea, Duke systems to generate a virtual multi camera system. The two main Univ.; S. Pomerantz, D. Nave, Pittsburgh Supercomputing Ctr.; N. advantages of mirror-based multiple-view vision systems can be seen Mistry, A. Johnson, Duke Univ. in the absence of synchronisation requirements and the costs of only one camera. The latter applies especially if high speed cameras are Mouse models of cardiac disease provide detailed information about required to capture the dynamics of an application. The orientation live heart function that cannot be obtained in humans. Quantitative parameters of the fixed camera mirror system can be determined in a measurements of left ventricular (LV) volume at multiple contractile previous step. This is a very important fact for industrial applications phases are particularly important. However, the mouse heart s small where a stability of control point for on the job calibration can not be size and rapid motion present challenges for precise measurement in ensured. These advantages have to be weighted against the live animals. Researchers at Duke University’s Center for In Vivo disadvantages of a reduced active image format per view and the Microscopy (CIVM) have developed noninvasive time-gated restrictions to the imaging geometry. microcomputed tomography (micro-CT) techniques providing the The accuracies achieved in practical tests and in applications are temporal and spatial resolutions required for in vivo characterization of almost comparable to those obtained with multi camera systems. 3D cardiac structure and function. This paper describes analysis of the object coordinates of signalled markers can be determined by the resulting reconstructions to produce volume measurements and fixed mirror system with an accuracy of 0.5 mm in lateral direction and corresponding models of heart motion. We believe these are the most 3.5 mm in depth direction. Small Styrofoam spheres and be located precise noninvasive estimates of in vivo LV volume currently available. with an accuracy of 0.15 mm in lateral components and 0.6 mm in Our technique uses binary mixture models to directly recover volume depth components. estimates from reconstructed datasets. Unlike methods using segmentation followed by voxel counting, this approach provides statistical error estimates and maintains good precision at high noise 6491-22, Session 5 levels. This is essential for long term multiple session experiments that must Gait analysis in forensic medicine simultaneously minimize contrast agent and x-ray doses. The analysis P. K. Larsen, N. Lynnerup, E. B. Simonsen, Univ. of Copenhagen tools are built into the Pittsburgh Supercomputing Center’s Volume (Denmark) Browser (PSC-VB) that provides networked multi-site data sharing and collaboration including analysis and visualization functions. We have combined the basic human ability to recognize other individuals with functional anatomical and biomechanical knowledge, in order to analyze the gait of perpetrators as recorded on surveillance video. The perpetrators are then compared with similar analyses of

74 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6491: Videometrics IX suspects. At present we give a statement to the police as to whether 6491-24, Session 6 the perpetrator has a characteristic gait pattern compared to normal gait, and if a suspect has a comparable gait pattern. We have found Three-dimensional body scanning technology for fashion agreements such as: limping, varus instability in the knee at heel and apparel industry strike, larger lateral flexion of the spinal column to one side than the other, inverted ankle during stance, pronounced sagittal head- N. D’Apuzzo, Homometrica Consulting (Switzerland) movements, and marked head-shoulder posture. Based on these This paper presents an overview of the 3D body scanning characteristic features, we state whether suspect and perpetrator technologies with applications to the fashion and apparel industry. could have the same identity but it is not possible to positively identify Complete systems for the digitization of the human body exist since the perpetrator. Nevertheless, we have been involved in several cases more than ten years. One of the main users of this technology with where the court has found that this type of gait analysis, especially application in the textile field was the military industry. In fact, body combined with photogrammetry, was a valuable tool. The primary scanning technology is being successfully employed since many requisites are surveillance cameras recording with sufficient years in military bases for a fast selection of the correct size of frequency, ideally about 15 Hz, which are positioned in frontal and uniforms for the entire staff. Complete solutions were especially preferably also in profile view. developed for this field of application. Many different research projects were issued for the exploitation of 6491-23, Session 6 the same technology in the commercial field. Experiments and start- up projects were performed and are to time running in different parts Developing measurement tools for the future: of the world by installing full body scanning systems in various commercial innovation in the videometrics industry locations such as shopping malls, boutiques or dedicated scanning R. Godding, AICON 3D Systems GmbH (Germany) centers. Everything is actually ready to be exploited and all the required hardware, software and solutions are available: full body The use of optical 3d measurements systems in industrial scanning systems, software for the automatic and reliable extraction environments has drastically increased within the last years. of body measurements, e-kiosk and web solutions for the Especially in automotive and aerospace industry there is a much presentation of garments, high-end and low-end virtual-try-on better acceptance of such systems like some years ago. Nevertheless systems. However, complete solutions in this area have still not yet there seems to be a much bigger potential for the next years, found the expected commercial success. Today, with the on-going especially in the field of inline measurements. large cost reduction given by the appearance of new competitors, Taking into account that background the development of enhanced methods for digitization of the human body becomes more interesting software and hardware tools has an important role, for the commercial for the fashion and apparel industry. Therefore, a large expansion of system manufacturer as well as for the industrial users of such human body digitization is expected in the near future. systems. To date, different technologies are used commercially for the The paper describes some important influence factors for the measurement of the human body. They can be divided into three development under a commercial point of view. On the one hand the distinguished groups: laser-scanning, projection of light patterns, requirements of the system users must be taken into account. Here combination modeling and image processing. The different solutions we have the requirement to get systems which are easy to use, even have strengths and weaknesses that profile their suitability for specific from users without any background in optical measurement applications. This paper gives an overview of their differences and techniques. Most of the systems have to work more or less characteristics and expresses clues for the selection of the adequate automatically (especially for inline solutions) or as so called “one method. A special interest is given to practical examples of the button solutions”. For those reasons a lot of development is commercial exploitation of human body digitization with applications necessary in the field of automation of reliable algorithms and in the to the fashion and apparel industry. design of user friendly user-interfaces, which in principle have to hide the high complexity of most of the used algorithms. In addition the systems have to fulfil defined accuracy requirements, results must be 6491-25, Session 6 traceable according to different regulations. Single CMOS sensor system for high-resolution double- On the other hand there are requirements from the side of the system volume measurement applied to membrane distillation manufactures, too. One of the main targets is to design the systems system as easy as possible regarding to support tasks and future developments, taking into account the above described solutions. M. G. Lorenz, Univ. Carlos III de Madrid (Spain); M. A. Izquierdo- Similar conditions are valid for the development and use of hardware Gil, Univ. Complutense de Madrid (Spain); R. Sanchez-Reillo, components. It is an advantage to use technologies, which fulfil Univ. Carlos III de Madrid (Spain); C. Fernandez-Pineda, Univ. industrial standards, which are compatible to other used hardware Complutense de Madrid (Spain) and which are as open as possible for future developments. To fulfil most of that demands a co-operation between specialists from Membrane distillation (MD) is a relatively new process that is being computer science, image processing and photogrammetry is very investigated world-wide as a low cost, energy saving alternative to helpful in commercial development. conventional separation processes such as distillation and reverse osmosis (RO). This process offers some advantages compared to In addition to a discussion of the different requirements and solutions other more popular separation processes, such as working at room the paper will show different examples for optical measurement conditions (pressure and temperature); low-grade, waste and/or systems, which successfully have been integrated in industrial alternative energy sources such as solar and geothermal energy may environments. One example is a multi-camera system with up to 16 be used; a very high level of rejection with inorganic solutions; small cameras, which is used for the inspection of tubes under industrial equipment can be employed, etc. The driving force in MD processes conditions in a lot of installations. The focus will be a description of is the vapour pressure difference across the membrane. A the hardware concept and the image processing and computation temperature difference is imposed across the membrane, which algorithms on the one hand and the methods to make such a complex results in a vapour pressure difference. The lower vapour pressure on system usable for a fully automatic measurement as an inline- the permeate side can be set up in various measurement system for a 100% control on the other hand. ways: direct contact MD (DCMD), osmotic MD (OMD), sweeping gas An overview about current and future developments will be given. MD (SGMD), vacuum MD (VMD), and air gap MD (AGMD). In DCMD, which is the subject of this paper, hot temperature solution (feed) is brought into contact with one side of the membrane and cold solution (permeate) into contact with the other, so that the vapour pressure is different at each side of the membrane. This pressure difference drives the vapour through the membrane pores, and then the vapour

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 75 Conference 6491: Videometrics IX

condenses in contact with the cold solution on the other side. This The laser light scanner is a direct 3D shape acquisition system based condensation produces an increase in the volume in the recipient that on the observation of the distortion of a laser beam when shined onto contents the cold solution. an object. A scanner is used as a measurement tool to quantitative The principal problem in this kind of system is the accurate estimate the deformation of an art work or investigate the possible measurement of the recipient volume change, especially at very low restoration work. flows. A cathetometer, with up to 0,05 mm resolution, is the On the other hand, cameras provide 2D information that needs a instrument used to take these measurements, but the necessary mathematical model to be converted into 3D data. Images can be human intervention makes this instrument not suitable for automated acquired in the visible of IR spectrum, revealing different information systems. according to the used band. Filters and UV lamps, custom designed In order to overcome this lack, a high resolution systems is proposed, to block stray visible light, can be combined to reveal and document that makes automatic measurements of the volume of both recipients, details which are not perceivable to the normal eye. Using a passive cold and hot, at a rate of up to 10 times per second. Only one 3.1 thermography measurement, the material surface will naturally radiate, Mpixel CMOS area sensor is used, connected to an FPGA (Field scatter and reflect IR energy that can be imaged by the thermal Programmable Gate camera. Differences in thermal emissivity are then imaged as differences in brightness in the thermal map. In this way a Array) Board. The FPGA is programmed with a specially tailored discontinuity shows a thermal emissivity that is different from that of control system. This complex control system includes: camera the substrate material, allowing voids and anomalies to be detected control, image acquisition, image processing, image data extraction, and characterized. The multispectral data achieved can be also used linearization of data, and statistical processing of the data. A specially to estimate surface spectral reflectance, which is an inherent physical designed line-laser illumination system makes possible the use of only property of the art work surface. In this way we infer information about one area sensor for both recipients, making the system simpler and the nature and the chemical composition of the materials on them, cheaper. The system is easily calibrated using a VGA monitor where and with this system this information can be acquired at the same the measurement image is presented. The measurements have an time on the whole surface of the object. accuracy and resolution, better than the cathetometer (0,025mm). However higher resolution can be achieved with sensors with more For Cultural Heritage applications combining the 3D relief to the resolution or stacking additional sensors. During operation water image data or thermal map, it is possible to use the results as surface tilts caused by vibration, these tilts are canceled trough quantitative measurement tool or to exploit the 3D model for virtual statistical treatment of the measurement data. restoration and fruition. The accumulator-multipliers embedded in the FPGAs make possible to design a system, which calculates the statistical treatment at higher 6491-28, Session 7 throughput than conventional microcontroller or DSP (Digital Signal Processor). The potential of 3D techniques for cultural heritage reconstruction and visualization 6491-26, Session 6 G. Bitelli, V. A. Girelli, Univ. degli Studi di Bologna (Italy); F. Remondino, ETH Zürich (Switzerland); L. Vittuari, Univ. degli Studi Beyond inspection: the promise of videometrics for di Bologna (Italy) industry and engineering The paper deals with the creation of the 3D model of an important K. Edmundson, Geodetic Systems, Inc. archaeological find. In fact this kind of product have become very important in the Cultural Heritage field, offering the possibility of The commercial landscape of industrial metrology has changed studying and inspecting the object, even if it is not really available, significantly in recent years. Suppliers of industrial videometrics because kept in the original country. systems face increasing competition from each other and from existing and emerging alternative technologies. Portable coordinate In our case study photogrammetric and laser scanning surveys were measurement machines in the form of articulated arms, laser trackers, conducted on a faithful mould of a Stela, in order to evaluate the and laser scanners have all established a secure niche for themselves. obtainable products and the problems related to these As in the past, the aircraft and aerospace industries continue to be the methodologies. The 3D photogrammetric model produced by means primary users of metrology systems of all types. The reverberations of some amateur images of the real object, acquired during the from the impact of the terrorist attacks of September 11, 2001 on excavation, presented data processing problems related to camera these industries, felt throughout the metrology sector, are easing. For calibration, unfair image network configuration, surface modelling. photogrammetrists, innovation is the key to future commercial The accurate laser scanning survey required specific procedures for success in the field of videometrics. The strengths and advantages of merging of large number of point clouds and surface generation. digital photogrammetry, commonly known as vision metrology (VM), The work underlines the potentialities of the two techniques, ensure that it will remain a highly competitive, three-dimensional especially if integrated together; the power of the active sensors is coordinate measurement tool. Used predominantly for inspection and confirmated, but it’s demonstrated that, with the support of specialist reverse engineering, it is increasingly utilized for building within the procedures and with the utilization of software often ad-hoc manufacturing process. This paper presents an overview of innovation developed, digital images, also acquired with relatively cheap and the current status of VM in industry and engineering. A systems, contain all the information necessary for the production of description of the state-of-the-art is followed by summaries of five detailed and textured 3D models. unique applications illustrating the basic types of available VM systems. In concluding remarks, selected emerging trends are discussed. 6491-29, Session 7 Wood artworks dimensional monitoring through high- 6491-27, Session 7 resolution 3D cameras Active and passive sensors for art works analysis and G. Guidi, Politecnico di Milano (Italy); C. Atzeni, Univ. degli Studi investigations di Firenze (Italy); J. Beraldin, National Research Council Canada (Canada) A. Pelagotti, Art-Test s.a.s. (Italy); A. Del Mastio, Univ. degli Studi di Firenze (Italy); A. V. Razionale, Consultant (Italy) This paper shows the applicability of non-contact 3D imaging technology to the dimensional monitoring of wooden artworks. The Nowadays inspections, documentations and digital preservations are results of a study on a wooden test artifact submitted to sudden generally performed with active (e.g. laser scanner) and passive (e.g. environmental parameters changes (temperature and humidity) are cameras) sensors. presented. It was possible to verify that the range maps generated by

76 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6491: Videometrics IX a 3D camera based on optical triangulation have the necessary using extended Laws’ convolution kernels. Laws’ kernels are well resolution to show dimensional variations in the order of a few tenths known for 2D texture analysis, and have been used for various pattern of a millimeter. 3D models generated before and after the parameter recognition applications. Although typical Laws’ convolution kernels variation were processed with a specific software in order to highlight are represented in 2D masks, we have extended the kernels to form the amount of deformation through a color coded image. The results 3D masks. The three dimensional extension of the masks allows a demonstrate that such technique represents a unique instrument to pattern recognition system to handle 3D volume data, whereas a capture and track the deformations of wooden artifacts before they traditional approach can handle only 2D image data. Also our become permanent. This study has been conducted in collaboration approach can be extended for use with various lengths of kernels to with the laboratories for restoration of the “Opificio delle Pietre Dure” generate multiple resolutions of masks. In our experiments, mask at Florence, Italy. resolutions of 3x3x3, 5x5x5, 7x7x7 and 9x9x9 were tested.

6491-30, Session 7 6491-33, Session 8 Digital documentation of complex architectures by Acquisition of three-dimensional coordinates of objects integration of multiple techniques: the case study of from Axi-Vision image data and application to Valer Castle stereoscopic display using integral photography system F. Voltolini, ITC-IRST (Italy); S. F. El-Hakim, National Research T. Aida, A. Yoneda, T. Uragaki, Osaka City Univ. (Japan); M. Council Canada (Canada); F. Remondino, ETH Zürich Kawakita, Japan Broadcasting Corp. (Japan) (Switzerland); M. Pontin, S. Girardi, A. Rizzi, L. Gonzo, ITC-IRST Versatile three-dimensional (3D) image data as in the form of I(X,Y,Z) (Italy) defined in the camera coordinate system are acquired from sets of This paper proposes an integrated approach based on the color image and depth image, which are simultaneously captured at a combination of different 3D modelling techniques for the virtual video frame rate by a 3D camera named Axi-Vision camera. The X and reconstruction of complex architectures. The need of combining Y coordinates are calculated from the image coordinates (x, y) on the multiple techniques, like terrestrial laser scanning, photogrammetry image plane of CCD based on the perspective projection theory. The and digital surveying comes from the complexity of some structures Z coordinate can be obtained for each pixel of the color image directly and by the lack of a single technique capable of giving satisfactory from the depth image without any complex processing. Thus, the 3D results in all measuring conditions. image data of objects, I(X,Y,Z), is obtained from the relation between the camera coordinate and the image coordinate systems. A This paper will address modelling issues related to the automation of stereoscopic video image display is demonstrated using the 3D image photogrammetric methods and to the fusion of 3D models acquired data using an integral photography (IP) system, which combines a 4- with different techniques, at different point densities and inch VGA liquid crystal color display panel and a pinhole array. measurement accuracies. Aliasing, an issue from mismatch of image qualities between the 3D The test bench is a medieval castle placed in Trentino A.A., a tiny image data and the IP system, is effectively suppressed by pre- region in Northern Italy. processing the color Axi-Vision image data using a two-dimensional Depending on the required spatial resolutions, different techniques lowpass filter, which is designed based on analyses of the maximum (3D laser scanner, close-range and aerial photogrammetry, surveying) spatial frequency of the 3D image transformed appropriate for the can be used and at the end all the acquired data have to be integrated LCD size and the Nyquist spatial frequency of the IP system. into a single 3D model. The overall model of the castle was done using images acquired from 6491-34, Session 8 an helicopter while the detailed inner models were created by means of close-range photogrammetry and 3D laser scanner. Crack propagation imaging by the ISIS camera and a In addition to the model of the castle, we have also created a DSM of video trigger system the surroundings in order to visualize the castle within its environment. T. Okinaka, P. Karimov, T. G. Etoh, Kinki Univ. (Japan) An ultra-high-speed video camera of 1 Mfps (Million frames per 6491-31, Session 8 second) was applied to visualize crack propagation by means of conventional photo-elastic imaging and a newly-developed video Performance evaluation of a coded structural light trigger system. system for cultural heritage applications In 2001, we have developed a video camera of 1 Mfps with a D. Akca, F. Remondino, D. Novak, T. Hanusch, G. Schrotter, A. dedicated image sensor ISIS, the in-situ storage image sensor, Gruen, ETH Zürich (Switzerland) designed by the authors. Each pixel of the ISIS has more than one hundred CCD memory elements. The image signals are continuously 3D documentation and visualization of Cultural Heritage objects is an recorded in the memory area in all pixels in parallel. As the sensor has expanding application area. The selection of the right technology for 81,120 pixels (260x312), the parallel recording in all the pixels these kinds of applications is very important and strictly related to the achieves 81,120 time faster image capturing than the recording of project requirements, budget and user’s experience. Active sensors, image signals captured by a standard image sensor with one signal i.e. triangulation based laser scanners and structured light systems readout line and an memory device outside the sensor. are used for many kinds of 3D object reconstruction tasks and in particular for 3D documentation of cultural heritage objects. This During the image capturing phase, old image signals are drained out study presents some experiences in the results of two case studies in of the sensor and the latest image signals for consecutive images of which a close-range structured light system is used for the 3D more than one hundred frames are always kept in each pixel. The digitization. The paper includes all necessary steps of the 3D object continuous overwriting mechanism makes it easy to synchronize modeling pipeline from data acquisition to 3D visualization. image capturing timing to the occurrence of a target event. The image signals more than one hundred frames reproduces a continuous moving image at 10 fps for more than 10 seconds. 6491-32, Session 8 Almost only one disadvantage of the camera is relatively small frame numbers which sometimes makes the synchronization for the image A solid texture analysis based on three-dimensional capturing difficult, even with the continuous overwriting mechanism. convolution kernels We can use some trigger signals for the synchronization, such as a M. T. Suzuki, National Institute of Multimedia Education (Japan) sound trigger for explosion, collision, etc, a laser light trigger for flying objects passing through the beam, etc. However, there are some This paper describes techniques for analyzing 3D volume data by phenomena which do not release such a clear signal for

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 77 Conference 6491: Videometrics IX synchronization. We typically meet this difficulty in ultra-high speed planar and no-planar grids with circular, square or cross shaped image capturing in microscopic observation. targets in order to calibrate close range cameras. The approach has Generation of a crack is usually captured, employing as the trigger been performed on several calibration sessions each of them followed signal the sound released at the initiation of the crack propagation, or by the three-dimensional reconstruction of a reference object that was abrupt increase of electronic resistance due to separation or break of previously measured through a coordinate measuring machine. The a conductive strip pasted at the crack root. However, if the load to the performance of the proposed approach has been expressed in terms specimen is gradually increased, sharp noises are sometimes emitted of deviation between the reference coordinates and the corresponding before crack generation and the strip breaks with some randomness, ones provided by triangulation. The results obtained have also been which makes the synchronization unstable. compared with those achieved through ShapeCaptureTM, a software widely used in photogrammetry for camera calibration and 3D model We developed a video trigger system, which is very sensitive and reconstruction from images. works accurately. It monitors the central part in a view area of a scope and detects a sudden change in the local area. The system is efficiently applied to ultra-high-speed image capturing of generation 6491-42, Poster Session of a crack under the increasing static load and the propagation. At the edge of the initial crack artificially prepared before the experiment, the Design of an image sensor for an ultra-high-speed and stress concentration produces very clear and dense stripes before the ultra-high-sensitive video microscope generation of the crack; then, the stripes suddenly change at the instance of initiation of propagation of the crack. N. Ohtsuka, Kinki Univ. (Japan) Some interesting mechanism of the crack propagation were detected. This paper outlines a special microscope under development, named Among them, the propagation speed of a crack fluctuates during one “Ultra-high-speed bionanoscope” for ultra-high-speed imaging in propagation process, seemingly relating to the propagation and biological applications, and preliminary design of the image sensor, reflection of the stress wave. which is the key component in the system. The ultra-high-speed bionanoscope consists of two major subsystems: a video camera operating at more than 10 Mfps with 6491-39, Poster Session ultra-high-sensitivity and the special microscope to minimize loss of Variability of bodily measures of normally dressed people light for seriously reduced illumination light energy due to the ultra- high-speed imaging. The ultra-high-frame rate is achieved by using PhotoModeler(r) introducing a special structure of a CCD imager, the ISIS, In-situ P. K. Larsen, N. Lynnerup, Univ. of Copenhagen (Denmark); L. Storage Image Sensor, invented by Etoh and Mutoh. The ISIS has an Hansen, National Institute of Occupational Health (Denmark) array of pixels each of which equips with a slanted linear CCD storage area for more than 100 image signals for reproduction of smoothly Since the 1990s photogrammetry has been used in forensic medicine moving images. The ultra-high-sensitivity of the sensor of less than 10 to help identifying perpetrators from crime scenes covered with photons is achieved by introducing three existing technologies, surveillance video. Some case studies have shown promising results backside-illumination, cooling, and the CCM, Charge Carrier measuring both heights and segment lengths. Photogrammetry is Multiplication invented by Hynecek. used with high precision when measuring clearly defined points but less is known about the reproducibility of hidden body-points. In this study we quantified the inter- and intra-observer variability of bodily 6491-43, Poster Session measures with the software PhotoModeler(r) Pro in low resolution images. The study showed that the body height and shoulder height Performance analysis and parameter optimization for iris were reproducible within ± 1 cm and ± 2 cm respectively. The recognition using Log-Gabor wavelet measurement error were markedly higher for all measurements between points hidden by clothes and measurements of segment V. A. Pozdin, Y. Du, Indiana Univ.-Purdue Univ. Indianapolis lengths were only usable in the intra-observer situation when flexion in The field of biometrics has become a popular field of research as it the joints were present. In the inter-observer study the measurement delivers increased security and convenience. Biometrics is a method error was so high that few measurements could be used beside of recognizing individuals based on their unique physiological or shoulder and body height. Measurements are suggested to be done behavioral characteristics. Among numerous implemented methods of with the perpetrator and suspect in the same pose. recognition, iris recognition has been shown to be the most accurate. The focus of this paper is the improvement of accuracy of the iris recognition system, which is based on a 1-D Log-Gabor wavelet. 6491-40, Poster Session Dependence of accuracy on wavelength, phase quantization, and radial resolution is studied. In this paper, a unique approach has been Subpixel location of discrete target images in close- developed for phase quantization, feature encoding, and matching range camera calibration: a novel approach based on signed integer phase representation. The performance of iris R. Anchini, C. Liguori, Univ. degli Studi di Salerno (Italy); J. recognition system is evaluated based on the Equal Error Rate from Beraldin, National Research Council Canada (Canada) the Receiver Operating Characteristic curves from over 20 million matches. The achieved results reveal a linear dependence between The recent and continuous advancements in machine vision have filter wavelength and the normalized length in lateral dimension. In allowed a relevant increase in the number of measurement addition, the optimum phase number of 8 and radial resolution of 40 applications based on the processing of digital images acquired by are extrapolated. In our experiments using CASIA database and close range cameras. Regardless of the method chosen, camera optimal parameters, an accuracy of above 99.5% has been achieved. calibration needs to be addressed up-front. Neglecting self-calibration schemes, this stage is always preliminary at any measurement since it allows one to find the vision system parameters (internal and external) necessary to infer 3-D information from 2-D images acquired by cameras. The calibration phase uses spatial objects called targets that are arranged in ad hoc rigid structures (3-D or planar grids). It means that the relative position of each target with respect to the others is known because it has been measured through traditional techniques. The overall accuracy of final measurements is heavily dependent on the initial step of the calibration e.g. the precise location of each target within acquired images. A novel approach is presented for the location of discrete targets images with subpixel accuracy. It can be adopted as first stage of any method that uses

78 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492: Human Vision and Electronic Imaging XII Monday-Thursday 29 January-1 February 2007 Part of Proceedings of SPIE Vol. 6492 Human Vision and Electronic Imaging XII

6492-01, Session 1 and in artificial movies derived from simulated environments. Natural movies represent the full complexity of retinal image motion, whereas Painterly rendered portraits from photographs using a artificial movies provide complete knowledge of the relationship knowledge-based approach between retinal image motion and motion in the environment. Here we consider the problem of coding retinal image motion when an S. DiPaola, Simon Fraser Univ. (Canada) observer moves through an environment. Natural movies were created Portrait artists using oils, acrylics or pastels use a specific but open by moving a calibrated camera along a known trajectory at a constant human vision methodology to create a painterly portrait of a live sitter. speed. Simulated environments were created by combining the range When they must use a photograph as source, artists augment their statistics of natural scenes with the spatial statistics of natural process, since photographs have: different focusing - everything is in images. The natural movies allow us to determine whether the artificial focus or focused in vertical planes; value clumping - the camera movies are correctly capturing relevant statistical properties. The darkens the shadows and lightens the bright areas; as well as color artificial movies allow us to determine how sensors should be and perspective distortion. In general, artistic methodology attempts designed to optimally estimate local retinal velocities. We find that the following: from the photograph, the painting must ‘simplify, across a range of environments the optimal integration area of local compose and leave out what’s irrelevant, emphasizing what’s motion sensors increases logarithmically with the speed to which the important’. While seemingly a semantic goal, artists use known sensor is tuned. This finding makes predictions for neurons involved in techniques such as relying on source tone over color to indirect into a heading perception and may find use in robotics applications. semantic color temperature model, use brush and tonal “sharpness” to create a center of interest, lost and found edges to move the viewers gaze through the image towards the center of interest as well as other 6492-04, Session 2 techniques to filter and emphasize. Our work attempts to create a Bilinear models of natural images knowledge domain of the portrait painter process and incorporates this knowledge into a multi-space parameterized system that can create an B. A. Olshausen, Univ. of California/Berkeley array of NPR painterly rendering output by analyzing the photographic- based input which informs the semantic knowledge rules. In previous work, we have shown that if one attempts to model the statistics of natural images using a basis function decomposition, where the basis function coefficients are assumed to be sparse 6492-02, Session 1 (mostly zero), then the optimal basis resembles the receptive fields of neurons in the visual system (in addition to the bases of certain Nonlinear encoding in multilayer LNL systems optimized wavelet transforms). Here we attempt to build upon this work by for the representation of natural images showing that one can achieve independent representations of ‘what’ (form) and ‘where’ (transformations) in natural images through the use C. Zetzsche, Univ. Bremen (Germany); U. Nuding, Ludwig- of a bilinear model in which coefficients are multiplied by each other. Maximilians-Univ. München (Germany) Our goal here is to understand how invariant representations are We consider the coding properties of multilayer LNL (linear - nonlinear formed in the visual system, but the model also shows promise for -linear) systems. They consist of interleaved layers of linear filter use in a variety of practical pattern analysis tasks where it is desired operations, nonlinear transformations, linear filter operations, etc. to factor apart the shape of an object from its transformation. These systems can be used as models of visual processing in higher cortical areas (V2, V4), and are also interesting with respect to image processing and coding. The linear filter operations in the different 6492-05, Session 2 layers are optimized for the exploitation of the statistical redundancies Locally adaptive image representation of natural images. We explain why even simple nonlinear operations - like ON/OFF rectification - can convert higher-order statistical E. P. Simoncelli, New York Univ. dependencies remaining between the linear filter coefficients of the No abstract available first layer to a lower order. The resulting nonlinear coefficients can then be linearly recombined by the second-level filtering stage, using the same principles as in the first stage. In order to obtain insights into the 6492-06, Session 2 coding efficiency of these systems we investigate how the information is distributed among the output coefficients. Our investigations Spatiotemporal power spectra of motion parallax: the indicate that (a) information can be preserved across layers, and (b) case of cluttered 3D scenes although the state space is formally expanded (the dimensionality is increased) higher-level neurons can yield an information-theoretically D. Rivait, M. S. Langer, McGill Univ. (Canada) efficient representation of the input. (Supported by DFG:SFB/TR-8) We examine the spatiotemporal power spectra of image sequences that depict dense motion parallax, namely the parallax seen by an observer moving laterally in a cluttered 3D scene. Previous models of 6492-68, Session 1 the spatiotemporal power have accounted for effects such as a static New vistas in image and video quality 1/f spectrum in each image frame, a spreading of power at high spatial frequencies in the direction of motion, and a bias toward either A. C. Bovik, The Univ. of Texas/Austin lower or higher image speeds depending on the 3D density of objects in the scene. Here we use computer graphics to generate a No abstract available parametrized set of image sequences and qualitatively verify the main features of these models. The novel contribution is to show how 6492-03, Session 2 failures of 1/f scaling can occur in cluttered scenes. Such failures have been described for the spatial case, but not for the spatiotemporal Optimal sensor design for estimating local velocity in case. natural environments We find that when objects in the cluttered scene are visible over a T. Tversky, W. Geisler, The Univ. of Texas/Austin wide range of depths, and when the image size of objects is smaller than the image width, failures of 1/f scaling tend to occur at certain Motion coding in the brain undoubtedly reflects the statistics of retinal critical frequencies, defined by a correspondence between object size image motion occurring in the natural environment. To characterize and object speed. these statistics it is useful to measure motion both in natural movies

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 79 Conference 6492: Human Vision and Electronic Imaging XII

6492-07, Session 2 The best result on MNIST has an error rate of 0,43%. This result, however, has been obtained by using explicit knowledge that is Feature category systems for 2nd order local image specific to the data (elastic distortion model for digits). Currently we structure induced by natural image statistics and obtain the second best result with an error rate of 0,55% and thereby otherwise outperform all methods that do not use data-specific knowledge. L. D. Griffin, M. Lillholm, Univ. College London (United Kingdom) We report progress on an approach (Geometric Texton Theory - GTT) 6492-10, Session 2 that like Marr’s ‘primal sketch’ aims to describe image structure in a Unsupervised learning of a steerable basis for invariant way that emphasises its qualitative aspects. In both approaches, image representations image description is by labelling points using a vocabulary of feature types, though compared to Marr we aim for a much larger feature M. Bethge, S. Gerwinn, Max Planck Institut für biologische vocabulary. Kybernetik (Germany) We base GTT on the Gaussian derivative (DtG) model of V1 While the pixel representation of an image changes dramatically under measurement. Marr’s primal sketch was based on DtG filters of affine transformations such as translation, rotation, and scaling, the derivative order up to 2nd, for GTT we plan to extend to the image content is preserved to a large extent. Specifically, if the physiologically plausible limit of 4th. This is how we will achieve a changes in a D-dimensional vector of pixel intensities can be larger feature vocabulary (we estimate 30-150) than Marr’s ‘edge’, described as a one-parametric Lie group, the concept of steerable ‘line’ and ‘blob’. The central requirement of GTT then is for a filters allows one to find a lossless image representation with 1 procedure for determining the feature vocabulary that will scale up to component describing the Lie group parameter and the other D-1 4th order. We have previously published feature category systems for components being invariant under the Lie group action. We describe 1-D 1st order, 1-D 2nd order, 2-D 1st order and 2-D pure 2nd order. In this representation in terms of a generative model from which we this paper we will present results of GTT as applied to 2-D mixed 1st derive an unsupervised learning algorithm for steerable filters. We use + 2nd order features. the algorithm to determine a complete steerable basis for a rotation We will review various approaches to defining the feature vocabulary, invariant representation of 32x32 image patches. A good performance including ones based on (i) purely geometrical considerations, and (ii) of the learned basis is demonstrated by measuring the variance of the natural image statistics. amplitude and phase spectrum in response to analytically defined steerable filters during rotation. 6492-08, Session 2 6492-11, Session 2 The independent components of natural images are perceptually dependent Analysis of segment statistics for semantic classification M. Bethge, T. Wiecki, F. A. Wichmann, Max Planck Institut für of natural images biologische Kybernetik (Germany) D. Depalov, T. N. Pappas, Northwestern Univ. The independent components of natural images are a set of linear A major challenge facing content-based image retrieval is bridging the filters which are optimized for statistical independence. With such a gap between low-level image primitives and high-level semantics. We set of filters images can be represented without loss of information. have proposed a new approach for semantic image classification that Intriguingly, the filter shapes are localized oriented, and bandpass, utilizes the adaptive perceptual color-texture segmentation algorithm resembling important properties of V1 simple cell receptive fields. by Chen et al., which segments natural scenes into perceptually Here we address the question whether the independent components uniform regions. The color composition and spatial texture features of natural images are also perceptually less dependent than other image representations. We have compared the pixel basis, the ICA of the regions are used as medium level descriptors, based on which basis and the discrete cosine basis by asking subjects to interactively the segments are classified into semantic categories. The segment predict missing pixels (for the pixel basis) or the coefficients of ICA features consist of spatial texture orientation information and color and DCT basis functions in patches of natural images. Like Kersten composition in terms of a limited number of spatially adapted (1987) we find the pixel basis to be perceptually highly redundant but dominant colors. The feature selection and the performance of the perhaps surprisingly, the ICA basis showed significantly higher classification algorithms are based on the segment statistics. In perceptual dependencies than the DCT basis. this paper we investigate the dependence of the segment statistics on the segmentation algorithm. For this, we extract and compare statistics of the segment features obtained using the Chen et al. 6492-09, Session 2 algorithm to those that correspond to human segmentations. Our Learning optimal features for visual pattern recognition findings indicate that although segmentations of the same image by different humans appear to be quite different, the resulting K. Labusch, Univ. zu Lübeck (Germany); U. Siewert, PLANET statistics are consistent. Moreover, the statistics are similar to those intelligent systems GmbH (Germany); T. Martinetz, E. Barth, Univ. obtained when the Chen algorithm is used. We also compare the zu Lübeck (Germany) segment statistics when other algorithms are used for segmentation. The optimal coding hypothesis proposes that the human visual system has adapted to the statistical properties of the environment by the use of relatively simple optimality criteria. 6492-12, Session 3 We here (i) discuss how the properties of different models of image The role of spatially adapted versus non-spatially coding, i.e. sparseness, decorrelation and statistical independence adapted structural distortion in supra-threshold are related to each other (ii) propose to evaluate the different models compression by verifiable performance measures (iii) analyze the classification performance on images of handwritten digits (MNIST data base). M. D. Gaubatz, S. Kwan, S. S. Hemami, Cornell Univ. We first employ the SPARSENET algorithm (Olshausen, 1996) to In wavelet-based image coding, a variety of masking properties have derive a local filter basis (on 13x13 pixels windows). We then filter the been exploited to create spatially-adaptive quantization schemes. images in the database (28x28 pixels images of digits) and reduce the Previous improvements in supra-threshold compression have been dimensionality of the resulting feature space by selecting the locally achieved through analyses of the visibility of wavelet distortions that maximal filter responses. We then train a support-vector-machine on a resulted in non-spatially-adaptive quantization schemes. More recent training set to classify the digits and report results obtained on a studies have illustrated how spatially-adaptive properties such as separate test set. contrast and structural class of natural image data influences masking

80 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492: Human Vision and Electronic Imaging XII properties at threshold. Though these results have been extended in a 6492-15, Session 3 number of ways to achieve supra threshold compression schemes, the relationship between inter-band and intra-band masking at supra- Visual ergonomic aspects of computer displays: glossy threshold rates is not well understood. This work aims to quantify the screens and angular dependence importance of spatially-adaptive distortion as a function of compressed target rate. K. E. Brunnström, B. Andren, Z. Konstantinides, L. Nordström, Acreo AB (Sweden) Two experiments are performed that require the subject to specify the optimal balance between spatially-adaptive and non-spatially- Recently computer displays and laptop computer are designed with a adaptive distortion. Analysis of the resulting data indicates that on so called glare panel i.e. highly glossy screens. The shiny look of the average, the preferred balance between spatial and non-spatial display appeals to the costumers, also there are arguments that the distortion is almost constant across all tested rates. Furthermore, contrast, colour saturation etc improves by using a glare panel. though it is known that mean-squared-error alone is not a good LCD displays suffer often from angular dependent picture quality. This indicator of image quality, it can be used to fairly accurately predict has been even more pronounced by the introduction of Prism Light the outcome of this experiment. This result has convenient Guide plates into displays for laptop computers. implications for image coding that are also discussed. The TCO label is the leading labelling system for computer displays. It has currently about 50% of all computer displays on the market today. 6492-13, Session 3 The requirements are periodically updated to keep up with the technical development and the latest research in e.g. visual Image compression using sparse color sampling ergonomics. The gloss level of the screen and the angular combined with nonlinear image processing dependence has recently been investigated by conducting user studies. S. Brooks, Dalhousie Univ. (Canada); I. Saunders, N. A. Dodgson, Univ. of Cambridge (United Kingdom) A study of the effect of highly glossy screens compared to matt screens has been performed. The results show a slight advantage for We apply two recent non-linear, image-processing algorithms to the glossy screen when no disturbing reflexes are present, however colour image compression. The two algorithms are colorization and the difference was not statistically significant. When disturbing joint bilateral filtering. Neither algorithm was designed for image reflexes are present the advantage is changed into a larger compression. Our investigations were to ascertain whether their disadvantage and this difference is statically significant. Another study mechanisms could be used to improve the image compression rate of angular dependence has also been performed. for the same level of visual quality. Both show interesting behaviour, with the second showing a visible improvement in visual quality, over JPEG2000, at the same compression rate. In both cases, we store 6492-16, Session 3 luminance as a standard, lossily compressed, grey scale image and store colour at a very low sampling rate. Each of the non-linear The blur effect: perception and estimation with a new no- algorithms then uses the information from the luminance channel to reference perceptual blur metric determine how to propagate the colour information appropriately to F. Crete, STMicroelectronics (France) and Lab. des Images et des reconstruct a full colour image. Signaux (France); M. Nicolas, STMicroelectronics (France); T. Dolmiere, P. Ladret, Lab. des Images et des Signaux (France) 6492-14, Session 3 To achieve the best image quality, noise and artifacts are generally removed at the cost of a loss of details generating the blur effect. To Compression of image clusters using Karhunen Loeve control and quantify the emergence of the blur effect, blur metrics transforms have already been proposed in the literature. By associating the blur M. Kramm, Technische Univ. München (Germany) effect with the edge spreading, these metrics are sensitive not only to the threshold choice to classify the edge, but also to the presence of In large-scale multimedia archive systems, it’s often desirable to apply noise which can mislead the edge detection. image compression not only to individual images, but also to groups Based on the observation that we have difficulties to perceive of images, in order to gain better compression results by exploiting differences between a blurred image and the same re-blurred image, inter-image redundancies, and mimimize the overhead of we propose a new approach which is not based on transient compression header information like coding trees and lookup tables. characteristics but on the discrimination between different levels of Algorithms for compressing such image databases need to first divide blur perceptible on the same picture. the image data into manageable blocks (clusters), and then apply a Using subjective tests and psychophysics functions, we validate our given comression algorithm (extended to a multi-image blur perception theory for a set of pictures which are naturally unsharp version) to each cluster. or more or less blurred through one or two-dimensional low-pass In this paper, we analyse the Karhunen-Loeve image compression filters. Those tests show the robustness and the ability of the metric to algorithm for its multi-image compression and clustering properties. evaluate not only the blur introduced by a restoration processing but also focal blur or motion blur. Requiring no reference and a low cost We therefore extend the algorithm so that the Karhunen-Loeve implementation, this new perceptual blur metric is applicable in a transform is derived from the blocks of a given number of cluster large domain from a simple metric to a means to fine-tune artifacts sample images, instead of only a single image, and apply the same corrections. transform to all the images in the cluster. Experimental results show that for most images, the compression rates of the individual images decrease slightly if the KLT Matrix is 6492-17, Session 3 derived from a larger set of images, but the fact that the transformation matrix only needs to be stored once per cluster Perceptual quality evaluation of geometric distortions in compensates that drawback. images Algorithms are presented for on-the-fly cluster creation and matching A. D’Angelo, G. Menegaz, M. Barni, Univ. degli Studi di Siena of images to clusters. (Italy) Research on human perception of image quality has been widely performed through the years, however, it has not dealt with distortions due to geometric operations. This is an important problem in many applications such as digital watermarking in the presence of de- synchronization attacks.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 81 Conference 6492: Human Vision and Electronic Imaging XII

Like all the others distortions that affect digital images, geometric IFD or SNR scalability are used. In this experiment we tested the distortions are also bounded by the maximum visual quality relation between the double-stimulus continuous quality scale degradation they induce before the distorted image loses its (DSCQS) and the double-stimulus impairment scale (DSIS) for two commercial value. The goal of this work is twofold: first to design a video sequences, coded with either IFD or SNR scalability and send mathematical model to describe geometric distortions, then to over a wireless in-home network. Results show that the correlation measure the impact of the parameters of the model from a perceptual between DSIS and DSCQS are linear for SNR scalability, but harder to point of view. Specifically, the maximum admissible distortion that can interpret for IFD. be applied before the distortion becomes visible is measured thus leading to the definition of the perceptually admissible subset of distortions. 6492-21, Session 4 The first part of the paper describes the proposed model. The second Can you see me now? An objective metric for predicting part regards the investigation of the perception of the geometrical intelligibility of compressed American Sign Language distortion introduced by the model. In particular, both objective and subjective tests are performed to characterize the visibility of the video corresponding artifacts. F. Ciaramello, S. S. Hemami, Cornell Univ. For members of the Deaf Community in the United States, current 6492-18, Session 4 communication tools include TTY/TTD services, video relay services, and text-based communication. However, none of these provides the Color preference, color naturalness, and annoyance of personal, instant communication that mobile phones currently offer to compressed and color-scaled digital videos hearing users. With the growth of cellular technology, mobile sign language conversations are becoming a possibility. Proper coding C. C. Koh, J. M. Foley, S. K. Mitra, Univ. of California/Santa techniques must be employed to compress American Sign Language Barbara (ASL) video for low-rate transmission while maintaining the quality of Naïve subjects numerically rated either the level of color preference, the conversation. In order to evaluate these techniques, an color naturalness, or overall annoyance produced by the combination appropriate quality metric is needed. of compression and color (chroma) scaling in digital videos. The Traditional video quality metrics, such as PSNR, are measures of objectives were to determine how varying the chroma of compressed aesthetics, while the quality of an ASL video should be related to its videos affected the mean rating for preference, naturalness, and intelligibility. This work demonstrates an objective intelligibility metric annoyance across subjects and to determine how the ratings were for coded ASL video and shows that it correlates well with subjective related. Preference and naturalness increased to a maximum and observers’ ratings. The intelligibility metric also gives insight into decreased as the mean chroma increased for most of the videos and techniques that can be used in the coding algorithm. compression bit rates. As bit rate decreased, preference and naturalness achieved maximum at increasing lower values. Naturalness peaked at a lower mean chroma than preference, 6492-22, Session 4 indicating that high mean chroma was preferred but may not be considered natural. Annoyance for all compression bit rates Price-dependent quality: examining the effects of price decreased to a minimum and increased as mean chroma increased. on multimedia quality requirements This effect suggests that it may sometimes be possible to reduce the D. S. Hands, R. Jacobs, K. Cheng, British Telecommunications annoyance produced by compression by adjusting chroma. However, plc (United Kingdom); C. Partridge, Univ. of Essex (United the chroma at which annoyance is minimal was less than the most Kingdom) preferred chroma. As bit rate decreased, annoyance increased as expected. For individual videos, preference and naturalness were Traditionally, subjective quality assessments are made in isolation of approximated relatively well by simple Gaussians while inverted mediating factors (e.g. interest in content, price). This approach is Gaussians provided a good account of more than half of the useful for determining the pure perceptual quality of content. There is annoyance ratings with mean chroma as the independent variable. now a growing interest in understanding users’ quality of experience. With total squared error as the independent variable, annoyance was To move from perceptual quality assessment to quality of user characterized by S-shaped forms while preference and naturalness experience assessment, factors beyond reproduction quality must be were described by linear functions. considered. From a commercial perspective, content and prices could determine the success or failure of the business. This paper investigates the relationship between price and quality. Subjects 6492-19, Session 4 selected content that was of interest to them. Subjects were given a budget of ten pounds at the start of the test. When viewing content, Relation between DSIS and DSCQS for temporal and subjects were free to select different levels of quality. The lowest spatial video artifacts in a wireless home environment quality was free (and subjects left the test with ten pounds). The N. Van den Ende, Philips Research Labs. (Netherlands) and highest quality used up the full budget (and subjects left the test with Technische Univ. Eindhoven (Netherlands); L. Meesters, no money). A range of pricing tariffs was used in the test. During the Technische Univ. Eindhoven (Netherlands); R. Haakma, Philips test, subjects were allowed to prioritise on quality or on price. The Research Labs. (Netherlands) results of the test found that subjects tended to prioritise quality over price across all tariff levels. At the higher pricing tariffs, subjects Perceived video quality depends on factors such as image distortion, became more price sensitive. Using data from a number of subjective image content, user preference, judgement strategies and judgement tests, a utility function describing the relationship between price and scales. Here, we concentrate on a combination of image distortion quality was produced. and judgement scales. Image distortion can be caused by a wireless in-home network context, where perceived video quality can suffer from temporal and spatial video artifacts. Video coding adaptation 6492-23, Session 5 methods can remedy wireless network problems to maintain the best possible perceived video quality. We concentrate on two video coding Correcting spurious resolution in defocused images adaptation methods: I-Frame Delay (IFD), dropping least important J. I. Yellott, Univ. of California/Irvine; J. W. Yellott, Consultant frames while introducing temporal artifacts and Signal-to-Noise Ratio Scalability (SNR scalability), dividing video up in a base layer and at Optical modeling suggests that levels of retinal defocus routinely least one enhancement layer while introducing spatial artifacts. To created by presbyopia should produce phase reversals (spurious judge the perceived video quality various measures are used. resolution—SR) for spatial frequencies in the 2 cycles/letter range However, different scales might result in different results when either known to be critical for reading. Simulations show that such reversals can have a decisive impact on character legibility, and that correcting

82 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492: Human Vision and Electronic Imaging XII only this feature of defocused images (by re-reversing spectral sign- In this study spatial contrast sensitivity is investigated, using defined errors caused by defocus) can make unrecognizably blurred letters targets at which observers are asked to point to perceived threshold completely legible. This deblurring effect of SR correction is locations. Several pixel artifacts are considered in an image, including remarkably unaffected by the magnitude of defocus, as determined by pixels with one-half of the typical sensitivity and permanently stuck blur circle size. Both the deblurrring itself and its robustness can be pixels. This experimental study also points towards investigating understood from the effect SR correction has on the defocused variability of Contrast Sensitivity Function (CSF) across different panel pointspread function, which changes from a broad flat cake to a sharp of observers. These visual psychophysical judgements (detection, impulsive peak. This SR-corrected pointspread acts like a delta discrimination and preference) are made under controlled viewing function, preserving image shapes during convolution regardless of conditions (fixed lighting, viewing distance etc) to generate close to blur-disk size. Curiously, such pointspread functions always contain a repeatable and highly reliable data, which can be used to optimise the thin annulus of negative light-intensity values whose radius equals the design of imaging systems. The validity of this collected statistical diameter of the blur circle. We show that these properties of SR- data also depends on methods used to characterise the stimuli, correction all stem from the mathematical nature of the Fourier inverse decisions made by the observers and analysis of the recorded values of the sign of the optical transfer function, which also accounts for the which will be presented in the paper. inevitable low contrast of images pre-corrected for SR. The goal of this work is to quantify quality for a batch of given sensors and/or to simulate noise in a camera imaging system simulator. This experimental data also helps to identify visual signals/pixel artifacts 6492-24, Session 5 that annoy customers and improve on designs to minimise these Do focus measures apply to retinal images? visual signal content. Y. Tian, C. F. Wildsoet, Univ. of California/Berkeley Focus measures have been successfully used in some auto-focusing 6492-56, Poster Session systems. Accommodation/emmetropization represents an ocular Making flat art for both eyes auto-focusing mechanism, which requires analysis of blur at the retinal level. The purpose of the current study is to investigate whether S. Mason, Yavapai College any of the focus measures can be applied to retinal images. Ten focus By expanding the way art is made, possibilities open up in visual measures were chosen for analysis, six in the spatial domain and four experience and thus, to what we might become. transform-based. The performance of the focus measures in the presence of astigmatism and higher order aberrations (HOAs) was First, this is achieved by creating images in 3-D. The work aligns with examined using distributed computing on a small computer cluster. natural physiology, that is, it’s seen with both eyes. Traditionally, it is Aberrations (some randomly generated and some from human eyes) color and rules of perspective that trick the viewer into thinking 3-D. were combined with positive and negative defocus. For each By adding a stereoscopic element, an object is experienced in a combination, a wavefront was reconstructed and a PSF computed naturally three dimensional space with the use of two eyes. using FFT. A blurred image was obtained as the convolution of the Second, further visual expansion is achieved with the use of PSF and a perfect image. A focus measure curve was derived from ChromaDepth glasses to actually see the work in 3D as it is being these blurred images for each focus measure. In the presence of small created. This cannot be done with most other 3-D methods that amounts of astigmatism and HOAs, all focus measures showed require two images or special programming to work. Hence, the monotonic changes with defocus and the curves were unimodal in spontaneous creation of an image within a 3-D space becomes a shape, although there were significant differences in robustness, reality for the artist. defocus sensitivity and effective range. These focus measures need to Third, the connection is made between the 3-D process and the be further investigated to understand which, if any, might be used to creation of art. This is the “how we see” part. By working in a truly analyze retinal image defocus for accommodation/emmetropization. three dimensional space that depends on two eyes to experience, an artist gains a new perspective on color, transparency, overlapping, focus, etc. that allows him/her new ways of working and hence 6492-55, Poster Session seeing: a new form of expression. Contrast sensitivity measurements for quantifying defectivity and grading of image sensors 6492-57, Poster Session V. Sukumar, Micron Technology, Inc. A novel Bayer-like WRGB color filter array for CMOS Authors :*Vinesh Sukumar, *Doug Warner, *Patrick Doherty, **Herbert image sensors Hess, **Ken Noren, *** Steve Krone H. Honda, Y. Iida, G. Itoh, Y. Egawa, Toshiba Corp. (Japan) Affliations: * Micron Technology, Inc., 8000 S. Federal Way, Boise, Idaho 83707-0006 We have developed a CMOS image sensor with novel color filter array **Microelectronics Research and Communications Institute, University where one of the green pixels of the Bayer pattern was replaced with of Idaho, Moscow, Idaho, U.S.A. a white pixel. A transparent layer has been fabricated on the white pixel instead of a color filter to realize over 95% transmission for ***Department of Mathematics, University of Idaho, Moscow, Idaho, visible light whose wavelengths are 400-700 nm. Pixel pitch of the U.S.A device was 3.3 um and number of pixels was 2 million (1600H x Over the years, their has been an increased need to come up with an 1200V). objective way to measure perceived image quality in a camera The novel Bayer-like WRGB CFA realized higher SNR of interpolated imaging system. This is limited not only by the physical parameters of R/G/B values in low illumination (3lux) by 6dB/1dB/6dB, respectively, the image forming system, like resolution and contrast, but also on compared with that of the Bayer pattern, with the low-noise pre-digital the impression of the image formed by the eye of the observer which signal process. Furthermore, there was no degradation of either has is own limitations. This is gaining significant importance for resolution or color representation for the interpolated image. engineers working in CMOS Imaging industry. Design for test/pixel defect correction schemes becomes easier to implement as the The low-noise pre-digital signal process converted WRGB values into limitations for perceived quality of an image are better understood. the RGB values by referring to the R/G/B values surrounding the W pixel to separate the W value into new R/G/B values. Afterwards, The scope of this paper is to introduce a unique defect detection interpolation using both the R/G/B raw values and the R/G/B values model at a system level which is based of weighted combination of was done at every pixel. the physical parameters of the image forming system and psychophysical parameters of the human visual system. Emphasis will The SNR of G increased because of higher sensitivity of W than G, be placed on the experimental study of detection threshold of a and the SNR of R/B increased because of doubling of the effective R/ human visual processing system based of contrast sensitivity models. B pixel number.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 83 Conference 6492: Human Vision and Electronic Imaging XII

6492-59, Poster Session temporal axis, different ways of varying temporal frequency of the stimulus exists. It can be adjusted by flickering the stimulus (spatio- Improving video captioning for deaf and hearing- temporal CSF) or by moving it with a velocity which depends on its impaired people based on eye movement and attention spatial frequency (spatio-velocity CSF). However, supra-threshold overload sensitivity of the HVS to moving stimuli is not well-known. In this paper, a new supra-threshold experiment using moving stimuli is C. Chapdelaine, M. Beaulieu, L. Gagnon, CRIM (Canada) described. It shows that this supra-threshold spatio-velocity CSF Deaf and hearing-impaired people capture information in video have globally the same band-pass shape as Kelly’s threshold spatio- through visual content and captions. Those activities require different velocity CSF for low reference contrasts and tend to a low-pass visual attention strategies and up to now, little is known on how shape when increasing reference contrast. This behaviour appears as caption readers balance these two visual attention demands. well along spatial frequencies as along temporal ones. These results Understanding these strategies could suggest more efficient ways of would enable the enhancement of existing CSF models which are producing captions. Eye tracking and attention overload detections used in different applications such as video psychovisual coding and were used to study these strategies. Eye tracking was monitored development of objective video quality metrics. using a pupil-center-corneal-reflection apparatus. Afterward, gaze fixation was analyzed for each region of interest. Those regions included caption zone, high motion areas and faces location. This 6492-63, Poster Session data was also used to identify the scanpaths. Attention overload was Third- and first-party ground truth collection for auto key monitored through a tactile secondary task to assess the impact of frame extraction from consumer video clips caption rate on viewing scanpaths. The collected data was used to establish specifications for caption adaptation strategies based on the K. Costello, J. Luo, Eastman Kodak Co. location of visual action and presence of character faces. The strategies were implemented in a computer-assisted captioning Extracting key frames from video is of great interest in many software which uses a face detector and a motion detection algorithm applications, such as video summary, video organization, video based on the Lukas-Kanade optical flow algorithm. The different compression, and prints from video. KF extraction is not a new scanpaths obtained among the subjects provided us with alternatives problem. However, current literature has been focused mainly on for conflicting caption positioning. This implementation is now sports or news video. In the consumer video space, the biggest undergoing a user evaluation with hearing impaired participants to challenges for key frame selection from consumer video are its validate the efficiency of our approach. unconstrained content and lack of any pre-imposed structure. In this study, we conduct ground truth collection of key frames from video clips taken by digital cameras (as opposed to camcorders) using both 6492-60, Poster Session third-party and first-party judges. The goals of this study are: (1) to create a reference database of video clips reasonably representative Comparison of methods for the simplification of mesh of the consumer video space, (2) to identify associated key frames by models using quality indices and an observer study which automated algorithms can be compared and judged for effectiveness, and (3) to uncover the criteria used by first and third S. Silva, B. Sousa Santos, C. M. Ferreira, J. Madeira, Univ. de party human judges so they may influence algorithm design. Aveiro (Portugal) Meshes are currently used to model objects in many application domains (e.g., CAD/CAM, and physically-based modeling). However, 6492-64, Poster Session if they have a large number of triangles, rendering times may not be Quantifying the use of structure in cognitive tasks adequate to allow interactive visualization. A possible solution consists in showing a simplified version while the user chooses the D. M. Rouse, S. S. Hemami, Cornell Univ. viewpoint and, then, a fully detailed version of the model to enable its Modern algorithms that process images to be viewed by humans analysis. analyze the images strictly as signals, where processing is typically While several simplification methods have been developed and limited to the pixel and frequency domains. The continuum of visual reported in the literature, only a few studies compare them concerning processing by the human visual system (HVS) from signal analysis to the perceived quality of the obtained meshes. Even though observer cognition indicates that the signal-processing based model of the studies provide a ground truth, they are costly to perform. Thus, it is HVS could be extended to include some higher-level, structural of paramount importance to explore how available quality indices processing. An experiment was conducted to study the relative relate to those results. importance of higher-level, structural representations and lower-level, A comparison among mesh simplification methods is presented, signal-based representations of natural images in a cognitive task. which was performed through quality indices and a controlled Participants viewed sequences of images from either structural or experiment involving 65 observers and a set of 5 mesh models. Three signal-based representations, where subsequent images in the simplification methods were compared at two simplification levels: sequence reveal additional detail or visual information from the source 50% and 20% of the original number of faces. Obtained results were image. When the content was recognizable, participants were compared and show that some quality indices (e.g., geometric instructed to provide a description of that image in the sequence. The deviation) can be used as reasonable estimators of the observers’ descriptions were subjectively evaluated to identify a participant’s performance in specific circumstances. recognition threshold for a particular image representation. The results from this experiment suggest that signal-based representations possess meaning to human observers when the proportion of high 6492-61, Poster Session frequency content, which conveys shape information, exceeds a seemingly fixed proportion. Additional comparisons among the Influence of motion on contrast perception: supra- representations chosen for this experiment provide insight toward threshold spatio-velocity measurements quantifying their significance in cognition and developing a rudimentary measure of visual entropy. S. Tourancheau, P. Le Callet, D. Barba, Univ. de Nantes (France) Contrast sensitivity functions (CSF) describe the sensitivity of the human visual system (HVS) to the contrast of a stimulus under different viewing conditions and for different spatial and temporal characteristics of this stimulus. Typically, visibility threshold detection measurements and supra-threshold perception measurements are used in order to determine CSF. To measure the sensitivity along

84 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492: Human Vision and Electronic Imaging XII

6492-65, Poster Session the features to be matched. Once an object’s position in the image has been determined, its 2D pose is found by approximating the Quality metric for H.264/AVC scalable video coding with geometrical transformation most consistently mapping the locations full scalability of prototype’s subwindows to their matching locations in the input image. C. Kim, D. Suh, T. Bae, Y. Ro, Information and Communications Univ. (South Korea) Scalable Video Coding (SVC) is one of the promising techniques to 6492-26, Session 6 ensure quality of service(QoS) in diverse multimedia environments. Aperture and object mode appearances in images SVC compress a raw video into multi bitstream composed of a base and an enhancement bitstream. Therefore SVC supports multi J. J. McCann, McCann Imaging scalability such as spatial, temporal and SNR, and extracts a new Vision scientists have segmented appearances into aperture and different one from original coded bitstream without re-encoding to object modes, based on observations that scene stimuli appear adapt video in user environments. Existing studies of the subjective different a black -no light- surround. This is a 19th century assumption quality in SVC are focused to mainly temporal and/or SNR scalability. that the stimulus determines the mode, and cognitive feedback If we control spatial resolution in addition to the SNR and frame rate in determines the appearance. Since the 1960’s there have been SVC, it would be easier to manage QoS of streaming video and to innumerable experiments on spatial vision following the work of Hubel meet the wide range of bitrate budget. and Wiesel, Campbell, Gibson, Land and Zeki. The modern view of In this paper, we present the effect of frame rate, SNR, spatial vision is that appearance is generated by spatial interactions, or scalability and motion characteristics by the subjective quality contrast. Here, new experiments provide a significant increment of assessment, and then a new video quality metric with full scalability is new data on the effects of contrast, overall luminance and constancy proposed. Experimental results show that this quality metric has high over a wider range of luminances than previously studied. (7 correlation coefficient with subjective preference. Because this metric observers, 3,150 matches, 6log-unit range, 3log-unit targets, 3D is able to measure quality according to the variation of scalability, it targets). Matches were not consistent with the predictions of plays a role in the extraction point determining the quality of the svc constancy implied by discounting the illuminant. The observers’ based video as well as will be a useful video quality metric in scalable matches fit the simple two-step physical description: Maxima are video environments. dependent on luminance and darker areas are dependent on spatial contrast. Spatial interactions provide a more explicit description of appearance using the above observations. The need to rely on 6492-66, Poster Session unspecified cognitive processes of aperture mode, verses object mode, is no longer necessary. Simple rules of maxima and spatial Temporal relation between bottom-up versus top-down interactions account for all matches. strategies for gaze prediction S. Krishna, P. Kuchi, V. Balasubramanian, J. A. Black, Jr., N. Chatapuram-Krishnan, S. Panchanathan, Arizona State Univ. 6492-27, Session 6 Much research has been focused on the study of bottom-up, feature- Visibility improvement based on gray matching based visual perception, as a means to generate salience maps, and experiment between dark and ambient condition in predict the distribution of fixations within images. However, it is mobile display plausible that the eventual perception of distinct objects within a 3D scene (and the subsequent top-down effects) would also have a I. Kim, H. Ok, D. Park, SAMSUNG Advanced Institute of significant effect on the distributions of fixations within that scene. Technology (South Korea) This research is aimed at testing a hypothesis that there exists a Mobile displays allow users to view contents of interest anywhere. switching from feature-based to object-based scanning of images, as However, an inevitable downside is that environmental conditions the viewer gains a higher-level understanding of the image content, greatly influence the quality of the rendered contents. Luminance and that this switching can be detected by changes in the pattern of under sunny daylight or shade measures between 20,000 and eye fixations within the image. An eye tracker is used to monitor the 100,000 lux. Under these conditions a user oftentimes squints or fixations of human participants over time, as they view images, in an casts a shadow over the display to obtain better rendering. These effort to answer questions pertaining to 1) the nature of fixations phenomena became motivations for our experiment to overcome during bottom-up and top-down scene scan scenarios 2) the ability of visibility deterioration in mobile displays. We focus on gray visibility to assessing whether the subject is perceiving the scene content based overcome visibility deterioration in mobile and target luminance is on low-level visual features or distinct objects, and 3) identification of 20000 lux (shadow condition) for experiment as table 1. the participant’s transition from a bottom-up feature-based perception to a top-down object-based perception. The purpose of this study is to examine gray matching between dark and ambient condition and to improve visibility using result of gray matching experiment in mobile display. First of all, for measuring 6492-67, Poster Session visibility on ambient condition, the patch count experiment is conducted by investigating that how many patches can be seen at Eigen local color histograms for object recognition and original images under the ambient light. Next, the gray matching orientation estimation experiment is conducted by comparing gray patches between dark and ambient condition using method of adjustment. To confirm the D. Muselet, Univ. des Sciences et Technologies de Lille (France); visibility improvement through the result of gray matching experiment, B. V. Funt, Simon Fraser Univ. (Canada); L. Macaire, Univ. des visibility is measured under the ambient light after simple Sciences et Technologies de Lille (France) implementation. Color has been shown to be an important clue for object recognition and image indexing. We present a new algorithm for color-based recognition of objects in cluttered scenes that also determines the 2D 6492-28, Session 6 pose of each object. As with so many other color-based object Multispectral color constancy: real image tests recognition algorithms, color histograms are also fundamental to our new approach; however, we use histograms obtained from M. Mosny, B. V. Funt, Simon Fraser Univ. (Canada) overlapping subwindows rather than the entire image. An object from Experiments using real images are conducted on a variety of color a database of prototypes is identified and located in an input image constancy algorithms (Chromagenic, Greyworld, Max RGB, and an whenever there are many good histogram matches between the extension to Maloney-Wandell called Subspace Testing) in order to respective subwindow histograms of the input image and the image determine whether or not extending the number of channels from 3 to prototype from the database. In essence, local color histograms are 6 to 9 enhances the accuracy with which they estimate the scene electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 85 Conference 6492: Human Vision and Electronic Imaging XII illuminant color. To create the 6 and 9 channel images, filters where the visual response to each particular viewing condition. One of these placed over a standard 3-channel color camera. In an earlier study, mechanisms is chromatic adaptation, where colored objects get the Mosny and Funt [CIC°Ø14 Nov. 2006] examined the effect of sensor same appearance under different light sources, i.e. a white paper dimensionality on the accuracy of illumination estimation for color appears white under tungsten, fluorescent and day light. In the constancy. The results reported indicated that increasing the number process of judging printed color images, paper whiteness will be part of sensor channels provided surprisingly little benefit to the of the chromatic adaptation. This implies that variations in paper illumination estimation algorithms that were tested. Those results were whiteness would be discounted by the human visual system. In this based on synthetic image data. We report on the performance of the study, a set of images were printed on pilot papers with similar same algorithms on data from real images. Although minor properties except for the whiteness level and evaluated by visual improvements are found with 6 channels, the results confirm that the assessment studies. The results show an improved color rendering extra channels do not help. quality with increased CIE whiteness value to a certain level. Further increase in paper whiteness does not contribute to an improved color reproduction quality. 6492-29, Session 6 Color balancing based upon gamut and temporal 6492-33, Session 7 correlations Higher-order image representations for hyperresolution S. Kim, H. Lee, B. Kang, S. Lee, D. Park, C. Kim, SAMSUNG image synthesis and capture Advanced Institute of Technology (South Korea) B. Watson, North Carolina State Univ. Image acquisition devices inherently do not have color constancy mechanism like human visual system. Machine color constancy Real time imaging applications such as interactive rendering and problem can be circumvented using a white balancing technique video conferencing face particularly challenging bandwidth problems, based upon accurate illumination estimation. Unfortunately, previous especially as we attempt to improve resolution to perceptual limits. study can give satisfactory results for both accuracy and stability Compression has been an amazing enabler of video streaming and under various conditions. To overcome these problems, we suggest a storage, but in interactive settings, it can introduce application-killing new method: spatial and temporal illumination estimation. This latencies. Rather than synthesizing or capturing a verbose method, an evolution of the Retinex and Color by Correlation method, representation and then immediately converting it into its succinct predicts on initial illuminant point, and estimates scene-illumination form, we should generate the concise representation directly. Several between the point and sub-gamuts derived by from luminance levels. projects have begun taking the first steps down this research path. The method proposed can raise estimation probability by not only Bala et al. combine sparse color samples with complete silhouette detecting motion of scene reflectance but also by finding valid scenes and shadow edge information to synthesize high-quality interactive using different information from sequential scenes. This proposed imagery. Tumblin et al. propose gradients in image capture. Our own method outperforms recently developed algorithms. renderer uses spatio-temporally adaptive gradient sampling, reconstructing just-in-time imagery from a frameless sample buffer. Resulting RMS qualities are equivalent to traditionally synthesized 6492-30, Session 6 imagery with 10 times more samples. We are currently attempting to reconstruct imagery using color samples and incomplete edge Visibility of hue, saturation, and luminance information, and plan to experiment with other higher order sampling nonuniformities in LEDs primitives including corners, curves and Laplacians. We believe such P. J. H. Seuntiens, I. Vogels, Philips Research Labs. (Netherlands) higher-order representations will be a crucial enabler of interactive, hyper-resolution image capture, synthesis and display. The market for LED applications continues to grow rapidly with improved technology leading to new application areas. It is well known that LEDs have problems with color consistency and color 6492-34, Session 7 stability over time. A perception experiment was conducted in order to determine guidelines for the allowed color and luminance variations Semantic photosynthesis between LEDs. The experiment focused on the visibility threshold of M. A. Johnson, Univ. of Cambridge (United Kingdom) hue, saturation, and luminance non-uniformities in homogeneous light distributions using LEDs. A tuning test was used to determine the Composite images are synthesized from existing photographs by visibility thresholds of the observers for a hue, saturation, and artists who make concept art, e.g., storyboards for movies or luminance shift in the primaries of the LEDs for several colors. The architectural planning. Current techniques allow an artist to fabricate results of the experiment show that the visibility threshold for a hue such an image by digitally splicing parts of stock photographs. While non-uniformity is 0.007 ?u’v’ for red, green, and blue and 0.004 ?u’v’ these images serve mainly to quickly convey how a scene should for white. The threshold for saturation is 0.015 ?u’v’ for red, green, look, their production is laborious. This talk will explore a new area of and blue and 0.007 ?u’v’ for white. Luminance non-uniformities in research in intelligent image stitching which utilises the state of the art primaries are not critical for red, green, and blue colors, but are very in computer vision to enable users to automatically compose new critical for white (luminance degradation of a primary gives a color images from a very large database of stock photographs using only a shift in white). The threshold for a luminance non-uniformity in one of few, intuitively supplied semantic constraints. The system on display the primaries in the case of a white color is 0.005 ?u’v’. allows the user to paint and label a canvas to provide these constraints, or even to simply write a caption which describes their desired result. It then automatically interprets these constraints to 6492-31, Session 6 provide a range of images which satisfy them from which the user can choose. The talk will focus on the current system and its development Paper whiteness and its effect on the reproduction of and then examine areas of improvement, ending in a discussion about colors some of the the possibilities of computer-vision informed image manipulation tools. O. Norberg, Mid Sweden Univ. (Sweden) The whiteness level of a printing paper is considered as an important quality measure. High paper whiteness improves the contrast to 6492-35, Session 7 printed areas providing a more distinct appearance of printed text and colors as well as increases the number of reproducible colors. Revealing pentimenti: the hidden history in a painting Although paper whiteness is considered as an important paper quality A. Gooch, Univ. of Victoria (Canada) criterion, its influence on printed color images is complicated. The dynamic mechanisms of the human visual system strive to optimize Investigation of paintings using X-ray imagery allows conservators to explore the pentimenti, or hidden aspects, beneath the first surface of

86 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492: Human Vision and Electronic Imaging XII a painting. X-ray image analysis of paintings can be difficult to contrast LCD displays, not to mention next generation HDR displays, interpret because an X-ray image represents the relative density of all that can easily create bright highlights and the impression of self- paint layers superimposed. This talk will present current research in luminous colors. We argue that high dynamic range representation identifying visually significant differences between X-ray images and can encode images regardless of the technology used to create and visible light images, i.e. photographs, of historically significant works display them, with the accuracy that is only constrained by the of art. limitations of the human eye. To facilitate the research on high dynamic range imaging, we have created a software package (http:// pfstools.sourceforge.net/), capable of handling HDR data on all stages 6492-36, Session 8 of image and video processing. The software package includes solutions for high quality image acquisition from multiple exposures, a Self-calibrating wide color gamut high-dynamic-range range of tone mapping algorithms and a visual difference predictor for display HDR images. We demonstrate how the software can be used for H. Seetzen, Brightside Technologies, Inc. (Canada) processing single images as well as video sequences. No abstract available 6492-41, Session 8 6492-38, Session 8 Veiling glare: the dynamic range limit of HDR images Tone mapping for high-dynamic range displays J. J. McCann, McCann Imaging; A. Rizzi, Univ. degli Studi di Milano (Italy) L. Meylan, S. E. Süsstrunk, École Polytechnique Fédérale de Lausanne (Switzerland); S. J. Daly, Sharp Labs. of America, Inc. Multiple exposure techniques for capturing High Dynamic Range (HDR) images, are limited by the interplay of the scene and camera Recently, new HDR monitors have been developed, which allow glare. We used a single test target (40 luminance patches DR displaying a much larger dynamic range than standard CRTs or LCDs. =18,619:1) on a lightbox (2049 cd/m2)in white & black surrounds. We As these displays will become more widely available, we will be made CCD&film photographs, with multiple exposure times to confronted with the need for new tone-mapping algorithms adapted measured the accuracy of multiple-exposure-HDR. We found scene to these new conditions of vision. We address the problem of re- dependent errors in luminance estimates (max 300% in black and rendering legacy images that have already been tone-mapped to max 10,000% in white surrounds). Camera digits from multiple standard displays. We propose to use a tone-scale function that exposures cannot accurately measuring HDR scene flux because of allocates more display range to the specular highlights than what was veiling glare that is scene-, exposure-, lens- and camera-dependent. allocated in the image tone-mapped to standard displays. How well can humans see hypothetical true HDR images? In fact, the The shape of the proposed tone-scale depends on a segmentation of stray light problems of the human eye are more limiting that the veiling the input image into its diffuse and specular components. In this glare found in cameras. Stiehl et. al. showed that a lightness scale on article, we present a method to perform this segmentation a white background with greater than 4log units dynamic range ends automatically. Our method detects specular highlights by using two up with a 1.5log range on the retina. Magnitude estimation results low-pass filters of different sizes combined with morphological show that ocular-media glare controls the retinal image. HDR images operators. The results show that our method successfully detects are superior to conventional images. The multiple exposure technique small and middle size specular highlights. The location of specular does significantly improve digital quantization. Humans see details in highlights defines a mask used for the segmentation and thus for the shadows because of discriminable spatial information at low construction of the tone-scale function. We also present an alternative luminances, and not from accurate luminance reproduction. way of applying the tone-scale function, which uses the mask to perform a spatially variant treatment on the image. 6492-42, Session 9 6492-39, Session 8 Hidden Markov model-based face recognition using selective attention Tracking gaze positions during evaluating high-dynamic A. A. Salah, Bogaziçi Univ. (Turkey); M. Bicego, Univ. degli Studi di range images Sassari (Italy); L. Akarun, Bogaziçi Univ. (Turkey); E. Grosso, M. J. Liu, W. Cheng, National Chiao Tung Univ. (Taiwan) Tistarelli, Univ. degli Studi di Sassari (Italy) Where are people looking at when evaluating high-dynamic range Sequential methods for face recognition rely on the analysis of local (HDR) images? What makes an HDR algorithm better than another? facial features in a sequential manner, typically with a raster scan. To answer these questions, we used an eye tracker to record the gaze However, the distribution of discriminative information is not unifom positions and durations during psychophysical experiments of over the facial surface; for instance the eyes and the mouth are more evaluating HDR image pairs. Via this study, we want to learn what informative than the cheek. We propose an extension to the regions on an image are important for evaluating HDR images and sequential approach, where we take into account local feature how to improve on the current HDR algorithms. saliency, and replace the raster scan with a guided scan that mimicks the scanpath of the human eye. The selective attention mechanism that guides the human eye operates by coarsely detecting salient 6492-40, Session 8 locations, and directing more resources (the fovea) at interesting or informative parts. We simulate this idea by employing a High-dynamic range imaging pipeline: perception- computationally cheap saliency scheme, based on Gabor wavelet motivated representation of visual content filters. Hidden Markov models are used for classification, and the R. Mantiuk, G. Krawczyk, Max-Planck-Institut für Informatik observations, i.e. features obtained with the simulation of the (Germany); R. Mantiuk, Szczecin Univ. of Technology (Poland); H. scanpath, are modeled with Gaussian distributions at each state of the model. We show that by visiting important locations first, our Seidel, Max-Planck-Institut für Informatik (Germany) method is able to reach high accuracy with much shorter feature The advances in high dynamic range (HDR) imaging, especially in the sequences. We compare several features in observation sequences, display and camera technology, have a significant impact on the among which DCT coefficients result in the highest accuracy. existing imaging systems. The assumptions of the traditional low- dynamic range imaging, designed for paper print as a major output medium, are ill suited for the range of visual material that is shown on modern displays. For example, the common assumption that the brightest color in an image is white can be hardly justified for high

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 87 Conference 6492: Human Vision and Electronic Imaging XII

6492-43, Session 9 essentially based on segmentation, have been proposed. It has been shown that under the assumption of unspecified image content, The role of eye movement signals in dorsal and ventral segmentation becomes unstable. According to recent psychological processing theories, certain image regions attract the attention of human observers more than others and, generally, the image main meaning is J. A. Black, Jr., N. Chatapuram-Krishnan, S. Panchanathan, Arizona concentrated in those regions. Initially, regions attracting our attention State Univ. are perceived as a whole and hypotheses on their content are As human eyes scan an image, each fixation captures high resolution formulated; successively the components of those regions are visual information from a small region of that image. The resulting carefully analyzed and a more precise interpretation is reached. It is intermittent visual stream is sent along two visual pathways to the visual interesting to observe that an image decomposition process centers of the brain, along with eye movement information. The ventral performed according to these psychological visual attention theories (what) pathway is associated with object recognition, while the dorsal might present advantages with respect to a traditional segmentation (where) pathway is associated with spatial perception. This research approach. In this paper we propose the implementation of an employs two experiments to compare the relative importance of eye automatic procedure generating image decomposition according to a movement information within these two visual pathways. During tree structure based on the detection of visual attention regions. A Experiment 1 participants examine (1) outdoor scenery images, and (2) new clustering algorithm taking advantage of the Delaunay-Voronoi object images, while their fixation sequences are captured. These diagrams for achieving the decomposition target is proposed. fixation sequences are then used to derive a foveated sequence of According to a top-down approach, by applying recursively that images. In Experiment 2, this sequence of images is viewed by another algorithm starting from the whole image, a region tree is obtained set of participants, who simply stare at the center of a visual display. In (Attention Tree). Currently we are investigating the way for exploring doing so, participants in Experiment 2 experience the same sequence of the Attention Trees, both by using a Neural Tree in order to associate foveal stimuli as those in Experiment 1, but do not experience the semantic values to the Attention Tree nodes and by exploiting an corresponding eye movement signals. The subsequent ability of the Ontology Net for checking the decomposition of complex objects. Experiment 2 participants to (1) describe the outdoor scenes, and (2) to Several experiments of Attention Tree construction have been recognize the objects, provide information about the relative importance performed and partial results will be shown. of eye movement information in the two visual pathways. 6492-47, Session 9 6492-44, Session 9 Motion integration in visual attention models for Variable resolution images and their effects on eye predicting simple dynamic scenes movements during free viewing A. Bur, H. Hügli, Univ. de Neuchâtel (Switzerland) M. Nystrom, Lunds Univ. Visual attention is the ability of a visual system, be it biological or Earlier studies have shown that people free-viewing images tend to gaze artificial, to rapidly detect potentially relevant parts of a scene, on at regions with a high local density of bottom up features such as which higher level tasks such as object recognition, can focus. Given contrast and edge density. In particular, this tendency seems to be more the high relevance of temporal aspects in the visual attention emphasized during the first few fixations after image onset. In this paper, mechanisms, dynamic information as well as static information must we further investigate how gaze locations are chosen by introducing be considered in the computer vision model for low-level attention. A varying image resolution, and see how it affects eye-movement behavior few works have extended the classical static model of visual attention during free viewing. Results show that gaze density overall is shifted for motion. However, a comparison of the performances of models toward regions presented in high resolution over those degraded in integrating motion in different manners is not available. In this paper, resolution. However, certain image regions seem to attract early fixations we present a comparative study of various visual attention models regardless of display resolution. combining static and dynamic features. The models are considered and compared by measuring their respective performance with respect to the eye movement patterns of human subjects. Simple 6492-45, Session 9 synthetic video sequences that combine static and moving objects are used for analyzing the model suitability. Qualitative and Hierarchy visual attention map quantitative results provide a ranking of the different models. K. Yang, P. K. Das, C. C. Guest, Univ. of California/San Diego A new framework for establishing human attention maps has been 6492-48, Session 10 proposed. It allows more flexibility for features in pooling stage. The basic idea is that human eyes have differing sensitivities for each feature Motion of specularities on low-relief surfaces: frequency with different video content. Therefore, the weights of each feature domain analysis should be varied with different video shooting scenarios for feature Y. Farasat, M. S. Langer, McGill Univ. (Canada) integration. Hence, several indicators of different video content are extracted and utilized as important cues to determine the weight for Typical studies of the visual motion of specularities have been each attention feature. Based on this concept, a human visual attention concerned with how to discriminate the motion of specularities from map has been implemented and the experimental results show very the motion of surface markings, and how to estimate the underlying promising results. In future work, a more thorough analysis of indicators surface shape. Here we take a different approach and ask whether a of video shooting scenarios will be developed. Also, eye tracking data field of specularities gives rise to motion parallax that is similiar to that will be used to determine the weights for each feature for different video of the underlying surface. The idea is that the caustics that are defined content. by specularities exist both in front of and behind the underlying surface and hence define a range of depths relative to the observer. Furthermore, this attention map will be applied to video quality We asked whether this range of depths leads to motion parallax. Our assessment research. experiments are based on image sequences generated using computer graphics and Phong shading. Using low relief undulating 6492-46, Session 9 surfaces and assuming a laterally moving observer, we compare the specular and diffuse components of the resulting image sequences. In Attention trees and semantic paths particular, we compare the image power spectra. We find that, as long as the undulations have a sufficiently large amplitude, the range of C. Giusti, G. G. Pieroni, Univ. degli Studi di Udine (Italy); L. Pieroni, speeds indicated in the power spectra of the diffuse and specular Univ. degli Studi di Roma/La Sapienza (Italy) components is similar. This suggests that specularities could provide In the last few decades several techniques for image content extraction, motion parallax information to a moving observer.

88 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6492: Human Vision and Electronic Imaging XII

6492-49, Session 10 designed to be an effective approximation of the noncomputable but universal Kolmogorov distance between two strings. We compare the Fully automatic perceptual modeling of near-regular effectiveness of different types of compression algorithms in textures predicting image distinguishability when they are used to compute G. Menegaz, A. Franceschetti, A. Mecocci, Univ. degli Studi di the NCD between a sample of images and their watermarked Siena (Italy) counterparts. Our experiment shows that, as predicted by Li’s theory, the NCD is largely Near regular textures feature a relatively high degree of regularity. They can be conveniently modeled by the combination of a suitable independent of the underlying compression algorithm. set of textons and a placement rule. The main issues in this respect However, in some cases the NCD fails as a predictor of image are (i) the selection of the minimum set of textons bringing the distinguishability, since it is designed to measure the more general variability of the basic patterns; (ii) the identification and positioning of notion of similarity. We propose and study a modified version of the the generating lattice; and (iii) the modelization of the texture NCD to model the latter, which requires that not only the change be variability. Capturing the local variations in the texton structure as well small but also in some sense random with respect to the original as the global deviation from periodicity of the underlying lattice is image. essential for reproducing the “naturalness” of the texture through a synthesis algorithm. In this contribution, we provide a fully automatic solution to both the 6492-52, Session 10 analysis and the synthesis issues leading to the generation of texture Cortical network and dynamics of figure-ground samples that are perceptually indistinguishable from the original ones. The definition of an ad-hoc periodicity index allows predicting the organization suitability of the model for a given texture. The model is validated L. T. Likova, C. W. Tyler, The Smith-Kettlewell Eye Research through a psycho-visual experiments proving the subjective Institute equivalence of the original and synthetic textures while allowing to determine the minimum number of textons to be used to meet such a Introduction: Segmentation and figure-ground organization are basic requirement for a given texture class. This is of prime importance in mechanisms involved in perception of visual scene. Use of model-based coding applications, as is the one we foresee, as it asynchronous figure-ground structure has revealed for the first time in allows to minimize the amount of information to be transmitted to the the human brain by means of fMRI a mechanism of figure-ground receiver. organization through top-down suppression of the stimulated background representation in V1/V2, with the activation pattern implying hMT+ as the source of suppression(Likova&Tyler,2005). 6492-50, Session 10 Purpose and Methods: To dissect the processing stream and understand the neural dynamics of the figure/ground network, we Machine perception using the five human defined forms employed a novel paradigm that creates distinct and prolonged plus infrared percept of spatial structures by instantaneous refreshes in random dot P. J. DeRego, Honeywell, Inc. fields. Four different classes of spatial configurations were generated (i) in the figure region only, (ii) in the background only, (iii) in the whole A machine vision system observing how an object behaves within its stimulus field, (iv) figure-ground defined by brief temporal asynchrony. environment can extract valuable clues about the object’s identity and FMRI responses were obtained throughout the brain in a GE Signa 3T intentions. We are using this approach in developing a surveillance scanner at 1.5 sec TR. The dynamics of the neural activation were system having the goal of extracting foreign objects from a baseline determined by deconvolution of the BOLD signal with a standard environment, classifying object identities, and perceiving their estimate of the blood-specific hemodynamic response function (HRF). behavior as a threat or non-threat. The approach follows biological Results: A distributed network extending beyond the occipital lobe perception systems which collect data on both the environment and into the parietal and frontal cortex was activated by the figure-ground observed objects within the environment. Observations of the object, configuration. Also, comparison with three control conditions revealed its behavior, and the environment must conform to strict constraints of remarkable differences in one or more response properties (latency, a perception model. Choosing the appropriate perception model and amplitude, sign, and width) within the individual areas. The verifying correct behavior in the environment is required for instantaneous stimulus paradigm generated a wide variety of HRF and recognition. The weakness of this type of perception mechanism is its corresponding neural responses estimates throughout the network. In vulnerability to optical illusions as well as deliberate deception. many cases the estimated neural response parameters varied by a Biological systems attempt to reduce this vulnerability by using factor of 2 to 4 within a single area. Such expressively different multiple mechanisms for extracting and tracking objects. The responses evoked by differential stimulation of the identical cortical investigation in progress uses extraction filters based on motion, regions assure that the differences could be securely attributed to the color, texture, luminance, binocular disparity, and thermal imaging. neural dynamics, not to spatial variations in the HRF. Extracted objects in the environment are evaluated across these six visual defined forms for the likelihood of being foreign to the Conclusions: The activation pattern for figure-ground implies a widely environment and a potential threat. distributed recurrent architecture with functional feedback loops, distinct from the control conditions. Even when activations were partially overlapping, the integrated analysis of the hemodynamic 6492-51, Session 10 response properties, as well as of the neural response reconstruction, enabled us to distinguish and more quantitative characterize the The normalized compression distance and image functional specificity or generality of the involved cortical areas. distinguishability N. Tran, Santa Clara Univ. 6492-53, Session 10 We use an information-theoretic distortion measure called the Normalized Compression Distance (NCD), first proposed by M. Li et Comparing realness between real objects and images at al., to determine whether two rectangular gray-scale images are various resolutions visually K. Masaoka, M. Emoto, M. Sugawara, Y. Nojiri, NHK Science & distinguishable to a human observer. Image distinguishability is a Technical Research Labs. (Japan) fundamental constraint on operations carried out by all players in an image watermarking system. Image resolution is one of the important factors for visual realness. We performed subjective assessments to examine the realness of images The NCD between two binary strings is defined in terms of at six different resolutions, ranging from 19.5 cpd (cycles per degree) compressed sizes of the two strings and of their concatenation; it is to 156 cpd. A paired-comparison procedure was used to quantify the

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 89 Conference 6492: Human Vision and Electronic Imaging XII

realness of six images plus the real object. Three real objects were used. Both real objects and images were viewed through a synopter, which removed horizontal disparity and presented the same image to both eyes. Sixty-five observers were asked to choose the viewed image which was closer to the real object and appeared to be there naturally for each pair of stimuli selected from the group of six images and the real object. It was undisclosed to the observers that real objects were included in the stimuli. The paired comparison data were analyzed using the Bradley-Terry model. We found that realness of an image increased as the image resolution increased up to about 40-50 cpd, which corresponded to the discrimination threshold (based on the observers’ visual acuity or minimum separable acuity), and remained constant above this threshold, including two of the three real objects.

6492-54, Session 10 Navigation based on a sensorimotor representation: a virtual reality study C. Zetzsche, K. Schill, G. Christopher, Univ. Bremen (Germany) We investigate the hypothesis that the main representation which underlies human navigation does not resemble an image-like map but is of an inherently sensorimotor nature (i.e. results from a combination of sensory features and motor actions). This hypothesis is already supported by results in psychology and neurobiology, which indicate that the traditional strict separation of sensory and motor systems is no longer tenable. Here we present further support from a behavioural study in a VR environment. We tested the influence of physically inconsistent configurations which can not be represented in a map- like fashion. Since performance is not influenced by these inconsistencies, a map-like representation can not be the major basis of human navigation. We also used a simulated agent that operates in a VR environment to demonstrate that navigation can be based on a sensorimotor representation which is not affected by the restrictions of map-like representations. We designed a hybrid architecture which integrates a bottom-up processing of sensorimotor features with a top-down reasoning that is based on the principle of maximum information gain. As a result, the agent is able to use the minimum number of exploratory actions to orient itself within the VR environment. (Supported by DFG: SFB/TR-8)

90 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

Tuesday-Thursday 30 January-1 February 2007 Part of Proceedings of SPIE Vol. 6493 Color Imaging XII: Processing, Hardcopy, and Applications 6493-01, Session 1 algorithm consists of two components: Cost Optimized Segmentation (COS) and Connected Component Classification (CCC). The COS On the behavior of spatial models of color algorithm is a blockwise segmentation algorithm formulated in a global A. Rizzi, Univ. degli Studi di Milano (Italy); J. J. McCann, McCann cost optimization framework, while CCC is based on feature vector classification of connected components. In the experimental results, we Imaging show that the new algorithm achieves the same accuracy of text There is a growing family of algorithms that treat/modify/enhance detection but with lower false detection of non-text features, as color information in its visual context, also known as spatial color compared to state-of-the-art commercial MRC products. This results in methods. Judging their performance is a challenging task and still an high quality MRC encoded documents with fewer non-text artifacts, open problem. Two main variables affect the final result of these and lower bit rate. algorithms: their parameters and the visual characteristics of the input image. With the term visual characteristics we refer not only to their imaging parameters, but also to their spatial distribution in the image. 6493-04, Session 2 This paper does not deal with parameter tuning, but aims at discussing the visual configurations in which a spatial color method A new approach to JBIG2 binary image compression show interesting or critical behavior. A survey of the more significant M. Figuera, Purdue Univ.; J. Yi, SAMSUNG Electronics Co., Ltd. visual configurations will be presented and discussed. These (South Korea); C. A. Bouman, Purdue Univ. configurations include phenomena, such as color constancy, color contrast, and color assimilation. The discussion will present strength The JBIG2 binary image encoder dramatically improves compression and weakness of different algorithms, hopefully allowing a deeper ratios over previous encoders. The effectiveness of JBIG2 is largely due understanding of their behavior and stimulating discussions about to its use of pattern matching techniques and symbol dictionaries for finding a common judging ground. the representation of text. While dictionary design is critical to achieving high compression ratios, little research has been done in the optimization of dictionaries across stripes and pages. 6493-02, Session 1 In this paper we propose a novel dynamic dictionary design that substantially improves JBIG2 compression ratios, particularly for multi- Effect of time spacing on the percieved color page documents. This dynamic dictionary updating scheme uses S. Roch, J. Y. Hardeberg, Gjøvik Univ. College (Norway) caching algorithms to more efficiently manage the symbol dictionary memory. Results show that the new dynamic symbol caching technique At the core of the concept of soft proofing (simulation on a computer outperforms the best previous dictionary construction schemes by screen of the appearance of a print) is the problem of identifying, for between 13% and 46% for lossy compression when encoding multi- any printed color, the most similar color a monitor can display. This page documents. In addition, we propose a fast and low-complexity problem is made difficult by such factors as varying viewing pattern matching algorithm that is robust to substitution errors and conditions, color gamut limitations, or time spacing. This paper achieves high compression ratios. attempts to shed more light on the difference between simultaneous and time-spaced color matching, in order to contribute to improving the accuracy of soft proofs. A color matching experiment setup has 6493-05, Session 2 been established in which observers were asked to match, by adjusting its RGB values, a color patch displayed on a LCD monitor to Removal of artifacts from JPEG compressed document another color patch printed on paper. In the first part of the images experiment the two colors were viewed simultaneously. In the second part, the observers were asked to produce the match according to a B. Oztan, Univ. of Rochester; A. Malik, Z. Z. Fan, R. Eschbach, previously memorized color. According to the obtained results, the Xerox Corp. color appearance attributes lightness and chroma were the most We present a segmentation-based post-processing method to remove difficult components for the observers to remember, generating huge compression artifacts from JPEG compressed document images. JPEG differences with the simultaneous match, whereas hue was the compressed images typically exhibit ringing and blocking artifacts, component which varied the least. This indicates that for soft which can be objectionable to the viewer above certain compression proofing, getting the hues right is of primordial importance. levels. The ringing is more dominant around textual regions while the blocking is more visible in natural images. Despite extensive research, reducing these artifacts in an effective manner still remains challenging. 6493-03, Session 2 Document images are often segmented for various reasons. As a result, Segmentation for MRC compression the segmentation information in many instances is available without requiring additional computation. We have developed a low E. Haneda, Purdue Univ.; J. Yi, SAMSUNG Electronics Co., Ltd. computational cost method to reduce ringing and blocking artifacts for (South Korea); C. A. Bouman, Purdue Univ. segmented document images. The method assumes the textual parts and pictorial regions in the document have been separated from each Mixed Raster Content (MRC) is a standard for efficient document other by an automatic segmentation technique. It performs simple compression which can dramatically improve the compression/quality image processing techniques to clean out ringing and blocking artifacts tradeoff as compared to traditional lossy image compression from these regions. algorithms. The key to MRC’s performance is the separation of the document into foreground and background layers, represented as a binary mask. Typically, the foreground layer contains text colors, the 6493-06, Session 2 background layer contains images and graphics, and the binary mask layer represents fine detail of text fonts. Digital images for eternity: color microfilm as archival The resulting quality and compression ratio of a MRC document medium encoder is highly dependent on the segmentation algorithm used to C. M. Normand, R. Gschwind, Univ. of Basel (Switzerland) compute the binary mask. In this paper, we propose a novel segmentation method based on the MRC standards (ITU-T~T.44). The In the archiving and museum communities, the long-term preservation

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 91 Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications of artworks has traditionally been assured by making duplicates of the smaller LCD images. Conversely, the contrast of projected high original. For photographic reproductions, digital imaging devices have frequency images were reduced for the same images on LCD with a now become standard, providing better process control and lower smaller size. The effect was more pronounced in the matching of costs than film photography. However, due to the very short life cycle projected image to the smaller images on the LCD display. Compared of digital data, losses are unavoidable without repetitive data to the mean luminance level of the LCD images, a reduction of the migrations to new file formats and storage media. We present here a mean luminance level of the adjusted images was observed for low solution for the long-term archiving of digital images on color frequency noise patterns. This decrease was more pronounced for microfilm (Ilfochrome Micrographic). This extremely stable and high- smaller images with lower contrast and high mean luminance level. resolution medium, combined with the use of a novel laser film recorder is particularly well suited for this task. Due to intrinsic limitations of the film (dynamic range, limited color gamut) colorimetric 6493-09, Session 3 reproductions of the originals are not achievable, and the microfilm must be first considered as an information carrier and not primarily as Modeling for hue shift effect of human visual system on an imaging medium. Color transformations taking into account the high-luminance display film characteristics and possible degradations of the medium due to T. Lee, M. Lee, K. Park, Y. Ha, Kyungpook National Univ. (South aging are investigated. An approach making use of readily available Korea) color management tools is presented which guarantees the recovery of the original colors after re-digitization. An extension of this project This paper proposes a color correction method based on modeling considering the direct recording of digital information as 3D bit-code the hue shift phenomenon of human visual system (HVS). Observers on the film is also introduced. tend to perceive same color stimuli, but of different intensity, as different hues, what is referred to as the hue shift effect. Although the effect can be explained with the Bezold-Brücke(B-B) effect, it is not 6493-07, Session 3 enough to apply the B-B model on high luminance displays because most displays have a broad-band spectrum distribution and results A study on color appearance for large size display under vary according to type of display. In the proposed method, the various illumination conditions quantities of hue shift between a high luminance display and a normal S. Y. Choi, R. M. Luo, Univ. of Leeds (United Kingdom); M. R. luminance display were first modeled by a color matching experiment Pointer, National Physical Lab. (United Kingdom) with color samples along the hue angle of the LCH color space. Based on the results, the hue shift was then modeled piecewise and This paper focuses on the change in colour appearance of a large size was finally applied to the inverse characterization of display to display arising from the various illumination conditions. Nine compensate the original input image. From evaluating the proposed experimental phases were conducted using surrounds including dark, method using the psychophysical experiment with some test images, indoor and outdoor conditions. These surrounds were divided into two we confirmed that the proposed modeling method is effective for categories. In the first category, the surround had no effect on the color correction on high luminance displays. displayed colours, but observers could still sense the different brightness levels of the surround. In the second category, surround introduced flare to the displayed colours together. Additionaly, the 6493-10, Session 3 effect due to different sizes of stimuli and surround size difference was investigated by controlling the viewing distance. For the first category, Optimizing color edge contrast in presence of noise and little visual lightness difference was shown between different amplitude quantization errors surrounds, unlike the expectation that the perceived lightness contrast F. Lebowsky, STMicroelectronics (France); Y. Huang, may increase as the surround becomes brighter. The lightness STMicroelectronics (Singapore) dependency of colourfulness, however, was found to change. For the second category, as the surround became brighter, the perceived Large-scale, direct view TV screens, in particular those based on dynamic range of visual lightness decreased but high chromatic liquid crystal technology, are beginning to use structures with more colours appeared more colourful. In the investigation of the change of than three subpixels to further reduce the visibility of defective stimulus size under different surround conditions, visual colour subpixels, increase spatial resolution, and perhaps even implement a appearance was not affected by the stimulus sizes of 2° and 0.6° in multi-primary display with up to six different primaries. This newly the dark surround. However, the difference was found in the very dark available “subpixel” resolution enables us to improve color edge colours with a dim surround. Finally, all of visual colour appearance contrast thereby allowing for shorter viewing distances through a data were used to test the performance of the colour appearance reduction of perceived blur. However, not only noise, but also model CIECAM02. Minor modification was accomplished to improve amplitude quantization can lead to undesirable, additional visual the colourfulness predictor, especially for the black background. artifacts along contours. Using our recently introduced method of contour phase synthesis in combination with non-linear, color channel processing, we propose a simple method that maximizes color edge 6493-08, Session 3 contrast while maintaining or suppressing visual artifacts as well as noise below the threshold of visual perception. To demonstrate the An investigation of the effect of image size on the color advantages of our method, we compare it with some classical appearance of softcopy reproductions using a contrast contrast enhancement techniques such as cubic spline interpolation matching technique and color transient improvement. M. Nezamabadi, R. S. Berns, E. D. Montag, Rochester Institute of Technology 6493-11, Session 4 Many art objects have a size much larger than their softcopy reproductions. In order to develop a multiscale model that accounts Correlating 2D NTSC gamutrRatio to its 3D gamut for the effect of image size on image appearance, a digital projector volume and LCD display were colorimetrically characterized and used in a P. Sun, Shih Hsin Univ. (Taiwan) contrast matching experiment. At three different sizes and three levels of contrast and luminance, a total of 63 images of noise patterns were The present study provides functions to correlate gamut size across rendered for both displays using three cosine log filters. Fourteen different color spaces including 2D planes - (x,y) and (u’,v’) and 3D observers adjusted mean luminance level and contrast of images on spaces - CIELAB, CIECAM02 JCh and QMh. All gamut size must be the projector screen to match the images displayed on the LCD. The converted to NTSC gamut ratio before using the functions. As viewing contrasts of the low frequency images on the screen were boosted conditions influence 3D gamut ratio significantly, predict 3D gamut while their mean luminance values were decreased relative to the ratio in high precision is not easy. However, their means, medians, standard deviations, and confidence intervals can be predicted

92 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications accurately. In terms of viewing parameters, we modeled the effects of contrast of image can be adjusted with this curve. The optimization of ambient level, display luminance level, contrast ratio and color gamma curve in a digital imaging device has not been temperature individually under IEC reference condition successfully. comprehensively studied yet, and the gamma curve is usually The resulted functions would be a good reference to derive more manually tuned by manufacturers, which is heavily based on their versatile functions to predict gamut ratio under complex viewing long-term experiences. In this paper, the gamma curve is optimized in conditions. In this study, we also found that (u’,v’) gamut ratio did not a systematic approach by taking into account the color reproduction well correlate to the ratios of any 3D color space, whereas (x,y) gamut accuracy, the propagated noise minimization and the image contrast ratio fit their data well. As all color science literatures taught us the preservation. While the first two have been discussed relatively more, (u’,v’) chromaticity diagram is more perceptual uniform than the (x,y) the contrast is usually not considered for this purpose in the literature. diagram, our results would be interesting for display industry to think In this paper the criterion of image contrast is defined first, followed over. by the proposal of a joint quality metric, which is applied onto the optimization of parameters for the nonlinearity curve. 6493-12, Session 4 6493-15, Session 4 Device calibration method for 1D light modulator Y. Bang, A. Baik, D. Park, SAMSUNG Advanced Institute of Reducing rainbow effect for filed sequential color LCDs Technology (South Korea) P. Sun, Shih Hsin Univ. (Taiwan) Due to subtle misalignment of optical components in the fabrication The paper presents two models to reduce the visibility of rainbow process, images projected by an optical light modulator have effect for field sequential color LCDs. The basic idea of the two undesirable line artifact along the direction of the optical scan. In this models were first changing the LED backlight state, introducing paper, we propose a novel methodology to calibrate the modulator crosstalk among the R/G/B LEDs to reduce inter-sub-field color and generate the compensate image for the misaligned optical variations and finally modifying the RGB signals to minimize its image modulator in order to eliminate line artifact. A camera system is color differences to its original state. The results of our simulations employed to construct Luminance Transfer Function (LTF) that suggested that the proposed methods would reduce the visibility of characterizes the optical modulator array. Spatial uniformity is rainbow effect to some extent. However, its performance was quite obtained by redefining the dynamic range and compensating the image-dependent. If no saturate color in an image, the rainbow effect characteristic curvature of the LTF for each optical modulator array could be removed successfully. However, if an image contains variety element. Simulation results show significant reduction in the visibility of colors, no gamut reduction will be made, and thus the rainbow of line artifact. effect cannot be reduced. The two models require costly image analysis to determine the magnitude of LED backlight mixing. How to reduce its computational cost and calibrate the LED backlight 6493-13, Session 4 accurately is a great challenge to implement the models on real field sequential color LCDs. A LCD TV comparable to a CRT TV for moving image quality K. Kim, LG. Philips LCD Co., Ltd. (South Korea) 6493-16, Session 4 The world’s best motion fidelity LCD, which is based on scanning Design and implementation of optimal color analyzers for backlight technique, was developed. In order to improve the motion TFT-LCDs blur, we have accomplished a series of tasks from visual perception G. Chang, C. Liao, Y. Lin, National Taiwan Normal Univ. (Taiwan) tests to practical developments. Moving image quality on an LCD TV could be improved by developing and adapting fast LC driving, For many years, the technology of TFT-LCDs (thin-film-transistor liquid scanning backlight, high frame rate, and image processing technique. crystal displays) has played an important role in the development of As a result, the MPRT (Motion Picture Response Time) of 4.0 ms for flat panel displays industries. To characterize the color display quality an LCD is achieved while a CRT has about 3.7 ms and a PDP still has of TFT-LCDs, it is important to design and important a simple and about 8.1 ms. It means that the last weakness bothering an LCD so cost-effective color measuring device, e.g., color analyzer. The far compared with a PDP and a CRT was cleared. Moreover, 4.0 ms objective of this paper is to propose a color analyzer with optimal MPRT is a number at which people cannot perceive motion blur for colorimetric accuracy, under the condition of a given collection of moving images under most conditions and therefore the MBR LCD is physically realizable (or commercially available) spectral filters. In this superior to a PDP by above 50% and equal to a CRT in motion blur. paper, to faithfully convey color information, the optical probe of the color analyzer contains a set of photoelectric detectors (e.g., photodiodes, PDs) associated with well-designed (or optimally 6493-14, Session 4 selected) spectral filters for different color channels. In our approach, a monochromator (or a multiple-channel filtering system) is developed Nonlinearity curve optimization in a digital imaging to measure the PDs¡õ spectral responses. According to the Luther- device Ives (LI) condition, a spectral-vector model in an infinite-dimensional S. Quan, X. Jiang, Qualcomm, Inc. space for color vision is first briefly presented to establish a theoretical basis for the optimal filter selection. Then, an optimization Due to the nonlinearity of traditional CRT display, which is factored algorithm is proposed to efficiently select the spectral filters from a into the widely accepted standard color space, sRGB, a digital given collection of spectral filters. As a result, the overall spectral imaging device will need an inverse nonlinear transformation to responses of the probe can be as close as possible to a set of color compensate the nonlinearity of a sRGB output device. The nonlinear mixing curves derived from the LI condition. Experimental results transformation is also often referred as gamma, as the nonlinearity of show that all of the colormetric quality factors corresponding to CRT can be fairly approximated by a power function. The nonlinearity individual color channels are quite close to 1. Also, the comparison transformation of imaging devices is also needed for color encoding between the measured color data and those from a spectrometer for compression, since the human visual system perceives the confirms that the objective of this work is achieved. luminance level differently, which is close to a logarithmic function. The logarithmic function is approximated with power function in this scenario. The nonlinearity curve can do more than just compensate the nonlinearity of output devices. Applications such as high dynamic range compression can be combined with this purpose. Further, the

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 93 Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

6493-17, Session 4 6493-53, Poster Session Third-time lucky: why a series of ISO display standards A GIS-based sensitivity mapping as a baseline for deserves extensive coverage at EI conferences coastal resort planning F. L. van Nes, Technische Univ. Eindhoven (Netherlands) and M. N. Hegazy, H. M. A. Effat, National Authority for Remote ErgoNes (Netherlands) Sensing and Space Sciences (Egypt) From cell phone and GPS to beamer and PC, almost all IT systems use Environmental sensitivity mapping is a tool for guiding the process of visual displays as output medium - so their utility hinges on the utility of landscape planning. these displays. That utility is a function of the display properties, which It provides the planner with a holistic view of the constraints and should fulfil essential visual requirements. The requirements, as well as opportunities of the site in an integrated vision. It is also an excellent methods to measure display properties and to verify whether a tool for land use decision-making. This paper aims at producing a particular display complies with the requirements, are accurately sensitivity map for an area to be designed as a coastal resort at Marsa specified in a new series of standards from ISO, the International Wazer, Red sea in Egypt. Organization for Standardization. These standards, ISO 9241-300 to 307, will provide knowledge for the manufacturers of displays, the Environmental sensitivity of the site was developed based on satellite display users, and those that procure the displays for these users. A imageries and several Geographic Information Systems operations progress report on the making of the new ISO display standard will be and models. These comprise building separate vector layers for the presented here for the third time, after the Electronic Imaging different physical environment baseline components, buffering of conferences in 2003 and 2005. The processes involved in creating certain features based on legislative standards, clipping and standards will be elucidated in some detail. Standard creation is a rasterization of layers and finally running the weighted overlay matter of teamwork, and participation in such a multidisciplinary team sensitivity model in Arc GIS 9. can be shown to widen the team members’ vision. The seven-part Different GIS layers outlines the ranking of the area under standard has the ambition to cover all visual displays and their investigation based on different terrain characteristics and its applications, so that it in principle caters for the needs of all fields, from suitability for different types of land use. Taking in consideration the medicine to telecommunication. effect of natural hazard that the different parts of the area can be subjected to, the area have been classified into classes based on vulnerability of different degrees of different natural hazards. Natural 6493-51, Poster Session hazards in the area include flash flooding, earthquakes, sand encroachment and mass movement. Study of digital watermarking for halftoneiImages The resultant map is expected to assist in the formulation of an X. Wan, D. Xie, Wuhan Univ. (China) environmentally sensitive conceptual master plan for the site. It can Abstract: This paper proposes a new watermarking algorithm be used as a prototype for coastal resort pre-planning analysis in based on Dot-to-Dot error diffusion technique in spatial domain, the similar coast zones as a base for creating sustainable land use plans. watermarking is embedded when the grayscale image is changed into the halftone image. Before embedded, the watermark image is permuted to enhance the capacity against geometry attacks. Because 6493-55, Poster Session of Dot-to-Dot error diffusion producing a intermediate bi-level image, Clapper-Yule reflectivity of halftone color fluorescent the watermark image can be hidden within it suitably, then after the imaging process of error diffusion, the halftone image embedded of watermark is made in the end. The research result indicates that under the strict Y. Zhang, Y. Du, Southern Yangtze Univ. (China) controlled experimental conditions, this algorithm can resist the distorts The color prediction law of fluorescent ink halftone is a pivotal during print and scan course, the watermark image can be extracted problem in the region of image processing. The thesis generalizes the successfully. The watermarking algorithm for halftone images has reflective low of color fluorescent ink halftone by utilizing the idea that intrinsic property for the copyright protection of printing products. the reflected light by fluorescent print is divided two parts: the primary Keywords:Digital watermarking, Digital halftoning , Halftone streams which consist of originally incident light and fluorescent images, Dot-to-Dot error diffusion, Copyright protection streams which are created by absorption of the primary stream, and a matrix exponential modelizing the layer of fluorescent ink is utilized to compute the energy of fluorescence. We introduce the weighted 6493-52, Poster Session factor to describe the phenomenon that the proportion is different between the exited light from the same colorant and from whole Compression of color images using a hologram of gray surface by taking into account the fact that proportionally more tones incident light through a given colorant surface is reflected back onto A. Restrepo-Martínez, R. Castañeda, Univ. Nacional de Colombia the same colorant surface than onto other colorant surfaces. And an (Colombia) improved spectral reflection Clapper-Yule model is presented for the fluorescent ink halftone. A strategy for to compress color images in digital holograms of gray tones was developed. The procedure codifies the information of each channel of the RGB model in a system of fringes, it is a gray image 6493-56, Poster Session denominated “hologram”. The fringes in their intensity of gray tone carry the signal of the channel, in this manner the amplitude Implement of FGS video encoding based on H.264 information for each channel of color image is stored. The angles of Q. Lin, Huaqiao Univ. (China) fringes define how the information of each channel will be packaged. The sum of the different gray fringes images is the hologram, it is the FGS encoding scheme may achieve consecutive enhancement layer “object” for a digital holographic system. The RGB channels are high rate control, so it has the capability to adapt the varying of network intensity peaks of information in the hologram’s Fourier space, and bandwidth. when the peaks are filtered each channel can be extracted. In H.264 video coding standard, the combination encoding frame was Parameters such as: space frequency, visibility, direction and quality of adopted. It introduces some new algorithms, and modifies several the fringes affect the quality of the reconstructed image. However, the aspects of the encoding scheme. So the encoding scheme improves propose methodology allow a radius 3:1 for the compression of color the encoding efficiency obviously. But the H.264 standard is not image, too with this process is possible the compression of different supporting FGS encoding. So a H.264 based self- adaptive FGS spectrum in a one color image. encoding scheme is proposed in this paper. In this encoding scheme,

94 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

the base layer of encoder is keeping H.264 encoder architecture, 6493-20, Session 6 which consists of the motion estimation, motion compensation, intra predictive, integer transformation, loop filtering, content based Adaptive color artwork arithmetic encoding, and etc. In the base layer generated block we G. B. Beretta, Hewlett-Packard Labs. obtain base code flux of FGS. Subtracting the original image from the reconstruction image of the base layer, we get the residual error. Then The words in a document are often supported, illustrated, and after the DCT transform and the variable length encoding enriched by visuals. When color is used, some of it is used to define compresses, we obtain the enhanced code flux of FGS. the document’s identity and is therefore strictly controlled in the Compared with the original MPEG-4 FGS encoding scheme, the design process. The result of this design process is a “color proposed FGS encoding scheme has the feature of increasing specification sheet,” which must be created for every background encoding efficiency by 1~3 dB and keep the all properties that MPEG- color. While in traditional publishing there are only a few backgrounds, 4 FGS encoding technology provided. in variable data publishing a larger number of backgrounds can be used. We present an algorithm that nudges the colors in a visual to be distinct from a background while preserving the visual’s general color 6493-18, Session 5 character. A novel color mapping method for preferred color reproduction 6493-21, Session 6 K. Kim, M. Jang, H. S. Oh, S. H. Kim, D. Choi, SAMSUNG Peteye detection and correction Electronics Co., Ltd. (South Korea) J. Yen, Toshiba America; H. Luo, Yahoo! Research Berkeley; D. R. We propose a novel color mapping method that generates smooth Tretter, Hewlett-Packard Labs. color transition in gamut boundaries and can accommodate the color preference. When colors are transferred from one device into another, Redeyes are caused by the camera flash light reflecting off the retina. such as from a monitor to a printer, gamut mapping is an essential Peteyes refer to similar artifacts in the eyes of other mammals caused step to adapt colors from one gamut shape to another. However, it is by camera flash. In this paper we present a peteye removal algorithm not trivial to produce smooth maps for gamut surface colors since the for detecting and correcting peteye artifacts in digital images. Peteye gamut shapes and the lightness of primaries of two devices are removal for animals is significantly more difficult than redeye removal normally very different. Moreover, it is difficult to apply color for humans, because peteyes can be any of a variety of colors, and preferences to the mapping. To overcome these problems, the human face detection cannot be used to localize the animal eyes. In developed method defines target colors of HSV 12x5x5 LUT lattice many animals, including dogs and cats, the retina has a special points and finds the best mapping points to reproduce them. The reflective layer that can cause a variety of peteye colors, depending “target color” means a preferred reproduction of an input color and on the animal’s breed, age, or fur color, etc. This makes the peteye can be obtained from various color books such as Munsell book, NCS correction more challenging. We have developed a semi-automatic Index, Pantone color and so on. Color preferences can be considered algorithm for peteye removal that can detect peteyes based on the when target colors are defined. Since the mapping relationships of the cursor position provided by the user and correct them by neutralizing gamut boundary colors (the most saturated and brightest color of the colors with glare reduction and glint retention. each hue) affect the mapping of the remaining LUT lattice points, it is very important where to map the gamut boundary colors. The proposed method generates N candidate colors for the gamut 6493-23, Session 7 boundary color of each hue and then selects one of them which is Modeling an electro-photographic printer, part I: visually the most similar with the target color. After finishing the monochrome systems mapping of the boundary colors, it adjusts the target colors of remaining LUT lattice points using the previous mapping relationships M. A. Kriss, Consultant and then finds the best mapping points of them. The results show that This paper will outline a simplified model for the development of toner this method can utilize the gamut of the output device fully and dots on a reflective support. Using this model and the interaction of reproduce the preferred colors. light between the reflective support and the dot’s micro-structure, the physical, optical and total dot gain will be calculated, along with the 6493-19, Session 5 resulting tone scales, for a variety of digital halftone patterns. The resulting tone reproduction curves and dot gain will be compared with Gamut mapping method for ICC saturated intent the classical literature on dot gain and tone reproduction curves, more modern approaches and experimental data from the literature. A M. Cho, Y. Bang, Y. Kim, S. Kim, H. Choh, SAMSUNG Advanced comparison to a well defined experimental system will be shown. Institute of Technology (South Korea) The same image on the display and color printer isn’t the same. Firstly, this is due to the bit depth difference for representing the color 6493-24, Session 7 of a pixel. The display uses the color data of the eight or more bits, Modeling an electro-photographic printer, part II: color but the color printer uses just 1bit for representing color of a pixel. So, systems the display can reproduce smoother image than the color printer. Secondly, the display gamut is larger than the printer gamut, so the M. A. Kriss, Consultant display color is brighter and more saturate than the printer color. For This paper will outline a simplified model for the development of toner minimizing the problems due to these differences, many halftoning dots on a reflective support in a color electro-photographic system. A and gamut mapping techniques have been developed. For the gamut model developed for a monochrome system will be adapted to a color mapping, color management standard organization, ICC, imaging system where four pigments, each capable of scatting light, recommended 2 gamut mapping methods, HPMINDE and SGCK. But is used to form a digital halftone image. The combination of physical the recommended methods by ICC have some weak points; and optical dot gains, interlayer scattering, on-dot and off-dot digital contouring (HPMINDE), paled pure color reproduction (SGCK) and too halftones will be explored and the results demonstrated in terms color reddish hair color reproduction (HPMINDE, SGCK). This paper shifts due to layer order and dot gain due to halftone geometry. introduces a gamut mapping method that can reproduce smooth gradation, pure colors with high saturation and natural hair color. The proposed method is developed for optimal reproduction of graphic image, and it also gives good results for pictorial image.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 95 Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

6493-25, Session 7 the actual printer and media combination. The new transform can be computed independently of the image conversion, or at run-time by a Printer color calibration using an embedded sensor dynamic or programmable CMM. Once the transform is computed it Y. Wu, Hewlett-Packard Co. can be written in the form of an ICC profile, making the approach fully compatible with current colour management workflows. The purpose of printer color calibration is to ensure accurate color This approach avoids the necessity of printing and measuring a test reproduction and to maintain the color consistency. A closed loop chart for each printer / media combination and this permits both color calibration system has the ability to perform the whole printer and media vendors to supply a single profile for each product. calibration work automatically without any user intervention. A color calibration process consists of several steps: print a pre-designed The accuracy of the final transform depends on the methods used to target; measure the printed color patches using an color measuring compute it from the reference printer and media transforms. Some instrument; run a set of algorithms to calibrate the color variations. In methods of computing this transform are described and results given this paper, we present a closed loop color calibration system. And we for several different printer and media combinations. will show in particular how to use a low cost optical sensor to get Although the approach may not always be sufficiently accurate for accurate color measurement. Traditional, low cost optical sensors are applications where high levels of colorimetric accuracy are required, only used to measure the voltage data or density data. The novelty of such as graphic arts proofing, it represents a reasonable trade-off for our approach is that we can also use a low cost optical sensor to situations where the user needs to simplify the process of profile measure colorimetric data. Using the colorimetric measurement, we creation and selection. can perform more complicate color calibration works for color printing Accuracy can be increased by adding further reference media so that systems. the gamut differences between the reference print and the actual print are minimised. 6493-26, Session 7 Color measurements on prints containing fluorescent 6493-28, Session 8 whitening agents Efficient color printer characterization based on M. Andersson, O. Norberg, Mid Sweden Univ. (Sweden) extended Neugebauer spectral models Papers with a slightly blue shade are, at least among a majority of P. Soler, J. Arnabat, Hewlett-Packard Co. (Spain) observers being perceived as whiter than papers having a more In order to print accurate colors on different substrates, color profiles neutral color. Therefore, practically all commercially available printing must be created for each specific ink-media combination. We tackled papers contain bluish dyes and fluorescent whitening agents (FWA) to the problem of creating such color profiles from only few color give the paper a whiter appearance. As an effect of the FWA, samples, in order to reduce the needed time of operation. Our spectrophotometer measurements of printed matter are sensitive to strategy is to use a spectral reflectance prediction model in order to the ultraviolet (UV) content of the light source used in the instrument. estimate a large sampling target (e.g. IT8.7/4) from only a small subset Standard spectrophotometers used in the paper industry are of color patches. In particular, we focused on the so-called Yule- equipped with an adjustable filter for calibrating the UV-content of the Nielsen modified Spectral Neugebauer model, proposing a new area illumination from a light source of type C. The graphical arts industry coverage estimation. We reviewed the basis of such model, interpret it on the other hand, typically measures with spectrophotometers under the perspective of generalized averaging, and derived new having a light source of type A, with only limited possibilities to adjust expressions to decouple optical and mechanical dot gain effects. The the UV-content by the use of different weighting filters. The results proposed area coverage estimations are based on assumptions of the demonstrate how the use of spectral weighting functions for printing process, and characterized through few extra color samples. simulating standard illuminants works properly on non-fluorescent We tested the models with thermal ink-jet printers on a variety of material. However, when FWA is present, disparities in UV content media, with dye-based and pigment-based inks. Our results show between the light source and the simulated illuminant will result in that we can effectively reduce in more than 50% the number of color differences. Therefore, comparisons between measurements on patches, without a significant loss of colorimetric accuracy. In all substrates containing FWA from two instruments having light sources cases, spectral models clearly outperformed linear interpolation in the with different UV-content are complicated. Finally, in many printing colorimetric space leading to a reduced sampling color target. processes, some of the used inks are UV-blocking. This further complicates the effect of FWA in printed material. 6493-29, Session 8 6493-27, Session 8 Digital camera characterization for color measurements on printed-paper media A reference printer and color management architecture M. Andersson, Mid Sweden Univ. (Sweden) P. J. Green, London College of Communication (United Kingdom) The media dependency must be considered in color calibrations of The concept of a reference printing condition has been introduced in flatbed scanners and digital cameras, where the intention is to recent publications. This paper explores an application of a reference measure color in printed matter. Even if the calibration problem is printer and proposes an associated colour management architecture. restricted to measurements on prints produced in a single printer, it is Colour management based on ICC profiles is a well-established still tedious work to make a separate calibration for each new paper mechanism for transforming images from a CIE-based colour space grade to be used in the printer. Therefore, it would be of interest to to a corresponding device encoding. For hard copy output devices a find a method where it is sufficient to characterize for only one or a profile is required for each combination of printer setting and media, few papers within a grade segment and then be able to apply a which results in a proliferation of profiles and practical difficulties in correction based on measurable optical paper properties. However, creating, managing and selecting appropriate profiles. before being able to make any corrections, the influence of A reference printer definition specifies the colorimetric properties of measurable paper properties on color calibrations must be studied. the primary colorants, the colour gamut, the colorant combinations Fluorescence has been mentioned as a potential error source in color required to produce a neutral scale, a set of characterization data, a calibrations of trichromatic capturing devices intended for reference medium, and procedures for metrology and encoding. measurements on printed matter. In this study, the influence of fluorescent whitening agents in printing paper on the color calibration A colour management architecture using a reference printer is of a digital camera for color measurements on printed targets is described, in which transforms are built for the reference printer on discussed. The results show that varying amounts of fluorescent any actual medium, and any printer on the reference medium. These whitening agents between printing papers, as well as differences in transforms are used in estimating a PCS - to - device transform for

96 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

UV-content between the light sources used in the image-capturing 6493-32, Session 8 system and the reference spectrophotometer, can introduce significant calibration errors. Production planning and automated imposition C. Tuijn, Agfa-Gevaert Group (Belgium) 6493-30, Session 8 Printed products such as books, magazines, leaflets etc. all consist of sections that will be printed on a press sheet. After printing, the User preferences in color enhancement for unsupervised sheets will be folded and cut. As a last step, the different sections will printing methods be collected and bound together (glued, stapled, stitched etc.). In the C. E. Parraman, Univ. of the West of England (United Kingdom); A. prepress environment, one traditionally uses special imposition Rizzi, Univ. degli Studi di Milano (Italy) templates to identify how the pages should be imposed on the press sheet. The main drawback of this approach is that one needs to In order to obtain a good quality image for inkjet printing, the process remake imposition templates each time a parameter has been of adjusting images can be a time consuming and a costly procedure. changed. A new approach to overcome this problem consists of In this paper, we consider the use of an unsupervised color specifying a so-called stripping specification of the sections on the enhancement method as final automatic pre-processor for printing. sheet. In addition to the stripping information, one also can specify Other unsupervised colour enhancement methods are utilsed: how the sections can be combined physically. This is done in the so- Retinex, RSR, ACE, Histogram Equalization, Auto Levels. Test images called assembly specification. Both stripping and assembly allow are subjected to all of the enhancement methods, which are then defining unambiguously how a product can be produced. In the first printed. Users are asked compare each of the printed images. part of the paper, we will explain how the stripping and assembly In all cases, the results are dependent on the image. Thus we have specification can be used to further automate the prepress, printing selected a series of test images: photographic - well balanced, as well and finishing processes. In the second part, we will discuss how the as imperfect ones; reproduction of prints - how pigment inks, surface production planning itself can be automated. This assumes an in- marks or texture can be reproduced; collaged artworks - combination depth knowledge of the available equipment and resources. of drawing, photography, scanning; black and white photographs and artworks; digitally generated artworks 6493-33, Session 8 Some of the tested methods are parameter dependent. We do not intend to consider fine tuning for each of the techniques, rather to The perfect photo book: hints for the image selection consider an average parameter set for each one and then test if this process approach can go in the user preference direction. R. Fageth, W. Schmidt-Sacht, CeWe Color AG & Co. OHG Three user groups are employed: the general user, commercial (Germany) photographer expert and fine artist. Groups are asked to make a blind evaluation of a range of images (the original and the colour enhanced People take more and more digital images. This has several reasons. by the different methods); these are randomly placed. All images are They are afraid to miss the right moment and pressing the button is printed on the same printer using the same settings. Users are asked not linked to costs as it ha been with silver halide photography. This to identify their preferred print in relation to lightness, tonal range, behaviour seems to be convenient but ends up in a dilemma for the colour range, quality of detail and overall subjective preference. The consumer. results are presented and discussed. He is forced to analyze these images before printing them. This is immanent for all printing possibilities; at home, at retail or via wholesale finishing. This task is time consuming and leads to loss of 6493-31, Session 8 access to important images. Images one cannot find is like not having A user-friendly digital image processing methodology taken them. R. Shaw, White Rose Digital This paper will present a novel approach of enabling the consumer to present a folder or folders with numerous images to a software. The Last year at this conference the author described a consumer digital- desired result is a photo book produced by electro photography image interface based entirely on an intuitive image-choice-only digital presses. Getting lots of different images onto one product such operation. This was made possible by the definition of a set of as a photo book is one of the benefits of using electro photography orthogonal image vectors, and defining all excursions in terms of a for consumer products and additionally offering the industry to fixed linear visual-pixel model, independent of the image vector. address consumers who do want to have separate prints. During the following months the author has extended this same methodology both to the more routine image editing features, and to advanced image enhancement algorithms based on conditional pixel 6493-34, Session 9 mapping that are necessary to address the need for images requiring Color recovery from gray image based on analysis of sharpening or similar special treatments. wavelet packet subbands The current contribution describes these features in further detail, and will demonstrate the user-interface as applied to a variety of typical K. Ko, O. Kwon, C. Son, E. Kwon, Y. Ha, Kyungpook National consumer digital images. Also the fundamental image vectors and Univ. (South Korea) their natural ranges, and the underlying linear-visual pixel model, have This paper proposes a colorization method that uses wavelet packet been both extended and refined, based on practical experience with sub-bands to embed color components. The proposed method, more than ten thousand typical consumer images. firstly, involves a color-to-gray process, in which an input RGB image A further novel facility based on the absolute nature of this is converted into Y, Cb, and Cr images, and a wavelet packet enhancement methodology concerns the relative ease of transform (WPT) applied to Y image to divide it into 16 sub-bands. incorporating artificial intelligence into the natural choice hierarchy. The Cb and Cr images are then embedded into two sub-bands that Already promising results have been obtained relating to the specific include minimum information on the Y image. Once the inverse WPT characteristics of image-acquisition devices, similar scene-types, is carried out, a new gray image with texture is obtained. Secondly, a individual user preferences, the specific lighting conditions and gray-to-color process is performed. The printed textured-gray image display-technology defining the viewing environment, right through to is scanned and divided into 16 sub-bands using a WPT to extract the individual user preferences. Cb and Cr components, and an inverse WPT is used to reconstruct the Y image. At this time, the original information is lost in the color- to-gray process. Nonetheless, the details of the reconstructed Y image are almost the same as those in the original Y image because it uses sub-bands with minimum information to embed the Cb and Cr

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 97 Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

components. In addition, to recover color saturations more accurately, platform. Subsequently, the generated light is used to excite the gray patches for compensating the characteristics of printers and spectral characteristics of the camera under test. It is found that the scanners are used. Consequently, the proposed method can improve system signal to noise ratio (SNR) is fairly low since the strength of the both the boundary details and the color saturations in recovered color monochromatic light is very weak relative to that of the light source, images. due to the presence of the optical signal only in a extremely narrow bandwidth or wavelength division. To increase the SNR, a digital LIA technique is devised and implemented in the PC-based platform. The 6493-35, Session 9 LIA technique primarily performs the demodulation for the digital signals contaminated with noise, in response to the monochromatic Model-based deduction of CMYK surface coverages light modulated at the same frequency as that of the light source, from visible and infrared spectral measurements of such that the effect of noise can be effectively suppressed. To halftone prints demonstrate the effectiveness of this approach, the experimental T. Bugnon, M. Brichon, R. D. Hersch, École Polytechnique results from a multichannel-filtering system with broadband filters is compared to those from the proposed system. From the experiments, Fédérale de Lausanne (Switzerland) it is found that the measurement accuracy is fairly satisfactory. The Yule-Nielsen modified Spectral Neugebauer reflection prediction model enhanced with an ink spreading model provides high accuracy when predicting reflectance spectra from ink surface coverages. In 6493-37, Session 10 the present contribution, we try to inverse the model, i.e. to deduce the surface coverages of a printed color halftone patch from its Interpolation for nonlinear Retinex-type algorithms measured reflectance spectrum. This process yields good results for D. Shaked, Hewlett-Packard Labs. (Israel) cyan, magenta, and yellow inks, but unstable results when simultaneously fitting cyan, magenta, yellow, and black inks due to In this paper we propose a method to speed up non-linear Retinex- redundancy between these four inks: black can be obtained by type algorithms. Those algorithms consist of a computationally printing either the black ink or similar amounts of the cyan, magenta, intensive non-linear illumination estimation module followed by a and yellow inks. To overcome this problem, we use the fact that the relatively simple manipulation module. Non-linear Retinex type black pigmented ink absorbs light in the infrared domain, whereas algorithms span dynamic range compression, spatial gamut mapping, cyan, magenta, and yellow inks do not. Therefore, with reflection and other image enhancement algorithms. In different algorithms the spectra measurements spanning both the visible and infrared domain, intermediate image is a mask or another fuzzy segmentation image it is possible to accurately deduce the black ink coverage. Since there instead of a proper illumination image. In all cases, speed up is is no redundancy anymore, the cyan, magenta, yellow, and pigmented obtained by way of computing the illumination on a sub-sampled black ink coverages can be recovered with high accuracy. image. The challenge is to interpolate a piece-wise smooth low resolution image. We present and analyze the trade-of between two types of interpolation methods. On one hand, regular illumination 6493-57, Session 9 interpolation, which preserves the Retinex-type output quality, however may result in artifacts. On the other hand a detail preserving Automatic processing of images for visual preference interpolation which removes artifacts, however may compromise the output quality. M. Bressan, D. Arregui, H. Poirier, Xerox Research Ctr. Europe (France); C. R. Dance, Xerox Research Ctr. Europe (United Kingdom) 6493-38, Session 10 We introduce a novel algorithm for local contrast enhancement. Our Omnidirectional scene illuminant estimation using a algorithm exploits a background image which is estimated with an edge-preserving filter. The background image controls a gain which multispectral imaging system enhances important details hidden in underexposed regions of the S. Tominaga, T. Fukuda, Osaka Electro-Communication Univ. input image. Our designs for the gain, edge-preserving filter and (Japan) chrominance recovery avoid artifacts and ensure the superior image quality of our results, as extensively validated by user evaluations. A method is developed for estimating an omnidirectional distribution Unlike previous local contrast methods, ours is fully automatic in the of the scene illuminant spectral distribution, including spiky sense that it can be directly applied to any input image with no fluorescent spectra. First, we show a measuring apparatus, consisting parameter adjustment. This is because we exploit a trainable decision of the mirrored ball system and the imaging system using a LCT filter mechanism which classifies images as benefiting from enhancement (or color filters), a monochrome CCD camera, and a personal or otherwise. Finally, a novel windowed TRC mechanism based on computer. Second, the measuring system is calibrated and images monotonic regression ensures that the algorithm takes only 0.3 s to representing the omnidirectional light distribution are created. Third, process a 10 megapixel image on a 3 GHz Pentium. we present an algorithm for recovering the illuminant spectral-power distribution from the image data. Finally, the feasibility of the proposed method is demonstrated in an experiment on a classroom scene with 6493-36, Session 10 different illuminant sources such as fluorescent light, incandescent light, and daylight. The accuracy of the estimated scene illuminants is Accurate spectral response measurement system for shown in the cases of the 6-channel multi-band camera, 31-channel digital color cameras spectral camera, and 61-channel spectral camera. G. Chang, C. Liao, National Taiwan Normal Univ. (Taiwan) In machine vision systems, color imaging plays an essential role in 6493-39, Session 10 conveying and recording visual information from the real world. To Deducing ink-transmittance spectra from reflectance accurately represent colors acquired from digital cameras, a spectral response measurement system is proposed for the imaging devices in and transmittance measurement of prints this paper. The proposed system primarily contains a grating-based H. Mathieu, R. D. Hersch, École Polytechnique Fédérale de spectral filtering module (called monochromator), a digital lock-in Lausanne (Switzerland) amplification (LIA) module, a system light source modulated at a given frequency by an optical chopper, and a PC-based platform to perform The spectral transmittance of inks are important for predicting the measurement and control. The grating-based module is developed by reflectance spectra of halftone prints. We propose a new method for filtering the light emerging from the light source, to produce a determining ink transmittances from solid ink prints, which relies on a collection of monochromatic light with different wavelengths in the recently developed reflectance and transmittance prediction model for visible range, through the rotation of the grating controlled by the prints [1] incorporating the ink transmittances as spectral parameters. 98 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications

In order to deduce the transmittance of an ink, we take benefit of the We observed that the sVED of a spectral image generates poor quality different measuring configurations supported by the model: reflectance image when its gamut is bigger than the printer gamut.The error of the solid print, transmittance of the solid print with the ink layer facing diffusion is visible, a spreading of the colorant in the image appears and the light source, and transmittance of the solid print with the ink layer we loose information from the original image. sVED of image inside the facing the detector. Each of these three configurations provides a printer gamut does not bring such problem. But a change in reflectance different equation from which the ink transmittance can be in the image is still visible after sVED due to the slowness of error independently deduced. The three independently deduced diffusion in this halftoning technique. The error accumulation changes transmittance spectra are compared with one another for checking the enormously the reflectance in the image during this halftoning process. accuracy of the method. We consider prints on a plastic white diffusing There is a need for limiting the error accumulation before performing substrate, where the inks do not penetrate the substrate. From the three sVED on the image and during the sVED. measuring configurations, we deduce three nearly identical ink In this paper we propose two methods to reduce the slowness of the transmittance spectra. sVED: one by gamut mapping and a second by error thresholding during When comparing the ink transmittance spectra deduced from our the sVED. Our gamut mapping is based on the inversion of the spectral reflectance and transmittance model with the ones deduced from the Neugebauer printer model. This model says that the reflectance Clapper-Yule model [2], the Clapper-Yule model yields slightly different estimation of a colorant combination is the weighted summation of the ink transmittance spectra. This is due to the fact that the Clapper-Yule NP. By optimization technique we calculate the weights of the NP given model does not take into account the obliqueness of light traversing the the closest reflectance estimation the printer can reproduce. We then ink layer. However, the error is relatively small, as suggested by Clapper replace the original reflectance value by its closest estimation. For the and Yule in their original article. threshold operation we calculate how far is the original reflectance from [Ref 1:] M. Hebert and R. D. Hersch, “A reflectance and transmittance the current pixel with the added error from the previous pixels, when a model for recto-verso halftone prints,” J. Opt. Soc. Am. A, doc. ID threshold is reached the error is reset to zero. The two methods are 67306 (posted 11 May 2006, in press). experimented independently and combined. Two sort of spectral images are used in our experiments: the ColorChecker testchart and spectral [Ref 2:] Clapper F.R. and Yule J.A.C. (1953), “The Effect of Multiple test image. To evaluate the proposed methods we compared for each Internal Reflections on the Densities of Halftone Prints on Paper”, image the resulting halftoned image by classic sVED and the halftoned Journal of the Optical Society of America, 43, 600-603. image with the proposed methods.

6493-40, Session 11 6493-42, Session 11 Halftone independent methods for color correction Holladay halftoning using superresolution encoded V. Monga, R. Bala, S. Wang, Xerox Corp. templates Similar to an orchestra which needs to tune all music instruments to J. S. McElvain, C. M. Hains, Xerox Corp. match a reference pitch for a harmonic performance, the uniform- A new method for halftoning using high resolution pattern templates is rosette halftone method requires all halftone frequencies involved to described, that expands the low level rendering capabilities for engines match two common fundamentals for a harmonic halftoning. By that support this feature. This approach, denoted super resolution selecting halftone frequencies from high-order harmonics of two encoded halftoning (SREH) is an extension of the Holladay concept, and common rosette fundamentals for all color separations, the interference provides a compact way to specify high resolution dot growth patterns between any two frequency components, fundamentals or high-order using a lower resolution Holladay brick. Fundamentally, this new harmonics, of different colors will also result in a linear combination of halftoning method involves using the SRE patterns as building blocks the two rosette fundamentals. Thereby, no visible interference, or moiré, for constructing clustered dot growth assemblies. Like the traditional at all will be shown in the output. The halftone outputs are two- Holladay dot description, the SRE halftone is characterized by a size, dimensionally repeated patterns, as uniform rosettes, which are visually height, and shift, all of which are specified at the lower resolution. Each pleasant, robust to mis-registration and truly moiré free. The uniform- low resolution pixel position in the SRE halftone brick contains a pair of rosette configurations can be implemented easily for digital halftoning. lists. The first of these is a list of digital thresholds at which a transition Several such designs have been successfully applied to Xerox iGen3 in SRE patterns occurs for that pixel position, and the second is the color printers for high-quality color reproduction. corresponding list of SRE codes. For normal cluster dot growth sequences, there will be at most LxM list elements, creating a simple 6493-41, Session 11 and compact mechanism for specifying higher resolution halftones. Techniques for emulating traditional high resolution Holladay dots using Controlling the error in spectral error diffusion SREH are discussed, including mechanisms for choosing substitutions for patterns that do not exist among the available SRE patterns. J. Gerhardt, Hogskolen I Gjovik (Norway); J. Y. Hardeberg, Gjøvik Univ. College (Norway) We aim to print spectral images. In a spectral image each pixel is a 6493-44, Session 12 reflectance. We use a digital multi-ink printer, seven colorants are The hybrid screen: improving the breed available. The printer needs an image made of n binary colorant channels, n for the number of available colorant. In this colorant image C. Lee, J. P. Allebach, Purdue Univ. each pixel value {0, 1} is a command to print or not to print a drop of ink The hybrid screen is a halftoning method that generates stochastic of a colorant channel. The spectral image has to be converted in a multi dispersed dot textures in highlights and periodic clustered dot textures binary-layer image. Spectral vector error diffusion (sVED) is an halftoning in midtones. Each tone level is sequentially designed from highlight to technique which performs this operation. midtone by applying an iterative halftoning algorithm such as direct The sVED is based on the error diffusion (ED) algorithm. As in ED an binary search (DBS). By allowing random seeding and swap-only DBS in input pixel is binarised and the difference (i.e. the error) input pixel minus a predefined core region within each microcell, we design each level the output pixel is diffused to the neighboring pixels. The error is while satisfying the stacking constraint and guaranteeing a smooth weighted and diffused with a filter weight by wavelength. This operation transition between different levels. This paper focuses on a number of is performed in a raster scan mode. enhancements to the original hybrid screen and their impacts on print The output pixel value is decided after the calculation of the Euclidean quality. These include analytical determination of the human visual distance from the input pixel to the Neugebauer primaries (NP). The NP system filter in the spatial domain for DBS and multilevel screen design are all the possible binary combination within the available colorants. A either by extending a bilevel screen or by directly generating a multilevel seven inks printer will produce 128 NP. The primary given the smallest screen on the high resolution grid. Our results show that the multilevel distance to the pixel sets the output pixel values. screen design method has a direct impact on hybrid screen design parameters such as the optimal core size. We also extend the whole electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 99 Conference 6493: Color Imaging XII: Processing, Hardcopy, and Applications design process to color by jointly optimizing the color screens using resolution would increase data size to be transmitted and memory color DBS. Our results demonstrate a significant improvement in the size at printer. For a high-speed printer, increased transmission time highlights over halftones generated by independently designed screen. may limit the throughput in imaging chain. One of possible solutions to this problem is to develop resolution enhancement techniques. In this paper, a fast and efficient spatial resolution enhancement 6493-45, Session 12 technique is proposed. Objectives of the proposed technique are to reduce the data size for transmission and minimize image quality Ranked dither for robust color printing deterioration. In the proposed technique, number of black pixels in the M. R. Gupta, J. J. Bowen, Univ. of Washington halftoned data is binary coded for data reduction. At printer, black pixel placement algorithm is applied to binary coded data. For non- A spatially-adaptive method for color printing is proposed that is robust edge area, screen order is utilized for the black pixel placement. When to printer instabilities, reproduces smooth regions with the quality of identified as edge area, locations of black pixels are selected by the ordered dither, reproduces sharp edges significantly better than ordered edge order designed by genetic algorithm. dither, and may be less susceptible to moire. The new method acts in parallel on square, non-overlapping blocks of each color plane of the image. For blocks with low spatial activity, standard ordered dither is 6493-49, Session 12 used, which ensures that smooth regions are printed with acceptable quality. Blocks with high spatial activity are halftoned with a proposed Contribution to quality assessment of digital halftoning variant of dither, called ranked dither. Ranked dither uses the the same algorithms ordered dither matrix as standard dither, but the ranks of the thresholds are used rather than the thresholds themselves. Ranked dither is more F. Cittadini, S. J. Berche, Océ Print Logic Technologies (France); sensitive than ordered dither to edges and more accurately reproduces M. Ben Chouikha, G. Alquié, Univ. Pierre et Marie Curie (France) sharp edges. Many new proposals are continually published in the halftoning Experiments were done with standard ordered dither masks of size 130, domain. Alas, the demonstration of the interest of the proposed 130, 128, 144 for the cyan, magenta, yellow, and black planes methods is often limited to a few favourable tests for the proposed respectively. Both on-screen and in-print, the results were sharper methods, and images showing the defects of the other halftoning halftones. The entire process can be implemented in parallel and is not methods. computationally expensive. The halftoning community needs to be able to compare a halftoning method with the innovations that appear in this domain. A complete and measured evaluation of quality is necessary through to a well 6493-46, Session 12 defined set of test images and metrics to evaluate the algorithm. Rank-ordered error diffusion: methods and applications This paper proposes a first sketch of a protocol for the quality assessment of digital halftoning algorithm that can be used to B. Xu, R. P. Loce, Xerox Corp. compare one algorithm to another. No abstract available It discusses the assessment of halftoner quality. It analyzes the perceived image quality concepts and defines the technical criteria that a good halftoner must match. A simple quality assessment 6493-47, Session 12 protocol is proposed. It is composed of test images and quality metrics. AM/FM halftoning: improved cost function and training framework This protocol could be used to provide new proposed halftoning algorithms with objective results. S. W. Han, Purdue Univ. FM halftoning generates good tone rendition but it is not appropriate for electrophotographic (EP) printers due to the inherent instability of the EP 6493-50, Session 12 process. Although AM halftoning yields stable dots, it is susceptible to Uniform rosette for moire-free color halftoning moiré and contouring artifacts. To combine the strengths of AM and FM halftoning, the AM/FM halftoning algorithm exploits each advantage of S. Wang, Xerox Corp. AM and FM halftoning. The resulting halftone textures have green noise Similar to an orchestra which needs to tune all music instruments to spectral characteristics. In this paper, we present an improved training match a reference pitch for a harmonic performance, the uniform- procedure for the AM/FM halftoning algorithm. Since most of the green rosette halftone method requires all halftone frequencies involved to noise energy is concentrated in the middle frequencies, the tone match two common fundamentals for a harmonic halftoning. By dependent error diffusion (TDED) parameters (weights and thresholds) selecting halftone frequencies from high-order harmonics of two are optimized using a new cost function with normalization to distribute common rosette fundamentals for all color separations, the the cost evenly over all frequencies. With the new cost function, we can interference between any two frequency components, fundamentals obtain image quality that is very close to that provided by the direct or high-order harmonics, of different colors will also result in a linear binary search (DBS) search-based dispersed-dot halftoning algorithm. combination of the two rosette fundamentals. Thereby, no visible The cost function for training the AM part is also modified by penalizing interference, or moiré, at all will be shown in the output. The halftone variation in measured tone value across the multiple printer conditions outputs are two-dimensionally repeated patterns, as uniform rosettes, for each combination of dot size and dot density. which are visually pleasant, robust to mis-registration and truly moiré free. The uniform-rosette configurations can be implemented easily for digital halftoning. Several such designs have been successfully 6493-48, Session 12 applied to Xerox iGen3 color printers for high-quality color Resolution enhancement techniques for halftoned images reproduction. B. T. Ryu, J. O. Lee, C. Kim, Inha Univ. (South Korea); S. H. Kim, SAMSUNG Electronics Co., Ltd. (South Korea) Recently, speed and resolution of electrophotographic printer engine have been significantly improved. Printing speed is determined by the processing time at computer, data transmission time between computer and printer, and processing and printing time at printer. When halftoning is performed at computer side, halftoned data would be compressed and sent to printer. In this case, increase in the spatial and bit-depth

100 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6494: Image Quality and System Performance IV Tuesday-Thursday 30 January-1 February 2007 Part of Proceedings of SPIE Vol. 6494 Image Quality and System Performance IV

6494-28, Poster Session MSE (the Mean Square Error), SNR ( Signal to Noise Ratio ) and WSNR(Weight Signal to Noise Ratio ) were often used to evaluate the Accurate and cost-effective MTF measurement system common con-tone images and the halftones. But these methods do not for lens modules of digital cameras suit to evaluating the quality of the halftones because of the special properties of the halftones by different halftoning algorithms and G. Chang, S. Pan, C. Su, National Taiwan Normal Univ. (Taiwan) limitation of assumption of these methods themselves according to For many years, the widening use of digital imaging products, e.g., many researches. So a series of halftonig algorithm-based methods are digital cameras, has given rise to much attention in the market of proposed, which adapt to the special properties of halftoning consumer electronics. However, it is important to measure and algorithms. All of those methods were not adaptive. In the last part of enhance the imaging performance of the digital ones, compared to that this paper, an adaptive method was propose to evaluate the halftoning of conventional cameras. For example, the effect of diffraction arising algorithms and the corresponding halftones, which is based on the from the miniaturization of the optical modules tends to decrease the statistical features of the residual image between the original image and resolution of the image. As a figure of merit, modulation transfer the corresponding halftone on the retinal of human eye. function (MTF) has been broadly employed to estimate the image quality. Therefore, the objective of this paper is to design and implement an accurate and cost-effective MTF measurement system 6494-01, Session 1 for the camera lens modules. In this paper, a theoretical framework is first briefly presented and then, a cost-effective approach is proposed. Measuring user experience: theoretical and In this approach, a spatial light modulator (SLM) is employed to methodological issues spatially modulate the light, which is emitted from a white-light LED, J. M. Takatalo, Univ. of Helsinki (Finland); J. P. Häkkinen, Nokia with the slit patterns of variable widths. The modulated light going Research Ctr. (Finland); H. Särkelä, J. Komulainen, J. Kaistinen, G. through the lens module under test is consecutively detected by the S. Nyman, Univ. of Helsinki (Finland) imaging sensor of the proposed system. The corresponding images formed by the optical modules are acquired by a computer and then, There are innumerable concepts, terms and definitions for user they are processed by an algorithm for computing the MTF. Obviously, experience. Few of them have a solid empirical founda-tion. In trying to the mapping from the collection of the modulated light to the output understand user experience in interactive technologies such as images is treated as a computation kernel for the MTF measurement. computer games and virtual environments, reliable and valid concepts Through the investigation on the accuracy and cost of the are needed for measuring relevant user reactions and experiences. Here measurement systems using various methods, such as the methods of we present our approach to create both theoretically and point spread function, line spread function, edge spread function, and methodologically sound methods for quantification of the rich user bar target, it appears that our approach give quite satisfactory results. experience in different digital environments. Our approach is based on the idea that the experience received from a content presented with a specific technology is always a result of a complex psychological 6494-29, Poster Session interpretation process, which components should be understood. The main aim of our approach is to grasp the complex and multivariate Quality improvement by selective regional slice coding nature of the experience and make it measurable. We will present our implementation in H.264/AVC two basic measurement frameworks, which have been developed and H. Ryu, W. Cheong, S. Jeong, K. Moon, Electronics and tested in large data set (n=2182). The 15 measurement scales extracted from these models are applied to digital gaming with two different Telecommunications Research Institute (South Korea) displays. The results show how it is possible to map between In this paper, we propose a new selective regional slice coding method experience, technology variables and the background of the user. This for H.264/AVC that can improve the quality of the decoded video. In approach can help to optimize, for example, the contents for specific the proposed method, a picture is divided into foreground and viewing devices or viewing situations. background regions then the two regions are coded differently. In the most of regional coding method, the main purpose of the methods is to improve the quality of the foreground area. Thus, they cause annoying 6494-02, Session 1 boundary effect on the boundary between foreground and background area, since the foreground is coded to have much better quality than Audio-visual quality estimation of mobile phone video the background area. To solve this problem, we propose a new cameras with interpretation-based quality approach selective regional slice coding method that can improve the picture J. E. Radun, T. Virtanen, G. S. Nyman, Univ. of Helsinki (Finland); J. quality by reducing the boundary effects between the foreground and Olives, M. Vaahteranoksa, T. Vuori, Nokia Research Ctr. (Finland) background area. In the proposed method, the boundary effect is removed by assigning different QP values to the macroblocks in the We present an effective method for comparing subjective audiovisual foreground area. And FMO (Flexible Macroblock Ordering) is used for quality and the features related to the quality changes of different video transmitting the region information. Therefore, the proposed method cameras. Both quantitative estimations of overall quality and qualitative has two advantages over conventional regional coding methods. One descriptions of critical quality features are achieved by the method. The is that the proposed method has better picture quality than the aim was to combine two image quality evaluation methods, the conventional methods by new allocation of QP value, and the other is quantitative Absolute Category Rating (ACR) method with hidden compatibility with the H.264/AVC since it uses FMO. reference removal and the qualitative Interpretation-Based Quality (IBQ) method in order to see how they complement each other in audiovisual quality estimation tasks. 26 observers estimated the audiovisual quality 6494-32, Poster Session of six different cameras, mainly mobile phone video cameras. In order to achieve an efficient subjective estimation of audiovisual quality, only two Quality evaluation of the halftone by halftoning algorithm- contents with different quality requirements were recorded with each based methods and adaptive method camera. The results show that the subjectively important quality X. Wan, D. Xie, Wuhan Univ. features were more related to the overall estimations of cameras’ visual video quality than to the features related to sound. The data Digital halftoning algorithm is a operation, converting the captured con- demonstrated two significant quality dimensions related to visual tone images to the corresponding binary images supported by most quality: darkness and sharpness. We conclude that the qualitative output devices, which makes the tow kinds of images similar as methodology can complement quantitative quality estimations also with possible. In order to evaluate the halftoning algorithms and the audiovisual material. The IBQ approach is valuable especially, when the corresponding halftones, a criterion must be needed. In the literature, induced quality changes are multidimensional.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 101 Conference 6494: Image Quality and System Performance IV

6494-03, Session 1 6494-05, Session 2 Modeling color quality difference for mobile displays Color differences without probit analysis J. J. Yoo, Y. J. Kim, R. M. Luo, Univ. of Leeds (United Kingdom); N. Moroney, Hewlett-Packard Co. S. Lee, D. Park, C. Kim, SAMSUNG Advanced Institute of Color science generally considers color differences from the Technology (South Korea) standpoint of distance metrics. These distance metrics are typically A series of psychophysical experiments using paired comparison experimental and are based on many paired comparisons and probit method was performed to investigate various visual attribute affecting analysis. The predominant focus is on the derivation of a uniform image quality of a mobile display. An image quality difference model metric that is optimized for small color differences around the just- was developed to show high correlation with visual results. The result noticeable difference limit. Increasingly sophisticated mathematical showed that Naturalness and Clearness are the most significant modeling is then used to fit a range of laboratory data sets. While this attributes among the perceptions. A colour quality difference model work has yielded invaluable industrial applications, it has perhaps left based on image statistics was also constructed and it was found certain aspects of color differences under explored. For example how colour difference and colour naturalness are important attributes for do non-experts typically describe color differences? What are the predicting image colour quality difference. natural language characteristics of the description of color difference? This paper considers color differences specifically from the nominal or linguistic perspective. 6494-30, Session 1 Threshold value for acceptable video quality using 6494-06, Session 2 signal-to-noise ratio Web-based versus controlled-environment M. Vaahteranoksa, Nokia Corp. (Finland) psychophysics experiments In the near future for consumers high quality still-images in the mobile S. Zuffi, C. Brambilla, Consiglio Nazionale delle Ricerche (Italy); G. phones are not enough, there has to be also a high quality video B. Beretta, Hewlett-Packard Labs.; P. Scala, Consiglio Nazionale mode available. Video quality of camcorders will probably be the delle Ricerche (Italy) reference as to what people expect at the least. Noise decreases video quality considerably, particularly in the dark environments with A recent trend in psychophysics experiments related to image quality small pixel size. In the video, noise can be seen as an unwanted is to perform the experiments on the World Wide Web with a large spatial or temporal variation in pixel values. number of observers instead of in a laboratory under controlled The object of the study was to find a threshold value for signal-to- conditions. This method assumes that the large number of noise ratio (SNR) when video quality is perceived to be good enough. participants involved in a Web investigation “averages out” the Different illumination levels were studied using subjective testing and parameters that the experiments would require to keep fixed in the objective SNR measurements. same experiment performed, following a traditional approach, under controlled conditions. We present the results of two experiments we Five camcorders were studied. Camcorders were selected to cover have conducted to assess the minimum value of color contrast to different sensor technologies, recording formats and price categories. ensure readability. The test material for the subjective test was recorded in an The readability test consisted in a character counting task. The first environment simulator, where it was possible to adjust lighting levels. experiment was conducted in a controlled environment designed to The studied illuminance levels were from 0 to 1000 lux. Illumination match the sRGB specification. The second was performed on the levels were selected to cover indoor scenarios. The video content Web. included, e.g. small details (textiles and coils of wool strings), memory colors (fruits and baby face) and uniform field (gray walls). In order to The Web experiment and the experiment conducted in the laboratory generate movement on the video, a back and forth moving fan was yielded the same results in terms of readability and indicate that a used. Also randomly moving strings were set in front of the fan. suitable threshold in lightness difference for readability could be about 27 CIELAB units. Ten second test stimuli were presented in PAL/VGA resolution on a calibrated 21" Sony FD Trinitron CRT computer monitor. Interlacing The comparison of the times employed to perform the counting task was removed. Double staircase test was used as the test method in the two experiments separately for each lightness difference do not because it is a relatively efficient test in finding threshold values. The indicate any significant difference, suggesting that Web experiments test persons were semi-experts who were used to working with image could be a good strategy to perform studies related to visual tasks quality. 14 test subjects took part in the test. and that they could very well complement, if not replace, experiments conducted in laboratory under controlled conditions. The test videos for objective measurements were recorded using an ISO 15739 based environment. The setup for transmissive test chart was selected because by using it, it was possible to reach a wide 6494-07, Session 3 luminance range. Uniform surface of the diffuser was used as the test chart. SNR was calculated by dividing the mean signal value by total Video quality assesment using M-SVD noise. Calculated SNR was average of several frames. The luminance P. Tao, A. M. Eskicioglu, The City Univ. of New York levels correspond with the illuminance levels used in the subjective test material. Objective video quality measures play an important role in a variety of In this study the correlation between objective and subjective video processing applications ranging from lossy compression to measurements were found between measured SNR and perceived printing. The quality measures in the literature can be classified into quality. According to results good enough video quality was reached two groups: Subjective and objective. There is an increasing need to between SNR values of 16.8 dB and 17.3 dB. Canon MVX250i good develop an objective quality measure that is able to predict the enough video quality was reached at the SNR level of 15.3 dB. 3CCD perceived video quality automatically. and super HAD-CCD technologies video was brighter, less noisy, and In this paper, we present a new objective video quality assessment SNR was better in low light conditions than conventional CCDs. algorithm, M-SVD, which can express the quality of distorted video Arranging extensive subjective tests such as the one conducted in sequence numerically or graphically. Based on the singular value this study, requires time, resources, and several test persons. decomposition, we designed and tested three models to obtain the However the threshold value found in this research can be used to correlation with subjective evaluation. In the first model, M-SVD/Y, evaluate the subjective performance of videos in low light scenarious only the luminance Y component is used in the computations. In the based on SNR measurements. A certain amount of subjective testing, second model, denoted by M-SVD/YCbCr, the frame level quality e.g. short expert evaluation, is still recommended.

102 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6494: Image Quality and System Performance IV error measure is a weighted summation of quality error indexes of Y, Cb 6494-10, Session 4 and Cr components, each making a contribution with an assigned weighting factor. In the third model, namely M-SVD/Edge Detection, an Image quality and automatic color equalization edge function is used to detect edges, which returns a binary image M. Chambah, Univ. de Reims Champagne-Ardenne (France); A. containing 1’s where edges are found and 0’s elsewhere. This output binary image allows us to assign an edge index to each block within a Rizzi, Univ. degli Studi di Milano (Italy); C. Saint-Jean, Univ. de La frame. Those blocks in an image that correspond to object boundaries Rochelle (France) receive higher edge index. In the professional movie field, image quality is mainly judged visually. The proposed three models are applied to the VQEG Phase I test In fact, experts and technicians judge and determine the quality of the dataset for FR-TV video quality assessment. We follow the procedures film images during the calibration (post production) process. As a employed in the VQEG Phase I objective video quality model test plan consequence, the quality of a restored movie is also estimated to evaluate the performance of our system. Four metrics are used in the subjectively by experts. evaluation of objective results. Metric 1 is the correlation coefficient On the other hand, objective quality metrics do not necessarily between objective and subjective scores after variance-weighted correlate well with perceived quality. Moreover, some measures regression analysis, including a test of significance of the difference. assume that there exists a reference in the form of an “original” to Metric 2 is the correlation coefficient between objective and subjective compare to, which prevents their use in digital restoration field, where scores after non-linear regression analysis. Metric 3 is the Spearman often there is no reference to compare to. That is why subjective rank-order correlation coefficient between the objective and subjective evaluation is the most used and most efficient approach up to now. scores. Metric 4 is outlier ratio of “outlier-points” to total number of points. But subjective assessment is expensive, time consuming and does not respond, hence, to the economic requirements of the field. Thus, reliable automatic methods for visual quality assessment are needed 6494-08, Session 3 in the field of digital film restoration. Ideally, a quality assessment system would perceive and measure Method of estimating perceived video quality for mobile image or video impairments just like a human being. The ACE multimedia application based on full reference framework method, for Automatic Color Equalization, is an algorithm for digital O. Sugimoto, S. Sakazawa, A. Koike, KDDI R&D Labs. (Japan) images unsupervised enhancement. Like our vision system ACE is able to adapt to widely varying lighting conditions, and to extract The authors study a method of estimating perceived picture quality for visual information from the environment efficaciously. We present in multimedia applications based on the full reference framework. Since this paper is the use of ACE as a basis of a reference free image multimedia applications usually have less capability in display and quality metric. ACE output is an estimate of our visual perception of a communication channels than television broadcasting, an objective scene. The assumption, tested in other papers, is that ACE enhancing quality model should be developed considering a low-resolution, low- images is in the way our vision system will perceive them, increases frame-rate video format and low-bit-rate video coding that applies a their overall perceived quality. The basic idea proposed in this paper, high compression ratio. The proposed method therefore applies the is that ACE output can differ from the input more or less according to blockiness of the picture, time variance of MSE and temporal PSNR the visual quality of the input image In other word, an image appears degradation as indices of objective picture quality. Computer simulation good if it is near to the visual appearance we (estimate to) have of it. shows that the proposed method can estimate perceived picture quality Reversely bad quality images will need “more filtering”. Test and at a correlation coefficient of above 0.94. results are presented.

6494-09, Session 4 6494-11, Session 4 Performance evaluation of digital still camera image A unified framework for physical print quality processing pipelines A. Eid, B. E. Cooper, Lexmark International, Inc. D. W. Hertel, E. T. Chang, L. C. Shih, Cypress Semiconductor In this paper we present a unified framework for physical print quality. Corp.; J. Sproul, Boston Scientific Corp. This framework includes a design for a testbed, testing methodologies Although its lens and image sensor fundamentally limit a digital still and quality measures of physical print characteristics. An automatic camera’s imaging performance, image processing can significantly belt fed flatbed scanning system is calibrated to acquire L* data for a improve the perceived quality of the output images. A well-designed wide range of flat field imagery. Testing methodologies based on processing pipeline achieves a good balance between the available wavelet pre-processing and spectral/statistical analysis are designed. processing power and the image yield (the fraction of images that meet We apply the proposed framework to three common printing artifacts: a minimum quality criterion). banding, jitter, and streaking. Since these artifacts are directional, This paper describes the use of subjective and objective measurements wavelet based approaches are used to extract one artifact at a time to establish a methodology for evaluating the image quality of and filter out other artifacts. Wavelets at different levels are applied to processing pipelines. The test suite contains images both of analytical the input images in different directions to extract each artifact within test targets for objective measurements, and of scenes for subjective specified frequency bands. evaluations that cover the photospace for the intended application. Following wavelet reconstruction, images are converted into 1 D Objective image quality metrics correlating with perceived sharpness, signals describing the artifact under concern. Accurate spectral noise, and color reproduction were used to evaluate the analytical analysis using a DFT with Blackman-Harris windowing technique is images. An image quality model estimated the loss in image quality for used to extract the strength of periodic signals (banding and jitter). each metric, and the individual metrics were combined to estimate the Since streaking is an aperiodic signal, a statistical measure is used to overall image quality. The model was trained with the subjective image quantify the streaking strength. quality data. Experiments on 100 print samples scanned at 600 dpi from 10 The test images were processed through different pipelines, and the different printers show high correlation (75% 88%) between the overall objective and subjective data was assessed to identify those ranking of these samples by the proposed metrologies and experts’ image quality metrics that exhibit significant correlation with the visual ranking. perception of image quality. This methodology offers designers guidelines for effectively optimizing image quality.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 103 Conference 6494: Image Quality and System Performance IV

6494-13, Session 5 6494-15, Session 6 Information distance-based selective feature for iris image Appearance can be deceiving: using appearance models in quality measure color imaging C. S. Belcher, Y. Du, Indiana Univ.-Purdue Univ. Indianapolis G. M. Johnson, Apple Computer, Inc. and Rochester Institute of Technology Iris recognition systems have been tested to be the most accurate biometrics systems. However, poor quality images greatly affect As color imaging has evolved through the years, our toolset for accuracy of iris recognition systems. Many factors can affect the quality understanding has similarly evolved. Research in color difference of an iris image, such as blurriness, resolution, image contrast, iris equations and uniform color spaces spawned tools such as CIELAB, occlusion, and iris deformation, but blurriness is one of the most which has had tremendous success over the years. Research on significant problems for iris image acquisition. In this paper, we propose chromatic adaptation and other appearance phenomena then extended a new method to measure the blurriness of an iris image called CIELAB to form the basis of color appearance models, such as information distance based selective feature clarity measure. Different CIECAM02. Color difference equations such as CIEDE2000 evolved to from any other approach, the proposed method automatically selects reconcile weaknesses in areas of the CIELAB space. Similarly, models portions of the iris with most changing patterns to measure the level of such as S-CIELAB were developed to predict more spatially complex blurriness based on their frequency characteristics. Log-Gabor wavelet color difference calculations between images. Research in all of these is used to capture the features of the selected portions. By comparing fields is still going strong and there seems to be a trend towards the information loss from the original features to blurred versions of the unification of some of the tools, such as calculating color differences in a same features, the algorithm decides the clarity of the original iris color appearance space. Along such lines, image appearance models image. The preliminary experiment results show that this method is have been developed that attempt to combine all of the above models effective. and metric into one common framework. The goal is to allow the color imaging research to pick and choose the appropriate modeling toolset for their needs. 6494-33, Session 5 Along these lines, the iCAM image appearance model framework was Measurement-based objective metric for the printer developed to study a variety of color imaging problems. These include resolution image difference and image quality evaluations as well gamut mapping and high-dynamic range (HDR) rendering. It is important to stress that J. Hasegawa, SAMSUNG Electronics Co., Ltd. (South Korea) iCAM was not designed to be a complete color imaging solution, but rather a starting point for unifying models of color appearance, color We propose a new measurement-based objective metric for the printer difference, and spatial vision. As such the choice of model components resolution. The proposed metric takes account of the position is highly dependent on the problem being addressed. For example, with reproduction about vector graphic objects and the detection limit CIELAB it clearly evident that it is not necessary to use the associated caused by the noise. The measurement system is comprised of a newly color difference equations to have great success as a device- designed stripe chart, a scanner device and a calculation program for independent color space. Likewise, it may not be necessary to use the the contrast transfer function (CTF). The information about the noise is spatial filtering components of an image appearance model when used to calculate the detection limit of the contrast. The detection limit performing image rendering. value is calculated by the variances of reflectance values and the numbers of the pixels corresponding to the light region and the dark This paper attempts to shed some light on some of the confusions region according to the Welch test. We define the significant contrast involved with selecting the desired components for color imaging transfer function (SCTF) as the CTF subtracted by the detection limit research. The use of image appearance type models used for calculating and the significant limit of resolution (SLR) as the maximum resolution image differences, like S-CIELAB and those made be CIE TC8-02 will be value which causes the greater contrast than the detection limit. In discussed. Similarly the use of image appearance for HDR applications, addition to the stripe pattern, we introduce the line pattern for as studied by CIE TC8-08, will also be examined. As with any large evaluation of the line reproduction, the dot pattern for the dot project, the easiest way to success is in understanding and selecting the reproduction and the checker pattern for the halftone reproduction. right tool for the job. Those patterns also are evaluated by the exactly same calculation procedure and produce various aspects of the printer resolution attribute. 6494-16, Session 6 Applying and extending ISO/TC42 digital camera 6494-14, Session 6 performance standards to mobile imaging products Driving color management into the office D. R. Williams, P. D. Burns, Eastman Kodak Co. T. Newman, Canon Development Americas, Inc. There are no fundamental differences between today’s mobile telephone cameras and consumer digital still cameras that suggest many existing CIE Division 8 is trying to spread color management from the graphic ISO imaging performance standards do not apply. To the extent that they arts market into the broader office and home markets. The paper have lenses, color filter arrays, detectors, apertures, image processing, studies how the automobile industry takes technology from the and are hand held, there really are no operational or architectural professional level of automobile racing and brings it into the consumer differences. Despite this, there are currently dramatic differences in the market. The key to success seems to be to make the technology easier levels of imaging performance. These are driven by physical and to use and more robust under uncontrolled conditions. These become economic constraints. Several ISO standards for resolution, dynamic the goals for bringing technology from the professional printing market range, noise and color imaging, well established for digital consumer into the office. digital cameras, require care when applied to current generation of cell The paper details CIE work in three areas. First, CIE is starting an phone cameras. In particular, accommodation of optical flare, shading international study on office lighting for imaging. The goal is to find out non-uniformity and distortion are recommended. We offer proposals for typical spectral power distributions and illumination levels in offices. the application of existing ISO imaging performance standards to mobile The second area is the use of standard color encodings for sharing imaging products, and suggestions for extending performance standards color images. The third area is the adaptation of the visual system when to the characteristic behavior of camera phones. comparing softcopy images and hardcopy images. CIE Division 8 members hope that by applying the research results from our technical committees, color engineers will be able to help their customers get pleasing results in the uncalibrated, unmeasured, unpredictable environment that is the office workplace.

104 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6494: Image Quality and System Performance IV

6494-17, Session 6 6494-20, Session 7 Differential gloss quality scale experiment update: an Scanners for analytic metrology: the devil in the details appearance-based image quality standard initiative E. K. Zeise, NexPress Solutions, Inc.; D. R. Williams, P. D. Burns, (INCITS W1.1) Eastman Kodak Co.; W. C. Kress, Toshiba Y. S. Ng, C. Kuo, Eastman Kodak Co.; E. Maggard, Hewlett Inexpensive and easy-to-use linear and area-array scanners have Packard Co.; D. Mashtare, Xerox Corp.; O. Morris, Hewlett frequently substituted as densitometers for low-frequency (i.e., large- Packard Co. area) hard copy image metrology. A number of workers and standards As part of the US National Standard (INCITS W1.1) initiative on committees1 have in fact advocated such use. Increasingly, scanners developing an appearance-based image quality standard, the gloss are also being tasked for high spatial frequency, image microstructure uniformity sub-team has conducted a differential gloss quality scale metrology (which is usually reserved for high-performance round-robin experiment to obtain just noticeable difference (JND) microdensitometers that use microscope optics, photomultiplier tubes relating to differential gloss. Three different images that contains (PMT), and log amps). It is hard to resist the adoption of commercial off- image elements with potentially different types of differential gloss the-shelf (COTS) scanners for such use in analytic metrology (such as gloss gradient, high frequency and low frequency gloss applications, given their convenience level, their high speed, large scan difference features, colorant coverage/substrate gloss difference, etc) areas, auto-focus, discomfiting low cost, and low operator skill were selected and printed with seven types of printing technologies requirements. Is this too good to be true? Does the devil lie in the (including inkjet systems with pigment/dye colorants on various paper details? substrates, toner-based electro-photographic systems with matte and In this paper we will address characteristics of flatbed reflection glossy paper and various fusing systems/conditions) to form a wide scanners in the evaluation of geometric distortion, geometric range of differential gloss test suite. These 21 images were used in a repeatability, as well as the effects of interpolation and sampling on round-robin psychophysics experiment that consists of 210 pair analytic measurement. comparisons to obtain a just noticeable difference (JND) for differential gloss. Correlation with objective differential gloss measurements and subjective quality ruler measurements are also 6494-21, Session 8 being made. The result of this experiment will be reported. Paper roughness and the color gamut of printed images J. S. Arney, Rochester Institue of Technology 6494-18, Session 7 Common experience indicates the quality of a printed image depends W1.1 macro-uniformity and color fidelity on the choice of the paper used in the printing process. In the current R. Rasmussen, Xerox Co.; D. Gusev, NexPress Solutions, Inc. report, we have used a recently developed device called a micro- goniophotometer to examine toner on a variety of substrates fused to To address the standardization issues of perceptually based image varying degrees. The results indicate that the relationship between the quality for printing systems, ISO/IEC JTC1/SC28, the standardization printed color gamut and the topography of the substrate paper is a committee for office equipment chartered the W1.1 project with the simple one for a color electrophotographic process. If the toner is fused responsibility of drafting a proposal for a multi-part International completely to an equilibrium state with the substrate paper, then the Standard for the evaluation of printed image quality. In this paper, we toner conforms to the overall topographic features of the substrate. For will summarize the current status and technical progress of this rougher papers, the steeper topographic features are smoothed out by development effort in the areas of appearance-based quantitative the toner. The maximum achievable color gamut is limited by the evaluation of realistic macroscopic uniformity defects as well as topographic smoothness of the resulting fused surface. Of course, evaluation of the relative importance of a wide range of color achieving a fully fused surface at a competitive printing rate with a distortions on the appearance-based evaluation of color fidelity. minimum of power consumption is not always feasible. However, the only significant factor found to limit the maximum state of fusing and the ultimate achievable color gamut is the smoothness of the paper. 6494-19, Session 7 Recent progress in the development of INCITS W1.1: 6494-22, Session 8 appearance-based image quality standards for printers Investigation of two methods to quantify noise in digital T. Bouk, Eastman Kodak Co.; E. N. Dalal, Xerox Corp.; K. D. images based on the perception of the human eye Donohue, Univ. of Kentucky; S. Farnand, Rochester Institute of Technology; F. Gaykema, OCE Technologies BV (Netherlands); D. J. Kleinmann, Univ. of Applied Sciences Köln (Germany); D. Wueller, Gusev, Eastman Kodak Co.; A. Haley, Monotype Imaging; P. L. Image Engineering (Germany) Jeran, Hewlett-Packard Co.; W. C. Kress, Toshiba America Since the signal to noise measuring method as standardized in the Business Solutions; O. Martinez, Hewlett-Packard Co. (Spain); D. normative part of ISO 15739:2002(E)1 does not quantify noise in a way Mashtare, Xerox Corp.; A. L. McCarthy, Lexmark International, that matches the perception of the human eye, two alternative methods Inc.; Y. S. Ng, Eastman Kodak Co.; D. R. Rasmussen, Xerox have been investigated which may be appropriate to quantify the noise Corp.; M. Robb, Lexmark International, Inc.; E. H. Barney-Smith, perception in a physiological manner: Boise State Univ.; M. Tse, Quality Engineering Associates, Inc.; E. - the model of visual noise measurement proposed by Hung et al2 (as K. Zeise, Eastman Kodak Co.; S. Zoltner, Xerox Corp. described in the informative annex of ISO 15739:20021) which tries to To address the standardization issues of perceptually based image simulate the process of human vision by using the opponent space and quality for printing systems, ISO/IEC JTC1/SC28, the standardization contrast sensitivity functions and uses the CIEL*u*v*1976 colour space committee for office equipment chartered the W1.1 project with the for the determination of a so called visual noise value. responsibility of drafting a proposal for a multi-part International - The S-CIELab model and CIEDE2000 colour difference proposed by Standard for the evaluation of printed image quality. In this paper, we Fairchild et al3 which simulates human vision approximately the same will summarize the current status and technical progress of this way as Hung et al2 but uses an image comparison afterwards based on development effort in the areas of appearance-based evaluation of CIEDE2000. microscopic uniformity defects, development of a meaningful With a psychophysical experiment based on just noticeable difference appearance-based measure of effective resolution, and we will (JND), threshold images could be defined, with which the two examine the relative importance and utility of text reproduction approaches mentioned above were tested. The assumption is that if the characteristics in the appearance-based evaluation of text quality. method is valid, the different threshold images should get the same ‘noise value’.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 105 Conference 6494: Image Quality and System Performance IV

Both models, are reliable to quantify at least the JND for noise in proposed in this paper includes a simple model of human visual uniform areas of digital images. While the visual noise measurement masking characteristics to improve theoretical defect hiding model can only evaluate uniform colour patches in images, the S- effectiveness. A visual experiment indicated all three methods CIELab model can be used on images with spatial content, but in this provided significant improvement over uncompensated sub-pixel case the “noise value” varies depending on the contrast of the spatial defects across all color patches and images tested. The masking- content. based method and an empirically optimized PSF method were more effective due to the masking-type patterns generated. Hiding effectiveness was linearly related to the inverse of the lightness error 6494-23, Session 8 generated by a defect. For moderate lightness errors, both the PSF and masking-based methods completely hid the sub-pixel defects, The effective pictorial information capacity of JPEG 6b as with decreasing effectiveness for larger lightness errors. Similar results an image quality metric were found for images and corresponding color patches, though R. B. Jenkin, Cranfield Univ. (United Kingdom); S. Triantaphillidou, some dependency on the image content was observed for two of the Univ. of Westminster (United Kingdom); M. A. Richardson, five images. With the addition of a simple visual masking effects Cranfield Univ. (United Kingdom) model, the iCAM Image Difference Model was found to predict the general performance trends of the three methods with reasonable The measurement of the MTF of JPEG 6b and similar compression accuracy. systems has remained challenging due to their non-linear and non- stationary nature. Previous work has shown that it is possible to estimate the effective MTF of the system by calculating an ‘average’ 6494-27, Session 9 MTF using a noise based technique. This measurement essentially provides an approximation of the linear portion of the MTF and has Scanner motion error detection and correction been argued as being representative of the global pictorial effect of the L. C. Cui, Lexmark International, Inc. compression system. By design, a typical desktop scanner scans a document, line by line. This paper presents work that calculates an effective point spread The final image is formed by stacking up each scan line assuming function for the compression system by utilizing the derived MTFs for ideal relative motion of document to the sensor. Therefore, in addition JPEG 6b. These PSFs are then combined with estimates of the noise in to scanner lens distortion, motion distortion or non-ideal motion can the compression system to yield an estimate for the Effective Pictorial create scan distortion and artifacts. In particular, when a paper Information Capacity (EPIC) of the system. Further modifications are document is fed via an automatic document feeder at high speed, made to the calculations to allow for the size and viewing distances of such motion errors can be large and may cause unpleasant visible the images to yield a simple image quality metric. The quality metric is artifacts in the form of jaggy oblique edges, uneven compressed compared with previous data generated by Ford using Bartens’ Square horizontal lines, unpleasant moiré patterns, local color mis- Root Integral with Noise and Jacobson and Topfers’ Perceived registration, and etc. Such errors are ubiquitous to most document Information Capacity. The approach is further tested using categorical scanners available on the market and not well studied and scaling. Despite its simplicity, EPIC is shown to provide credible documented. In this paper, we report a method to measure and correlation. characterize scanner motion errors via a diagonal line across the page. We showed that the motion profile can be extracted by characterizing the errors of the diagonal line. Examples of actual 6494-24, Session 8 motion error data are also presented based on actual measurement. Objective video quality assessment method for evaluating We further demonstrated the principle and method to correct such effects of freeze distortion in arbitrary video scenes errors via software based on the measured motion profile. Such correction is also possible to be included in a document scanner to K. Watanabe, J. Okamoto, T. Kurita, Nippon Telegraph and provide motion-error-free document scanning. Telephone Corp. (Japan) With the development of the broadband IP networks, video distribution, streaming, and communication services over networks are rapidly becoming common. To provide these services appropriately, we must design, monitor, and manage their quality based on subjective video quality. Therefore, we need an objective quality assessment method that enables video quality to be evaluated from video characteristics easily and quickly. We have already proposed an objective video quality assessment method for freeze distortion in network video services by considering the perceptual characteristics of short freeze distortion. In this previous method, we derived the objective video quality under the assumption that all freeze lengths in the video are constant. However, it was impossible to derive objective video quality when arbitrary freeze lengths occur because of network degradation in real services. We propose an objective video quality assessment method for arbitrary freeze lengths by extending the previous method and confirm that it can estimate subjective video quality with good accuracy for real applications. The correlation coefficient between subjective and objective quality is 0.94.

6494-26, Session 9 Evaluation of algorithms for visual masking of defective LCD sub-pixels J. W. Stellbrink, Hewlett-Packard Co. The potential use of vision-based algorithms for hiding defective display pixels is quite appealing. Two prior approaches utilized either the point spread function (PSF) or contrast sensitivity functions to represent effects of the human visual system. A third approach

106 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6495: Visualization and Data Analysis 2007 Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6495 Visualization and Data Analysis 2007

6495-01, Session 1 perhaps only viewed for a moment and thrown away) during the solution of analysis tasks. This paper discusses what fraction of these A visualization method for ontology based distance visualization instances are worth keeping and why, and argues that measure on relation network keeping more (or all) visualization instances has high value and very low cost. Even if a small fraction is retained the result over time is still S. Cho, J. Park, Information and Communications Univ. (South a large number of visualization instances and the issue remains, how Korea) can users utilize them? This paper describes what new functionality Relation network is constructed by discovering relations between users need to utilize all those visualization instances, illustrated by objects. Discovering relations is challenging and usually time examples using an information workspace tool based on zoomable consuming job. For example, most relation in protein-protein user interface principles. The paper concludes with a concise set of interaction networks has been discovered one by one empirically. principles for future analysis tools that utilize spatial organization of large numbers of visualization instances. However, if we know some objects have similar functions, we can make inference of the relationship between objects. And these inferences can avoid false trial and errors in discovering relations. 6495-04, Session 1 Ontology is a structured representation of conceptual knowledge. This hierarchical knowledge can be applied at inference of relation A visual analysis of multi-attribute data using pixel matrix between objects. displays Objects with similar functions share similar ontology terms. M. C. Hao, U. Dayal, Hewlett-Packard Labs.; D. C. Keim, T. Therefore, combining relation network with ontology makes it possible Schreck, Univ. Konstanz (Germany) to reflect that kind of knowledge and we can infer unknown relations. Charts and tables are commonly used to visually analyze data. These In this paper, we propose a visualization method in 3D space, to graphics are simple and easy to understand, but charts show only examine specific relation network based on a proper ontology highly aggregated data and present only a limited number of data structure. To gather related ontology terms, we added a degree of values while tables often show too many data values. As a freedom to conventional layered drawing algorithm so that the consequence, these graphics may either lose or obscure important position of the term in an ontology tree can move like a mobile. And information, so different techniques are required to monitor complex we combined it with modified spring embedder model to map relation datasets. Users need more powerful visualization techniques to digest network onto the ontology tree. We have used protein-protein and compare detailed multi-attribute data to analyze the health of interaction data from Ubiquitination Information System for relation their business. This paper proposes an innovative solution based on network, and Gene Ontology for ontology structure. The proposed the use of pixel-matrix displays to represent transaction-level method lays out the protein relation data in 3D space with a information. With pixel-matrices, users can visualize areas of meaningful distance measure. Finally, we have designed experiments importance at a glance, a capability not provided by common charting to verify the relationship between Euclidean distance of each protein techniques. We present our solutions to use colored pixel-matrices in and existence of interaction. The results support that our method (1) charts for visualizing data patterns and discovering exceptions, (2) provides a means to discover new relation based on visualization. tables for visualizing correlations and finding root-causes, and (3) time series for visualizing the evolution of long-running transactions. The solutions have been applied with success to product sales, Internet 6495-02, Session 1 network performance analysis, and service contract applications Incorporating highlighting animations into static demonstrating the benefits of our method over conventional graphics. visualizations The method is especially useful when detailed information is a key part of the analysis. J. Woodring, H. Shen, The Ohio State Univ. Rendering a lot of data results in cluttered visualizations. It is difficult 6495-05, Session 2 for a user to find regions of interest from contextual data especially when occlusion is considered. We incorporate animations into An interactive visual exploration tool for Northern visualization by adding positional motion and opacity change as a California’s water-monitoring system highlighting mechanism. By leveraging our knowledge on motion perception, we can help a user to visually filter out her selected data J. Sreevalsan-Nair, Univ. of California/Davis; E. van by rendering it with animation. Our framework of adding animation is Nieuwenhuyse, Bureau of Reclamation; I. Hotz, Univ. of California/ the animation transfer function, where it provides a mapping from Davis; L. Linsen, Ernst Moritz Arndt Univ. Greifswald (Germany); data and animation frame index to a changing visual property. The B. Hamann, Univ. of California/Davis animation transfer function describes animations for user selected regions of interest. In addition to our framework, we explain the The water monitoring network in Northern California provides us with implementation of animations as a modification of the rendering an integrated flow and water-quality dataset of the Sacramento-San pipeline. The animation rendering pipeline allows us to easily Joaquin Delta, the reservoirs, and the two main rivers feeding the incorporate animations into existing software and hardware based Delta, namely the Sacramento and the San Joaquin rivers. volume renderers. Understanding the dynamics and complex interactions among the components of this large water supply system and how they affect the water quality, and ecological conditions for fish and wildlife requires 6495-03, Session 1 the assimilation of large amounts of data. As a solution to this challenge, we have developed a multivariate, time-series data Re-using millions of visualizations visualization tool which encompasses various components of the system, in a geographical context. The visualization gives an abstract R. D. Rimey, D. S. Bolme, Lockheed Martin Corp. representation of the water system, which uses various information Our goal is to enable an individual analyst to utilize and benefit from visualization techniques, like focus+context techniques, graph millions of visualization instances created by a community of analysts. representation, 3D glyphs, and color-mapping, to visualize time-series A visualization instance is the combination of a specific set of data data of multiple parameters. The tool allows the user to control the and a specific configuration of a visualization providing a visual number of variables one can see at a time, and the visibility of various depiction of that data. As the variety and number of visualization components, like, the Delta, the rivers, the reservoirs, and the flows in techniques and tools continues to increase, and as users increasingly the Delta. The unique contribution of this work is a flexible adopt these tools, more visualization instances will be created (today, visualization tool tailored to the needs of scientists whose focus is to

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 107 Conference 6495: Visualization and Data Analysis 2007 study the ecological impact of the flows in the water system in present a prototype of a system that implements these new ideas. Northern California. This paper describes work in progress.

6495-06, Session 2 6495-29, Session 3 PerfViz: a visualization tool for analyzing, exploring, and Visual analytics and knowledge visualization comparing storage controller performance data W. Ribarsky, The Univ. of North Carolina at Charlotte A. P. Sawant, North Carolina State Univ.; M. Vanninen, Network No abstract available Appliance, Inc.; C. G. Healey, North Caroline State Univ. This paper presents a technique that allows viewers to visually analyze, explore, and compare a storage controller’s performance. We 6495-09, Session 4 present an algorithm that visualizes storage controller’s performance A study on multiple views for tree visualization metrics along a traditional 2D grid or a linear space-filling spiral. We use graphical “glyphs” (simple geometric objects) that vary in color, S. T. Teoh, San José State Univ. spatial placement and texture properties to represent the attribute values contained in a data element. When shown together, the glyphs Different visual representations of a tree can provide different views of form visual patterns that support exploration, facilitate discovery of the same data, leading the viewer to obtain different information, and gain different knowledge. Any visual representation of a tree, data characteristics, relationships, and highlight trends and exceptions. therefore, may potentially obscure some important aspects of the data, or sometimes even mislead the user. We create Tree-Panels, a We identified four important goals for our project: tree visualization system that provides four simultaneous 1. Design a graphical glyph that supports flexibility in its placement, visualizations of a tree. Our user study shows that different tree and in its ability to represent multidimensional data elements. representations used in Tree-Panels can uncover different and complementary information about the data. 2. Build an effective visualization technique that uses glyphs to represent the results gathered from running different tests on the storage controllers by varying their performance parameters. 6495-10, Session 4 3. Build an effective representation to compare the performance of storage controller(s) during different time intervals. Task-based evaluation of multirelational 3D and standard 4. Work with domain experts to select properties of storage controller 2D parallel coordinates performance data that are most useful to visualize. C. Forsell, Uppsala Univ. (Sweden); J. Johansson, Linköpings Univ. (Sweden) 6495-07, Session 2 Multivariate data sets exist in a wide variety of fields and parallel coordinates visualizations are commonly used for analysing such Supporting interactive graph exploration using edge data. This paper presents a usability evaluation where we compare plucking three types of parallel coordinates visualization for exploratory N. Wong, M. S. T. Carpendale, Univ. of Calgary (Canada) analysis of multivariate data. We use a standard parallel coordinates display with manual permutation of axes, a standard parallel Excessive edge density in graphs can cause serious readability coordinates display with automatic permutation of axes, and a multi- issues, which in turn can make the graphs difficult to understand or relational 3D parallel coordinates display with manual permutation of even misleading. Recently, we introduced the idea of providing tools axes. We investigate whether a 3D layout showing more relations that offer interactive edge bending as a method by which edge simultaneously, but distorted by perspective effects, is advantageous congestion can be disambiguated. We extend this direction, when compared with a standard 2D layout. The evaluation is presenting a new tool, Edge Plucking, which offers new interactive accomplished by means of an experiment comparing performance methods to clarify node-edge relationships. Edge Plucking expands differences for a class of task known to be well-supported by parallel the number of situations in which interactive graph exploration tools coordinates. Two levels of difficulty of the task are used and both can be used to address edge congestion. In this paper, we discuss require the user to find relationships between variables in a the design criteria and algorithm of Edge Plucking. We also provide multivariate data set. Our results show that for the manual exploration examples and a case study explaining how Edge Plucking can be of a complex interrelated multivariate data set, the user performance used to address edge congestion issues in graphs. with multi-relational 3D parallel coordinates is significantly faster. In simpler tasks, however, the difference is negligible. The study adds to the body of work examining the utility of 3D representations and what 6495-08, Session 2 properties of structure in 3D space can be successfully used in 3D representations of multivariate data. Exploratory simulation for astrophysics R. Walker, P. Kenny, J. Miao, Univ. of Kent at Canterbury (United Kingdom) 6495-11, Session 4 Exploratory simulation involves the combination of computational Evaluation of mesh simplification algorithms using steering and visualization at interactive speeds. This presents a PolyMeCo: a case study number of challenges for large scientific data sets, such as those from S. Silva, Instituto de Engenharia Electrónica e Telemática de astrophysics. A computational model is required such that steering Aveiro (Portugal); F. Silva, Univ. da Beira Interior (Portugal); J. the simulation while in progress is both physically valid and scientifically useful. Effective and appropriate visualization and Madeira, B. Sousa Santos, Instituto de Engenharia Electrónica e feedback methods are needed to facilitate the discovery process. Telemática de Aveiro (Portugal) Smoothed Particle Hydrodynamics (SPH) techniques are of interest in Polygonal meshes are used in many application scenarios. Often the the area of Computational Fluid Dynamics (CFD), notably for the generated meshes are too complex not allowing proper interaction, simulation of astrophysical phenomena in areas such as star visualization or transmission through a network. To tackle this formation and evolution. This paper discusses the issues involved problem, simplification methods can be used to generate less with creating an exploratory simulation environment for SPH. We complex versions of those meshes. introduce the concepts of painting and simulation trails as a novel solution to the competing concerns of interactivity and accuracy, and For this purpose many methods have been proposed in the literature and it is of paramount importance that each new method be

108 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6495: Visualization and Data Analysis 2007 compared with its predecessors, thus allowing quality assessment of Chunking has been demonstrated to improve the typical access to the solution it provides. This systematic evaluation of each new such data, but it requires a complete re-organization of the file and method requires tools which provide all the necessary features sometimes efficient access is only achieved if multiple different (ranging from quality measures to visualization methods) to help users chunking organizations are maintained simultaneously. Our approach gain greater insight into the data. can be thought of as on-the-fly chunking, but it does not require This article presents the comparison of two simplification algorithms, physical re-organization of the data or multiple copies with different NSA and QSlim, using PolyMeCo, a tool which enhances the way formats. We also describe an implementation of our technique and users perform mesh analysis and comparison, by providing an provide some performance results that are very promising. environment where several visualization options are available and can be used in a coordinated way. 6495-15, Session 6 6495-12, Session 5 NeuroVis: combining dimensional stacking, and pixelization to visually explore, analyze, and mine Situation analysis at a digital situation table with Fovea- multidimensional multivariate data Tablett J. T. Langton, Brandeis Univ. E. Peinsipp-Byma, R. Eck, N. Rehfeld, J. Geisler, Fraunhofer The combination of pixelization and dimensional stacking uniquely Institut für Informations-und Datenverarbeitung (Germany) facilitates the visualization and analysis of large, multidimensional A digital situation table which allows a team of experts to databases. Pixelization is the mapping of each data point in some set cooperatively analyze a situation has been developed. It is based on a to a pixel in a 2D image. Dimensional stacking is a layout method horizontal backward projected work table providing a general where N dimensions are projected onto the axes of an information overview of the situation. Tablet PCs referenced precisely to the scene display. We have combined and expanded upon both methods in an image using a digital image processing algorithm display a detailed application named NeuroVis that supports interactive, visual data view of a local area of the image. In this way a see-through effect mining. Users can spontaneously perform ad hoc queries that color the providing high local resolution at the position of the tablet PC is visualization, cluster the results through dimension reordering, and established. Additional information not fitting the bird’s eye view of the execute analyses on selected pixels. While NeuroVis is not intrinsically work table is displayed on a vertical screen. All output devices restricted to any particular database, it is named after its original (horizontal work table and vertical screen) can be controlled using function: the examination of a vast neuroscience database. Images tablet PCs where each team member has his own tablet PC. An produced from its approaches have now appeared in the Journal of interaction paradigm has been developed allowing each team Neurophysiology \cite{taylor06:structure} and NeuroVis itself is being member to interact with a high degree of freedom and ensuring used for educational purposes in neuroscience classes at Emory cooperative teamwork. University. In this paper we detail the theoretical foundations of NeuroVis, the interaction techniques it supports, an informal evaluation of how it has been used in neuroscience investigations, and a 6495-13, Session 5 generalization of its utility and limitations in other domains. Parallel unstructured volume rendering in ParaView K. Moreland, Sandia National Labs.; L. Avila, Kitware Inc.; L. A. 6495-16, Session 6 Fisk, Sandia National Labs. A modular extensible visualization system architecture for ParaView is a popular open-source general-purpose scientific culled prioritized data streaming visualization application. One of the many visualization tools available J. P. Ahrens, N. Desai, P. S. McCormick, Los Alamos National Lab.; within ParaView is the volume rendering of unstructured meshes. K. Martin, Kitware Inc.; J. Woodring, The Ohio State Univ. Volume rendering is a technique that renders a mesh as a translucent solid, thereby allowing the user to see every point in three- Massive dataset sizes can make visualization difficult or impossible. dimensional space simultaneously. Because volume rendering is One solution to this problem is to divide a dataset into smaller pieces computationally intensive, ParaView now employs a unique parallel and then stream these pieces through memory, running algorithms on rendering algorithm to speed the processes. The parallel rendering each piece. This paper presents a modular data-flow visualization algorithm is very flexible. It works equally well for both volumes and system architecture for culling and prioritized data streaming. This surfaces, and can properly render the intersection of a volume and streaming architecture improves program performance both by opaque polygonal surfaces. The parallel rendering algorithm can also discarding pieces of the input dataset that are not required to complete render images for tiled displays. In this paper, we explore the the visualization, and by prioritizing the ones that are. The system implementation of parallel unstructured volume rendering in ParaView. supports a wide variety of culling and prioritization techniques, including those based on data value, spatial constraints, and occlusion This work was done in part at Sandia National Laboratories. Sandia is tests. Prioritization ensures that pieces are processed and displayed a multiprogram laboratory operated by Sandia Corporation, a progressively based on an estimate of their contribution to the resulting Lockheed Martin Company, for the United States Department of image. Using prioritized ordering, the architecture presents a Energy’’s National Nuclear Security Administration under contract DE- progressively rendered result in a significantly shorter time than a AC04-94AL85000. standard visualization architecture. The design is modular, such that each module in a user-defined data-flow visualization program can cull 6495-14, Session 6 pieces as well as contribute to the final processing order of pieces. In addition, the design is extensible, providing an interface for the Spatial prefetching for out-of-core visualization of addition of user-defined culling and prioritization techniques to new or multidimensional data existing visualization modules. D. R. Lipsa, Armstrong Atlantic State Univ.; P. J. Rhodes, Univ. of Mississippi; R. D. Bergeron, T. M. Sparr, Univ. of New Hampshire 6495-28, Session 7 In this paper we propose a technique called storage-aware spatial Maps as visualization systems prefetching that can provide significant performance improvements for out-of-core visualization. This approach is motivated by file D. Rumsey, Luna Imaging chunking in which a multidimensional data file is reorganized into The technology of mapping has changed radically not only with the multidimensional sub-blocks that are stored linearly in the file. This advent of greater computing power, but equally important, the advent increases the likelihood that data close in the n-dimensional volume of the Web. Maps are no longer static items, limited to print-on- paper. represented by the file will be closer together in the physical file. Now they are dynamic displays of geospatial and temporal information,

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 109 Conference 6495: Visualization and Data Analysis 2007

assemblies of complex layers of data and image capable of We introduce an intuitive and effective novel approach for projection- representing all four dimensions simultaneously. With the introduction based similarity visualization for interactive discrimination analysis, of free, Web based applications such as 3-D Google Earth, every man, data exploration, and visual evaluation of similarity space woman, and child can use the globe as armature for cultural, effectiveness. The approach is based on the convex hull metaphor for educational, or personal information. Search engines are increasing aggregating sets of points in projected space, and can be used with a using maps and location filtered search results that encourage us to variety of projection techniques. The effectiveness of the approach is think geographically in our quest for information. Just as the demonstrated by application on two benchmark data sets. Statistical revolution of the printing press resulted in the use of text as the evidence supporting the validity of the hull metaphor is presented. We preeminent method of communication, so these trends in mapping advocate our approach over the standard symbol-based approach to will revolutionize the way maps are used by us and challenge the projection visualization, as it allows more effective perception of printed page as the dominant way to convey information - maps will similarity relationships and class distribution characteristics. become armatures for much of the information of our culture. Novel as these uses may seem, traditional map-making offers crucial insights into the ways these new maps will affect our lives. When 6495-19, Session 8 Lewis and Clark explored the American Northwest 200 years ago, Visualization of rule productivity in deriving nonstandard their paper map of the journey was a “location-based technology” spellings that in one glance revealed as much or more than hundreds of pages of their written journal. Their map changed forever that part of the S. Kempken, T. Pilz, W. Luther, Univ. Duisburg-Essen (Germany) word - politically, culturally, and economically - by allowing new ways of seeing, of knowing, of imagining. The same kind of impact will The paper describes several treemap techniques used to visualize occur from the location-based technologies now emerging on the among other aspects the productivity of rule sets in deriving non- Web: by displaying data spatially, people will become their own standard spellings in old German texts. The treemap visualization will explorers and mappers. help finding typical rule sequences depending on the localization of the spellings and their epoch. First evaluation results and user Drawing on his personal collection of over 150,000 historical maps as comments are displayed. well as his work with geographic information systems and image database display systems, David Rumsey will show how information of all kinds has been mapped and will be mapped in the future, 6495-20, Session 8 including his most recent work on the use of historic maps in three- dimensional spaces such as Google Earth. Wavelet-based data-hiding of DEM in the context of real- time 3D visualization 6495-17, Session 8 K. Hayat, W. Puech, Lab. d’Informatique de Robotique et de Microelectronique de Montpellier (France); G. Gesquière, Univ. de Patterns of coordination in improvise visualizations Provence (France); M. Chaumont, Lab. d’Informatique de C. Weaver, The Pennsylvania State Univ. Robotique et de Microelectronique de Montpellier (France) Visualization systems have evolved into live-design environments in The use of aerial photographs, satellite images, scanned maps and which users explore information by constructing coordinated digital elevation models necessitates the setting up of strategies for multiview visualizations rapidly and interactively. Although these the storage and visualization of these data in an interactive way. In systems provide built-in support for well-known coordinations, they order to obtain a three dimensional visualization it is necessary to map do not allow invention of novel coordinations or customization of the images, called textures, onto the terrain geometry computed with existing ones. This paper organizes 29 coordination patterns that have Digital Elevation Model (DEM). Practically, all of these informations are proven to be broadly useful for visual data analysis into six categories: stored in three different files: DEM, texture and geo-localization of the Navigation, Selection, Ordering, Containment, Mutation, and data. In this paper we propose to save all this information in a single Compound. By coupling coordination with visual abstraction, it has file for the purpose of synchronization. For this, we have developed a been possible to realize 27 of these patterns in Improvise, wavelet-based embedding method for hiding the data in a color demonstrated here with five example visualizations. Although the image. The texture images containing hidden DEM data can then be effective visual analysis space in Improvise is limited in practice by its sent from the server to a client in order to effect 3D visualization of library of implemented views, visual attributes, and data processing terrains. The embedding method is integrable with the JPEG2000 algorithms, this library can be extended in a modular fashion to coder to accommodate compression and multi-resolution encompass the wide variety of techniques coming out of the visualization. visualization community every year. Realizing new patterns in Improvise is a matter of combining modules, implementing new ones when necessary. As such, it is likely that many more useful 6495-21, Session 8 coordination patterns have yet to be discovered. Analyzing sampled terrain volumetrically with regard to error and geologic variation 6495-18, Session 8 T. Butkiewicz, R. Chang, Z. J. Wartell, W. Ribarsky, The Univ. of North Carolina at Charlotte A new metaphor for projection-based visual analysis and data exploration Most terrain models are created based on a sampling of real-world terrain, and are represented using linearly-interpolated surfaces such T. Schreck, Univ. Konstanz (Germany); C. Panse, ETH Zürich as triangulated irregular networks or digital elevation models. The (Switzerland) existing methods for the creation of such models and representations In many important application domains such as Business and of real-world terrain lack a crucial analytical consideration of factors Finance, Process Monitoring, and Security, huge volumes of complex such as the errors introduced during sampling and geological data are collected. Strong efforts are underway developing automatic variations between sample points. We present a volumetric and interactive analysis tools for mining useful information from such representation of real-world terrain in which the volume encapsulates data repositories. Many analysis algorithms require the appropriate both sampling errors and geological variations and dynamically definition of similarity between data instances for meaningful changes size based on such errors and variations. We define this clustering, classification, and retrieval, among other tasks. Projection- volume using an octree, and demonstrate that when used within based data visualization is highly interesting (a) for visual applications such as line-of-sight, the calculations are guaranteed to discrimination analysis of a data set within a given similarity space, be within a user-defined confidence level of the real-world terrain. and (b) for comparative analysis of similarity characteristics of a given data set represented under different similarity definitions.

110 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6495: Visualization and Data Analysis 2007

6495-22, Poster Session graphmotion/). We present a general discussion of games as a test- bed for empirical experiments in graph comprehension, and explain GrSim: a system for network visualization and why they might, in particular, provide a useful way to do exploratory exploration experiments. We then discuss the requirements for games that can be used and describe some individual games including those we have Y. Chiricota, Univ. du Quebec a Chicoutimi (Canada) tried on our own site. We are presenting a system dedicated to the visualization and The main part of the paper describes our own experiment in setting exploration of graphs. It is based on clustering and visualization up a graph-gaming site where we have carried out tests on the through a spring-embedder-type algorithm. A particularity of the benefits of different graph visualizations. We discuss the design of the program is the control given to the user during actual simulations. We site, describe its underlying architecture and present a promising also introduce heuristics that help increase the processing speed of initial trial on movement in graph visualization that gathered the the simulation to rapidly yield useful information relative to the results from over 70,000 played games. We then present some topological organization of graphs and the information they contain. statistical conclusions that can be drawn from the resulting data. Finally, we summarize the lessons that have been learned and discuss ideas for future work. 6495-23, Poster Session Extension of star coordinates into three dimensions 6495-26, Poster Session N. D. Cooprider, The Univ. of Utah; R. P. Burton, Brigham Young Univ. Visualizing computer lab usage at Indiana University/ Bloomington Traditional Star Coordinates displays a multi-variate data set by mapping it to two Cartesian dimensions. This technique facilitates K. Hanks, Indiana Univ.; J. E. Judd, Columbia Univ.; M. Henry, K. cluster discovery and multi-variate analysis, but binding to two Brunette, Indiana Univ. dimensions hides features of the data. Three-dimensional Star Understanding usage patterns of various university resources is Coordinates spreads out data elements to reveal features. This allows important when making budget and departmental allocations. the user more intuitive freedom to explore and process the data sets. Computer labs are one of the most highly used classrooms on Three-dimensional Star Coordinates is implemented by extending the campus. In order to best make use of them, IT professionals mus t data structures and transformation facilities of traditional Star know how the variables of platform, seat count, lab location, and Coordinates. We have given high priority to maintaining the simple, departmental association might influence usage patterns. After traditional interface. We simultaneously extend existing features, such conducting user studies and developing and getting feedback on as scaling of axes, and add new features, such as system rotation in several iterations of visualizations the client’s goals were discussed. three dimensions. These extensions and additions enhance data Key goals in this process include seeing trends over time, detailed visualization and cluster discovery. We use three examples to usage reports, aggregate data viewing, and being able to detect demonstrate the advantage of three-dimensional Star Coordinates outliers. Four visualization techniques, consisting of geospatial maps, over the traditional system. First, in an analysis of customer churn tree maps, radial maps, and spectrum maps were created to handle data, system rotation in three dimensions gives the user new insight these goals. It is evident that a number of different visualizing into the data. Second, in cluster discovery of car data, the additional techniques are needed, including static and interactive versions. dimension allows the true shape of the data to be seen more easily. Third, in a multi-variate analysis of cities, the perception of depth increases the degree to which multi-variate analysis can occur. 6495-27, Poster Session Multiple oil and gas volumetric data visualization with 6495-24, Poster Session GPU programming A user-driven interface for exploring visualizations J. C. Lin, Landmark Graphics Corp. D. H. Hepting, P. Schmiedge, Univ. of Regina (Canada) In recent years, multi-volume visualization has become an industry standard for analyzing and interpreting large surveys of seismic data. There is presently a variety of methods by which to create Advances made in computer hardware and software have moved visualizations, and many of these require a great deal of manual visualization from large, expensive visualization centers to the intervention. Even with those methods by which it is easy to create a desktop. Two of the greatest factors in achieving this have been the single visual representation, understanding the range of possible rapid performance enhancements to computer processing power and visual representations and exploring amongst them is difficult. We increasing memory capacities. In fact, computer and graphics present a generalized interface, called cogito, that permits the user to capabilities have tended to more than double each year. At the same control exploration of the visualization output of various manual tools, time, the sizes of seismic datasets have grown dramatically. all without the requirement to modify the original tool. Programming Geoscientists regularly interpret projects that exceed several within the cogito API is required to connect to each tool, but it is not gigabytes. Geoscientists need to interpret prospects quickly and onerous. efficiently, and they expect their desktop workstations and software We consider that the exploratory experience or activity is valuable, applications to be as performant as possible. Interactive, multi-volume and that it is possible to easily create this experience for standard visualization is important to rapid prospect generation. tools that do not normally permit exploration. We illustrate this Consequently, the ability to visualize and interpret multiple seismic approach with several examples from different kinds of manual and attribute volumes enhances and accelerates the interpretation interfaces and discuss the requirements of each. process by allowing geoscientists to gain a better understanding of the structural framework, reservoir characteristics, and subtle details of their data. Therefore, we analyzed seismic volume visualization and 6495-25, Poster Session defined four levels of intermixing: data, voxel, pixel, and image A method for testing graph visualisations using games intermixing. Then, we designed and implemented a framework to accomplish these four levels of intermixing. To take advantage of J. D. Bovey, P. J. Rodgers, Univ. of Kent at Canterbury (United recent advancements in programmable graphics processing units Kingdom) (GPUs), all levels of intermixing have been moved from the CPU into We describe a system that enables us to perform exploratory the GPU, with the exception of data intermixing. We developed a empirical experiments with graph visualization techniques by prototype of this framework to prove our concept. This paper incorporating them into games that can be played on an internet site, describes the four levels intermixing, framework, and prototype; it also the Graph Games Website (http://www.cs.kent.ac.uk/projects/ presents a summary of our results and comments made by geoscientists and developers who evaluated our endeavor.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 111 Conference 6496: Real-Time Image Processing 2007 Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6496 Real-Time Image Processing 2007

6496-01, Session 1 then tracked on a ground plane map. If an object being tracked separates into multiple objects, a feed-forward neural network is used Two-dimensional statistical linear discriminant analysis to classify as person or luggage. When an individual separates from for real-time robust vehicle-type recognition the luggage that they entered the scene with, an event timer begins. If they do not return within a preset amount of time, a high resolution E. A. Edirisinghe, I. Zaffar, Loughborough Univ. (United Kingdom) image of the abandoned item is taken. The final system we present is Vehicle Make and Model Recognition (MMR) techniques can provide able to detect abandonment events, and using a high resolution effective functional enhancements to security and access control camera mounted on a motorized mount, is able to take images of the systems operated solely by Automatic Number Plate Recognition luggage and flag them for a human supervisor to investigate. (ANPR) systems. Algorithms that use gradients, edge maps and line segments as significant features in MMR have been proposed in literature. However the use of such features in MMR is not reliable as 6496-24, Session 1 the accuracy of their detection depends upon the accuracy of the Camera position estimation method based on matching feature extractor used. Therefore using feature based techniques under extreme illumination conditions can be challenging. As a of top-view images for running vehicle solution to this problem we propose the use of two dimensional Linear T. Teshima, H. Saito, S. Ozawa, Keio Univ. (Japan); K. Yamamoto, Discriminant Analysis (2D-LDA), a statistical approach that maximizes T. Ihara, Mitsubishi Fuso Truck and Bus Corp. (Japan) the ratio of between-class variance to the within-class variance, in appearance based vehicle MMR. The use of 2D-LDA resolves the Camera movement estimation in real-time process is especially singularity and computational cost related shortcomings of the important for ITS, but has a search area of 6 dimensions. The straightforward use of 1D-LDA in MMR. The proposed algorithm is proposed method is based on the matching of the top-view images. tested on a database of 200 car images of 25 classes and is Two consecutive frames are converted to the top-view images based compared with the performance of a Principle Component Analysis on the hypothesis of the movement of the car. Instead of searching in (PCA) based approach. We show that the proposed 2D-LDA based 6 degrees, we use a new assumption which constrains the search approach has an identification accuracy of 90% as compared to 83% area in 1 dimension. Thus, the proposed method shows much less accuracy of the PCA based approach, for the given data set. calculate cost and more short time than usual. The top-view images are created based on the calibration information of the camera. The camera is modeled by an affine pinhole camera 6496-02, Session 1 and assumed to have the world coordinate origin at the just under the camera, and has a parallel axis against the world X axis. By using this Determination of vehicle density from traffic images at information, the input images can be converted to top-view images. day- and nighttime Thus, the estimation becomes just a registration of the 2 top-view M. Mehrubeoglu, L. McLauchlan, Texas A&M Univ. images between the consecutive frames. This process will drastically decrease the computation cost. In this paper we propose a method, based on various directional edges, that addresses the issue of vehicle differentiation in traffic density computations. The main goal of this extended work is to 6496-05, Session 2 create vehicle density history for a given road under different weather conditions and at different times of the day. Vehicle differentiation is Print-from-video: computationally efficient outlier addressed to account for connected or otherwise long vehicles, such reduction pattern filtering as trucks or tankers, to obtain more accurate vehicle counts per frame. Average vehicle size in pixels, given the magnification within Q. Peng, N. Kehtarnavaz, Univ. of Texas/Dallas the field of view for a particular camera, is used to separate regular Print-from-video can be achieved via super-resolution (SR) cars and long vehicles. techniques, which involve merging information from multiple low A separate algorithm and procedure have been developed to resolution images to generate a high resolution image. Due to determine vehicle density after dark when the headlights are turned inaccuracies of sub-pixel motion estimation and motion modeling, on. Night-time vehicle recognition utilizes template matching. The undesired artifacts or outliers are produced when using such vehicle headlights are used as matched patterns to recognize and techniques. This paper discusses the use of the direct SR approach count vehicles. The template matching technique utilizes headlight for the print-from-video application and introduces an outlier shapes as well as their high intensity in each frame to identify and reduction algorithm, named pattern filtering, as part of the super- count vehicles at night time. resolution reconstruction process. This filter exploits the alternating outlier pattern appearing along edges. Its performance is compared to The stationary traffic frames are downloaded from the internet as they another non-iterative outlier reduction algorithm in terms of are updated. The procedures are implemented in MATLAB. The results computation cost, mean square error, and qualitative visual of both night-time and day-time vehicle density decision algorithms appearance. It is shown that this algorithm in addition to be are presented in this paper. computationally efficient for deployment on digital camera platforms generates more visually acceptable outcome. 6496-03, Session 1 Dual-camera system for acquisition of high-resolution 6496-06, Session 2 images of moving targets A fast contour descriptor algorithm for supernova image J. Papon, R. W. Ives, R. P. Broussard, U.S. Naval Academy classification Video surveillance is becoming increasingly prevalent in today’s public C. R. Aragon, Lawrence Berkeley National Lab.; D. B. Aragon, areas, but as the number of cameras increases, so are the human DCA resources required to monitor them. With this in mind, we have We describe a fast contour descriptor algorithm and its application to developed a surveillance system that is able to autonomously monitor a distributed supernova detection system (the Nearby Supernova a scene for suspicious events, which for our purposes is defined as a Factory) that processes 600,000 candidate objects in 80 GB of image person leaving a piece of luggage unattended for an extended period data per night (one classification per 150 msec assuming 24-hour of time. Foreground and background items are separated initially operation). Our shape-detection algorithm reduced the number of using adaptive median filtering. Items identified as foreground are false positives generated by the supernova search pipeline by 38%

112 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6496: Real-Time Image Processing 2007 while producing no measurable impact on running time. 6496-09, Session 2 Good candidate supernovae are distinguished from poor candidates partly by their shape. Fourier descriptors are an established method of Measuring the complexity of design in real-time imaging numerically describing the shapes of object contours. We devised a software fast contour descriptor implementation for supernova candidates that R. S. Sangwan, P. A. Laplante, The Pennsylvania State Univ. meets the tight processing budget of the application. Noting that the lowest-order descriptors (F1 and F-1) convey the circular aspect of We review and describe the methodology for measuring complexity in the contour, our features are based on those terms and the total imaging systems. We then apply a new complexity measurement variance in the contour. Because the number of Fourier terms to be methodology to several sets of imaging and non-imaging code in calculated is fixed and small, the algorithm runs in linear time, rather order to compare the complexity differences between the two types of than the O(n log n) time of an FFT. Constraints on object size allow applications. further optimizations so that the total cost of producing the required contour descriptors is about 4n addition/subtraction operations, where n is the length of the contour. 6496-10, Session 3 A generic software-framework for distributed, high- 6496-07, Session 2 performance processing of multiview video D. Farin, P. H. N. de With, Technische Univ. Eindhoven Computationally scalable fast distance transform (Netherlands) computation F. M. Porikli, Mitsubishi Electric Research Labs.; T. Kocak, We propose a generic framework for distributed real-time processing of video data in a cluster of computers. Algorithm parallelization is Polytechnic Univ. achieved by splitting the processing into a set of processing tiles. We propose a fast and elegant solution to distance transformation The inputs and outputs of the processing tiles can be connected to computation, which is a popular tool to compare binary shape build arbitrary processing graphs, describing the order and representations for several pattern recognition applications. Distance dependencies of processing. The framework software takes care transformation requires finding of the minimum distance to an image about the parallelization and the control of execution, which makes point from any of the shape points in the binary mask. The minimum the parallelization transparent to the user. distance computation is repeated for all image points, thus, it is very The processing graph is not limited to a certain topology and demanding. Instead, we accomplish extraction of all minimum processing-tile outputs can be connected to the inputs of several distances without computing any distance norm by using scan-line other tiles. While the processing within one processing tile is propagation techniques. Iteratively applying scan-line evaluations in considered an atomic operation, parallelism is obtained by running varying opposite directions, we approximate the linear distance from processing tiles in parallel. The framework also allows to process data the nearest shape point. We use an integral image for further local from different time instances in parallel, such that not only horizontal refinement of the transform. In addition, the precision of our method parallelism (concurrency of independent tasks), but also vertical can be adjusted according to the computational power and parallelism (pipelining) is exploited. constraints. Our simulations show that the speed of the distance transformation application can be accelerated in the order of We further describe an example application of the framework to hundreds (thousands) in comparison to fast Voronoi maps based implement a real-time view-interpolation system for multiple input (conventional) solutions. cameras. In our setup, some computers were used for capturing, rectifying, and estimating the depth-image for two cameras each, while additional computers were used to estimate the depth between 6496-08, Session 2 cameras connected to different computers. View-interpolation and display is carried out on another computer. Quantification of IR images for sparking defects in resistance welds through segmentation and marphological processing 6496-11, Session 3 N. M. Nandhitha, N. Manoharan, B. Sheela Rani, Sathyabama Real-time stabilization of long-range observation system Deemed Univ. (India); B. Venkatraman, P. Kalyanasundaram, B. turbulent video Raj, Indira Gandhi Ctr. for Atomic Research (India) B. Fishbain, L. P. Yaroslavsky, I. A. Ideses, A. Shtern, O. Ben-Zvi, Zircalloy-2 thin walled tubes are used as cladding tubes to contain the Tel Aviv Univ. (Israel) nuclear fuel in Pressurized Heavy Water Reactors and are closed with endcaps to make fuel elements. Resistance welding is employed for This paper presents a real-time algorithm that compensates images welding these endcaps to the nuclear fuel elements. Sparking is distortion due to atmospheric turbulence in video sequences, while inherent in the resistance spot welding process. It determines the keeping the real moving objects in the video unharmed. The algorithm electrical and mechanical contact condition and thus strongly affects is based on three building blocks: Estimates of the stable scene, real the resistance spot weld quality. A practical approach to minimizing motion extraction, and generation of stabilized frames. To preserve the sparking effect is to compensate the welding current. However, real motion in the scene, the moving objects are located and the the existing methods for welding current compensation rely on results compensation for the distortion of the turbulence is applied only to from visual inspection methods or the results from NDT techniques the static areas of images. such as 100% helium leak testing, ultrasonic testing and For each pixel in the incoming frame it is needed to decide if it is of a metallography on statistical basis. This paper presents an automatic moving or a stationary object. The algorithm suggests a hierarchical spark detection and quantification algorithm with advanced image decision mechanism. The first step presents lightweight processing techniques. The work involved processing of the computational decision mechanism, and extracts most stationary thermographic image of Resistance Spot welded endcaps. Sparks in areas. The second step improves accuracy by more computationally the endcaps are then identified and quantified. The proposed complex algorithms, applying optical flow for motion segmentation. approach also compares the severity of defect with that obtained from Discriminating real motion is achieved through statistical analysis of line profile and the deviation is less than 1%. Results have been the motion field. Finally, all areas in the incoming frame, that were validated with outputs from metallography. The results have shown tagged as stationary, are replaced with an estimation of the stationary that the proposed algorithm is robust under various process scene, which are computed using fuzzy logic algorithms. conditions including both welding current and electrode force. These The restored videos exhibit excellent stability for stationary objects quantified defect features are then used in further processing of the while retaining real motion. This is achieved in real-time on standard development of intelligent on-line weld monitoring system. computer hardware.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 113 Conference 6496: Real-Time Image Processing 2007

6496-13, Session 3 processing needs to be parallelized in several threads that have least possible dependencies. Development of new image compression algorithm We consider that we have achieved a good implementation regarding (Xena) to the practical use of all the resources (and specifically the cpu Y. Sugita, A. Watanabe, Fuji Photo Film Co., Ltd. (Japan) resource). International standard technologies of JPEG_LS, JPEG2000, etc. are widely used for lossless compression of image data. However, they 6496-15, Session 4 were really not designed for handling enormous amount of data, and therefore cannot be supported by the applications which require high Hardware-based JPEG2000 video coding system compression capability and speed. We also have been dissatisfied A. Schuchter, Univ. Salzburg (Austria) with the compression speed of the software and have been pursuing of an entirely new alternative to replace the existing JPEG standard. In this paper, we mainly discuss a hardware based low complexity Our ultimate goal is the development of a technology which achieves JPEG2000 video coding system. The hardware system is based on a both maximum compression capability and faster processing speed. software simulation system, which on the software side the temporal Xena, the new technology described in this paper. redundancy is exploited by coding of differential frames which are This paper provides an overall description of new image compression arranged in an adaptive GOP structure and the GOP structure itself is technology, Xena, and the strengths in its lossless compression determined by statistical analysis of differential frames which is capability and speed in comparison to JPEG2000 and JPEG_LS, the referred to in [1]. world standards in the continuous tone image compression field. Xena has achieved an extremely high compression speed over 20 times faster than that of JPEG2000, while the compression capability 6496-16, Session 4 remains almost the same. Three-dimensional color image processing procedures using DSP 6496-23, Session 3 V. I. Ponomaryov, Instituto Politécnico Nacional (Mexico) Real-time aware rendering of scalable arbitrary shaped Image processing of the vector image information is seemed very MPEG-4 decoder for multiprocessor systems important because multichannel sensors in different applications. We introduce novel algorithms to process color images that are based on M. Pastrnak, P. H. N. de With, LogicaCMG Nederland B.V. order statistics and vectorial processing techniques: Video Adaptive (Netherlands) and Technische Univ. Eindhoven (Netherlands); J. Vector Directional (VAVDF) and the Vector Median M-type K-Nearest van Meerbergen, Philips Research Labs. (Netherlands) and Neighbour (VMMKNN) Filters presented in this paper. These Technische Univ. Eindhoven (Netherlands) algorithms apply the adaptive vector processing, directional processing and RM-estimators. It has been demonstrated that novel Arbitrary-shaped video objects are an essential part of the object- algorithms suppress effectively an impulsive noise in comparison with based video signals. The MPEG-4 video standard extends the different other methods in 3D video color sequences. The direction traditional frame-based processing towards the composition of processing in the 3D window data is defined selecting image vector several video objects superimposed on a background sprite image. In that gives the minimum angle deviation value among all the vectors. this paper we present real-time aware rendering of the individual video At final stage the vectors pass through a magnitude processing filter objects mapped on multiprocessor Network-on-Chip. The traditional where the noise detector is used finishing the VAVDF processing. worst-case approach for run-time implementation is too pessimistic in Simulation results have been obtained using video sequences “Miss terms of over-dimensioning in resource allocation and is too slow in America” and “Flowers”, which were corrupted by noise. The filters: reaction to the change of resource usage. In our previous work, we KNNF, VGVDF, VMMKNN, and, finally the proposed VAVDATM have have proposed a Quality-of-Service (QoS) based approach that been investigated. The criteria PSNR, MAE and NCD demonstrate efficiently combines the fixed-reservation of resources with the best- that the VAVDATM filter has shown the best performances in each a effort computing. However, while the fixed-reservation of resources criterion when intensity of noise is more that 7-10%. DSP that is based on predictable principles does not require addition run- implementation of the proposed filters demonstrates the possibility to time control, the best-effort enhanced modes need extension for the process the video sequence in real time mode. data availability control. The control model, the data presence availability and the synchronization aspects are presented and evaluated by arbitrary-shaped MPEG-4 decoding application. 6496-17, Session 4 Light-sensitive high-speed line scan camera with digital 6496-14, Session 4 time delay integration Real-time 3D video conference on generic hardware E. Bodenstorfer, J. Fürtler, J. Brodersen, K. J. Mayer, Austrian X. Desurmont, Multitel (Belgium) Research Ctrs. Seibersdorf Research GmbH (Austria); C. Eckel, K. Gravogl, Oregano Systems Design & Consulting GesmbH Nowadays, video-conference tends to be more and more (Austria); H. Nachtnebel, Technische Univ. Wien (Austria); P. advantageous because of the ecological cost of transport. The goal of Rössler, Fachhochschule Technikum Wien (Austria) the TIFANIS immersive platform is to let users interact as if they were physically together but using generic hardware. The basic functions of Dealing with high speed image processing systems, the speed of the system are to capture the scene, transmit it through digital operation is often limited by the amount of available light, due to short networks to other partners and then render it according to each exposure times. Therefore, high speed applications often use line partner viewing characteristics. The image processing part should run scan cameras, based on charge-coupled device (CCD) sensors with in real-time (20 fps) with low latency (50 ms). The basic functions are time delayed integration (TDI). Synchronous shift and accumulation of the acquisition, the 3D reconstruction, the detection and tracking of photoelectric charges on the CCD chip according to the objects’ the face, the sending of all this data to a remote place and finally the movement results in a longer effective exposure time without display of the data. introducing additional motion blur. The paper presents a high speed line scan camera based on a commercial complementary metal oxide We propose to analyze the whole system. We see that they can be semiconductor (CMOS) area image sensor with a Bayer filter matrix split in different services like cpu, graphical rendering, DMA and for color vision and a field programmable gate array (FPGA). The communications trough the network.Most of the processing is done camera implements a digital equivalent to the TDI effect exploited with by the cpu resource. It is composed of the 3D reconstruction and the CCD cameras. The proposed design benefits from the high frame detection and tracking faces from the video stream. However, the rates of CMOS sensors and from the possibility of arbitrarily

114 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6496: Real-Time Image Processing 2007 addressing the rows of the sensor’s pixel array. By reading out only a 6496-21, Poster Session few rows of the sensor, a digital version of TDI is implemented by shifting and accumulating the images according to the movement of Real-time quadtree analysis using HistoPyramids transportation of the inspected objects. The paper gives a detailed G. Ziegler, R. Dimitrov, C. Theobalt, H. Seidel, Max-Planck-Institut description of the digital TDI algorithm implemented on the FPGA. Relevant aspects for the practical application are discussed and key für Informatik (Germany) features of the camera are listed. Region quadtrees are convenient tools for hierarchical image analysis. Like the related Haar wavelets, they are simple to generate within a fixed calculation time. The clustering in each resolution level requires 6496-19, Poster Session only local data, yet they deliver intuitive classification results. The region quadtree partitioning is admittedly very rigid, but can be rapidly Motion object detection and tracking for traffic video computed from arbitrary imagery. This research result demonstrates analysis now how graphics hardware can be utilized to build region quadtrees at I. Rajapandi, Anna Univ. (India) unprecedented speeds. To achieve this, a data-structure called HistoPyramid sums up the number of desired image features in a Video-based surveillance systems have a wide range of applications pyramidal 2D array. Then, the newly-created HistoPyramid is traversed for traffic monitoring. In this work, the behaviour and activity detection from top down, creating lists of detected feature regions directly in in traffic videos obtained from stationary video cameras are done. GPU memory, thus making the HistoPyramid serve as an implicit Moving targets are segmented from the images and tracked in real indexing data structure, and minimizing bus transfers between CPU time. These are classified into different categories, which makes use and GPU. With this novel twist, quadtrees can be used in real-time of image features and image-sequence-based tracking results for video processing on standard PC hardware. A multitude of applications robust classification. Further the tracking and classification results are in image and video processing arises, since region quadtree analysis to be used in a programmed context to analyze behaviour of the becomes a light-weight preprocessing step for feature clustering in vehicles. vision tasks, motion vector analysis, PDE calculations, or data The system segments a traffic video to identify the moving targets and compression. Finally, we outline how this algorithm can be applied to they are classified into different categories. Initially a background 3D volume data, effectively generating region octrees. image for the video was modeled. This is done by constructing histograms for each pixel position for a group of frames. This helps to identify the stationary and slow moving vehicles. For foreground 6496-22, Poster Session segmentation background subtraction method is employed. Then shadow detection was done to eliminate shadow pixels. The Tracking object with radical color changes using modified segmented image consists of regions other than vehicles and the mean shift targets are distorted. So there is a need to enhance the segmented I. Whoang, Chung-Ang Univ. (South Korea); S. H. Chang, Spatial output. For effective object segmentation image subdivision, blob Integrated Systems, Inc.; K. Choi, Chung-Ang Univ. (South Korea) formation and region grouping are performed. This provides better results and the objects within the field of view are tracked effectively. This paper presents a new algorithm for color-based tracking of objects Then target features like width, height and area are calculated based with radical color using modified mean shift. Conventional color-based on the image coordinates. Camera calibration is done to obtain the object tracking using mean shift does not provide appropriate result world coordinates corresponding to the image coordinates. Using the when initial color distribution disappears. In this proposed algorithm, world coordinates the exact features of the target is obtained. Using mean shift analysis is first used to derive the object candidate with the target features they are classified into small, medium and heavy maximum increase of density direction from current position. Then the vehicles. Occlusion detection is performed to extract the individual proposed algorithm as the modified mean is used iteratively to update targets and then the moving vehicles are tracked effectively. Finally the object color information regardless of the object color is changed the behavior of the vehicle is analyzed from the vehicle features or not. The implementation of the new algorithm achieves effective obtained through the frame sequence. real-time tracking of objects with complete color changed by time. The subject of real-time object tracking can be described as a correspondence problem involved with object finding from frame to 6496-20, Poster Session frame in a video image sequence. Normally the time interval between Real-time speckle and impulsive noise suppression in 3D two successive frames is small and constant. Therefore the inter-frame changes can be limited and the correspondence can be created based imaging based on robust linear combinations of order on object features extracted from the image frames. However, object statistics features extraction can be very sensitive with variations of illumination, J. L. Varela-Benítez, F. J. Gallegos-Funes, V. I. Ponomaryov, view-point and scale occlusion of objects and complexity of Instituto Politécnico Nacional (Mexico) background. The appearances of object features can be changed under the working condition inconstancy. To keep tracking of an object The possibility to process 3-D images presents a new application when its features change, the important issue is to enhance and update where it is necessary to improve the quality of 3-D objects inside the object feature information from noisy and dynamic background. images, suppressing noise of different nature that always affects the In this paper, we propose the modified mean shift with color-based communication or acquisition process. Recently, we introduced the tracking algorithm resulting a robustness tracking of object with radical RM L-filters that can provide impulsive and speckle noise suppression and complete color changes. Our approach is: First, the color with small image feature preservation in 2-D image processing distribution of an object is initialized. Second, the object position is applications. These algorithms are based on the L-filter and the robust determined using color distribution in each frame using mean shift RM-estimators with different influence functions. analysis. Third, the color distribution of the object is updated using the In this paper, we present the 3-D image processing capability of the object position. In this algorithm, the proposed algorithm was used RM L-filters for the removal of impulsive and speckle noise. Extensive iteratively to update density function for tracking in order to ensure the simulation results in video sequences and ultrasound medical imaging robustness of the tracking from radical change of color distribution. have demonstrated that the proposed 3-D RM L-filters consistently Conventional tracking algorithm by using Mean shift is based on color outperform other 3-D filters by balancing the tradeoff between noise distribution. This approach is appropriate for tracking non-rigid object suppression and detail preservation. The real-time implementation of that has fixed color distribution. But, when a target object changes its proposed 3-D algorithms was realized on the DSP TMS320C6701. color distribution, Mean shift tracking algorithm could have error. In this The processing time of proposed filters includes the time of data case, the best solution is to detect change of color distribution of the acquisition, processing and store data. The processing time values target. If we detect the change, color distribution of the target has to be depend of the calculation of influence functions and distribution updated as changed that. But, this method has a problem when the functions used to calculate the coefficients of the RM L-filter. target object is non-rigid object that has a difficulty to separate from

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 115 Conference 6496: Real-Time Image Processing 2007

background. Because, target and background color distribution are 6496-27, Poster Session mixed. Therefore, we proposed modified Mean shift algorithm. The proposed algorithm step is as following: A real-time hierarchical rule-based approach for scale independent human face detection 1) Initialize search window as density function of tracking target object, volume size and color bin size. J. Jun, H. H. S. Ip, City Univ. of Hong Kong (China) 2) i ? i + 1 In this paper, we present a scale independent automatic face location i : time sequence technique which can detect the locations of frontal human faces from images. Our hierarchical approach of knowledge-based face 3) Target location is obtained by Mean shift detection composed of three levels. Level 1 consists of a simple but 4) Update search window as current density function in obtained effective eyes model that generates a set of rules to judge whether or location. not there exists a human face candidate in the current search area in a 5) Perform Mean Shift procedure again by updated search window. scale-independent manner and in a single scan of the image. To utilize this model, we define a new operator - extended projection and define From first to third step is almost same as conventional Mean shift two new concepts: single projection line and pair projection line. At procedure. Fourth step is updating search window. This step does not level 2, an improved model of Yang’s mosaic image model is applied consider that change of target occurs. This step has two cases. One is to check the consistency of visual features with respect to the human change and the other is no change of a target color distribution. If a face within each 3x3 blocks of a candidate face image. At the third change occurs, we have a little different response as compared with level, we apply a SVM based face model, to eliminate the false response of 3th step by updated search window. The other case as no positives obtained from level 2. Experimental results show the change, we would have same result as 3th step. Because updated data combined rule-based and statistical approach works well in detecting is same except background as before. frontal human faces in uncluttered scenes. Therefore, our proposed tracking system could be obtained same result as well as conventional Mean shift and also applies for radical color distribution change of object. 6496-28, Poster Session Digital architecture for real-time processing in vision 6496-25, Poster Session systems for control of traffic lights The FPGA-based real-time image contrast adaptive linear J. Garcia-Lamont, Instituto Politécnico Nacional (Mexico) enhancement algorithm Digital architecture for real time processing in vision systems for Y. Wang III, D. Wang, Y. Hu, Huazhong Univ. of Science and control of traffic lights is presented. The main idea of this work is to Technology (China) identify cars on intersections, switching traffic lights in order to reduce traffic jam. The architecture is based on a color image segmentation Image enhancement based on histogram statistics is a popular and algorithm that comprises three stages. Stage one is a color space effective method, but the computation of this method and memory for transformation in order to measure the color difference properly, histogram information storage costs too much. So, the method based image colors are represented in a modified L* u* v* color space. Stage on histogram is difficult for hardware realization. In this paper, a multi- two consists in a color reduction, where image colors are projected level mean method is presented, it can obtain and mean value of the into a small set of prototypes using a self-organizing map (SOM). histogram. They are used for the linear transform of histogram. The Stage three realizes color clustering, where simulated annealing (SA) hardware system using this algorithm doesn’t need extra RAM seeks the optimal clusters from SOM prototypes. The proposed resource. The experiments show that the system is effective, adaptive. hardware architecture is implemented in a Virtex II Pro FPGA and The implementation of the algorithm using FPGA is simple and fast. The tested; having a processing time inferior to 25ms per 128x128 pixels. algorithm presented in this paper can be applied to real-time The implementation comprises 262,479 equivalent gates. application.

6496-26, Poster Session On the use of real-time agents in distributed video analysis systems B. Lienard, X. Desurmont, Multitel (Belgium) Today’ s technologies in video analysis use state of the art systems and formalisms like onthologies and datawarehousing to handle huge amount of data generated from low-level descriptors to high-level descriptors. In the IST CARETAKER project we develop a multi- dimensional database with distributed features to add a centric data view of the scene shared between all the sensors of a network. We propose to enhance possibilities of this kind of system by delegating the intelligence to a lot of other entities, also known as “Agent” which are specialized little applications, able to walk across the network and work on dedicated sets of data related to their core domain. In other words, we can reduce, or enhance, the complexity of the analysis by adding or not feature specific agents, and processing is limited to the data concerned by the processing. This article explains how to design and develop an agent oriented systems which can be used by a video analysis datawarehousing. We also describe how this methodology can distribute the intelligence over the system, and how the system can be extended to obtain a self reasoning architecture using cooperative agents. We will demonstrate this approach.

116 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6497: Image Processing: Algorithms and Systems V Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6497 Image Processing: Algorithms and Systems V

6497-01, Session 1 Min-Max method is used for score normalization to map the raw similarity scores to a common range. We apply a weighted sum rule to Multilevel surround inhibition: a biologically inspired combine GDSM and VNC similarity metrics. We have applied this contour detector approach to a database of images with translation, rotation, and lighting distortions. The results confirm that the two metrics extract G. Papari, Rijksuniv. Groningen (Netherlands); P. Campisi, Univ. different information from the images and hence can be fused so as to degli Studi Roma Tre (Italy); N. Petkov, Rijksuniv. Groningen obtain a more robust hybrid matching system. The threshold value (Netherlands) used on the combined similarity measures to establish a match is a Canny edge detector is based both on local statistics of the image - useful means for controlling the balance between the different types the gradient - and connectivity-based global analysis, present in the of error in corner matching. hysteretic thresholding. This contribution proposes a generalization of these ideas: Instead of the sole gradient magnitude, we consider several local statistics, inspired by biological principles, to take into 6497-04, Session 1 account how much texture is present around each pixel. Global Detection of junction in images analysis is also preformed, in a more general way with respect to the Canny edge detector, by introducing a long ray connectivity analysis. Y. Li, R. L. Stevenson, J. Gai, Univ. of Notre Dame A wide range of experimental results shows the effectiveness of our This paper proposes a junction detection method that detects approach, in comparison with other existing techniques. junctions as those points where edges join or intersect. The edges that form a junction are searched in a square neighbourhood, and the subtended angles among them are calculated by using edge 6497-02, Session 1 orientations. Local edge orientation at a pixel is estimated by utilizing Stochastic resonance investigation of object detection in those edge points close to the pixel. Based on the subtended angles, images the pixel is determined to be a junction candidates or not. Each actual junction is accurately localized by suppressing the D. W. Repperger, A. R. Pinkus, Air Force Research Lab.; J. A. candidates of non-minimum orientation difference. Skipper, Wright State Univ. The proposed method analyzes real cases of extracted edges, and Object detection in images was conducted using a nonlinear means estimates the change of orientations of edge segments in digital of improving signal to noise ratio termed “stochastic resonance” (SR). fields. The experimental results show that the proposed algorithm can In a recent patent application, it was shown that arbitrarily large signal robustly detect junctions in digital images. to noise ratio gain could be realized when the object recognition problem is cast within the context of a SR problem. In this study, we investigated a number of signal-to-noise ratio measures in a binary 6497-06, Session 2 object recognition task (friendly or hostile) by perturbing the recognition algorithm and noting the results of the computer Image mosaicking via multiresolution analysis and cut simulation. To fairly test the efficacy of the proposed algorithm, a line definition unique database has been constructed by taking intensity images of L. M. G. Fonseca, V. Blagi, Instituto Nacional de Pesquisas the two sample objects adjusting their brightness, contrast and size Espaciais (Brazil) via commercial software to gradually compromise their saliency in identification. The key to the use of the SR method is to produce a Image mosaicking can be defined as the registration of two or more small perturbation in the identification algorithm and then to threshold images that are combined to form a single image. The mosaicking the results, thus improving the overall system’s ability to discern process, in general, is performed in two stages: identification of objects. A background discussion of the SR methods is presented to control points and blending the intensities of the overlapped images. relate this nonlinear filtering concept to the image enhancement The objective of the blending process is to provide a smooth community. A standard speed-accuracy test is proposed in which transition between overlapping images and to eliminate visible seams, algorithms could be fairly compared with respect to their blurring, or ghosting. Therefore, this work aims to present a blending performance. technique based on multiresolution analysis to provide a seamless mosaicking of remote sensing images. The method automatically defines the transition zone size and the cut line through the gradient. 6497-03, Session 1 The method has been tested on satellite and aerial images with satisfactory results. Combining multiple similarity metrics for corner matching H. A. Khater, Univ. of Kent at Canterbury (United Kingdom) 6497-07, Session 2 Feature point similarity metrics play an important role in digital image Color filter array interpolation based on spatial adaptivity processing where they are used for a range of applications including D. V. Paliy, Tampere Univ. of Technology (Finland); B. C. Radu, stereo matching and image registration. However, any individual Nokia Research Ctr. (Finland); V. Katkovnik, Tampere Univ. of metric has a limited applicability depending on the content of images Technology (Finland); M. Vahvilainen, Nokia Research Ctr. to be registered and the different types of distortions that may be present. In this work we propose a new technique for combining (Finland) multiple corner matching metrics to produce a more accurate and Conventional approach in single-chip digital cameras is a use of color robust assessment of corner similarity. After computing image filter arrays (CFA) in order to sample different spectral components. gradients, the Harris corner detector is used to extract suitable Demosaicing algorithms interpolate these data to complete red, corners in each of the images. We have employed two similarity green, and blue values for each image pixel, in order to produce an measures with different underlying techniques and performance RGB image. In this paper we propose a novel demosaicing algorithm characteristics. They are Gradient Direction Similarity Measure for the Bayer CFA. It is assumed that the initial estimates of color (GDSM) and Variance Normalized Correlation (VNC).We treat the channels contain two additive components: the true values of color similarity value calculated by these techniques in a manner similar to intensities and the errors. The errors are considered as an additive scores of independent classifiers tasked with assessing whether each noise, and often called as a demosaicing noise, that has to be point pair belongs to the class of ‘’correctly matched pairs’’ or not. A removed. However, this noise is not white and strongly depends on a

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 117 Conference 6497: Image Processing: Algorithms and Systems V signal. Usually, the intensity of this noise is higher near edges of about the time and location of the image. In this paper, we propose a image details. Specially designed spatially-adaptive filters are used to novel method for the cloud classification that consists of two stages remove the noise. These filters are based on the local polynomial and utilize cloud movement as human experts adopt. We firstly approximation (LPA) and the paradigm of the intersection of classify the clouds into 20 classes according to their brightnesses of confidence intervals (ICI) applied for selection adaptively varying the two-band spectral images. We then closely analyze the classes scales (window sizes) of LPA. The LPA-ICI technique is nonlinear and according to five features such as the brightnesses, deviations of spatially-adaptive with respect to the smoothness and irregularities of brightness and cloud velocity estimated by varying window size the image. The efficiency of the proposed approach is demonstrated adaptively. The experimental results are shown to verify the proposed by simulation results. method.

6497-08, Session 2 6497-11, Session 3 Hierarchical texture motifs Histogram-based template matching for object detection S. D. Newsam, Lawrence Livermore National Lab. in images with varying contrast A fundamental challenge in analyzing spatial patterns in images is the C. D. Schrider, J. A. Skipper, Wright State Univ. notion of scale. Texture based analysis typically characterizes spatial We have developed a semi-automatic target detection algorithm that patterns only at the pixel level. Such small scale analysis usually fails is relatively insensitive to image brightness and contrast. For each to capture spatial patterns that occur over larger scales. This paper evaluated image region of interest (ROI), a sliding difference method presents a novel solution, termed hierarchical texture motifs, to this of histogram comparison is performed, wherein the sum of absolute texture-of-textures problem. Starting at the pixel level, spatial patterns histogram differences of a known object template and the ROI is are characterized using parametric statistical models and calculated. A new sum is calculated as the histograms incrementally unsupervised learning. Higher levels in the hierarchy use the same slide together and the minimum sum is stored in a corresponding analysis to characterize the motifs learned at the lower levels. This response plane. Local minima in the generated surface suggest target multi-level analysis is shown to outperform single-level analysis in locations. Whereas in the ideal situation image contrast would classifying a standard set of texture images. precisely match that of the template, in real world situations, contrast could be compromised by illumination conditions, background features, cloud cover, etc. A random contrast manipulation, which we 6497-09, Session 2 term ‘wobble’, is performed on the template histogram. To reduce computation time, a predicted range of possible wobbles based on Dust and scratch removal in scanned images image specific information is extracted using a subset of training R. Bergman, H. Nachlieli, G. Ruckenstein, Hewlett-Packard Labs. images acquired for testing. Applying a predicted set of wobbled (Israel); D. Greig, Hewlett-Packard Ltd. (United Kingdom) template histograms to the ROI and selecting the minimum difference offers a higher probability of matching the image contrast than would Dust, scratches or hair on originals (prints, slides or negatives) be expected using fixed parameters. distinctly appear as light or dark artifacts on a scan. These unsightly artifacts have become a major consumer concern. This report describes an algorithmic solution to the dust and scratch removal 6497-12, Session 3 task. The removal problem is divided into two phases: a detection phase Iris identification using contourlet transform and a reconstruction phase. Some scanners have dedicated hardware R. F. Zewail, M. K. Mandal, N. G. Durdle, Univ. of Alberta (Canada) to detect dust and scratch areas in the original. Without hardware assistance, dust and scratch removal algorithms generally resort to Biometric-based authentication is a rapidly evolving field of research. blurring, at the loss of image detail. We present a software-only Among different biometrics, authentication based on iris features has solution that does not affect the blur of the overall image. For received a lot of attention since its introduction in 1992. The wavelet detection we generate a detail-less image in which the defects are transform has been proposed by several researchers for extracting iris “erased”. We compare properties of the luminance channel of the features for authentication. In order to deal with inherent limitations in input image relative to the detail-less image to generate candidate classical wavelets, several extensions have been proposed in the defective pixels. The candidates pixels are then filtered using regional literature. In this work, we investigate the potentials of using the considerations. For reconstruction of the defective areas we present contourlet transform to represent the iris texture for identity algorithms that preserve image sharpness and texture better than authentication applications. A new iris representation and matching available alternatives. system based on contourlet transform is presented. The proposed system benefits from the high directionality and anisotropy properties The algorithms successfully remove such defects that have of contourlets. The iris is first localized, and normalized to a fixed size. reasonable gray-level contrast with the background. Compared with Next, the enhanced texture is subjected to a multi-scale, multi- other software-only dust and scratch solutions the results of the directional contourlet analysis step. Statistical features are then solution described here are less blurry. computed from the contourlet subbands and stored as a feature vector. The performance of the proposed system is evaluated in both identification and verification modes. We demonstrate the feasibility 6497-10, Session 3 and robustness of the proposed system to capture local and global Cloud classification of satellite image performed in two texture information, even with low quality images. Experimental results stages also show the performance gains of using contourlets to capture iris texture information, as compared to classical wavelets. H. Ikeda, M. Matsumoto, S. Hashimoto, Waseda Univ. (Japan) To understand a comprehensive atmospheric state, it is important to 6497-13, Session 3 classify clouds in satellite images into appropriate classes. Many researches utilizing various features concerning the cloud texture have Complex sinusoidally modulated imaging for real-time been reported in cloud classification. However, some clouds can not pixel-wise optical flow detection be classified uniquely only with the texture features. According to the knowledge of the experts, they classify the clouds in two stages. They D. Wei, P. Masurel, T. Kurihara, A. Shigeru, The Univ. of Tokyo firstly categorize the clouds into the provisional classes utilizing the (Japan) brightnesses of the satellite images. They then classify each In this paper, we study optical flow determination with complex provisional class into the objective class based on the texture, shape sinusoidally-modulated imaging using the Correlation Image Sensor. and velocity of the cloud employing the meteorological knowledge

118 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6497: Image Processing: Algorithms and Systems V

The proposed method is based on a complex Optical Flow equation In this paper, we introduce the multi-histogram equalization method, called optical flow identity (OFI) between intensity image and complex as well as local and adaptive multi-histogram equalization algorithms. sinusoidally-modulated image. This relation is a complex equation and The algorithms presented can enhance images with regions of the optical flow is a real vector. Therefore we get two linear equations inconsistent illumination or exposure. They utilize the characteristics on the optical flow. It is possible to compute the optical flow directly of the human visual system to achieve a new method of thresholding from one pixel neighborhood. Since the OFI does not present any time for multi-histogram equalization. derivatives, information on one frame only are required. Moreover, it is free from related sources of error such as the approximated differencing and the aliasing in the large motion, and the limitation on 6497-17, Session 4 the object’s velocity for it induced inaccuracy in approximation of time derivatives by frame differences is avoided. According to the analysis Using block-based spectral angle measure for iris of the aperture problem, we show that, in the locally one-dimensional pattern matching pattern moving case , we can compute the optical flow using OFI, Y. Du, D. D. Isklandar, Indiana Univ.-Purdue Univ. Indianapolis while, optical flow computing by traditional OFC is made impossible. Iris pattern matching is a key step in iris patter recognition system. This paper proposes a new approach for iris pattern matching using 6497-14, Session 4 block-based spectral angle measure (SAM). The iris patterns in this paper are generated using Log-Gabor wavelet method. The proposed Crop/weed discrimination in simulated agronomic images SAM method will capture not only the phase information, but also the G. Jones, C. Gee, Etablissement National d’Enseignement magnitude information. The block based method will enable the Superieur Agronomique de Dijon (France); F. Truchetet, Univ. de flexibility in the pattern matching step, which improves the accuracy Bourgogne (France) and is more tolerable of the pattern dilation/constrain resulted from imperfect preprocessing. The experimental results show that the In the context of site-specific weed management by vision systems, an proposed method has an encouraging performance. efficient image processing for a crop/weed discrimination is required in order to quantify weed infestation rates from image. This paper presents a modelling of crop field in presence of different weed 6497-18, Session 4 infestation rates and a set of simulated agronomic images is used to test and validate the effectiveness of any crop/weed discrimination Dynamic content-based vehicle tracking and traffic algorithms. For instance, an algorithm has been implemented to firstly monitoring system detect the crop rows in the field by the use of a Hough Transform and F. R. Bowen, Y. Du, Indiana Univ.-Purdue Univ. at Indianapolis; S. secondly to detect plant areas by a region based-segmentation on Li, Indiana Dept. of Transportation; Y. Jiang, Purdue Univ.; T. binary images. This image processing has been tested on simulated Nantung, S. Noureldin, Indiana Dept. of Transportation; M. J. cereal fields (i.e. wheat) with perspective effects. The vegetation in the Knieser, Indiana Univ.-Purdue Univ. at Indianapolis virtual field is modelled by a sowing pattern for crop plants and the weed spatial distribution is modelled by either a Poisson process or a In the past decades, traffic data has been widely used in Neyman-Scott cluster process. For each simulated image, a transportation planning, highway operations, traffic analysis, and comparison between the initial and the detected weed infestation rate performance measurement. Traffic data may be obtained from allow us to assess the accuracy of the algorithm. This comparison different sensors such as pneumatic sensors, loop detectors, or demonstrates an accuracy of better than 85% is possible, despite that cameras. Among them, loop detectors are often used to count intra row weeds can not be detected from this spatial method. vehicles. The loop detectors are limited to a spot where they are deployed. In addition, the cameras are easier to install as well as less costly. Video image processing based vehicle tracking and traffic 6497-15, Session 4 monitoring provides several advantages over traditional approaches. One of the challenges lies in the robust segmentation and region Image-based motion stabilization for maritime tracking. The traditional approaches have assumed 1) the region of surveillance the object of interest is uniform and homogeneous; 2) adjacent D. D. Morris, Northrop Grumman Corp.; B. R. Colonna, F. D. regions should differ significantly. These are often not true for a Snyder, General Dynamics Robotic Systems vehicle object. In this paper, we propose a dynamic content based image segmentation method for vehicle tracking and traffic Robust image-based motion stabilization is developed to enable visual monitoring. Only initial lane information is needed for camera surveillance in the maritime domain. The algorithm developed is neither calibration. The system will automatically detect the direction of the a dense registration method nor a traditional feature-tracking method, traffic flow for vehicle detection and traffic monitoring. The preliminary but rather it captures the best aspects of each of these approaches. It experimental results show that this method is effective. A larger set of avoids feature tracking and so can handle large intra-frame motions, traffic parameters such as lane changing, congestion, and accidents, and at the same time it is robust to large lighting variations and moving can be obtained and measures based on the information content clutter. It is thus well-suited for challenges in the maritime domain. The associated with image sequences. geometry of the maritime environment is leveraged, including the horizon and shoreline. Results of real-time operation on an in-water buoy are presented. 6497-19, Session 5 Edge and corner preserving smoothing for artistic 6497-16, Session 4 imaging Adaptive multi-histogram equalization using human vision G. Papari, N. Petkov, Rijksuniv. Groningen (Netherlands); P. thresholding Campisi, Univ. degli Studi Roma Tre (Italy) E. J. Wharton, Tufts Univ.; S. S. Agaian, The Univ. of Texas at San What visually distinguishes a painting from a photograph is absence Antonio; K. A. Panetta, Tufts Univ. of texture and presence of sharp edges: on paintings, edges are sharper than on photographic images and textured areas result Images that do not have uniform brightness pose a challenging relatively flatter. Therefore, such artistic effects can be achieved by problem for image enhancement systems. Histogram equalization has filters that smooth textured areas while preserving, or enhancing, been successfully employed for enhancing images that have overall edges and corners. However, hot all the edge preserving smoothers improper illuminated or are over/under exposed. However, this method are suitable for artistic imaging. This study, aimed to obtain artistic performs poorly for images that contain various regions of improper effects, presents a generalization of the well know Kuwahara filter. illumination or improper exposure. Theoretical limitations of the Kuwahara filter are discussed and solved

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 119 Conference 6497: Image Processing: Algorithms and Systems V by the new nonlinear operator proposed here. Experimental results images can be several thousands or even millions, whereas the shows that the proposed operator mimics very well the oil painting dynamic range of consumer imaging devices rarely exceeds 100; style and it is robust to corruption of the input image such as blurring. therefore some processing is needed in order to display a high Comparison with existing techniques shows situations where dynamic range image correctly. Global operators map each pixel traditional edge preserving smoothers, commonly used for artistic individually with the same nonlinear function; local operators use imaging, fail and our approach produces good results. spatially-variant functions in order to achieve a higher quality. The lower computational cost of global operators makes them attractive for real-time processing; the nonlinear mapping can however 6497-20, Session 5 attenuate the image details. In this paper we will define an expression which gives a quantitative measure of this artifact, and will compare Demosaicing of noisy data: spatially adaptive approach the performance of some commonly used operators. We will show D. V. Paliy, Tampere Univ. of Technology (Finland); M. Trimeche, that a modified logarithm we propose has a satisfactory performance Nokia Research Ctr. (Finland); V. Katkovnik, Tampere Univ. of for a wide class of images, and has a theoretical justification based on Technology (Finland); S. Alenius, Nokia Research Ctr. (Finland) some properties of the human visual system. We will also introduce a method for the automatic tuning of the parameters of our system, In this paper we propose processing of noisy Bayer-patterned images. based on the statistics of the input image. We will finally compare our The data is considered to be corrupted by the Gaussian or Poissonian method with others proposed in the literature. types of noise. The Poissonian noise is signal-dependent and common for CCD and CMOS digital image sensors. Demosaicing algorithms are used to reconstruct missed red, green, and blue values to produce an 6497-23, Session 5 RGB image. It can be seen as an interpolation problem usually called color filter array interpolation (CFAI). The conventional approach used Wavelet-based texture image classification using vector in image restoration chains for raw sensor data exploits successive quantization denoising and CFAI steps. The denoising step comes first. This is E. P. Lam, Raytheon Co. motivated by the fact that knowledge about the noise model is of great importance in denoising. In this paper we propose to combine the Classification of image segments on textures can be helpful for target filtering and CFAI in a one procedure. Firstly, we compute initial recognition. Sometimes target cueing is performed before target directional interpolated estimates of noisy color intensities. Afterward, recognition. Textures are sometimes used to cue an image processor these estimates are decorrelated and denoised by the special of a potential region of interest. In certain imaging sensors, such as directional anisotropic filters. This approach is found to be efficient in those used in synthetic aperture radar, textures may be abundant. The order to attenuate both noise and interpolation errors. The exploited textures may be caused by the object material or speckle noise. Even denoising technique is based on the local polynomial approximation speckle noise can create the illusion of texture, which must be (LPA). The adaptivity to data is provided by the multiple hypothesis compensated in image pre-processing. In this paper, we will discuss testing called the intersection of confidence intervals (ICI) rule which is how to perform texture classification but constrain the number of applied for adaptive selection of varying scales (window sizes) of LPA. wavelet packet node decomposition. The new approach performs a We show the efficiency of the proposed approach comparing it with the two-channel wavelet decomposition. Comparing the strength of each conventional algorithms. new subband with others at the same level of the wavelet packet determines when to stop further decomposition. This type of decomposition is performed recursively. Once the decompositions 6497-21, Session 5 stop, the structure of the packet is stored in a data structure. Using the information from the data structure, dominating channels are Prediction of signs of DCT coefficients in block-based extracted. These are defined as paths from the root of the packet to lossy image compression the leaf with the highest strengths. The list of dominating channels are N. N. Ponomarenko, National Aerospace Univ. (Ukraine); A. V. used to train a learning vector quantization neural network. Bazhyna, K. O. Egiazarian, Tampere Univ. of Technology (Finland) A practical unpredictability of signs of DCT coefficients is widely 6497-24, Session 6 accepted. Sign of each coded DCT coefficient occupies 1 bit of memory in compressed data. The size of all coded signs of DCT Classification SAR targets with support vector machine coefficients occupy 15-25 percents of a compressed bitstream, L. Cao, Chinese Leihua Electronic Technology Research Institute depending on used compression method. In this work we propose an (China) efficient method to predict signs of DCT coefficients in block based image compression. For that, values of pixels of already coded/ With the development of Synthetic Aperture Radar (SAR) technology, decoded neighbor blocks of the image are used. The method consist automatic target recognition (ATR) is becoming increasingly important. two stages. At first, values of pixels of a row and a column of currently In this paper, we proposed a 3-class target classification system in coded block, which both are the nearest to already coded neighbor SAR images. The system is based on invariant wavelet moments and blocks are estimated by a context-based adaptive predictor. At second support vector machine (SVM) algorithm. It is a two-stage approach. stage, these estimated row and column are used for prediction of the The first stage is to extract and select a small set of wavelet invariant signs of the DCT coefficients. Proposed method allows to compress moment features to indicate target images. The wavelet invariant signs of DCT coefficients to 60-85% from their original size. It moments take both advantages of the wavelet inherent property of corresponds to increase of compression ratio of the entire image by 3- multi-resolution analysis and moment invariants quality of invariant to 9 percents. The proposed method of the compression of the signs of translation, scaling changes and rotation. The second stage is DCT coefficients is independent from any other part of the coder. It classification of targets with SVM algorithm. SVM is based on the may be used for improving practically all existent image and video principle of structural risk minimization (SRM), which has been shown compression methods. better than the principle of empirical risk minimization (ERM), which is used by many conventional networks. To test the performance and efficiency of the proposed method, we performed experiments on 6497-22, Session 5 invariant wavelet moments, different kernel functions, 2-class identification, and 3-class identification. Test results show that wavelet Nonlinear mapping for dynamic range compression in invariant moments indicate the target effectively; linear kernel function digital images achieves better results than other kernel functions, and SVM G. Guarnieri, S. Carrato, G. Ramponi, Univ. degli Studi di Trieste classification approach performs better than conventional nearest distance approach. (Italy) The dynamic range of an image is defined as the ratio between the maximum and minimum luminance value it contains. This value in real

120 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6497: Image Processing: Algorithms and Systems V

6497-26, Session 6 its texture feature extracted and the neural network has to classify it as a predefined landmark. The video frames are also preprocessed to Adaptive color space transform using independent reduce their difference of scale and rotation from the satellite image component analysis before the texture feature extraction, so the UAV altitude and heading for each frame are considered as known. The neural network E. Vera, Univ. of Concepcion (Chile) techniques present the advantage of low computational cost, been In this paper, a novel color space transform is presented. It is an appropriate to real-time applications. Promising results were obtained, adaptive color space transform based on the application of mainly during flight over urban areas. independent component analysis to the RGB data of an entire color image. The result is a linear and reversible color space transform that provides three new coordinate axes where the projected data is as 6497-29, Session 7 much as statistically independent as possible, and therefore highly Autonomous learning in gesture recognition by using uncorrelated. Compared to many non-linear color space transforms such as the HSV or CIE Lab, the proposed one has the advantage of lobe component analysis being a linear transform from the RGB color space, much like the XYZ J. Lu, National Institute of Occupational Safety and Health (Japan) or YIQ. However, its adaptiveness has the drawback of needing an estimate of the transform matrix for each image, which is sometimes Gesture recognition is a new human/machine interface method computationally expensive for larger images due to the common implemented by pattern recognition(PR). iterative nature of the independent component analysis In order to assure robot safety when gesture is used in robot control, a implementations. Then, an image sub sampling method is also unified model is required to express gestures quantitatively and proposed to enhance the novel color space transform speed, accurately, and to implement the interface reliably. In addition, it is efficiency and robustness. The new color space is used for a large set required to evaluate the gesture design and the interface of test color images, and it is compared to traditional color space performance, before the interface can be actually used. Similar with transforms, where we can clearly visualize its vast potential as a other PR applications, 1) feature selection (or model establishment) promising preprocessing step for segmentation purposes for example. and 2) training from samples, affect the performance of gesture recognition largely. For 1), a simple model with 6 feature points at shoulders, elbows, and hands, is established. The gestures to be 6497-27, Session 7 recognized are restricted to still arm gestures, and the movement of arms is not considered. These restrictions are to reduce the Neural network-based multiscale image restoration misrecognition, but are not so unreasonable. For 2), a new biological approach network method, called lobe component, is used to learn in A. P. A. d. Castro, J. D. S. Silva, Instituto Nacional de Pesquisas unsupervised style. Lobe components, corresponding to high- Espaciais (Brazil) concentrations in probability of the neuronal input, are orientation selective cells follow Hebbian rule and lateral inhibition. Due to the This paper describes a neural network based multiscale image advantage of lobe component method for balanced learning between restoration approach. Multilayer perceptrons are trained with artificial global and local features, large amount of samples can be used images of degraded gray level circles, in an attempt to make the efficiently, and the learning converges fast. neural network learn inherent space relations of the degraded pixels. The present approach simulates the degradation by a low pass Gaussian filter blurring operation and the addition of noise to the 6497-30, Session 8 pixels at pre-established rates. The training process considers the degraded image as input and the non-degraded image as output for A new design approach to neural network pattern the supervised learning process. The neural network thus performs an recognition systems inverse operation by recovering a quasi non-degraded image in terms C. J. Hu, Southern Illinois Univ./Carbondale of least squared. The main difference of the approach to existing ones relies on the fact that the space relations are taken from different As we reported in the last two years in SPIE conferences, we can use scales, thus providing relational space data to the neural network. The a simple VB6 program to break the boundaries of some selected approach is an attempt to come up with a simple method that leads objects in an edge-detected binary picture into many simple branches to an optimum solution to the problem. Considering different window and reconstructed accurately the boundaries of the original objects sizes around a pixel simulates the multiscale operation. In the free of noise. By this means we can then program the computer to generalization phase the neural network is exposed to indoor, outdoor, automatically learn some standard objects and automatically and satellite degraded images following the same steps use for the recognize any test objects by a novel topological pattern recognition artificial circle image. (TPR) scheme. The learning and recognition are based on the topological connections of the simple, bi-directional graph of the selected object boundaries. It is very accurate, yet very robust way to 6497-28, Session 7 recognize the test objects, because it is like the design of an electric circuit. When the way of connection of an electric circuit, or the Landmarks recognition for autonomous aerial navigation topology of that circuit, is fixed, all electrical properties of that circuit by neural networks and Gabor transform are fixed. E. H. Shiguemori, M. P. Martins, M. V. T. Monteiro, F. L. L. Medeiros, M. A. P. Domiciano, Comando-Geral de Tecnologia 6497-32, Session 8 Aeroespacial (Brazil) Template matching in real-time is a fundamental issue in many Speckle noise filtering with edge preservation of medical applications in computer vision such as tracking, stereo vision and ultrasound images autonomous navigation. The goal of this paper is present a system for A. R. Upadhyay, R. K. Choudhary, Asia Pacific Institute of automatic landmarks recognition in video frames over a Information Technology (India); S. N. Talbar, Shri Guru Gobind georeferenced high resolution satellite image, for autonomous aerial Singhji Institute of Engineering & Technology (India) navigation research. The video frames employed were obtained from a camera fixed to a helicopter in a low level flight, simulating the vision Ultrasound images are the low cost, non-invasive imaging modality system of an unmanned aerial vehicle (UAV). The landmarks that has proved popular for many medical applications. Speckle noise descriptors used in recognition task were texture features extracted is an intrinsic property of medical ultrasound imaging; generally it by a Gabor Wavelet filters bank. The recognition system consists on a tends to reduce the image resolution and contrast, thereby reducing supervised neural network trained to recognize the satellite image the diagnostic value of this imaging modality. It also reduces the utility landmarks texture features. In activation phase, each video frame has of ultrasound for less than highly, trained users and also complicates

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 121 Conference 6497: Image Processing: Algorithms and Systems V image-processing tasks such as feature segmentation. The shown the robustness and effectiveness of color based tracking preprocessing allows filtering methods based on assuming the noise algorithms, especially for tracking tasks where object shape exhibits to be white and Gaussian, to perform in nearly optimal conditions. dramatic variability. This paper evaluates performances of different non-linear filters and There are two major ways how to adopt color and texture for the anisotropic diffusion method. The method is based on partial tracking task. First, previously accumulated foreground and differential equation [PDE], of which the initial data was the input background distributions are used to classify points of a newly images, transformed into differential forms and solved with iterations. coming frame into foreground and background sets. Second, the local After comparing the results with other filters, the anisotropic diffusion color/texture segmentation is carried out, starting from predicted method could smooth the speckles very well and the edges of the position, in order to refine object’s location. Some researches have image are also very clear. Anisotropic diffusion can remove the tried to combine both approaches into a single framework. speckle noise effectively and has great potential in filtering medical ultrasonic images. Tracking algorithm, presented in the article, can be related to the first category. The problem is solved by tracking color distributions of background and foreground (object) points simultaneously. The key 6497-33, Poster Session feature of proposed approach is careful selection of histogram resolution (or kernel size) on each frame of a sequence. Even being Green noise halftoning with dot diffusion used stand-alone, this method demonstrates relative accuracy and S. Lippens, Univ. Gent (Belgium) and Institute for the Promotion of stability in certain situations. Innovation through Science and Technology in Flanders (Belgium); W. R. Philips, Univ. Gent (Belgium) 6497-38, Poster Session Because a standard printer contains a limited set of inks, a continuous Background change detection in video with dynamic tone image has to be reduced to for example a binary pattern (black dots on white) prior to printing. This process is called halftoning. foreground activity using robust parameter estimation techniques Different halftoning techniques can be classified based on the characteristics of the binary patterns. C. A. Jenkins, J. B. Jordan, New Mexico State Univ.; J. J. Carlson, In this paper we focus on green noise halftoning, a technique that Sandia National Labs. arranges dot clusters in a stochastic way. This technique is valuable We propose a statistically robust parameter estimation technique to for print technologies with unreliable reproduction of isolated dots. automate video surveillance. The technique is capable of detecting The algorithm we use is based on dot diffusion, an error diffusion like changes in the background of extremely dynamic environments. For technique with a high degree of parallel processing, in contrast to the example covert placement of explosives (e.g., briefcase bombs) inside standard serial error diffusion algorithms. We explore two approaches crowded public facilities, such as an airport. A histogram of pixel to generate green noise halftones. The first approach uses hysteresis, intensity for each pixel over a fixed time span for a series of images analogous to standard green noise error diffusion techniques. We including background and dynamic foreground can be represented by observe that the resulting halftones are rather coarse and highly a mixture of two Gaussian functions. Extracting the foreground dependent on the used dot diffusion class matrices. In the second function from the mixture allows the background to be monitored approach we don’t limit the diffusion to the nearest neighbors. This despite considerable foreground activity. Pearson’s Method of leads to less coarse halftones, compared to the first approach. The Moments was used to separate the two probability distribution drawback is that it can only cope with rather limited cluster sizes. We functions. It has been determined that this method is effective when can reduce these drawbacks by combining the two approaches. the background in any given area is visible for at least ten percent of the time; only ten seconds of 30 frames per second video is needed to aptly depict the background. With time, changes such as lighting 6497-34, Poster Session are incorporated into the expected pattern. More drastic and permanent shifts, such as a personal item left in the field of view, are Fractal video textures synthesis-by-analysis with detected by our algorithm and flagged for investigation. Results directionality estimation indicate a significant potential for robust parameter estimation P. Campisi, E. Maiorana, A. Neri, Univ. degli Studi Roma Tre (Italy) techniques as applied to video surveillance. In this paper a model based synthesis-by-analysis approach for the generation of natural video textures with fractal properties is 6497-39, Poster Session presented. An elastic graph matching approach for motion Specifically, a video texture is modeled according to the three estimation on feature points in image sequences dimensional (3D) Extended Self-Similar (ESS) model. In the analysis stage, the AutoCorrelation Functions (ACFs) of the R. Feghali, Communications Research Ctr. Canada (Canada) increments of the original video texture are estimated according to the This study proposes a new approach to perform motion estimation on 3D-ESS model. a set of feature points via elastic graph matching. The approach starts Moreover a method to estimate the textures directionality is here by constructing a labeled graph on a set of feature points in the first proposed, thus refining the parameters analysis stage. image of a given sequence, then sequentially match the labeled graph with the remaining images in the sequence. The matching is based on The synthesis method generates 3D-ESS processes whose a similarity function that depends on image brightness and motion increments have the same ACFs of the original ones. characteristics on one side, and on geometric distortion on the other Experimental results show the effectiveness of the approach side. The graph matching throughout the sequence corresponds to a proposed for directionality estimation as well as for video texture motion field estimate on feature points that accounts for both synthesis of arbitrary spatio-temporal dimensions. geometric and photometric spatiotemporal consistency. The main advantage of the proposed approach is that it preserves high-level image characteristics as outlined by the geometrical structure of 6497-37, Poster Session moving objects and their relative positions in space and time, while it simultaneously accounts for low-level measurements such as motion Entropic tracker of color objects and brightness. A. Akhriev, Russian Scientific Ctr. of Roentgenology and Radiology (Russia); V. Somikov, ELVEES Research Ctr. (Russia) Color provides important visual information for real-time tracking of non-rigid and partially occluded objects. Recent developments have

122 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6497: Image Processing: Algorithms and Systems V

6497-40, Poster Session image features such as the mean, the standard deviation, Gabor filters features, and Wavelets features. These fuzzy sensors are applied to Algorithms for the resizing of binary and grayscale the different segments obtained using the Homogeneity Coefficient images using a logical transform Segmentation Algorithm. The resulted classification of the regions is compared against ground truth maps created using the actual E. E. Danahy, Tufts Univ.; S. S. Agaian, The Univ. of Texas at San procedure made by marine scientists; providing a benchmark to Antonio; K. A. Panetta, Tufts Univ. validate results of any future classifier. Image interpolation, both downsampling and upsampling, is necessary when resizing the data to match either the specifics of the communication channel or the output display. While it is more efficient 6497-43, Poster Session to transmit low-resolution versions to the client, the high-resolution Object classification, segmentation, and parameter version may be necessary when presenting the final visual data. There are several techniques for performing image resizing currently in use, estimation in multichannel images by classifier learning from the simple nearest neighbor to bi-linear and bi-cubic with clustering of local parameters interpolation. Many of the traditional methods either result in visually V. V. Lukin, N. N. Ponomarenko, A. A. Zelensky, National noticeable block artifacts and jagged lines, or suffer from blurring Aerospace Univ. (Ukraine); J. T. Astola, K. O. Egiazarian, Tampere issues particularly in edge regions. This paper presents a new Univ. of Technology (Finland) algorithm for image resizing based on manipulation of the Boolean minimized representation of binary image data. Nowadays multichannel (optical, radar, infrared, combined, etc.) Applicable to binary as well as grayscale images, this procedure imaging systems are widely used in such applications as remote performs the interpolation through the manipulation of the sum of sensing, non-destructive control, scene analysis, medicine and so on. primary implicants representation of low-resolution data generated via The final goal of using such systems is extraction of useful information a logical transform. This form is scaled up to produce the images of from data obtained. There are several obstacles preventing reliable higher resolution. In its most basic form, the algorithm outputs results and accurate retrieval of useful information. The basic of them are equivalent to nearest neighbor, while variations better approximate noise inherent for practically any kind of images and the problems missing pixels and reduce the blocking effects. Computer simulations with selecting a proper set of features to be used for classification are presented which demonstrate the abilities of the procedures as objects under interest and estimation of their characteristics. compared to other standard interpolation techniques. The proposed approach is based on learning with clustering that allows selecting the most informative local parameters and ignoring less informative or fictive ones. Moreover, it permits to automatically 6497-41, Poster Session adjust thresholds with which the local parameters are to be compared and to design a logic scheme for assigning a given set of parameters Texture classification by ICA to some cluster and, after this, to the corresponding class. We give a D. Coltuc, Univ. Politehnica Bucuresti (Romania) particular example of applying the designed approach to a task of substance content analysis using specially formed color images. The The ICA (Independent Component Analysis) is a mathematical tool substance contains four different components and the task was to traditionally employed for source separation. In this paper, we test its classify images with separating groups of pixels corresponding to ability for texture analysis, in order to provide a new method for the each class (segmenting images) as the first step. unsupervised classification of textures. Our approach is based on the assumption that a texture may be represented as a mixture of many random signals with a certain degree of statistical independence. 6497-44, Poster Session Under this hypothesis, we decompose the texture in a weighted sum of components, by using the ICA. From the multitude of the existing Automated image registration and fusion for thermal and ICA algorithms, we have chose FastICA, a version based on the fourth visible images order statistics. Since there is no reason to privilege a certain D. R. Deepti, Bharat Electronic Ltd. (India) component or a certain texture sample, FastICA is asked to start the components’ extraction from an initial mixing matrix constituted of Image fusion has been an interesting area of research that finds great ones. This guaranties a unique solution for the decomposition. Each application in varying fields like video surveillance, medical imaging, component is described by means of its negentropy, a measure of video compression, computer vision etc. Image fusion is the process component’s nongaussianity, also optimized by FastICA. We show, of combining the image data obtained from multiple sources such that experimentally, that the three most nongaussian components of each the resulting image is in some sense better than that provided by the analyzed texture are able to cluster the test textures. The influence on individual sources. The image data may differ in the sensor type, clustering of some parameters like the number and the size of texture viewing condition, camera position, capture time etc. Depending on samples are also analyzed. the synergy of the information inherent in the data, it may be possible to reduce noise, to extend the field of view beyond that of a single image, to restore high frequency content, and even to increase the 6497-42, Poster Session spatial resolution. The use of images obtained in different wave bands like thermal and visible images gives the advantages like availability of A fuzzy system approach for the classification of complementary information, improved performance in adverse underwater AUV color images environmental conditions etc. J. A. Díaz, R. E. Torres, Univ. de Puerto Rico Mayagüez One of the major tasks in image fusion is image registration also referred to as image alignment problem. Image registration is used to This paper presents a fuzzy system approach using texture to classify match two or more images of the same scene taken at different times, living coral cover in underwater color images. The monitoring of these from different viewpoints and from different sensors so that the kinds of ecosystems in deep coast waters is important for natural corresponding co-ordinate points in these images correspond to the resources protection. By automating the process, we are saving time same physical region of the scene being imaged. The accuracy of and human resources to help in the study of marine science. The image fusion depends on the accuracy with which image registration images used in this work were acquired by an autonomous is being carried out. underwater vehicle (AUV) to avoid poor spatial resolution introduced by satellite imagery of water depths. The AUV images have more In this paper an attempt has been made for the fusion of images spatial details in deep water objects difficult to detect by satellite. The obtained from thermal imagers and day vision cameras. Analyses main challenges, however, arise from the similarity of color statistics, were carried out on a set of image data generated at our lab. As the non-uniform illumination, and concentration of noise introduced by images obtained were slightly mis-aligned, an image registration scattering and pre-processing of the images. The proposed fuzzy method prior to image fusion was adopted. classifier consists in the assigning of fuzzy memberships to different

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 123 Conference 6497: Image Processing: Algorithms and Systems V

For image registration, a control point based method was used. This Gaussian distributions. Other RBF based algorithms were compared includes a contour extraction step followed by a feature point finding with our approach in pdf estimation of artificially generated data as step. Various methods such as wavelet based edge detection, edge well as in the microcalcification detection in mammographic image detection using Laplacian of Gaussian Edge detector etc were analysis. From simulation results we observe that the RMRBF gives adopted to extract the contour and directional variance was used for better estimation of the implied pdfs and has show better finding the feature points. The feature points obtained were further classification capabilities. matched in the two test images using normalized cross-correlation criteria. Image transformation was then carried out for the thermal image so as to make it aligned to the day vision image. Different 6497-47, Poster Session transformation functions like linear, affine, projective, non-linear etc were used to transform the thermal image and results are provided for Scanned image enhancement using multilayer neural each of these methods. Mean square error has been used as the networks criteria for the selection of the most suitable transformation method M. N. Ahmed, Lexmark International Inc. depending on the type of mis-alignment for the image under consideration. After the images are registered, fusion is being carried Document images typically contain a combination of text, out for the visible image and the registered thermal image. Image background, and halftoned images. Text, including line art and similar fusion has been carried out using various methods like Averaging, graphical content, is characterized by sharp, high-contrast edges and Laplacian pyramid, Contrast Pyramid, PCA method, Gradient Pyramid thin strokes. The background of the document is usually white or etc. The results obtained using the above methods are also presented nearly white, and it normally has a smooth texture. Halftoned images and discussed. consist of a pattern of small dots. In most cases, the dots are arranged in an ordered pattern and will vary slightly in size according to the darkness of the image they represent. Digital copying, in which 6497-45, Poster Session a digital image is obtained from a scanning device and then printed, involves a variety of inherent factors that compromise image quality. Motion deblurring for suppression of breathing Specifically, digitally copied documents usually suffer from blurring, distortions caused by camera shake flare, noise, and moire. Limited scan resolution blurs edges, I. Tsubaki, T. Komatsu, T. Saito, Kanagawa Univ. (Japan) degrading the appearance of text and other detail. Fine detail also suffers from flare, caused by the reflection and scattering of light from Video clips acquired by handheld cameras often include oscillatory the scanner’s illumination source. Flare blends together nearby motion caused by undesirable camera shakes. They are also colors, blurring the high-frequency content of the document. In degraded by motion blur over the image, which temporally varies in severe cases, it can extend even into mid-frequency content. Noise strength. of various forms results from the physical limitations of the Alternate blurred frames and sharp frames are included in such video mechanical and electrical components of the scanning device, as clips. This degradation is considered a kind of breathing distortion, well as imperfections in the source document itself. Finally, ordered which usually means oscillatory changes in the strength of focal blur halftone patterns in the original document interact with the periodic in old films. We describe the oscillatory changes of motion blur as a sampling of the scanner, producing objectionable moire patterns. breathing distortion. These are exacerbated when the copy is reprinted with an ordered halftone pattern. Digital video stabilization can remove the oscillatory motion. But the stabilized video images are still degraded by motion blur, which is not To address these distortions, digital filtering may be applied to the associated to the new image motion. Time-varying blur is scanned document image. Sharpening improves the appearance of conspicuous in the stabilized video clip, and can be perceived as text and fine detail, countering the effects of limited scan resolution flicker. and flare. Edges become clear and distinct. Of course, other artifacts such as noise and moire become sharper as well. Simple low-pass Various motion deblurring approaches have been studied to improve filtering can smooth higher-frequency noise in the background. Tone image sharpness. Most of them are considered a kind of intraframe level adjustment can map near-white backgrounds to solid white, approach, and the best image for each frame is restored individually. which is the intent for most documents. Although crude, a similar, In this paper, we propose an interframe motion deblurring approach small low-pass filter can correct the majority of moire artifacts. which aims that all frames achieve the same level of smoothness. A Unfortunately, low-pass filtering affects detail as well, blurring it even spatio-temporal regularization is introduced into the deconvolution further. approach to smooth the temporal change of blur. In order to perform the appropriate type of filtering, a preprocessing step is required to segment and label the document image into text, background, and halftoned image regions. With the segmentation 6497-46, Poster Session step, we can strongly sharpen text and similar edge detail, smooth Rank M-type radial basis functions network for medical and perhaps lighten document backgrounds, and descreen halftoned images using an appropriate low-pass filter. In this paper, Each pixel image processing applications is assigned a feature pattern consisting of a scaled family of J. A. Moreno-Escobar, F. J. Gallegos-Funes, V. I. Ponomaryov, differential geometrical invariant features and texture features Instituto Politécnico Nacional (Mexico) extracted from the cooccurence matrix. The invariant feature pattern is then assigned to a specific region using a multilayer neural network. Recently, we introduced the robust RM (Rank M-type) -estimators for In this paper, we demonstrate the power of the neural network image denoising applications. The combined RM-estimators use approach to segment document images into text, halftone, and different rank estimators such as the median, Wilcoxon and Ansari- background. Bradley-Siegel-Tukey estimators, and the M-estimator with different influence functions to provide better robustness. We demonstrated that the performances of combined RM-estimators are better in 6497-48, Poster Session comparison with original R- and M- estimators. In this paper we present the capability of the Rank M-Type Radial Identification of vegetable diseases using neural network Basis Function (RMRBF) Neural Network in medical image processing J. Zhang, J. Tang, L. Yao, Beijing Normal Univ. (China) applications. The proposed neural network uses the proposed RM- estimators in the scheme of radial basis function to train the neural Vegetables are widely planted all over China, but they often suffer network. The RMRBF-based training is less biased by the presence of from the some diseases. A method of major technical and economical outliers in the training set and was proved an accurate estimation of importance is introduced in this paper, which explores the feasibility of the implied probabilities. We provide the theoretical evaluation of the implementing fast and reliable automatic identification of vegetable bias for proposed algorithm in the case when estimating overlapping diseases and their infection grades from color and morphological

124 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6497: Image Processing: Algorithms and Systems V features of leaves. Firstly, leaves are plucked from clustered plant and pictures of the leaves are taken with a CCD digital color camera. Secondly, color and morphological characteristics are obtained by standard image processing techniques, for examples, Otsu thresholding method segments the region of interest, image opening following closing algorithm removes noise, Principal Components Analysis reduces the dimension of the original features. Then, a recently proposed boosting algorithm AdaBoost.M2 is applied to RBF networks for diseases classification based on the above features, where the kernel function of RBF networks is Gaussian form with argument taking Euclidean distance of the input vector from a center. Our experiment performs on the database collected by Chinese Academy of Agricultural Sciences, and result shows that Boosting RBF Networks classifies the 230 cucumber leaves into 2 different diseases (downy-mildew and angular-leaf-spot), and identifies the infection grades of each disease according to the infection degrees.

6497-49, Poster Session Fingerprint image enhancement and denoising A. R. Upadhyay, Asia Pacific Institute of Information Technology (India) Fingerprint images vary in quality. The performance of a fingerprint recognizer greatly depends on the fingerprint image quality. Different types of noises in the fingerprint images create greater difficulty for recognizers. Fingerprint image enhancement is an essential preprocessing step in fingerprint recognition applications. Numerous methods like Gaussian blurring, Median Filtering, Canny thresholding have been described in the paper. Generally noisy valley pixels and the pixels in the interrupted ridge may be observed. The primary goal of processing a noisy signal is to obtain a reconstruction as close to the original clean signal as possible. Scale-based analysis has recently played an increasingly important role in image analysis. Scale space is an isotropic method that smooths signals with Gaussian kernel uniformly by increasing scale. This paper describes a new approach to fingerprint image enhancement, the filter is based on Nonlinear Anisotropic Diffusion approach. Experimental results show that this enhancement algorithm improves the performance of the fingerprint verification system and makes it more robust with respect to the quality of input fingerprint images.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 125 Conference 6498: Computational Imaging V Monday-Wednesday 29-31 January 2007 Part of Proceedings of SPIE Vol. 6498 Computational Imaging V

6498-01, Session 1 6498-04, Session 1 Analysis of reconstructions in full view fluorescence Image reconstruction performance as a function of molecular tomography model complexity using information geometry: D. E. Hyde, Northeastern Univ. and Massachusetts General application to transmission tomographic imaging Hospital; A. Soubret, J. Dunham, T. Lasser, Massachusetts General J. A. O’Sullivan, L. Xie, D. G. Politte, B. R. Whiting, Washington Hospital; E. L. Miller, D. H. Brooks, Northeastern Univ.; V. Univ. in St. Louis Ntziachristos, Massachusetts General Hospital Models used to derive image reconstruction algorithms typically make The development of fluorescent probes capable of highly specific assumptions designed to increase the computational tractability of the targeting of cellular and sub-cellular processes has driven the algorithms while taking enough account of the physics to achieve development of new techniques for in-vivo tissue optical imaging. desired performance. As the models for the physics become more Fluorescence molecular tomography (FMT) is an emerging modality detailed, the algorithms typically increase in complexity, often due to which was developed to improve upon existing planar photographic increases in the number of parameters in the models. When imaging methods by quantitatively reconstructing three dimensional parameters are estimated from measured data and models of fluorochrome bio-distribution in tissues. FMT is a non-invasive method increased complexity include those of lower complexity as special and allows longitudinal imaging, potentially of several molecular targets cases, then as the number of parameters increases, model errors using different fluorochromes and spectral differentiation techniques. decrease and estimation errors increase. We adopt an information Here we demonstrate the development of 360 degree free-space geometry approach to quantify the loss due to model errors and collection geometries for FMT of entire animals. A volume carving Fisher information to quantify the loss due to estimation errors. These algorithm was employed to localize the animal’s surface. Prediction of are unified into one cost function. This approach is detailed in an X- photon propagation was accomplished by means of a normalized Born ray transmission tomography problem where all models are solution of the diffusion equation that offers high sensitivity in approximations to the underlying problem defined on the continuum. reconstructing fluorescence from optically heterogenous media. Computations and simulations demonstrate the approach. The Boundary modeling was based on the Kirchoff approximation. We analysis provides tools for determining an appropriate model investigated and compared the relative performance of two inverse complexity for a given problem and bounds on information that can solutions, ART (Algebraic Reconstruction Technique), and LSQR be extracted. algorithms. Results are presented for a variety of experimental settings, including 6498-50, Session 1 numerical simulations, physical phantoms, and in-vivo mouse models. These results are analyzed both qualitatively and quantitatively, and Progress in electrical impedance tomography compared against reconstructions in similar scenarios using a slab D. Isaacson, J. C. Newell, G. Saunier, Rensselaer Polytechnic geometry. Institute No abstract available 6498-02, Session 1 Image reconstruction for small animal SPECT with two 6498-05, Session 2 opposing half cones Y. Zheng, Univ. of Virginia; H. Li, The Univ. of Texas M.D. Anderson Creating panoramas on mobile phones Cancer Ctr.; J. Wang, A. V. Stolin, J. Pole, M. B. Williams, Univ. of J. J. Boutellier, O. J. Silvén, Oulun Yliopisto (Finland); M. Tico, M. Virginia Vehviläinen, Nokia Research Ctr. (Finland) Pinhole SPECT imaging offers high spatial resolution with the drawback Image stitching is used to combine several images into one wide- of low efficiency and small field of view. We developed a new system angled mosaic image. Traditionally mosaic images have been with two opposing pinhole-detector assemblies, each looking at half of constructed from a few separate photographs, but nowadays that the field of view. This configuration allows higher sensitivity and larger video recording has become commonplace even on mobile phones, it field of view. is possible to consider also video sequences as a source for mosaic We showed that this configuration provides complete data for image images. However, most stitching methods require vast amounts of reconstruction at the midplane. We deveoped 3D iterative image computational resources that make them unusable on mobile devices. reconstruction algorithms for this special half cone geometry and We present a novel panorama stitching method that is designed to demonstrated significantly improved image quality using real create high-quality image mosaics from both video clips and separate experimental data. images even on low-resource devices. The software is able to create both 360 degree panoramas and perspective-corrected mosaics. Features of the software include among others: detection of moving 6498-03, Session 1 objects, inter-frame color balancing and rotation correction. The application selects only the frames of highest quality for the final Spatio-temporal reconstruction of dynamic gated cardiac mosaic image. Low-quality frames are dropped on the fly while SPECT recording the frames for the mosaic. Y. Yang, M. Jin, E. Gravier, Illinois Institute of Technology The complete software is implemented on Matlab, but also a mobile phone version exists. We present a complete solution from frame In this paper we propose an image reconstruction procedure for gated acquisition to panorama output with different resource profiles that and dynamic SPECT images. We will divide the cardiac cycle into a suit various platforms. number of gate intervals as in gated SPECT, but treat the tracer distribution for each gate as a time-varying signal. By using motion- compensated spatial and temporal regularization, our reconstruction procedure will produce an image sequence that shows both cardiac motion and time-varying tracer distribution simultaneously. Our simulation results show that the proposed method can produce more accurate reconstruction of gated dynamic images than independent reconstruction of individual gate frames with spatial regularization alone.

126 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6498: Computational Imaging V

6498-06, Session 2 registration. Graph node weights are inversely proportional to the image overlap area. This allows predicting not yet discovered image Multidimensional image enhancement from a set of registrations by finding shortest paths by Floyd-Warshall algorithm. unregistered differently exposed images For example, if an image pair (A,B) is registered and there exists a path between B and C, the registration (or decision on non- P. Vandewalle, L. Meylan, S. E. Süsstrunk, École Polytechnique registration) between A and C is obtained using a sequence of Fédérale de Lausanne (Switzerland) transformations determined by the path between A and C. Super-resolution algorithms reconstruct a high-resolution image from a In the final step we search for image clusters, which are connected set of low-resolution input images. Similarly, a high dynamic range components in the registration graph. Images in each cluster are image can be constructed from a set of motionless input images using blended together to form an output of the system - a set of different exposure times. In this paper, we combine both techniques panoramas. and increase both spatial resolution and dynamic range from a set of We demonstrate our approach on typical image collections containing shifted images with varying exposure times. multiple panoramic shots. We show that the output of the algorithm With a varying exposure time approach, large parts of the images does not depend on image order: the results obtained from original typically have clipped values. We thus first detect the useable parts in ordered and randomly shuffled image collections are identical. each of the images and estimate the camera response function. Next, the images are aligned by computing the registration parameters between blocks of overlapping useable parts. After registration, the 6498-08, Session 2 pixels that are not clipped from each of the images are combined on a common grid. This results in an irregular grid of pixel values with Edge-directed interpolation using multiscale geometric varying pixel densities. A high-resolution high dynamic range image representations can be reconstructed using interpolation on a uniform grid. Typically, N. Mueller, Y. Lu, M. N. Do, Univ. of Illinois at Urbana-Champaign this results in images with a high resolution in the central part of the dynamic range, and lower resolution at both ends, where less useful TBD pixels were available. 6498-09, Session 3 6498-07, Session 2 Multiscale multigranular image segmentation Graph-based multiple panorama extraction from E. D. Kolaczyk, Boston Univ.; J. Ju, South Dakota State Univ.; S. unordered image sets Gopal, Boston Univ. A. Sibiryakov, M. Bober, Mitsubishi Electric ITE B.V. (United We describe a novel framework for producing labeled image Kingdom) segmentations that allows for adaptive choice of both the spatial This paper presents a multi-image registration method, which aims at resolution of subregions and the categorical granularity of labels. Our recognizing and extracting multiple panoramas in an unordered set of framework is based upon a class of models we call ‘mixlets’, a images. This allows users to input to the system the images on an blending of recursive dyadic partitions and finite mixture models. The entire flash card from a digital camera, and the system will first component of these models allows for the sparse representation automatically recognize and stitch panoramas without user input. of spatial structure at multiple resolutions. The second component provides a natural mechanism for capturing the varying degrees of A method for panorama recognition was first introduced in [Lowe and mixing of pure categories that accompany the use of different Brown, ICCV’03], where the method is based on a fast matching of resolutions, and enables them to be related to a user-specified scale invariant image features (SIFT), RANSAC, bundle adjustment and hierarchy of labels at multiple granularities in a straightforward multi-band image blending. The authors demonstrate their system in a manner. A segmentation is produced in our framework by selecting an many unordered datasets. Another method is described in [Brown optimal mixlet model, through complexity-penalized maximum et.al, CVPR’05]. It is based on matching of multi-scale image patches likelihood, and summarizing the information in that model with respect (MOPS) and can work with unordered sets. In these methods a full set to the categorical hierarchy. We motivate and illustrate the of invariant image descriptors is extracted at the location of interest performance of the proposed methodology in the context of remote points. The number of such descriptors is usually a few hundreds, and sensing land cover classification. this is computationally demanding step. Then these descriptors are matched using efficient nearest neighbour search in a high- dimensional descriptor space. 6498-10, Session 3 We propose a different approach, where we do not need to extract full set of image descriptors, and usually a small number of descriptors are Riemannian metrics on landmark-based shape spaces: required to register a pair of images. First, we extract interest points theory and applications from all available images and sort them according to corner strength M. Micheli, Brown Univ. measure. To register a pair of images, we consider only the point pairs, having similar corner strengths. Thus we reduce a number of potential The theory of Shape Spaces is closely connected to the theory of target point candidates for any source point. Then we compute image shape deformation, which is very central in computer vision. Direct region descriptors and store them for re-use with different point pairs. applications of such theory include object recognition, target Our image descriptor is normalized Fourier transform applied to the detection and tracking, classification of biometric data, and log-polar transformed image region centered at an interest point. Such automated medical diagnostics, to name a few. In this paper we shall descriptors are matched using 2D-phase correlation method give a brief introduction to the subject, focusing more on the general recovering unknown local scale difference and rotation between image principles and motivations rather than on mathematical detail. It turns regions. We update clusters in a rotation-scale accumulator with new out that Shape Spaces can be appropriately modeled as Riemannian successful point matches. Our extension of probabilistic approach manifolds, whence it is natural to study topics such as metrics, from [Lowe and Brown, ICCV’03] determines moments when these geodesics (paths of minimal length) between two shapes, and clusters are further processed with RANSAC estimator to find a set of curvature of the space. The last part of the paper summarizes some inliers of image homography. If number of inliers and global similarity recent results on curvature, emphasizing their importance in between images are sufficient, a fast geometry-guided point matching applications such as the computation of templates and doing is performed to finish image registration and improve the accuracy of statistics in medical imaging. the transformation parameters. A global registration graph is updated with each new image

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 127 Conference 6498: Computational Imaging V

6498-11, Session 3 6498-15, Session 4 4D segmentation of cardiac data using active surfaces Efficient multispectral imaging using compressive sensing with spatiotemporal shape priors D. Takhar, J. N. Laska, K. F. Kelly, R. G. Baraniuk, Rice Univ. A. Abufadel, A. J. Yezzi, Jr., Georgia Institute of Technology; R. W. TBD Schafer, Hewlett-Packard Labs. TBD 6498-16, Session 4 6498-12, Session 3 Characterizing compression efficiency of sparse image representation techniques Automated segmentation of alloy microstructures in A. Zakhor, Univ. of California/Berkeley serial section images In this paper we investigate methods of finding sparse representations J. Dumke, M. L. Comer, Purdue Univ. of a signal, i.e. one with the fewest non-zero coefficients. Finding a TBD sparse representation of a signal in an overcomplete dictionary is equivalent to solving a regularized linear inverse. For a given dictionary, finding the maximally sparse w is an NP-hard problem. A 6498-13, Session 3 great deal of recent research has focused on computationally feasible methods for determining highly sparse representations and is fueled Fast Mumford-Shah segmentation using image scale by applications in signal processing, compression and feature space bases extraction. C. V. Alvino, Univ. of Pennsylvania; A. J. Yezzi, Jr., Georgia In this paper we formulate the problem of finding a sparse inverse Institute of Technology solution and provide an overview of several popular techniques such as Method of Frames (MOF), Matching Pursuits (MP), Basis Pursuit (BP), Image segmentation using the piecewise smooth variational model Focal Underdetermined System Solution (FOCUSS), and Sparse proposed by Mumford and Shah is robust, but is also computationally Bayesian Learning (SBL). We then characterize the performance of expensive. Fortunately, both the intermediate segmentation functions each method, and compare their strengths and weaknesses in image computed in the process of the evolution, and the final segmentation compression applications. function itself have a common smooth structure. They are well approximated by linear combination of blurred versions of the original image. In this paper, we present methods for fast approximations to 6498-17, Session 4 Mumford-Shah segmentation using reduced image bases as an alternative to the standard method of considering each pixel in the Variational methods for compressive sampling image domain as its own basis function. Bases that are derived from J. Romberg, Georgia Institute of Technology; F. Park, Univ. of diffused version of the original image are considered. We show that, Michigan; E. J. Candes, California Institute of Technology when using such bases, the majority of the robustness of Mumford- Shah segmentation can be obtained without allowing each pixel to TBD vary independently in the implementation. In this fashion, we obtain most of the power of the original pixel-by-pixel Mumford-Shah model, while incurring only a small fraction of the computational cost. We 6498-18, Session 4 illustrate segmentations of real images that show how the proposed segmentation method is both computationally inexpensive, and has Compression of noisy Bayer pattern color filter array comparable performance to Mumford-Shah segmentations where images each pixel is allowed to vary freely. A. V. Bazhyna, K. O. Egiazarian, Tampere Univ. of Technology (Finland); N. N. Ponomarenko, V. V. Lukin, National Aerospace Univ. 6498-14, Session 4 (Ukraine) Color Filter Arrays (CFAs) are widely used in digital photo and video A conformal approach to shape representation cameras. Captured data is corrupted by a signal and exposure M. Feiszli, Brown Univ. dependent quantum noise. An automatic image processing carrying out in camera usually implies gamma and color corrections and Representation and comparison of shapes is a problem with many demosaicing. Thus, noise in images becomes of complex nature and applications in computer vision and imaging, including object spatially correlated. This results in drastic decrease of posterior noise recognition and medical diagnosis. We will discuss some reduction efficiency. Considerably better quality of final images is constructions from the theory of conformal mapping and hyperbolic achieved if non-processed CFAs (in RAW format) are extracted from geometry which provide ways to represent and compare planar camera and processed on PC with more efficient image reconstruction shapes. It is a remarkable fact that conformal maps from the unit disk algorithms. The major drawback of usage of RAW format is large size to a planar domain encode the geometry of the domain in useful and of file. Existing lossless image compression methods adapted for CFA tangible ways. Two examples of the relationship between conformal provide compression ratios of abn methods is defensible in this case mapping and geometry are provided by the medial axis and the while final decrease of effectiveness of noise reduction is inessential. boundary curvature of a planar domain. Both the medial axis and the The paper describes a method of adaptive selection of quantization boundary curvature can be used in applications to compare and step for each block of a CFA image for DCT based image compression. describe shapes and both appear clearly in the conformal structure. The achieve We’ll discuss some of the results from classical geometric function theory and modern work in hyperbolic geometry that demonstrate precisely how conformal mapping encodes the geometry of a planar 6498-19, Session 4 domain. In this conformal setting we can also construct metrics in order to compute distances which reflect the similarity or dissimilarity Multiscale reconstruction for computational spectral of shapes. imaging R. M. Willett, Duke Univ. Recent work in the emerging field of “compressed sensing” indicates that when the signal of interest is very sparse (i.e. zero-valued at most

128 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6498: Computational Imaging V locations) or highly compressible in some basis, relatively few Leibler divergence, and Bhattacharyya distance, to extract the incoherent observations are necessary to reconstruct the most structural damage related change. Damage indices are further significant non-zero signal components. In this work we develop a formulated for individual urban structures. We particularly focus on spectral imaging system and associated reconstruction methods that evaluating the performance of these dissimilarity-based change have been designed to exploit this theory. Conventionally, spectral detection methods. An empirical approach under the framework of imaging systems measure complete “data cubes” and are subject to pattern classification and cross-validation is adopted in this paper. performance limiting tradeoffs between spectral and spatial With the use of a pair of bi-temporal satellite images captured before resolution. We achieve single shot full 3D data cube estimates by and after a major earthquake in Bam, Iran, on Dec. 2004, the using compressed sensing reconstruction methods to process extracted damage indices calculated based on different types of observations collected using an innovative, real-time, dual-disperser dissimilarity measures are evaluated. The paper concludes that the spectral imager. The physical system contains a transmissive coding proposed change detection methods can be an effective component element located between a pair of matched dispersers, so that each in performing quick damage estimation for disaster-affected urban pixel measurement is the coded projection of the spectrum in the areas. corresponding spatial location in the spectral data cube. Using a novel multiscale representation of the spectral image data cube, we are able to accurately reconstruct 256Å~256Å~15 spectral image 6498-46, Poster Session cubes using just 256 Å~256 measurements. Superresolution single-particle Cryo-EM tomography of icosahedral virus using compressed sensing theory 6498-52, Session 4 M. Kim, J. Choi, L. Yu, J. Ye, Korea Advanced Institute of Science Importance prioritized coding of aerial surveillance video and Technology (South Korea) using H.264 Sparse object supports are often encountered in many imaging D. Butler, H. Cheng, J. Cook, Sarnoff Corp. problems. For such sparse objects, recent theory of compressed sensing tells us that accurate reconstruction of images or objects are No abstract available possible even from highly limited number of measurements drastically smaller than the Nyquist sampling limit. In this paper, we employ the compressed sensing theory for the high resolution 3-D reconstruction 6498-44, Poster Session of icosahedral virus using cryo-EM. Cryo-EM tomography is a nice fit for the compressed sensing theory because of the following reasons: A computational 3D microscope model for 1) due to the difficulty in sample collection, each experiment can reconstruction of translucent neural stem cells in bright- obtain micrographs with limited number of virus samples, providing field time-lapse microscopy severely undersampled projection data, and 2) the nucleic acid of a J. H. Degerman, Chalmers Tekniska Högskola (Sweden); E. viron is enclosed within capsid composed of a few proteins; hence its Winterfors, Univ. Pierre et Marie Curie (France); T. Gustavsson, support in 3-D real space is quite sparse. Three dimensional structure is obtained by solving L1 optimization problem based on a novel Chalmers Tekniska Högskola (Sweden) iterative algorithm that iterates between projection, backprojection This paper describes a computational model for image formation of and shrinkage operation. The experiments confirm that the in-vitro adult hippocampal progenitor (AHP) cells, in bright-field time- compressive sensing algorithm provides a superior reconstruction of lapse microscopy. Although these objectives barely generates the 3-D structures compared to the conventional reconstruction sufficient contrast for imaging translucent cells, we show that by using algorithms. a stack of defocused image slices it is possible to extract position and shape of spherically shaped specimens, such as the AHP cells. This inverse problem was solved by modeling the physical objects and 6498-47, Poster Session image formation system, and using an iterative nonlinear optimization algorithm to minimize the difference between the calculated and GPU-based visualization techniques for 3D microscopic measured image stack. By assuming that the position and shape of imaging data the cells do not change significantly between two time Q. Wang, Y. Sun, J. P. Robinson, Purdue Univ. frames, we can optimize these parameters using the previous time While computer volume visualization has advanced into a frame in a Bayesian estimation approach. The 3D reconstruction sophisticated level in medical applications, much fewer studies have algorithm settings, such as focal distance and CCD gain, were been made on data acquired by 3D confocal imaging. To optimize the calibrated and cross-validated using two different types of latex visualization of such data, it is important to consider the data spheres of known size and refractive index. Image intensity PSNR for characteristics. It is also interesting to apply the new GPU (graphics a sphere stack of 253 pixels was 28 dB after calibration. A qualitative processing unit) technology to interactive volume rendering. In this analysis was done using imuno-histochemically stained paper, we discuss several texture-based techniques to visualize and confocal scanned progenitor cells, showing that the cell shape confocal microscopy data by considering the data characteristics and can be retrieved as long as the cell is sufficiently spherical. with support of GPU. One simple technique generates one set of 2D textures along the axial direction of image acquisition. An improved technique uses three sets of 2D textures in the three principal 6498-45, Poster Session directions, and creates the rendered image via a weighted sum of the images generated by blending the individual texture sets. In addition, Empirical performance evaluation of dissimilarity we propose a new approach that textures are blended based on a measures for use in urban structural damage detection stencil control. Given the viewing condition, a texel needs to be drawn only when its corresponding projection on the image plane is inside a Z. Chen, T. C. Hutchinson, Univ. of California/Irvine stencil area. Finally, we have explored the use of multiple-channel Multi-temporal earth-observation imagery is now available at sub- datasets for flexible classification of objects. meter accuracy and has been found very useful for performing quick damage estimation for urban areas affected by large-scale disasters. In this paper, we propose a new perspective for performing change detection, where dissimilarity measures are used to characterize urban structural damage. We use the magnitudes of image gradients and local variances as the means to capture urban structural characteristics, and a family of distribution dissimilarity measures, including Euclidean distance, Cosine, Chi-square statistic, Kullback-

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 129 Conference 6498: Computational Imaging V

6498-48, Poster Session processing unit (SPU), a memory flow controller (MFC), and 256 kB of SRAM that are used as local store (LS) memory. All components, GICEB: automatic segmentation algorithm for biomedical SPU, SPEs, local store and the large XDR RAM that interfaces to the images CBE run at the full 3.2 GHz clock. We have implemented and optimized several tomographic image reconstruction components and Q. Wang, Y. Sun, Purdue Univ. benchmarked them against optimized PC-based implementations. Segmenting images and datasets is an essential problem in medical Among those components are parallel-beam and cone-beam imaging. Generally speaking, it is very challenging to obtain an forward- and backprojection, iterative cone-beam image automatic segmentation that works robustly and accurately. Some reconstruction, statistical cone-beam image reconstruction and semi-automatic techniques have been proposed, but they are slow approximate image reconstruction for spiral cone-beam CT. The and non-intuitive. In particular, these semi-automatic techniques rely modalities covered are clinical CT, C-arm CT and micro-CT. A speed- significantly on the user intervention to achieve accurate results, up factor of 10 to 20 was easily achieved with the cell processor. This especially when the datasets have complex structures. In this paper, new technology can therefore either be used to speed up image we propose a novel automatic segmentation algorithm called reconstruction or to improve image quality by allowing for the routine Gradient-Intensity Clusters and Expanding Boundaries (GICEB). This use of more complex reconstruction software. technique uses multiple parameters whose values depend upon the distinct object (i.e. tissue class). While most existing threshold-based techniques are global and use one single parameter in segmentation, 6498-23, Session 5 GICEB uses the intensity and gradient values along with the spatial connectivity information. This method can automatically segment the Iterative reconstruction of cone-beam CT data on a data into different tissues from the histogram obtained using the cluster intensity and gradient values. It identifies the locations with very high T. M. Benson, GE Global Research; J. Gregor, The Univ. of frequency in the histogram automatically. As a crucial component, our technique combines the two-dimensional histogram, the spatial Tennessee connectivity, and region growing to separate different segments from Three-dimensional iterative reconstruction of large CT data sets poses one another. several challenges in terms of the associated computational and memory requirements. In this paper, we present results obtained by implementing a computational framework for reconstructing axial 6498-20, Session 5 cone-beam CT data using a cluster of inexpensive dual-processor PCs. In particular, we discuss our parallelization approach, which uses Why do GPUs work so well for acceleration of CT? POSIX threads and message passing (MPI) for local and remote load K. D. Mueller, F. Xu, N. Neophytou, Stony Brook Univ. distribution, as well as the interaction of that load distribution with the implementation of ordered subset based algorithms. We also consider Commodity graphics hardware (GPUs) boards have enabled a heuristic data-driven 3D focus of attention algorithm that reduces remarkable speedups in various sub-areas of CT. This paper takes a the amount of data that must be considered for many data sets. deeper look into the GPU architecture and programming model to Furthermore, we present a modification to the SIRT algorithm that gain better insight which type of CT task and algorithm accelerates reduces the amount of data that must be communicated between better, and which accelerates less. Hallmarks for both types are given, processes. Finally, we introduce a method of separating the work in and it is shown that GPUs represent a commodity high-performance such a way that some computation can be overlapped with the MPI parallel architecture that resonates very well with the computational communication thus further reducing the overall run-time. We structure and operations inherent to CT. In fact, there is evidence summarize the performance results using reconstructions of suggesting that GPU-based ‘brute-force’ CT (i.e., CT at regular experimental data. complexity) can be significantly faster than CPU-based as well as GPU-based CT with optimized complexity, at least for practical data sizes. The performance gap is speculated to grow even more, since, 6498-51, Session 5 due to strong demands by the mass entertainment market, GPU performance has consistently doubled every six months, while CPU Evolution of computer technology for fast cone-beam performance traditionally doubles every 18 months. Finally, we will backprojection also have a brief look at the new Cell architecture in this regard. This new architecture, just as the GPU, is also a SIMD system, but is more I. Goddard, A. Berman, Mercury Computer Systems, Inc.; O. generalized than GPUs. Bockenbach, Mercury Computer Systems GmbH (Germany); F. P. Lauginiger, Mercury Computer Systems, Inc.; S. Schuberth, Mercury Computer Systems GmbH (Germany); S. Thieret, 6498-21, Session 5 Mercury Computer Systems, Inc. Hardware-accelerated deformable image registration for Cone-beam reconstruction (CBR) is useful for producing volume image-guided interventions images from projections in many fields including medicine, biomedical research, baggage scanning, paleontology, and nondestructive R. Shekhar, O. Dandekar, Univ. of Maryland Medical Ctr. manufacturing inspection. Most commonly used for CBR is the TBD Feldkamp algorithm; this involves filtering and cone-beam backprojection steps for each projection of the set. Often there is an economic or clinical reason for the reconstruction to be as fast as 6498-22, Session 5 possible, and performance optimization often involves a special computing platform. Tomographic image reconstruction using the cell Over the past decade we have observed or studied Feldkamp on broadband engine (CBE) general purpose hardware many different processing engines: M. Kachelriess, M. Knaup, S. Steckmann, Friedrich-Alexander- CPU (general purpose) Univ. Erlangen-Nürnberg (Germany); O. Bockenbach, Mercury ASIC Computer Systems GmbH (Germany) FPGA Cell processors are general purpose processors that combine a DSP PowerPC element (PPE) with eight synergistic processor elements (SPE). The SPEs are the most interesting feature of the Cell GPU (graphics processor) broadband engine (CBE), as they are the source of its processing Cell Broadband Engine(tm) power. A single chip contains eight SPEs, each with a synergistic

130 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6498: Computational Imaging V

We consider both single processor and parallel multiprocessor 6498-26, Session 6 architectures. Design factors include speed, scalability, computational accuracy, ease of programming, power, and cost; often the choice of Nonlocal evolutions for image regularization platform restricts the maximum practical size of volume that can be G. Gilboa, S. J. Osher, Univ. of California/Los Angeles reconstructed. Overall, it seems that special purpose processors (including GPU, A nonlocal quadratic functional of weighted differences is examined. FPGA, and ASIC) are best in terms of speed, power, and cost. Since The weights are based on image features and represent the affinity the algorithm is well-understood and stable, the programming effort of between different pixels in the image. By prescribing different the algorithm kernel is not a large factor unless time-to-market is formulas for the weights, one can generalize many local and nonlocal crucial. If image dimensions must be very large, a scalable linear denoising algorithms, including nonlocal means and bilateral multiprocessor engine may be indicated but is liable to be expensive. filters. The steepest descent for minimizing the functional can be interpreted as a nonlocal diffusion process. We show state of the art denoising results using the nonlocal flow. 6498-24, Session 6 Higher order bilateral filters and their applications 6498-27, Session 6 H. Takeda, S. Farsiu, P. Milanfar, Univ. of California/Santa Cruz Parametric point spread function modeling and reduction The bilateral filtering idea was first proposed by Tomasi-Manduchi of stray light effects (1998) as a very effective one pass adaptive filter for denoising B. Bitlis, Purdue Univ.; P. A. Jansson, The Univ. of Arizona; J. P. purposes while keeping the edges relatively sharp. Unlike Allebach, Purdue Univ. conventional filters, the bilateral filter defines the closeness of two pixels not only based on geometric distance but also based on In any real optical imaging system, some portion of the entering light radiometric distance. Later, Elad (2002) proved that such filter is a flux is misdirected to undesired locations due to scattering from single Jacobi iteration of a weighted least squares minimization, and surface imperfections and multiple reflections between optical suggested using more iterations to enhance the quality of estimation. elements. This unwanted light is called stray light. Its effects include In this paper, to further improve the performance and find new lower contrast, reduced detail and color inaccuracy. applications, we make contact with a classic non-parametric image There are methods for removing this contribution when the point reconstruction technique called kernel regression, which is based on spread function (PSF) characterizing the stray light effect is known. local Taylor expansions of the regression function. We extend and For digital still cameras, we assume a parametric, shift-variant, generalize the kernel regression method and show that bilateral rotationally invariant PSF model. For collection of data to estimate the filtering is a special case of this new class of adaptive image parameters of this model, we use a light source box that provides reconstruction techniques, considering a specific choice for weighting nearly uniform illumination behind a circular aperture. Several images kernels and zero order Taylor approximation. We show improvements of this light source are captured when it is at different locations in the over the classic bilateral filtering can be achieved by using higher field of view of the camera. Also, another exposure of each scene with order Taylor approximations. Moreover, the proposed framework different shutter speed is used to provide details in the darker regions. enables us to use bilateral filters for applications other than denoising, A subset of the data obtained from these images is used in a such as upscaling, interpolation, and deblurring. The resulting nonlinear optimization algorithm. algorithms and analyses are amply illustrated with practical examples. After estimating the parameters of the PSF model, we provide the results of applying the correction algorithm to the images taken of real 6498-25, Session 6 world scenes. The use of levelable regularization functions for MRF restoration of SAR images while preserving reflectivity 6498-28, Session 6 J. Darbon, EPITA (France); M. Sigelle, F. Tupin, École Nationale An image decomposition model using the total variation Supérieure des Télécommunications (France) and the infinity Laplacian A recurrent problem in image restoration using Markov Random Fields L. A. Vese, C. Elion, Univ. of California/Los Angeles; J. Morel, with Total Variation (TV) prior is the loss of contrast between restored École Normale Supérieure de Cachan (France) objects and their background. TV is the paradigm of those regularization functionals called levelable, which can be decomposed This paper is devoted to a recent topic in image analysis: the on discrete level sets. Levelable posterior energies can be exactly decomposition of an image into a cartoon or geometric part, and an minimized using maximum-flow techniques. We present here first oscillatory or texture part. Here, we propose a practical solution to the application to the denoising of Synthetic Aperture Radar (SAR) images (BV,G) model proposed by Yves Meyer. We impose that the cartoon is while preserving the reflectivity of each region of interest. a function of bounded variation, while the texture is represented as Laplacian of some function that belongs to Linfinity. The problem thus A nice-levelable function of two gray levels x1 and x2 writes as |S(x1) - becomes related with the absolutely minimizing Lipschitz extensions S(x2)|, where S is monotone increasing (TV corresponds to S(x) = x). and the infinity Laplacian. Experimental results together with details of Now consider piecewise constant object and background corrupted the algorithm will be presented. by overall noise. For the continuous quadratic + TV model, the loss of contrast between object and background decreases if both derivatives of S at original background and object levels are low. We 6498-29, Session 7 thus design a discrete levelable regularization function with low slope R(x) = S(x+1) - S(x) around these very gray levels and R(x) = 1 Object tracking using joint-visible and thermal-infrared everywhere else like TV! Extension to other types of noise follows video sequences readily. R. Péteri, O. Siler, E. Bichot, P. Courtellemont, Univ. de La This method was tested for a small circular object and gaussian noise Rochelle (France) such that low contrast with background occurs. We then tested a synthetic 1-look speckle image with four piecewise constant regions Particle filter methods based on color distribution can be used to on which a Rayleigh distribution was superimposed. Our nice- track non-rigid moving objects in color videos. They are robust in case levelable regularization allows spectacular preservation of contrast of noise or partial occlusions. However, using particle filters on color and original luminances on both images as compared to TV. videos is sensitive to changes in the lightening conditions of the scene. The use of thermal infrared image sequences can help the tracking process, as thermal infrared imagery is not sensitive to

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 131 Conference 6498: Computational Imaging V

lightening conditions. This paper presents a particle filter based 6498-32, Session 8 method for object tracking using automatic cooperation between the color and the infrared modalities. As the infrared and the visible image Multigrid optimal mass transport for image registration sequences are acquired with different cameras, a pre-step is spatio- and morphing temporal registration. After spatio-temporal registration, the proposed method is able to continuously track the target despite difficult T. U. Rehman, A. R. Tannenbaum, Georgia Institute of Technology conditions appearing in one of the modality. Our cooperative tracking In this paper we present a computationally efficient Optimal Mass method is successfully applied on several experimental datasets. Transport algorithm. This method is based on the Monge-Kantorovich Different test sequences are presented, including tracking in the theory and is used for computing elastic registration and warping visible video with the help of the infrared modality, or tracking in the maps in image registration and morphing applications. This is a infrared with the help of the visible modality. Comments and future parameter free method which utilizes all of the grayscale data in an prospects raised by this method are finally described. image pair in a symmetric fashion. No landmarks need to be specified for correspondence. In our work, we demonstrate significant improvement in computation time when our algorithm is applied as 6498-30, Session 7 compared to the gradient descent method originally proposed by Symmetry detection in 3D scenes Haker et al. The original algorithm was based on a gradient descent method for removing the curl from an initial mass preserving map T. Sawada, K. Sanik, Z. Pizlo, Purdue Univ. regarded as 2D vector field. This involves inverting the Laplacian in every iteration which is now computed using full multigrid technique Retinal image of a symmetric object is itself symmetric only for a small resulting in an improvement in computational time by a factor of two. set of viewing directions. Interestingly, human observers have little Greater improvement is achieved by decimating the curl in a multi- difficulty in determining whether a given retinal image was produced resolutional framework. The algorithm was applied to both synthetic by a symmetric object, regardless of the viewing direction. We tested and 2D short axis cardiac MRI images for testing and comparison. perception of planar (2D) symmetric figures (dotted patterns and polygons) when the figures were slanted in depth. We found that symmetry could be detected reliably with polygons, but not with 6498-33, Session 8 dotted patterns. Next, we tested the role image features representing the symmetry of the pattern itself (orientation of projected symmetry The effect of intergrid operators on multigrid axis or symmetry lines) vs. that representing the 3D viewing direction convergence (orientation of the axis of rotation). We found that symmetry detection is improved when the projected symmetry axis or lines are known to P. Navarrete Michelini, E. J. Coyle, Purdue Univ. the subject, but not when the axis of rotation is known. Finally, we We study the effect of interpolation and restriction operators on the showed that performance with orthographic images is higher than that convergence of multigrid algorithms for solving linear PDEs. Using a with perspective images. A computational model of symmetry modal analysis of a subclass of these systems, we determine how two detection will be presented. This model operates on shapes of groups of the modal components of the error are filtered and mixed at contours, rather than on corresponding vertices. Generalization of each step in the algorithm. We then show that the convergence rate of these results to the case of images of 3D symmetric objects will be the algorithm depends on both the properties of the interpolation and discussed. restriction operators and the characteristics of the system. The analysis opens the problem of optimization of these operators. By different choices of operators we show a trade-off between the 6498-31, Session 7 optimization of the convergence rate and the optimization of the Object recognition investigation using information- number of computations required per iteration. theoretic measures/metrics D. W. Repperger, A. R. Pinkus, Air Force Research Lab.; J. A. 6498-34, Session 9 Skipper, Wright State Univ. Using transmission electron microscopy to quantify the Discrimination of friendly or hostile objects is studied in an image that spatial distribution of nanoparticles suspended in a film has been compromised by a number of factors. The means by which the image objects can be degraded include reduced contrast, varying S. D. Newsam, E. Pernice, J. Jasinski, V. Leppert, Univ. of brightness levels, and relative object size changes. In aerial military California/Merced images, objects with different orientations can be accurately Advances in nanotechnology have resulted in a variety of exciting new represented by a single identification signature consisting of the nanomaterials, such as nanotubes, nanosprings and suspended average histogram of the object under rotation. Three different nanoparticles. Being able to characterize these materials is important information-theoretic measures/metrics are investigated as possible for refining the manufacturing process as well as for determining their criteria to classify the objects. The first measure is the standard optimal application. The scale of the nanocomponents makes high- mutual information (MI) between the sampled object and the library resolution imaging, such as electron microscopy, a preferred method object signatures. A second measure is based on information for performing the analyses. This work focuses on the specific efficiency, which differs from MI. Finally an information distance metric problem of using transmission electron microscopy (TEM) to quantify is defined which determines the difference, between the sampled the spatial distribution of nanoparticles suspended in a film. We object and the library object. It is shown the three information- investigate two approaches for estimating the distribution of the theoretic variables introduced here form an independent basis in the particles in the z-direction. First, we consider a direct approach in sense that any variable in the information channel can be uniquely which we use image segmentation to locate the particles in the tilted expressed in terms of the three parameters introduced here. The images. Then, by establishing correspondences between the methodology discussed is tested on a sample set of standardized segmented particles in differently tilted images using simple image images to evaluate their efficacy. Performance standardization is features, and by noting how these particles move in relation to each based on manipulation of contrast, brightness, and size attributes of other due to parallax, we estimate the spatial distribution of the the sample objects of interest. particles. We also investigate a second, indirect approach in which a Fourier-based spectral decomposition of the images is used to estimate the distribution of the particles in the z-direction.

132 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6498: Computational Imaging V

6498-35, Session 10 moving average (ARIMA) noise reveal the proposed procedure to be more sensitive and selective than conventional fMRI analysis methods Image inpainting based on energy minimization (reference set: principle component analysis, PCA; independent N. Kawai, T. Sato, N. Yokoya, Nara Institute of Science and component analysis, ICA; k-means clustering k=100; univariate t-test) in identification of active regions over the range of average contrast- Technology (Japan) to-noise ratios of 0.5 to 4.0. Results of analysis of extant human data Image inpainting techniques have been widely used to remove (for which the average contrast-to-noise ratio is unknown) are further undesired visual objects in images such as damaged portions of suggestive of greater statistical detection power. Refinement of this photographs and people who have accidentally entered into pictures. new procedure is expected to reduce both false positive and negative Conventionally, the missing parts of an image are completed by rates, without resorting to filtering that can reduce the effective spatial minimizing the energy function which is defined based on the sum of resolution. SSD(sum of squared differences). However, the naive SSD-based energy function is not robust against intensity change in an image. Thus, unnatural intensity change often appears in the missing parts. In 6498-38, Session 11 addition, when an image has continuously changing texture patterns, the completed texture in a resultant image sometimes blurs due to Fast joint estimation of magnitude, decay, and frequency inappropriate pattern matching. In this paper, in order to improve the from a single-shot MRI signal image quality of the completed texture, the conventional energy W. Tang, S. J. Reeves, Auburn Univ.; D. B. Twieg, The Univ. of function is newly extended by considering intensity changes and spatial Alabama/Birmingham locality to prevent unnatural intensity changes and blurs in a resultant image. By minimizing the extended energy function, the missing regions By acknowledging local decay and phase evolution, single-shot can be completed without unnatural intensity changes and blurs. In parameter assessment by retrieval from signal encoding (SS-PARSE) experiments, the effectiveness of our proposed method is successfully models each datum as a sample from (k, t)-space rather than k- demonstrated by applying our method to various images and space. This more accurate model promises better performance at a comparing the results with those obtained by the conventional method. price of more complicated reconstruction computations. Normally, conjugate-gradients is used to simultaneously estimate image magnitude, local decay, and frequency. Each iteration of the 6498-36, Session 10 conjugate-gradients algorithm requires several evaluations of the image synthesis function and one evaluation of gradients. Because of Signal reconstruction from a periodic nonuniform set of local decay and frequency and the non-Cartesian trajectory, fast samples using H_infinity optimization algorithms based on FFT cannot be directly used to accelerate the H. T. Nguyen, M. N. Do, Univ. of Illinois at Urbana-Champaign evaluation of the image synthesis function and gradients. This paper presents a fast algorithm to compute the image synthesis function We study the problem of signal reconstruction from a set of periodically and gradients by linear combinations of FFTs. By polynomial nonuniform samples. The considered system takes samples of delayed approximation of the exponential time function with local decay and versions of a continuous signal at low sampling rate, with different frequency as parameters, the image synthesis function and gradients fractional delays for different channels. Unlike traditional sampling and become linear combinations of non-Cartesian Fourier transforms. In interpolation methods, our approach uses techniques from control order to use the FFT, one can interpolate non-Cartesian trajectories. theory which suggests a nice connection from the continuous domain The quality of images reconstructed by the fast approach presented in and the discrete domain. We show that the system is equivalent to a this paper is the same as that of the normal conjugate-gradient discrete-time system with some analysis IIR filters in place of the method with significantly reduced computation time. fractional delays at channels. We then design IIR synthesis filters so that the overall system approximates an A/D converter of high sampling rate (with some delay) using techniques from model-matching problem 6498-49, Session 11 in control theory with read-made software in Matlab. The synthesis filters are designed so that they minimize the H infinity norm, i.e. Simultaneous surface and volume registration using maximum energy gain, of the error system. The experiments are also harmonic maps run for synthesized images. A. A. Joshi, D. W. Shattuck, Univ. of Southern California; P. Thompson, Univ. of California/Los Angeles; R. M. Leahy, Univ. of 6498-54, Session 10 Southern California Morphometric studies of anatomical changes in brain structures over A multiscale statistical model for time series forecasting time or across populations require that the brain data first be W. Wang, Purdue Univ.; V. Portnoy, Jefferies & Co.; I. Pollak, Purdue transformed to a common coordinate system in which the anatomical Univ. structures are aligned. In this paper, we describe a method for volumetric registration that also produces a bijective point No abstract available correspondence between cortical surfaces. We begin by assigning a 2D coordinate system to the cortical surfaces of each of the two brain volumes such that sulcal landmark coordinates are aligned. In order to 6498-37, Session 11 generate such a parameterization, we model the cortical surface as an elastic sheet and solve the associated Cauchy-Navier equation using Regional approach to event-related fMRI data analysis a finite element method in the intrinsic surface geometry. We map using hemodynamic response modeling each of the aligned surfaces to a unit sphere and then use a L. Liu, A. Rao, T. Talavage, Purdue Univ. constrained harmonic mapping to define a volumetric correspondence between the brains with sulcal landmark alignment constraints. The Analysis of functional magnetic resonance imaging (fMRI) data has intermediate spherical space constrains the surface flow to the been performed using both model-driven (parametric) methods and surface so that the two cortical surfaces remain aligned in the data-driven methods. An advantage of model-driven methods is volumetric harmonic maps. Finally, the correspondence is refined incorporation of prior knowledge of spatial and temporal properties of using an intensity-based warp. We show examples which the hemodynamic response (HDR). A novel analytical framework for demonstrate the cortical matching properties and the ability of this fMRI data has been developed that identifies multi-voxel regions of method to align surfaces as well as subcortical structures. activation through iterative segmentation-based optimization over HDR estimates for both individual voxels and regional groupings. Simulations using synthetic activation embedded in autoregressive integrated

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 133 Conference 6498: Computational Imaging V

6498-39, Session 12 6498-53, Session 12 High-resolution image reconstruction for PET using Statistical modeling of the myosin lattice in vertabrate detector response muscle M. S. Tohme, J. Qi, Univ. of California/Davis C. Yoon, R. P. Millane, N. D. Blakeley, A. Goyal, Univ. of Canterbury (New Zealand) The accuracy of the system model in an iterative reconstruction algorithm greatly affects the quality of reconstructed positron Muscle fibers contain the contractile proteins myosin and actin that emission tomography images. For efficient computation in are organized into long thin strands known as myofibrils. The cross- reconstruction, the system model in PET can be factored into a sections of individual myofibrils can be imaged by electron product of geometric projection and detector blurring, where the microscopy of thin transverse sections. Such micrographs show that former is often computed based on analytical calculation, and the within a single myofibril, the myosin filaments pack on a two- latter is estimated using Monte Carlo simulations. In this work, we dimensional triangular lattice, and that in some vertebrate muscles the propose a method to estimate the 2D blurring matrix from filament at each lattice site adopts one of two different orientations. experimental measurements. Point source measurements were The statistical characteristics of the spatial distribution of filament acquired with high count statistics in a relatively fine grid inside the orientations within a myofibril is of interest for a number of reasons. microPET II scanner using a high-precision 2-D motion stage. A First, the disorder affects the kinds of intermolecular contacts monotonically convergent iterative algorithm has been derived to between myosin and actin which is molecular driver of muscle estimate the detector blurring matrix from the point source contraction. Second, high resolution imaging of muscle structure is measurements. The algorithm also takes advantage of the rotational conducted using x-ray diffraction data from muscle specimens, and symmetry of the PET scanner with the modeling of the detector block interpretation of the diffraction data requires a model of the disorder structure. Since the resulting blurring matrix stems from actual to calculate the diffraction from model structures in order to solve the measurements, it can take into account the physical effects in the inverse problem. We have developed methods for automatic photon detection process that are difficult or impossible to model in a determination of the myosin filament orientations in electron Monte Carlo simulation, and hence provide a more accurate system micrographs, and here we describe analysis of the distribution of model. This method can be applied to clinical scanners. orientations to develop a model of the disorder. Analysis shows that the distribution has a number of distinct characteristics for nearest- and second-nearest-neighbor filaments. These characteristics 6498-40, Session 12 indicate a frustrated system which maps onto an antiferromagnetic Ising model. This is verified by conducting Monte Carlo energy Superresolution projection reconstruction MR imaging minimizations of this model on triangular lattices and comparing the using FOCUSS (PR-FOCUSS) second order statistics with those derived from the micrographs. We S. Tak, J. Ye, Korea Advanced Institute of Science & Technology find good agreement and are able to determine an effective (South Korea) temperature for the system that parameterizes the degree of disorder. This paper is concerned about super-resolution reconstruction of projection reconstruction MR imaging from angular under-sampled k- 6498-55, Session 13 space data. A similar problem has been recently addressed in the framework of compressive sensing theory. Unlike the existing Training-based algorithm for moiré suppression in algorithms used in compressive sensing theory, this paper employs scanned halftone images the FOCal Underdetermined System Solver (FOCUSS), which was H. Siddiqui, C. A. Bouman, Purdue Univ. originally designed for EEG and MEG source localization to obtain sparse solutions by successively solving quadratic optimization. We No abstract available show that FOCUSS is very effective for the projection reconstruction MRI, because the medical images are usually sparse in image domain, and the “soft’’ sparseness constraint used in FOCUSS avoids the visually unpleasant high frequency artifacts often observed in basis pursuit or matching pursuit reconstruction. Furthermore, the center region of the under-sampled radial k-space data still provides a meaningful low resolution image, which is essential for the convergence of FOCUSS. We applied FOCUSS for projection reconstruction MR imaging using single coil. Extensive experiments with synthetic data as well as in vivo brain data using 3.0T MRI scanner (ISOL Technology, Korea) confirms that high resolution reconstruction with virtually free of angular aliasing artifacts can be obtained from severely under-sampled k-space data.

6498-41, Session 12 Fast methods for tomographic reconstruction J. Thibault, GE Healthcare; Z. Yu, Purdue Univ.; K. Sauer, Univ. of Notre Dame; C. A. Bouman, Purdue Univ.; J. Hsieh, GE Healthcare TBD

134 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6499: Vision Geometry XV Thursday 1 February 2007 Part of Proceedings of SPIE Vol. 6499 Vision Geometry XV

6499-18, Poster Session 6499-01, Session 1 Perspex Machine IX: transreal analysis Perspex Machine VIII: axioms of transreal arithmetic A. A. Adams, J. A. D. W. Anderson, The Univ. of Reading (United J. A. D. W. Anderson, The Univ. of Reading (United Kingdom); N. Kingdom); N. Volker, Univ. of Essex (United Kingdom) Volker, Univ. of Essex (United Kingdom); A. A. Adams, The Univ. of Reading (United Kingdom) We introduce transreal analysis as a generalisation of real analysis. We find that the generalisation of the real exponential and logarithmic Transreal arithmetic is a total arithmetic that contains real arithmetic, functions is well defined for all transreal numbers. Hence, we derive but which has no arithmetical exceptions. The absence of exceptions well defined values of all transreal powers of all non-negative transreal makes it a more secure basis for computation than standard numbers. In particular, we find a well defined value for zero to the arithmetics, and allows the specification of the Universal Perspex power of zero. We then generalise all of the common, real, Machine which unifies geometry with the Turing Machine. Here we trigonometric functions to transreal functions and show that transreal axiomatise the algebraic structure of transreal arithmetic so that it (sin x)/x is well defined everywhere. This raises the possibility that provides a total arithmetic on any appropriate set of numbers. This transreal analysis is total, in other words, that every function and opens up the possibility of specifying a version of floating-point every limit is everywhere well defined. We go on to dispel all of the arithmetic that does not have any arithmetical exceptions and in which standard counter “proofs” that purport to show that division by zero is every number is a first-class citizen. This would be an improvement on impossible. This is done by carrying the proof through in transreal IEEE Floating Point Arithmetic where the objects NaN and +/- Inf are arithmetic or transreal analysis. We find that either the supposed not fixed numbers and which, therefore, have ill-defined semantics. counter proof has no content or else that it supports the contention Our axiomatisation is proved consistent by machine proof. It has that division by zero is possible. Finally, we show how to arrange that fewer and less restrictive side conditions than real arithmetic. a function has finite or else unmeasurable (nullity) values, but no infinite values. This arithmetical arrangement might prove useful in mathematical physics because it outlaws naked singularities in all 6499-02, Session 1 equations. Fidelity analysis of mechanically aided copying/enlarging of Jan van Eyck’s ‘Portrait of Niccolò Albergati’ 6499-19, Poster Session D. G. Stork, Ricoh Innovations, Inc.; M. F. Duarte, Rice Univ. Perspex Machine X: software development Artist David Hockney and scientist Charles Falco recently claimed that S. Noble, B. A. Thomas, J. A. D. W. Anderson, The Univ. of some artists as early as 1420 secretly used optical devices to project Reading (United Kingdom) images onto their canvases, traced them, and then applied paint, and that this was a source of the rise of realism in early Renaissance The Perspex Machine arose from the unification of computation with painting. They adduced as central evidence two works by Jan van geometry. We now report significant redevelopment of both a partial Eyck, a silverpoint study and a later, formal oil work, Portrait of C compiler that generates perspex programs and of a Graphical User Niccolò Albergati. Because of the high fidelity of the oil copy, they Interface (GUI). The compiler is constructed with standard compiler- claim the artist used an epidiascope, or primitive opaque projector, generator tools and produces both an explicit parse tree for C and an undocumented from that era. However, their optical explanation Abstract Syntax Tree (AST) that is better suited to code generation. suffers from a number of awkward implications, for instance van Eyck The GUI uses a hash table and a simpler software architecture to would surely have seen the “relative shift” between the two works and achieve an order of magnitude speed up in processing and, corrected it. Moreover, Ketelsen et al’s (2004) discovery of pinprick consequently, an order of magnitude increase in the number of holes in the silverpoint study very strongly indicates that van Eyck perspexes that can be manipulated in real time (now 6,000). Two used mechanical (not optical) methods. Such mechanical perspex-machine simulators are provided, one using trans-floating- Reductionszirkel or reducing compass was familiar to artists and point arithmetic and the other using transrational arithmetic. All of the documented from Roman times. software described here is available on the world wide web. In order to determine whether high fidelities could be achieved without optics, we had three professional artists copy/enlarge van Eyck’s 6499-20, Poster Session silverpoint study by mechanical methods known from the early 15th century and consistent with the number and location of the pinprick Aerial lidar data classification using expectation holes found by Ketelsen et al. The fidelities we found, expressed as a maximization Chamfer distance, were comparable to those achieved by van Eyck. We conclude that the most plausible explanation consistent with all S. K. Lodha, D. M. Fitzpatrick, D. P. Helmbold, Univ. of California/ evidence is that van Eyck used known, simple mechanical aids in his Santa Cruz work, rather than unknown complicated optical ones. We use the Expectation-Maximization (EM) algorithm to classify 3D aerial lidar scattered height data into four categories: road, grass, 6499-03, Session 1 buildings, and trees. To do so we use five features: height, height variation, normal Analysis of difference in orientations and focal lengths of variation, lidar return intensity, and image intensity. We also use only two arbitrary perspective viewing cameras lidar-derived features to organize the data into three classes (the road A. Ramachandran, Arizona State Univ.; R. Balasubramanian, and grass classes are merged). We apply and test our results using ten regions taken from lidar data collected over an area of Indian Institute of Technology Roorkee (India); K. Swaminathan, S. approximately eight square miles, obtaining higher than 94% Das, Indian Institute of Technology (India) accuracy. We also apply our classifier to our entire dataset, and For stereo imaging, it is a general practice to use two cameras of present visual classification results both with and without uncertainty. same focal lengths, with their viewing axis normal to the line joining We use several approaches to evaluate the parameter and model the camera centers. This paper analyses the result of difference in choices possible when applying EM to our data. We observe that our orientations and focal lengths of two arbitrary perspective viewing classification results are stable and robust over the various cameras, by deriving the epipolar lines and its corresponding subregions of our data which we tested. equations. This enables one to find the correspondence search space We also compare our results here with previous classification efforts in terms of focal length accuracies as well as camera orientation using this data.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 135 Conference 6499: Vision Geometry XV

parameters. Relevant numerically simulated results are also given. In the sensed data was sparse and so increase the density of various stereo imaging, two cameras of equal focal lengths are arranged with computed surface properties at such locations. Standard subdivision their axes parallel, to share a single common image plane. As a result, techniques are normally applied to the complete mesh and so add all points in the 3D space get projected into points in that joint image vertices and faces throughout the mesh. In modifying global adaptive plane. The join of corresponding image points of the two cameras is a subdivision schemes to perform local subdivision, it is necessary to line parallel to the base line joining the two camera centers. This line is guarantee smooth transition between subdivided regions and regions called the epipolar line, implying that, a pair of corresponding edges left at the original level so as to prevent the formation of surface needs to be searched, only within this same horizontal scanline. This artifacts at the boundaries between such regions. paper makes some contributions to this effect as an extension of work Moreover, the produced surface mesh needs to be suitable for of Sengupta for the different orientations and focal lengths of two successive local subdivision steps. We propose a novel approach for arbitrary perspective viewing cameras. incremental adaptive subdivision of triangle meshes which may be applied to multiple global subdivision schemes and which may be repeated iteratively without forming artifacts in the subdivided mesh. 6499-04, Session 1 The decision of where to subdivide in each iteration is determined Biologically motivated composite image sensor for deep- based on an error measure which is minimized through subdivision. field target tracking Smoothness between various subdivision levels is obtained through the postponement of local atomic operations. The proposed scheme P. Melnyk, R. A. Messner, Univ. of New Hampshire is evaluated and compared to known techniques using quantitative measures. The present work addresses the design of an image acquisition front- end for target detection and tracking within a wide range of distances. Inspired by raptor bird’s vision, a novel design for a visual sensor is 6499-07, Session 2 proposed. The sensor consists of two parts, each originating from the studies of biological vision systems of different species. The front end Three-dimensional mesh generation of an object from an is comprised of a set of video cameras imitating a falconiform eye, in image by shape-from-shading and anisotropic Delaunay particular its optics and retina. The back end is a software remapper triangulation that uses a popular in machine vision log-polar model of retino- cortical projection in primates. The output of this sensor is a M. Doi, Y. Takabe, Osaka Electro-Communication Univ. (Japan) composite log-polar image incorporating both near and far visual Three-dimensional (3D) modeling of object from two-dimensional (2D) fields into a single homogeneous image space. In such space it is image is important for many fields including computer graphics, easier to perform target detection and tracking for those applications image understanding and medical imaging. Shape-from-Shading is a that deal with targets moving along the camera axis. The target object popular method for 3D shape estimation from one image. For the 3D preserves its shape and size being handled seamlessly between mesh generation, Delaunay triangulation is effective. If depths for all cameras regardless of distance to the composite sensor. The pixels in an image are estimated by Shape-from-Shading and 3D prototype of proposed composite sensor has been created and is Delaunay triangulation is done for the depth data, it is not efficient. We used as a front-end in experimental mobile vehicle detection and propose an efficient method to generate 3D near-equilateral triangular tracking system. It has been tested inside a driving simulator and mesh from 2D image directly. In the proposed method, the projection results are presented. of 3D mesh on 2D image is generated by ellipsoidal bubble mesh method firstly. Then, the positions of mesh vertices in 3D space are 6499-05, Session 2 estimated from Lambertian model and the direction of axes of ellipsoidal bubble in ellipsoidal bubble method. We evaluated the Volumetric scattered data modeling based on modified proposed method by 3D mesh generation for a simulation image and Shepard’s method a picture taken by a digital camera. The results showed good estimation for the shapes of objects. K. Lee, J. M. Paik, Handong Global Univ. (South Korea) Alpha shapes are generalization of convex hull. The set of all real numbers alpha leads to a family of shapes capturing the intuitive 6499-09, Session 3 notion of “crude” versus “fine” shape of a point set. As alpha value Shear invariant 3D model retrieval decreases, the shape shrinks and gradually develops cavities. These cavities may join to form tunnels and voids. In reconstruction an S. S. Naik, B. Prabhakaran, The Univ. of Texas at Dallas interpolant from volumetric scattered from point data it is rarely the The increase in affordable computing power has fostered the creation case that the points are uniformly dense every in 3-D space. The of 3D models. The field of 3D models is on the rise with the advent of weighted alpha shapes method is defined for a finite set of weighted high speed Internet. Large repositories of 3D models are being points. It is a polytope uniquely determined by the points, their created and made public on the Internet. Searching among these weights, and a parameter alpha that controls the desired level of repositories is an obvious requirement. In this paper, we present a detail. Therefore, how to assign the weight for each point is one of the total shear invariant feature vector (SIFV) for searching a 3D model. main tasks to achieve the desirable volumetric scattered data This feature vector is robust to affine transformations of the original interpolation. In other words, we need to investigate the way to model. The proposed feature vector is Fourier transform of the achieve different levels of detail in a single shape by assigning weights histogram of the number of points in the concentric spheres to the data points. One of the ways to assign weight can be partitioning the point cloud data of a 3D model. considered by using Modified Shepard’s method. The following equation is a variation of the basic Shepard’s method. Trivariate quadratic function is approximated based on least-squares sense. 6499-10, Session 3 Reconstruction of shapes of 3D symmetrical objects by 6499-06, Session 2 using planarity and compactness constraints Incremental adaptive subdivision of mesh surfaces Y. Li, Z. Pizlo, Purdue Univ. G. Agam, S. Suresh, Illinois Institute of Technology Reconstruction of a 3D shape from its single image is an ill-posed problem. In order to produce a unique solution, the visual system has Subdivision of triangular meshes is a common technique for refining a to impose a priori constraints on the family of possible solutions. We given surface representation for various purposes in computer vision, describe a novel algorithm that uses symmetry, planarity and computer graphics, and finite element methods. Particularly, in the compactness constraints. Maximal compactness corresponds to processing of reconstructed surfaces based on sensed data, maximal volume for a given surface area. This constraint has not been subdivision can be used to add surface points at locations in which

136 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6499: Vision Geometry XV used before in algorithms of shape reconstruction. The algorithm was faces in the gallery, at great computational cost. We explore the tested on random symmetric polyhedra. The reconstructions are approach of registering the new facial surface to an average face usually accurate. Interestingly, the algorithm can reconstruct not only model (AFM), which automatically establishes correspondence to the the visible part of an opaque object, but also the invisible part. pre-registered gallery faces. Going one step further, we show that Preliminary psychophysical experiments showed that the using a couple of well-selected AFMs can trade-off computation time reconstructions produced by human observers are very close to the with accuracy. Drawing on cognitive justifications and the “other race algorithm’s reconstructions. This means that all three constraints are effect”, we use category-specific alternative average face models for most likely used by the human visual system. registration, which increases the accuracy of the subsequent recognition. We inspect rigid and deformable registration schemes under realistic assumptions on manual or automatic landmark 6499-11, Session 3 detection prior to registration. We evaluate several approaches for the coarse initialization. We propose a new algorithm for constructing an Recognizing point configurations in full perspective AFM, and show that it works better than a recent approach. Finally, Kevin Abbott and Peter F. Stiller (Texas A&M University, College we perform simulations with multiple AFMs that correspond to Station, TX 77843-3368) different clusters in the face shape space and compare these with gender and morphology based groupings. We examine two fundamental problems related to object recognition for point features under full perspective projection, i.e. a pinhole camera model of image formation. The first problem involves the 6499-14, Session 4 geometric constraints (object/image equations) that must hold between a set of object feature points and any image of those points Arithmetic characterization of polynomial-based discrete under a full perspective projection. These constraints are formulated curves in an invariant way, so that object pose, image orientation, or the C. Fiorio, J. Toutant, Univ. Montpellier II (France) choice of coordinates used to express the feature point locations on either the object or in the image are irrelevant. The second issue In the present paper, we investigate the arithmetic characterization of concerns the notion of shape and a description of the resulting shape the thinnest discrete approximations (GIQ, BBQ, OBQ) of curves spaces. These spaces aquire certain natural metrics which provide an defined by polynomials in the 2-dimensional space. even more robust approach to matching. For example, we might like to know if two configurations of a fixed number of points in 2D or 3D are the same if we allow projective transformations. If they are, then 6499-15, Session 4 we want a distance of zero, and if not, we want a distance that expresses their dissimilarity - always recognizing that we can Topological feature extraction using algebraic topology transform the points. Taken all together, the results in this paper S. Derdar, Univ. de Sherbrooke (Canada); M. Allili, Bishop’s Univ. provide an invariant and optimal way to measure the relationship that (Canada); D. Ziou, Univ. de Sherbrooke (Canada) exists between 3D geometry and its “residual” in a 2D image. Topological image feature extraction is very important for many high level tasks in image processing and for topological analysis and 6499-12, Session 3 modeling of image data. In this work, we use cubical homology theory to extract topological features as well as their geometric Three-dimensional face recognition based on geodesic representations in image raw data. Furthermore, we present two distances algorithms that will allow us to do this extraction task very easily. The S. Gupta, M. K. Markey, J. K. Aggarwal, A. C. Bovik, The Univ. of first one uses the elementary cubical representation to check the Texas/Austin adjacency between cubes in order to localize the connected components in the image data. The second algorithm is about cycle We propose a novel method to improve the performance of existing extraction. The .first step consists of finding cubical generators of the three dimensional (3D) human face recognition algorithms that employ first homology classes. These generators allow finding rough locations Euclidean distances between facial fiducial points as features. We of the holes in the image data. The second method localizes the further investigate a novel 3D face recognition algorithm that employs optimal cycles from the ordinary ones. The optimal cycles represent geodesic and Euclidean distances between facial fiducial points. We the boundaries of the holes in the image data. A number of demonstrate that this algorithm is robust to changes in facial experiments are presented to validate these algorithms on synthetic expression. Geodesic and Euclidean distances were calculated and real binary images. between pairs of 25 facial fiducial points. For the proposed algorithm, geodesic distances and ‘global curvature’ characteristics, defined as the ratio of geodesic to Euclidean distance between a pairs of points, 6499-16, Session 4 were employed as features. The most discriminatory features were selected using stepwise linear discriminant analysis (LDA). These were Complete boundary detection of textured objects via projected onto 11 LDA directions, and face models were matched deformable models using the Euclidean distance metric. With a gallery set containing one R. Dedic, Univ. de Sherbrooke (Canada); M. Allili, Bishop’s Univ. image each of 105 subjects and a probe set containing 663 images of (Canada) the same subjects, the algorithm produced EER=1.4% and a rank 1 RR=98.64%. It performed significantly better than existing algorithms Object recognition using the shape of objects boundaries and surface based on principal component analysis and LDA applied to face range reconstruction using slice contours rely on the identification of the images. Its verification performance for expressive faces was also correct and complete boundary information of the segmented objects significantly better than an algorithm that employed Euclidean in the scene. Geometric deformable models (GDM) using the level distances between facial fiducial points as features. sets method provide a very efficient framework for image segmentation. However, the segmentation results provided by these models are dependent on the contour initialization. Also, if there are 6499-13, Session 3 textured objects in the scene, usually the incorrect boundaries are detected. In most cases where the strategy is to detect the correct Alternative face models for 3D face registration boundary of all the objects in the scene, the results of the A. A. Salah, N. Alyüz, L. Akarun, Bogaziçi Univ. (Turkey) segmentation will only provide incomplete and/or inaccurate object’s boundaries. In this work, we propose a new method to detect the The accuracy of a 3D face recognition system depends on a correct correct boundary information of segmented objects, in particular registration that aligns the facial surfaces and makes a comparison textured objects. We use the average squared gradient to determine possible. The best results obtained so far use a one-to-all registration the appropriate initialization positions and by varying the size of the approach, which means each new facial surface is registered to all

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 137 Conference 6499: Vision Geometry XV test regions we create multiple images, that we will call layers, to determine the appropriate boundaries.

6499-17, Session 4 Finger-shape extraction and expansion by wavelet transform and hand-shape analysis and recognition C. Su, Da Yeh Univ. (Taiwan) This research uses the object extracting technique to extract the - thumb, index, middle, ring, and small fingers from the hand images. The algorithm developed in this research can find the precise locations of the fingertips and the finger-to-finger-valleys. The extracted fingers contain many useful geometry features. One can use these features to do the person identification. The geometry descriptor is used to transfer geometry features of these finger images to another feature-domain for image-comparison. Image is scaled and the reverse Wavelet Transform is performed to the finger image to make the finger image has more salient feature. Image subtraction is used to exam the difference of the two images. This research uses the finger-image and the palm image as the features to recognize different people. In this research, totally eighteen hundred and ninety comparisons are conducted. Within these eighteen hundred and ninety comparisons, two hundred and seventy comparisons are conducted for self-comparison. The other sixteen hundred and twenty comparisons are conducted for comparisons between two different persons’ finger images. The false accept rate is 0%, the false reject rate is 1.9%, and the total error rate is 1.9%.

138 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6500: Document Recognition and Retrieval XIV Tuesday-Thursday 30 January-1 February 2007 Part of Proceedings of SPIE Vol. 6500 Document Recognition and Retrieval XIV

6500-01, Session 1 and handwritten digits that demonstrate the viability of frequency coding in a classification task. Industrial OCR approaches: architecture, algorithms, and adaptation techniques 6500-04, Session 2 I. Marosi, Nuance-Recognita, Corp (Hungary) A multi-evidence multi-engine OCR system Optical Character Recognition is much more than character classification. An industrial OCR application combines algorithms I. Zavorin, E. Borovikov, A. Borovikov, CACI International Inc.; L. studied in detail by different researchers in the area of image Hernandez, Army Research Lab.; K. M. Summers, M. I. Turner, processing, pattern recognition, machine learning, language analysis, CACI International Inc. document understanding, data mining, and other, artificial intelligence domains. Most of these algorithms are run in an iterative process There is no OCR engine that performs equally well on all documents where the knowledge about the document increases with every for a given single language script. Most of the existing systems that iteration. There is no single perfect algorithm for any of the OCR use multiple OCR engines to boost output accuracy do not go beyond problems, so modern systems try to adapt themselves to the actual variations on voting. This approach has limitations, especially when features of the image or document to be recognized. This paper OCR technology for a given script has not yet fully matured. Our goal describes the architecture of a modern OCR system with an emphasis is to develop a system called MEMOE that combines output streams on this adaptation process. Adaptation in different areas is shown in of one or more OCR engines together with various types of evidence detail. For adaptive classification the use of a simple Nearest extracted from these streams as well as from original document Neighbor classifier with progressive training is suggested. It’s very images, to produce output of higher quality than that of majority important that the training set must be validated and any voting. We aim to improve OCR accuracy on images that might misrecognized patterns must be removed from it, therefore a voting otherwise significantly impact downstream processing. The MEMOE infrastructure is shown that gives good tools to do such validation. It system functions as an OCR engine taking document images and is also shown how the adaptive training set is used for doing adaptive configuration parameters as input and producing a single output text segmentation with baseline based image registration. Finally an stream. We describe the design of the system, various evidence types adaptive editor is described that trains from the user’s edits. and how they are incorporated into MEMOE in the form of filters. Results of initial tests involving two corpora of Arabic documents show that the system outperforms a voting algorithm and that even 6500-02, Session 2 more improvement may be achieved by incorporating additional evidence types into the system. A shape descriptor based on CSS M. Yang, K. Kpalma, J. Ronsin, Institut National des Sciences 6500-05, Session 3 Appliquées de Rennes (France) We propose a shape representation and description well adapted to Optimal interaction for style-constrained OCR pattern recognition, particularly in the context of affine shape G. Nagy, Rensselaer Polytechnic Institute; S. Veeramachaneni, transformations. The proposed approach operates from a single ITC-IRST (Italy) closed contour. The parameterized contour is convolved with a Gaussian kernel. The curvature is calculated to determine the inflexion OCR requires a delicate balance between economy of operator points and the main significant ones are kept by using a threshold intervention (labeling training samples, proofreading output, and defined by observing a segment-length between two curvature zero- correcting errors), acceptable throughput, and high accuracy. The crossing points. Then this filtered and simplified shape is registered error rate can be reduced on a style-consistent document if its style is with the original one. Finally, we separately calculate the areas identified and the right style-specific classifier is used. Since in some between the two segments corresponding to these two scale-space applications both machines and humans have difficulty in identifying representations. The proposed descriptor is a vector with components the style, we propose to improve the accuracy of style-constrained issued for each segment and the corresponding area. This article classification by enlisting the human operator to identify the labels of develops the new concepts: 1) compares the same segment under some characters selected by the machine. The key point of the paper different scales representation; 2) chooses the appropriate scales by is that labeling the appropriate patterns by class can help to identify applying a threshold to the shape shortest-segment; 3) then proposes their common style. We present an algorithm to select the characters the algorithm and the conditions of merging and removing the short- to be labeled. We demonstrate its efficacy on simulated data by segments. An experimental evaluation of robustness under affine comparing its error rate with the results of random selection and of transformations is presented on a shape database. selection of the most ambiguous patterns.

6500-03, Session 2 6500-06, Session 3 Frequency coding: an effective method for combining Reading text in consumer digital photographs dichotomizers V. Vanhoucke, S. B. Gokturk, Riya, Inc. S. Andra, G. Nagy, Rensselaer Polytechnic Institute; C. Liu, We present a distributed system to extract text contained in natural Institute of Automation (China) scenes within consumer photographs. The objective is to automatically annotate pictures in order to make consumer photo sets Binary classifiers (dichotomizers) are combined for multi-class searchable based on the image content. The system is designed to classification. Each region formed by the pairwise decision process a large volume of photos, by quickly isolating candidate text boundaries is assigned to the class with the highest frequency of regions, and successively cascading them through a series of text training samples in that region. With more samples and classifiers, the recognition engines which jointly make a decision on whether or not frequencies converge to increasingly accurate non-parametric the region contains text that is readable by OCR. In addition, a estimates of the posterior class probabilities in the vicinity of the dedicated rejection engine is built on top of each text recognizer to decision boundaries. The method is applicable to non-parametric adapt its confidence measure to the specifics of the task. The discrete or continuous class distributions dichotomized by either resulting system achieves very high text retrieval rate and data linear or non-linear classifiers (like support vector machines). We throughput with very small false detection rates. present a formal description of the method and place it in context with related methods. We present experimental results on machine-printed

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 139 Conference 6500: Document Recognition and Retrieval XIV

6500-07, Session 3 estimated by the condition derived from only parallelism of geodesics without the requirements for equal spacing. This method can estimate Adding contextual information to improve character the ruled surface model directly by numerical operations of recognition on the Archimedes Palimpsest differentiation, integration and matrix inversion without any iterative calculation. We also report on experiments that show the D. J. Walvoord, R. L. Easton, Jr., R. L. Canosa, Rochester Institute effectiveness of the proposed method. of Technology The objective of the character recognition effort for the Archimedes Palimpsest is to provide a tool that allows scholars of ancient Greek 6500-10, Session 4 mathematics to retrieve as much information as possible from the Multispectral pattern recognition applied to x-ray remaining degraded text. With this in mind, the current pattern recognition system does not output a single classification decision, as fluorescence images of the Archimedes Palimpsest in typical target detection problems, but has been designed to provide D. M. Hansen, R. L. Easton, Jr., Rochester Institute of Technology intermediate results that allow the user to apply his or her own decisions (or evidence) to arrive at a conclusion. To achieve this result, The Archimedes Palimpsest is one of the most significant texts in the a probabilistic network has been incorporated into our previous history of science. Much of the text has been read using images of recognition system, which was based primarily on spatial correlation reflected visible light and the visible light resulting from ultraviolet techniques. The system now provides the user with recognition fluorescence. However, these techniques do not perform well on the information from both the spatial characteristics of the images and the four pages of the manuscript that are obscured by forged icons that probabilities generated using an existing partial transcription. Using were painted over these pages during the first half of the 20th century. the probabilistic network, subsequent queries are performed by the X-ray fluorescence images of one of these pages have been user until he or she arrives at a final decision. We report on the revised processed using spectral pattern recognition techniques developed tool and the recent success in the transcription process. for environmental remote sensing to recover the original texts beneath the paint. 6500-08, Session 3 6500-11, Session 4 OCR result optimization based on pattern matching Degraded document image enhancement J. Shang, C. Liu, X. Ding, Tsinghua Univ. (China) G. Agam, G. Bal, Illinois Institute of Technology; G. Frieder, The Post-processing of OCR is a bottleneck of the document image George Washington Univ.; O. Frieder, Illinois Institute of processing system. Proof reading is necessary since the current Technology recognition rate is not enough for publishing. The OCR system provides every recognition result with a confident or unconfident label. Poor quality documents are obtained in various situations such as People only need to check unconfident characters while the error rate historical document collections, legal archives, security investigations, of confident characters is low enough for publishing. However, the and documents found in clandestine locations. Such documents are current algorithm marks too many unconfident characters, so often scanned for automated analysis, further processing, and optimization of OCR results is required. In this paper we propose an archiving. Due to the nature of such documents, degraded document algorithm based on pattern matching to decrease the number of images are often hard to read, have low contrast, and are corrupted unconfident results. If an unconfident character matches a confident by various artifacts. We describe a novel approach for the character well, its label could be changed into a confident one. enhancement of such documents based on probabilistic models Pattern matching makes use of original character images, so it could which increases the contrast, and thus, readability of such documents reduce the problem caused by image normalization and scanned under various degradations. The enhancement produced by the noises. We introduce WXOR, WAN, and four-corner based pattern proposed approach can be viewed under different viewing conditions matching to improve the effect of matching, and introduce confidence if desired. The proposed approach was evaluated qualitatively and analysis to reduce the errors of similar characters. Experimental compared to standard enhancement techniques on a subset of results show that our algorithm achieves improvements of 54.18% in historical documents obtained from the Yad Vashem Holocaust the first image set that contains 102,417 Chinese characters, and museum. In addition, quantitative performance was evaluated based 49.85% in the second image set that contains 53,778 Chinese on synthetically generated data corrupted under various degradation characters. models. Preliminary results demonstrate the effectiveness of the proposed approach. 6500-09, Session 4 6500-12, Session 5 Shape from parallel geodesics for distortion correction of digital camera document images Curvelets based tools for the characterization and the classification of Humanistic and Middle-Ages K. Fujimoto, J. Sun, H. Takebe, M. Suwa, S. Naoi, Fujitsu Labs., manuscripts Ltd. (Japan) G. Joutel, V. Eglin, H. Emptoz, Institut National des Sciences Distortion correction methods for digital camera document images of Appliquées de Lyon (France) thick volumes or curved papers become important for camera-based document recognition technologies. In this paper we propose a novel The aim of this scientific work is to propose a suitable assistance tool distortion correction method for digital camera document images for palaeographers and historians to help them in their intuitive and based on “shape from parallel geodesics.” This method considers the empirical work of identification of writing styles (for medieval following handwritings) and authentication of writers (for humanistic features: parallel lines corresponding to character strings or ruled lines manuscripts). We propose a global approach of writers’ classification of tables on extended surface become parallel geodesics on a curved based on Curvelets based features in relation with two discriminative paper surface and a smoothly curved paper can be modeled by a shapes properties, the curvature and the orientation. Those features ruled surface, which is sweep surface of rulings. The projected are revealing of structural and directional micro-shapes and also of geodesics and rulings exist in the input image derived from concavity that captures the finest variations in the contour. The perspective transformation. The presented method extracts the Curvelets based analysis leads to the construction of a compact Log- projected geodesics, estimates the projected rulings in the input polar signature for each writing. The relevance of the signature is image, estimates the ruled surface that models the curved paper, and quantified with a CBIR (content based image retrieval) system that generates the corrected image, in this order. The projected rulings are compares request images and database images candidates. The main

140 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6500: Document Recognition and Retrieval XIV experimental results are very promising and show 78% of good 6500-16, Session 6 retrieval (as precision) on the Middle-Ages database and 89% on the humanistic database. Making the public domain universally accessable A. Langley, D. S. Bloomberg, Google Inc. 6500-13, Session 5 Google Book Search is working with libraries and publishers around the world to digitally scan books. Some of those works are now in the Interactive training for handwriting recognition in public domain and, in keeping with Google’s mission to make all the historical document collections world’s information useful and universally accessible, we wish to allow D. J. Kennard, W. A. Barrett, Brigham Young Univ. users to download them all. For users, it is important that the files are as small as possible and of We present a method of interactive training for handwriting printable quality. This means that a single codec for both text and recognition in collections of documents. As the user transcribes images is impractical. We use PDF as a container for a mixture of (labels) the words in the training set, words are automatically skipped JBIG2 and JPEG2000 images which are composed into a final set of if they appear to match words that are already transcribed. By pages. reducing the amount of redundant training, better coverage of the data is achieved, resulting in more accurate recognition. Using word- We discuss both the implementation of an open source JBIG2 level features for training and recognition in a collection of George encoder, which we use to compress text data, and the design of the Washington’s manuscripts, the recognition ratio is approximately 2%- infrastructure needed to meet the technical, legal and user 8% higher after training with our interactive method than after training requirements of serving many scanned works. We also cover the the same number of words sequentially. Using our approach, less lessons learnt about dealing with different PDF readers and how to training is required to achieve an equivalent recognition ratio. A slight write files that work on most of the readers, most of the time. improvement in recognition ratio is also observed when using our method on a second data set, which consists of several pages from a diary written by Jennie Leavitt Smith. 6500-17, Session 6 Pixel and semantic capabilities from an image object 6500-14, Session 5 based document representation M. J. Gormish, K. Berkner, M. P. Boliek, G. Feng, E. L. Schwartz, Online handwritten mathematical expression recognition Ricoh Innovations, Inc. B. A. Yanikoglu, H. Buyukbayrak, A. Ercil, Sabanci Univ. (Turkey) This paper reports on novel and traditional pixel and semantic We describe a system for recognizing online, handwritten operations using a recently standardized document representation mathematical expressions. The system is designed with a user- called JPM. The JPM representation uses compressed pixel arrays for interface for writing scientific articles. A feed-forward neural network all visible elements on a page. Separate data containers called boxes recognizes symbols which are assumed to be single-stroke and a provide the layout and additional semantic information. JPM and recursive algorithm parses the expression by combining the output of related image-based document representation standards were the neural network and the structure of the expression. While there are designed to obtain the most rate efficient document compression. only 66 symbols in the system for now, the system can recognize The authors, however, use this representation directly for operations basic mathematical expressions as well as integrals, summations, other than compression typically performed either on pixel arrays or matrices etc. Furthermore, the parsing structure allows for easy semantic forms. This paper describes the image representation used addition of new symbols and expressions. Preliminary results show in the JPM standard and presents techniques to (1) perform traditional that writer-dependent recognition rates are very high (99.8%) while raster- based document analysis on the compressed data, (2) transmit writer-independent symbol recognition rates are lower (75%). The semantically meaningful portions of compressed data between interface associated with the proposed system integrates the built-in devices, (3) create multiple views from one compressed data stream, recognition capabilities of the Microsoft’s Tablet PC API for and (4) edit high resolution document images with only low resolution recognizing textual input and supports conversion of hand-drawn proxy images. figures into PNG format. This enables the user to enter text, mathematics and draw figures in a single interface. After recognition, all output is combined into one LATEX code and compiled into a PDF 6500-18, Session 6 file. Presentation of structured documents without a style sheet 6500-15, Session 5 S. J. Harrington, E. Wayman, Xerox Corp. Recognition of degraded handwritten digits using In order to present most XML documents for human consumption, dynamic Bayesian networks formatting information must be introduced and applied. Formatting is L. Likforman, M. Sigelle, École Nationale Supérieure des typically done through a style sheet, however, it is conceivable that Télécommunications (France) one could wish to view the document without having a style sheet (either because a style sheet does not exist, or is unavailable, or is We investigate in this paper the application of dynamic Bayesian inappropriate for the display device). This paper describes a method networks (DBNs) to the recognition of handwritten digits. The main for formatting structured documents without a provided style sheet. idea is to couple two separate HMMs into various architectures. First, The idea is to first analyze the document to determine structures and a vertical HMM and a horizontal HMM are built observing the evolving features that might be relevant to style decisions. A transformation streams of image columns and image rows respectively. Then, two can be constructed to convert the original document to a generic form coupled architectures are proposed to model interactions between that captures the semantics that will be expressed through formatting these two streams and to capture the 2D nature of character images. and style. In the second stage styling is applied to the structures and Experiments performed on the MNIST handwritten digit database features that have been discovered by applying a pre-defined style show that coupled architectures yield better recognition performances sheet for the generic form. . The document instance, and if available, than non-coupled ones. Additional experiments conducted on the corresponding schema or DTD can be analyzed in order to artificially degraded (broken) characters demonstrate that coupled construct the transformation. This paper will describe the generic form architectures better cope with such degradation than non coupled used for formatting and techniques for generating transformations to ones and than discriminative methods such as SVMs. it.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 141 Conference 6500: Document Recognition and Retrieval XIV

6500-19, Session 6 printing less as a discrete manufacturing process and more as a component of a complete marketing campaign. Each model Content selection based on compositional image quality approaches estimating differently. Understanding and predicting costs P. Obrador, Hewlett-Packard Co. can be extremely beneficial. Establishing a reliable system to estimate those costs can be somewhat challenging though. Unquestionably, Digital publishing workflows usually have the need for composition cost-estimating digital printing will increase in relevance in the years and balance within the document, where certain photographs will ahead, as margins tighten and cost knowledge becomes increasingly have to be chosen according to the overall layout of the document it more critical. is going to be placed in. i.e., the composition within the photograph will have a relationship/balance with the rest of the document layout. This paper presents a novel image retrieval method, in which the 6500-23, Session 8 document where the image is to be inserted is used as query. The algorithm calculates a balance measure between the document and A novel approach for nonuniform list fusion each of the images in the collection, retrieving the ones that have a W. Yan, Univ. of California/Irvine higher balance score. The image visual weight map, used in the balance calculation, has been successfully approximated by a new List fusion is a critical problem in information retrieval. The approach image quality map that takes into consideration sharpness, contrast using uniform weights for list fusion ignores the correctness, and chroma. This paper presents a method to help in the area of importance and individuality of various detectors for a concrete document creation where the user starts with a document, and needs application. In this paper, we propose a nonuniform and rational to select a photograph from a large image collection, i.e., given a optimized paradigm for TRECVid list fusion, which is expected to layout, automatically retrieve a photograph from an image database loyally preserve the precision in the outcomes and reach the that will balance that selected layout, or, retrieve the top N maximum Average Precision (A.P.). Therefore we exhaustively search photographs that would best balance that layout for the user to for the corresponding parametric set for the best A.P. in the space choose from. spanned by the feature vectors. In order to accelerate the fusion procedure of the input score lists, we train our model via training data sets, and use the learnt parameters to fuse those new vectors. We 6500-20, Session 7 take the nonuniform rational blending functions into account, the advantage of using this fusion is that the problem of weights selection Generic architecture for professional authoring is converted to the issue of parameters selection in the space related environments to export XML-based formats to the nonuniform and rational functions. The high precision and multiple resolution, controllable and stable attributes of rational F. Giannetti, Hewlett-Packard Co. (United Kingdom) functions are helpful in parameters selection. Therefore, the space for Professional authoring environments are used by Graphic Artists (GA) fusion weights selection becomes large. The correctness of our during the design phase of any publication type. With the increasing proposal is compared and verified with the average and linear fusion demand for supporting Variable Data Print (VDP) designs, these results. authoring environments require enhanced capabilities. The recurring challenge is to provide flexible VDP features that can be 6500-24, Session 8 represented using several VDP enabling XML based formats. Considering the different internal structure of the authoring Identification of comment-on sentences in online environments, a common platform needs to be developed. biomedical documents using support vector machines The solution must, at the same time, empower the GA with a rich VDP I. C. Kim, D. X. Le, G. R. Thoma, National Library of Medicine feature set, as well as, generating a range of output formats that drive their respective VDP workflows. MEDLINE is the premier bibliographic online database of the National We have designed a common architecture to collect the required data Library of Medicine, containing approximately 14 million citations and from the hosting application and a generic internal representation that abstracts from over 4,800 biomedical journals. This paper presents an enables multiple XML output formats. The proposed architecture is automated method based on support vector machines to identify a separated into three phases: “comment-on” list, which is a field in a MEDLINE citation denoting previously published articles commented on by a given online 1. the host dependent visitor;; biomedical article. Our strategy is first to detect all “citation 2. the host independent visitor; sentences” from the body text of a given article using hyperlink 3. the serialization task. information or specific tags indicating “in-text citations”, and then to recognize the “comment-on sentences” that mention “comment-on” Prototypal implementations have been carried out for both Adobe’s articles from these “citation sentences”. For comparative study, we InDesign CS2 (and planned for Quark xPress 7.0) to validate the also introduce another method based on scoring functions that approach. estimate the significance of each sentence in a given article. Preliminary experiments conducted on HTML-formatted online biomedical documents collected from 24 different journal titles show 6500-21, Session 7 that the support vector machine with polynomial kernel function Cost estimating for commercial digital printing performs best in terms of recall and F-measure rates. M. G. Keif, California Polytechnic State Univ. The purpose of this study is to document current cost-estimating 6500-25, Session 8 practices used in commercial digital printing. A research study was Combining text clustering and retrieval for corpus conducted to determine the use of cost-estimating in commercial adaptation digital printing companies. This study answers the questions: 1) What methods are currently being used to estimate digital printing? 2) What F. He, X. Ding, Tsinghua Univ. (China) is the relationship between estimating and pricing digital printing? 3) The application-relevant text data are very useful in various natural To what extent, if at all, do digital printers use fullabsorption, all- language applications. Using them can achieve significantly better inclusive hourly rates for estimating? Three different digital printing performance for vocabulary selection, language modeling, which are models were identified: 1) Traditional print providers, who supplement widely employed in automatic speech recognition, intelligent input their offset presswork with digital printing for short-run color and method etc. In some situations, however, the relevant data is hard to versioned commercial print; 2) “Low-touch” print providers, who collect. Thus, the scarcity of application-relevant training text brings leverage the power of the Internet to streamline business transactions difficulty upon these natural language processing. In this paper, only with digital storefronts; 3) Marketing solutions providers, who see

142 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6500: Document Recognition and Retrieval XIV using a small set of application specific text, by combining and security investigations. The indexing and analysis of large unsupervised text clustering and text retrieval techniques, the document collections is currently limited to textual features based proposed approach can find the relevant text from unorganized large OCR data and ignore the structural context of the document as well scale corpus, thereby, adapt training corpus towards the application as important non-textual elements such as signatures, logos, stamps, area of interest. We use the performance of n-gram statistical tables, diagrams, and images. Handwritten comments are also language model, which is trained from the text retrieved and test on normally ignored due to the inherent complexity of offline handwriting the application-specific text, to evaluate the relevance of the text recognition. We address important research issues concerning acquired, accordingly, to validate the effectiveness of our corpus content-based document image retrieval and describe a prototype for adaptation approach. The language models trained from the ranked integrated retrieval and aggregation of diverse information contained text bundles present well discriminated perplexities on the in scanned paper documents we are developing. Such complex application-specific text. The preliminary experiments on short document information processing combines several forms of image message text and unorganized large corpus demonstrate the processing together with textual/linguistic processing to enable performance of the proposed methods. effective analysis of complex document collections, a necessity for a wide range of applications. Our prototype automatically generates rich metadata about a complex document and then applies query tools to 6500-26, Session 9 integrate the metadata with text search. To ensure a thorough evaluation of the effectiveness of our prototype, we are developing a Document recognition serving people with disabilities test collection containing millions of document images. J. R. Fruchterman, The Benetech Initiative Document recognition advances have improved the lives of people 6500-29, Session 11 with print disabilities, by providing accessible documents. This invited paper provides perspectives on the author’s career progression from A statistical approach to line segmentation in document recognition professional to social entrepreneur applying handwritten documents this technology to help people with disabilities. Starting with initial thoughts about optical character recognition in college, it continues M. Arivazhagan, H. Srinivasan, S. N. Srihari, Univ. at Buffalo with the creation of omnifont character recognition that did not require A new technique to segment an handwritten document into distinct training. It was difficult to make a reading machine for the blind in a lines of text is presented. Line segmentation is the first and the most commercial setting, which led to the creation of a nonprofit social critical pre-processing step for a document recognition/analysis task. enterprise to deliver these devices around the world. This network of The proposed algorithm starts, by obtaining an initial set of candidate people with disabilities scanning books led to the creation of lines from the piece-wise projection profile of the document. The lines Bookshare.org, an online library of scanned books. Looking forward, traverse around any obstructing handwritten connected component the needs for improved document recognition technology to further by associating it to the line above or below. A decision of associating lower the barriers to reading are discussed. Document recognition such a component is made by (i) modeling the lines as bivariate professionals should be proud of the positive impact their work has Gaussian densities and evaluating the probability of the component had on some of society’s most disadvantaged communities. under each Gaussian or (ii)the probability obtained from a distance metric. The proposed method is robust to handle skewed documents and those with lines running into each other. Experimental results 6500-27, Session 10 show that on 200 documents contaning a total of 4,768 lines, 97.65% of the lines were segmented correctly. This accuracy corresponds to Title extraction and generation from OCR’d documents 98.81% of connected components associated to the correct lines. K. Taghva, J. Borsack, Univ. of Nevada/Las Vegas Extraction of meta-data from documents is a tedious and expensive 6500-30, Session 11 process. In general, documents are manually reviewed for structured data such as title, author, date, organization, etc. The purpose of Segmentation and labeling of documents using extraction is to build meta-data for documents that can be used in conditional random fields formulating structured queries. In many large document repositories such as the National Library of Medicine (NLM)[10] or university S. R. Shetty, H. Srinivasan, M. J. Beal, S. N. Srihari, Univ. at libraries, the extraction task is a daily process that spans decades. Buffalo For the Licensing Support Network (LSN) document repository, the The paper describes the use of Conditional Random Fields (CRF) task involves extraction of several fields from millions of OCR’d and utilizing contextual information in automatically labeling extracted electronic documents. Since this project has an expected certification segments of scanned documents as Machine-print, Handwriting and date of June 2007, automatic extraction turns out to be the only viable Noise. The result of such a labeling can serve as an indexing step for solution. There are more than a dozen fields associated with each a context-based image retrieval system or a bio-metric signature document that requires extraction. In this paper, we only report on the verification system. A simple region growing algorithm is first used to extraction and generation of the title field. segment the document into a number of patches, 23 state features are extracted for each of these patches. A maximum of 6 neighbors are identified for each of the segmented patches using the convex hull 6500-28, Session 10 distance and 4 transition features are extracted for each neighboring Content-based document image retrieval in complex pair. Maximum pseudo-likelihood estimates for the parameters of the document collections CRF model are learnt using conjugate gradient descent. A label for each such segmented patch is inferred by computing the probability G. Agam, S. Argamon, O. Frieder, D. Grossman, Illinois Institute of of the labels under the CRF model with Gibbs sampling. The Technology; D. Lewis, David D. Lewis Consulting robustness of the model is due to the inherent nature of modeling neighboring spatial dependencies in the labels as well as the We address the problem of content-based image retrieval in the observed data using CRF. Experimental results show that this context of complex document images. Complex documents typically approach provides for 95.75% of the data being assigned correct start out on paper and are then electronically scanned. These labels. The CRF based model is shown to be superior to Neural documents have rich internal structure and might only be available in Networks and Naive Bayes. image form. Additionally, they may have been produced by a combination of printing technologies (or by handwriting); and include diagrams, graphics, tables and other non-textual elements. Large collections of such complex documents are commonly found in legal

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 143 Conference 6500: Document Recognition and Retrieval XIV

6500-31, Session 11 Online medical journal article layout analysis J. Zou, D. X. Le, G. R. Thoma, National Library of Medicine We describe an HTML page physical and logical layout analysis algorithm, which is applied to segment and label online medical journal articles (regular HTML and PDF-Converted-HTML files). For these articles, the geometric layout of the Web page is the most important cue for physical layout analysis. The key to physical layout analysis is then to render the HTML file in a Web browser, so that the visual information in zones (composed of one or a set of HTML DOM nodes), especially their relative position, can be utilized. The recursive X-Y cut algorithm is adopted to construct a hierarchical zone tree structure. In logical layout analysis, both geometric and linguistic features are used. The HTML documents are modeled by a Hidden Markov Model with 16 states, and the Viterbi algorithm is then used to find the optimal label sequence, concluding the logical layout analysis. The important property of the proposed layout analysis algorithm is that it minimizes the dependence on HTML tags, and is therefore tolerant to variations in HTML implementation styles. Preliminary evaluation demonstrates that the algorithm performs successfully on HTML journal articles in various styles

6500-32, Session 11 Transcript mapping for handwritten Arabic documents L. M. Lorigo, V. Govindaraju, Univ. at Buffalo No abstract available.

6500-33, Session 11 Document image content inventories H. S. Baird, M. A. Moll, C. An, Lehigh Univ. We report an investigation into strategies, algorithms, and software tools for document image content extraction and inventory: that is, the location and measurement of regions containing handwriting, machine-printed text, photographs, blank space, etc. We have developed automatically trainable methods, adaptable to many kinds of documents represented as bilevel, greylevel, or color images, that offer a wide range of useful tradeoffs of speed versus accuracy using methods for exact and approximate k-nearest neighbor classification. We have adopted a policy of classifying each pixel (rather than regions) by content type: we discuss the motivation and engineering implications of this choice. We describe experiments on a wide variety of document-image and content types, and discuss performance in detail in terms of classification speed, per-pixel classification accuracy, per-page inventory accuracy, and subjective quality of page segmentation. These show that even modest per-pixel classification accuracies (of, e.g., 60-70%) support usefully high recall and precision rates (of, e.g., 80-90%) for retrieval queries of document collections seeking pages that contain a given minimum fraction of a certain type of content.

144 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X Wednesday-Thursday 31 January-1 February 2007 Part of Proceedings of SPIE Vol. 6501 Sensors, Cameras, and Systems for Scientific/ Industrial Applications VIII

6501-31, Poster Session A low-cost sensor platform (MORES Sensor) was combined with a microcontroller to build up an embedded solution which e.g. allows Pigment-dispersed colored photo resists for the new for a small hand-held Color Estimation System for blind people. The generation high-resolution color filters for CMOS image color sensor used here measures the intensity response of a surface sensors caused by radiation with a specific wavelength in the range of visible light. This radiation is realized by means of three LEDs, red, green, J. Kim, K. Lee, E. Jeong, C. Lee, Cheil Industries Inc. (South and blue, so that the response intensity values create an R’G’B’ color Korea) space, which differs from the standardized RGB color space due to The objective of our R&D project is to develop colored photo resists the wavelengths of the LEDs. By adjusting the measured response of which can be used to make color filters for high-end CMOS image the LEDs to the known spectral response of the individual color sensor arrays (CIS). Resolution of 10 million and higher pixels (sub- panels of the Macbeth Color Checker Chart (MCCC) a corresponding micron level of pixel dimension) are aimed at as the high-end set of coordinates can be constructed for this particular R´G´B´ color application of CIS. In order to adjust spectral properties of the color space. Owing to this approach, it is possible to obtain reasonable filters, the maximum transmittance of blue color filter should be color classification results, which can be compared to those of far shifted into a shorter wavelength region (a.k.a. blue shift), and that of more complex and expensive systems. The verification of the results green color filter should be changed to a longer wavelength (a.k.a. red was done by using the standardized MCCC together with commercial shift) with respect to the earlier generation of colored photo resists, vision solutions (RGB camera in combination with PC software). respectively. The spectral position of red remained unchanged. To Moreover, some comparison tests also prove the practicality of the obtain the desired spectral profiles, we had no choice but to take here described low-cost color sensor solution. advantage of another sorts of pigments and/or dyes with chromophores which had rarely been considered as color filter grades. In order to reduce the size of a unit pixel with a perfect square 6501-34, Poster Session dimension in color filters, we utilized two methods: (1) reduction of A CMOS vision chip for a contrast-enhanced image using secondary sizes of dispersed pigments and (2) synthesis of multi- a logarithmic APS and a switch-selective resistive components polymer resins. New dispersion systems including ultra fine milling technique were fabricated for the new pigments because network they showed dispersion-resistant property resulting from inadequate J. Kong, S. Kim, D. Sung, S. Seo, J. Shin, Kyungpook National surface treatment as well as undesirable geometry (rod-like not Univ. (South Korea) spherical) of primary particles. In this paper, a bio-inspired CMOS vision chip for a contrast- enhanced image using a resistive network is introduced. The key 6501-32, Poster Session advantage of the structure is high speed of operation, because of its parallel mechanism for signal processing. Previously reported results Qualitative measurement of average index profile of using a charge accumulation-type active pixel sensor (APS) shows symmetric microstructured optical fiber from that the time for charge accumulation limits the operation speed. In microscopic imaging order to enhance the operation speed, a logarithmic APS was applied into the vision chip. A logarithmic APS has superior dynamic range E. Saucedo, V. P. Minkovich, Ctr. de Investigaciones en Óptica, and speed, but it suffers from small output swing and large fixed- A.C. (Mexico) pattern noise (FPN). By applying a MOS-type photodetector with high A simple non-destructive and non-interferometric technique is quantum efficiency to a logarithmic APS, we could achieve sufficient presented. The technique allows us to obtain an approximation of the output swing for operating the vision chip. A CMOS buffer circuit, average index profile for symmetric air-silica microstructured optical which is embodied in the structure of a source follower amplifier, is fiber (MOF). This experimental technique deals with the transversal commonly used for both raw and smoothed images by adding gradient of the phase, which is obtained from microscopic bright field switches. By using the switch-selective resistive network, the total images. Then the inverse Abel transform is applied to receive an number of MOSFETs in the unit pixel and fixed-pattern noise was approximation of average index profile for experimental MOF with a reduced. In addition, in order to enhance the resolution, analog signal solid core surrounded by double microstructured cladding, which processing circuits were separated from a photo-sensing array. A consists of 4-ring structure of air channels. Bright field images of the vision chip with a 160Å~120 pixel array was fabricated using a 0.35 studied MOF were obtained with a DC3-163 Microscope using a 20 x µm double-poly four-metal standard CMOS technology and its 0.40 NA objective. We used a He-Ne laser, a pinhole and a collimator characteristics were measured. lens to obtain uniform illumination of the microstructured fiber samples. The amount of light intensity that goes through the studied fiber was adjusted using a polarizer into the experimental setup. Then, 6501-35, Poster Session the same scheme of image processing tools as for conventional fibers Highly sensitive and area-efficient CMOS image sensor was used to get final results. Results of our measurements are using a PMOSFET-type photodetector with a built-in presented. transfer gate S. Seo, K. Kim, J. Kong, J. Shin, P. Choi, Kyungpook National 6501-33, Poster Session Univ. (South Korea) A practicable method for color classification using low- This paper presents a highly sensitive and area-efficient CMOS image cost remission/reflection sensors sensor using a PMOSFET-type photodetector with a built-in transfer gate. The structure of the photodetector used in this CMOS image R. Hauser, T. Klinger, C. Madritsch, Fachhochschule Technikum sensor is similar to the structure of a gate/body-tied PMOSFET Kärnten (Austria); D. Roemhild, D. Starke, CiS Institut für photodetector. However, the proposed photodetector has another Mikrosensorik gGmbH (Germany); V. Mannheim, Fachhochschule gate that controls the generated photocurrent. This transfer gate of Technikum Kärnten (Austria) the proposed photodetector can perform the function of the transfer MOSFET in the 4-transistor active pixel sensor (APS). A prototype

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 145 Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X

CMOS image sensor was fabricated by standard 0.35 µm CMOS an array of autonomous, self-signaling pixels which asynchronously process. We measured the generated photocurrent of the proposed respond to the temporal changes (transients) of local brightness, and photodetector. The responsivity of the proposed photodetector is use sparse output representation of image information with minimum about 1.0 X 103 A/W. This value is considerably higher than that of a amount of redundancy while preserving precise timing information. conventional pn junction photodiode which is typically lower than 1 A/ The proposed algorithm processes AE-data generated by transient W. Therefore, smaller area is required for the same photocurrent by imager and take advantage of its focal pre-processing and efficient using the proposed photodetector, compared to the conventional pn asynchronous address-event communication framework. Due to these junction photodiode. We also measured the output waveform of the characteristics the hardware architecture of our sensor consists fabricated chip under different operation time of the transfer gate of essentially of only three devices: optical transient sensor as sensing the proposed photodetector using the white light source of 5 lux. element, buffer memory and a digital signal processor (DSP) as From the experimental results, we could confirm that the proposed processing unit, so that the complete system is implemented in form photodetector has a high sensitivity and the pseudo 3-transistor APS of small size, low power and cheap embedded system. Experimental performs the function of 4-transistor APS. The feasibility of an evaluation of our sensor show that is capable to detect moving application to a digital imager at a very low-illumination level for this homogeneous objects of different size and shape under industrial- image sensor was also confirmed. close conditions.

6501-37, Poster Session 6501-01, Session 1 Microscopic x-ray imaging system for biomedical 100-M pixel high-speed high-resolution CCD applications using synchrotron radiation K. L. Boggs, R. A. Bredthauer, G. R. Bredthauer, Semiconductor K. Umetani, Japan Synchrotron Radiation Research Institute Technology Associates Inc. (Japan); M. Kobatake, A. Yamamoto, T. Yamashita, S. Imai, A 111-Mega pixel, 92x92 mm2, full-frame CCD imager with 9x9 um2 Kawasaki Medical School (Japan) pixel size has been developed for use in scientific applications. Microscopic X-ray imaging with spatial resolution in the micrometer Recent interest for ultra-high resolution imagers for electronic imaging range was carried out for biomedical imaging using a high-resolution OEM customers in various scientific markets, including biotechnology, real-time detector and synchrotron radiation at SPring-8. An X-ray microscopy, crystallography, astronomy, spectroscopy, and digital direct-conversion type detector with a spatial resolution of ca. 10 photography markets has lead to the development of the STA1600A microns was developed. The X-ray image is directly converted into an 111-Mega pixel monochromatic charge-coupled device. Innovative electric signal in the photoconductive layer without image blur. In design techniques were utilized in the early development of this synchrotron radiation radiography, a long source-to-object distance device, yielding low RMS noise and high MTF for readout speeds and a small source spot can produce high-resolution images. Imaging ranging from 1 Mpixel/s to 10 Mpixel/sec. This paper will provide experiments were carried out for depicting tumor angiogenic vessels detailed information on the design and performance capabilities of the in a rabbit model of cancer. Sequential images of tumor angiogenic STA1600A, as well as background information on the commercial vessels induced by transplanted VX2 cancer cells in the rabbit auricle uses of this device. were obtained after iodine contrast agent injection to the auricular artery. Small blood vessels with diameters of 20-30 microns were visualized. The synchrotron radiation system is a useful tool for 6501-02, Session 1 evaluating micro-angioarchitecture of malignant tumors in animal Computation of dark frames in digital imagers models of cancer for preclinical testing. R. Widenhorn, Portland State Univ. and Digital Clarity Consultants; A. Rest, Cerro Tololo Inter-American Observatory; M. 6501-38, Poster Session Blouke, Ball Aerospace & Technologies Corp.; R. L. Berry, Digital Imaging based procedures to evaluate harrowing actions Clarity Consultants; E. Bodegom, Portland State Univ. and Digital Clarity Consultants on soil vegetal waste G. Bonifazi, Univ. degli Studi di Roma/La Sapienza (Italy); P. Dark current is caused by electrons that are thermally exited into the conduction band. These electrons are collected by the well of the Menesatti, M. Nillozza, Istituto Sperimentale per la CCD and add a false signal to the chip. We will present an algorithm Meccanizzazione Agricola (Italy) that automatically corrects for dark current. It uses a calibration Harrowing actions are usually addressed to fill in and distribute protocol to characterize the image sensor for different temperatures. vegetal wastes. This action is usually carried out by a drag or a brake For a given exposure time, the dark current of every pixel is usually driven by a tractor. The proposed approach is addressed to characteristic of a specific temperature. The dark current of every explore the possibility to develop a faster, automatic and reliable set pixel can therefore be used as an indicator of the temperature. Hot of imaging based procedures able to objectively evaluate such a kind pixels have the highest signal-to-noise ratio and are the best of soil processing. Tests have been carried out with reference to three temperature sensors. We use the dark current of a few hot pixels to different experimental contexts, that is: i) vegetal wastes presence sense the chip temperature and predict the dark current of all pixels evaluation on a “not worked soil”, ii) vegetal wastes presence on the chip. Dark current computation is not a new concept, but our evaluation on an harrowed soil and iii) evaluation of “vegetal wastes approach is unique. Some of the advantages of our method are the burying degree”. For each of the 3 previously defined contexts a applicability even for poorly temperature-controlled camera systems comparison among imaging based results and human expert based and the possibility of ex post facto dark current correction. evaluation was performed. 6501-03, Session 1 6501-39, Poster Session A novel pixel structure for 2-µm square pixel IT-CCD Sensor for moving foreground detection: an address G. Sakoda, M. Sato, D. Sugimoto, K. Maari, S. Miyazawa, T. event based embedded system Takeda, M. Kimura, T. Fujioka, Sony Corp. (Japan); M. Fujishima, N. Milosevic, Austrian Research Ctrs. GmbH (Austria) Y. Kani, Sony LSI Design Inc. (Japan); H. Kanbe, Sony Corp. (Japan) In this work, we present a prototype of a compact embedded system implementation of an address-event based algorithm for real-time A pixel structure suitable for IT-CCD image sensor with the pixel size 2 moving foreground detection in industrial environment. Our system ?m or under square was developed. The first feature relates to the uses an optical transient sensor as front-end element, containing of structure of pixel transfer electrode , which consists of island transfer 146 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X electrode and horizontally continuous transfer electrode alternately 6501-41, Session 1 placed in the form of a single layer. The island electrode is connected with polysilicon wiring formed on the horizontally continuous Orthogonal transfer arrays for the Pan-STARRS gigapixel electrode. This pixel structure eliminates one of two electrodes on the camera channel stop region in the conventional single layer electrode structure. Above 30% expansion of photo sensitive areas of 2?m B. E. Burke, MIT Lincoln Lab.; J. L. Tonry, Univ. of Hawaii; M. J. square pixel compared to the conventional design is achieved. Carrier Cooper, A. H. Loomis, P. Doherty, D. J. Young, T. A. Lind, MIT crosstalk between adjacent pixels, often caused by application of Lincoln Lab.; P. M. Onaka, Univ. of Hawaii; D. J. Landers, J. L. read-out voltage to the channel stop region, is well suppressed in the Daneu, MIT Lincoln Lab. new structure. The second feature comes from the photo-shield The orthogonal-transfer array is a new charge-coupled device (CCD) shape around the photodiode. By removing the extension of the for wide-field imaging under development for the University of Hawaii photo-shield and reducing the thickness of the lateral side of the Pan-STARRS program. As previously reported, prototype devices photo-shield, the fill factor expands by 70%. In addition the eclipse of revealed higher noise than desired, as well as spurious charge or incident light by the photo-shield above the vertical register is “glow” due to impact ionization in the NMOS depletion loads. In the suppressed. Our development achieved high performance in photo- redesigned device, the glow problem has been resolved by using two sensitivity, especially in the small F number region, sensitivity shading, load field-effect transistors in series to share the voltage. The noise and handling charge of the photo diode, and provides a desirable has likewise been reduced to our goal of less than 6 e-, compared solution for sub-2?m square pixel IT-CCD. with 10 - 20 e- on the prototype devices. Maintaining sufficient device flatness at low temperatures is another 6501-04, Session 1 issue. The coefficient of thermal expansion mismatch between Si and the Mo base of our package is sufficiently large that the CCD (which Smear noise and sensitivity estimation method for CCD measures 5x5 cm) and package become deformed to a dome shape image sensors having square cells less than 2.0 um by several tens of µm. This problem has been solved by lapping the Mo surface to produce a dish shape with about 15 m depth. This K. Kikuchi, S. Miyazawa, Y. Uchida, H. Kamata, T. Hirayama, Sony µ compensates for the warping so that the device surface remains flat Corp. (Japan) to within 15 µm down to -80°C. Since cell size shrinkage of Charge Coupled Device (CCD) image Currently we are midway through fabrication and packaging of the sensors, sensitivity of image sensors is decreased, which means a 22.6-megapixel redesigned device for the 64-chip Pan-STARRS signal noise ratio (SNR) is also degraded even though we could keep gigapixel camera. an amount of noise. In order to improve a SNR, we used Finite Difference Time Domain(FDTD) method to simulate light energy proportional to generated electrons in Si substrate. When the 6501-06, Session 2 dependence of the smear noise and the sensitivity on image height was investigated, it was found that they had accuracy within 3 dB and Improvement photoelectric conversion efficiency of red 3 % error respectively for all image heights. In order to reduce smear light in HARP film noise and increase sensitivity, we evaluated on-chip micro-lens (OCL) T. Matsubara, Y. Ohkawa, K. Miyakawa, S. Suzuki, M. Kubota, N. structure such as lens thickness and the distance between the Egami, K. Tanioka, NHK Science & Technical Research Labs. photodiode and the OCL. The optimal OCL shape is strongly dependent on device parameters. The optimal thickness of the OCL is (Japan); K. Ogusu, A. Kobayashi, T. Hirai, T. Kawai, Hamamatsu in regards to the smear noise is 0.2 um thinner than that for the Photonics K.K. (Japan) sensitivity. This result demonstrates that when designing the image We successfully enhanced the photoelectric conversion efficiency for sensor structure, including the OCL shape, we not only have to red light in a 15-micrometer-thick high-gain avalanche rushing consider the sensitivity but we also take the smear noise into amorphous photoconductor (HARP) film without deteriorating image consideration. pick-up characteristics and reliability. The conventional HARP film’s photoelectric conversion efficiency decreases gradually for incident light with wavelengths that exceed 450 nm due to the photocarrier 6501-05, Session 1 generation property and a band gap of 2.0 eV in a-Se which is main Development of fully depleted scientific CCD’s for constituent of HARP film. To achieve higher photoelectric conversion efficiency for red light, we designed a new film structure with an astronomy increased total amount of doped-Te with a narrower band gap of 0.34 K. L. Boggs, Semiconductor Technology Associates Inc.; M. P. eV than a-Se. On the other hand, the thickness of the LiF-doped layer Lesser, The Univ. of Arizona for trapping holes was increased than the conventional film to weaken the internal field enhanced by trapped electrons in extra doped-Te. Due to aggressive scientific specifications, Semiconductor The fabricated new red-extended HARP film achieved the Technology Associates and the University of Arizona’s Imaging photoelectric conversion efficiency for red light of about 24% at a Technology Laboratory have collaborated to develop a fully depleted wavelength of 620 nm, which is twice that of the conventional film. back illuminated CCD for scientific imaging. These devices are Finally, we confirmed the improvement in signal to shot noise ratio of designed to target increased quantum efficiency into the near-infrared, 3 dB and the drastic improvement of the color reproduction by using without reduction in the modulation transfer function, charge transfer an HDTV camera with a camera tube incorporating the new red- efficiency, or rms noise. The STA1700 series imagers are back extended HARP film. illuminated 100 micron thick devices with a 10 micron pixel pitch targeted to meet the requirements of the Large Synoptic Survey Telescope (LSST). Recent characterization results will be presented 6501-07, Session 2 including the point spread function of a 2 micron spot. Also discussed will be the thinning and packaging developments for the STA1700. Bifocal liquid lens zoom objective for mobile phone These efforts include the addition of a backside bias contact, invar applications package design with high density connectors, and etching and backside changing optimization for high resistivity silicon. F. C. Wippermann, P. Schreiber, A. H. Bräuer, Fraunhofer-Institut für Angewandte Optik und Feinmechanik (Germany); P. Craen, Varioptic SA (France) Commercially available solutions for mobile phone cameras with auto focus functionality without moving parts have been developed using the electrowetting technology. Here, the contact angle of a drop of a

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 147 Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X conductive or polar liquid placed on an insulating substrate can be scheme. In this system, Variance function is used as the coarse influenced by an electric field and finally the power of the lens can be evaluation function in a large scale, and Square-gradient function is adjusted. Besides virtually no deterioration, easy control electronics taken as the accurate evaluation function within a certain scale. and simple and therefore cost-effective fabrication, this type of liquid Fuzzy-control is adopted to find out the optimized focus. It’s proved lenses enables extremely fast settling times compared to mechanical by the experiment that the system can auto-focus accurately by using approaches. Mobile phones with zoom functionality are desired as a the focusing method based on fuzzy-control. And the sampled image next evolutionary step. With the recent commercially available liquid is so clear that it can satisfy the operating requirement. The auto- lens products a completely motionless and continuously adaptive focusing structure is quite simple and easy to be implemented. zoom system with market relevant optical performance is not feasible. The key factor that affects the automatic focusing speed is the This is due to the limited change in optical power the liquid lenses can redundancy movement of motor. So whether the motor can reach the provide. As an intermediate step we propose a bifocal liquid lens focus with least steps and rapidest speed or not is very important for zoom sufficient for toggling between two effective focal lengths improve the efficiency of focusing. While the key factor that affects without any moving parts. The system has its mechanical counterpart both parameters is the control arithmetic of auto focusing. in a bifocal zoom lens where only one lens group has to be moved. We present design considerations for a bifocal liquid lens zoom and While adopting the focusing evaluation function as our metewand, we propose an adapted image acquisition algorithm for simplifying the found that the focusing evaluation characteristic curve assumes the achromatization of the lens. parabola form. The focusing characteristic curve ascend or descend monotonously as a whole, and it varies slowly near the focus, while it varies quickly away from the focus, in the mean while, the polarity of 6501-08, Session 2 focusing function’s first derivative changes on the focus. Therefore we could locate the focus under the help of the mathematical method The intermediate size direct detection detector for that used to solve the extremum. But for the voltage surge of the electron microscopy lamp-house’s power supply, the vibration of step motor, the EMI of CCD and image adapter , as well as the straightness error of lens’s L. Jin, A. Milazzo, Univ. of California/San Diego; S. Kleinfelder, S. slideway , the image brightness varies so much in the actual Li, Univ. of California/Irvine; P. C. Leblanc, F. Duttweiler, J. C. environment of focusing that the smoothness of the curve is far from Bouwer, S. T. Peltier, M. H. Ellisman, N. Xuong, Univ. of California/ perfect, which is characterized by an large undulation and the San Diego interfering impulse on the top . presents the top segment of focusing In a longstanding effort to overcome limits of the current charge function’s characteristic curve, we can find that the automatic focusing function curve appears rather uneven , and companies with coupled devices (CCDs) systems in electron microscopy, we have developed a radiation-tolerant system that can withstand direct large undulation. As a result, it is unable to establish an accurate electron bombardment. A prototype system based on an Active Pixel mathematics model and it’s hard to solve the problem by classics control theory. In recent years, the intelligent control theory has been Sensor (APS) has already delivered unprecedented performance-in many areas exceeding the fundamental capabilities of CCD based developing quickly and it might provide us the possibility to solve the systems. This implementation features a 512Å~550 pixel format of above mentioned problems. Through the analysis of the curve, the variation of the curve can only be expressed by words that have fuzzy 5µm pitch with an excellent signal-to-noise ratio (approximately 13/1 for a single incident electron in the range of 200-400 keV) and a very concept, such as quicker and slower, etc. Therefore, we choose the fuzzy control theory to control the automatic focusing process. high spatial resolution (~2 µm). Radiation tolerance with high-energy exposure is also impressive, especially with cooling to -15 °C. Stable In this paper, CCD camera is taken as the pickup equipment, an performance has been demonstrated over a period of several months image card is used as the digital image sampling equipment, a of continuous use before a thermal annealing process is required to biological microscope is used as the imaging carrier, a Model- restore the device. The characteristics of this new detector have 36BF003 step motor together with a Model-XB6-25 harmonic reducer exciting implications for transmission electron microscopy, especially makes up the driver of the lens, the Variance function is adopted as for cryo-EM as applied to biomolecular structures. the coarse valuation function in large scale, and the Square-gradient function is adopted as the accurate valuation function within a certain scale. Sever experiments are performed under above conditions. The 6501-09, Session 2 results shows that within a focusing scale of 500μm, the focusing precision reaches 7μm. One CCD micro measure A novel image-acquisition system using the optical- system developed by HIT is used to test the repeated accuracy and multiplex system the data show that the non-repeated error is less than 0.67μm, T. Narabu, Sony Corp. (Japan) which can satisfy the requirement of the measurement well. We have developed for cameras a new system of capturing images called the optical-multiplex system. The system is comprised of an 6501-11, Session 2 image sensor and a multi-lens array, which has the advantage of being very compact and very light and can provide a deep depth of Position detection with hyperacuity using artificial field. In a model of the system for simulation purposes, light passes compound eyes through five pin holes in an “aperture sheet” and the object A. Brueckner, J. Duparré, A. H. Bräuer, A. Tuennermann, information is detected by an image sensor. What makes this system Fraunhofer-Institut für Angewandte Optik und Feinmechanik unique is that the object information passing through each of the pin (Germany) holes (or lenses) is a specific range of information, which is coordinated on the pixel array. This is in contrast to the current multi- Artificial apposition compound eyes are insect inspired imaging lens systems, in which the object information from each lens is systems which are best suited for miniaturization due to their completely independent. Our simulation model incorporating both fabrication with micro-optics technology. They combine the potential signal and noise suggests that some pixels in this new system will for a large field-of-view with a small volume and weight but they have an improved signal-to-noise ratio over conventional systems. exhibit a low space-bandwidth product. This means that the number of image details that can be transferred by the system is low because the diameter of the individual lenses is small and the fact that each 6501-10, Session 2 channel contributes to only one pixel. However, information e.g. movement or position of contrasting objects can be extracted locally A method of automatic focusing based on fuzzy control with sub-pixel accuracy. Inspired by the natural phenomenon of D. Shao, Harbin Institute of Technology hyperacuity, a redundant sampling in combination with the knowledge about the impulse response of the imaging system is used to extract This paper presents an automatic focusing method based on fuzzy- highly accurate information using a low resolving artificial apposition control. Using this method, we designed an optimized focusing

148 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X compound eye. Two methods are described which use linear system separated from other transistors secure sufficient amplifier gate length theory to calculate the position of simple objects like point sources without refining the minimum design. In this study, we have improved and edges from the measured intensities of adjacent channels. two characteristics of CIS having shared transistor architecture. The Therefore, the implementation of precise position detection with an QV conversion efficiency was increased by 10%, with a 30% accuracy increase by a factor of 25 compared to the conventional reduction in the diffusion capacitance. In addition, the number of image resolution is demonstrated. white blemishes decreased to 1% of the initial value when a weaker electric field was used. 6501-12, Session 2 6501-16, Session 3 High-dynamic-range video for photometric measurement of illumination Image sensor for objects detection using modulated light J. Unger, S. Gustavson, Linköpings Univ. (Sweden) H. Taruki, A. Ohno, F. Ono, T. Hamamoto, Tokyo Univ. of Science (Japan); T. Sasaki, T. Shirai, M. Sakai, The Nippon Signal Co., Ltd. We describe the design and implementation of a high dynamic range (Japan) (HDR) imaging system capable of capturing RGB color images with a dynamic range of 10,000,000:1 at 25 frames per second. We use a We have been investigating a system to detect moving objects highly programmable camera unit with high throughput A/D correctly at the place where luminous intensity is changing because of conversion, data processing and data output. the influence of incident light such as sunlight, fluorescent light and HDR acquisition is performed by multiple exposures, which are car headlight. The object detection system consists of a smart image processed and saved to a mantissa-exponent representation. The sensor and a modulated LED light, and it is possible to suppress the exposures are performed as a continuous rolling shutter progression influence of the change of background light by using a different value over the sensor. All rows are not imaged at the same time, which between two image values when the LED light turns on and off. yields a slight curtain effect for camera and scene motion, but in Because the speed of modulation is high for accurate detection of return all exposures for one particular row of pixels are acquired head moving objects, electric charges from a photodiode are distributed to tail within the frame time, which means that the entire frame time into two capacitors by switching in sync with the LED light in a pixel can be used for light integration, and the longest exposure is almost circuit of the sensor. Also, the sensor has a subtraction function by a the entire frame time. current mirror circuit to reduce the same charges from two capacitors before saturation. By the frequent subtractions, it is possible to The system is highly configurable, and there are trade-offs possible increase only the influence of the modulated light and reconstruct between dynamic range, number of exposures, image resolution and wide dynamic range images at outside of the sensor by using the frame rate. information of subtractions and the voltage value of each capacitor. We use this HDR video system in a setup where it is registering the scene through the reflection in a mirror sphere, producing a photometrically correct measurement of the incident illumination from 6501-17, Session 3 all directions. Linear-logarithmic image sensor with low noise and no flickers 6501-14, Session 2 S. Miyatake, M. Kusuda, T. Iwamoto, K. Kamon, K. Sato, Y. Serita, Thermal imaging based procedures for special concretes Konica Minolta Technology Ctr. (Japan) characterization Most logarithmic-response image sensors have been suffered from G. Bonifazi, M. Animali, L. Damiani, Univ. degli Studi di Roma/La poor performance at low light level because they can not afford buried Sapienza (Italy) photodiodes that reduce the dark current and kTC noise drastically in the four-transistor (4-TR) pixel CMOS image sensors. Another It is well known as the presence of fine and ultra-fine particulate solids drawback of them is that they output the logarithmically converted products (fillers), inside a concrete, strongly contributes to change value of the photocurrent only when the signal is readout without final manufactured products characteristics. The presence of “fillers”, integration, resulting in flicker noise. in fact, improve physical-chemical and mechanical properties of This paper proposes a linear-logarithmic image sensor whose pixel is concrete such as workability and water retention. In this paper an composed of one buried-photodiode, four n-channel MOS transistors, innovative approach, based on thermal-imaging, was applied in order two p-channel MOS transistors, and one capacitor. The sensor to establish a correlation, for the different utilized fillers, among fillers operates in both pure-linear mode and linear-logarithmic mode. When quantity and quality, concrete behaviour during ageing and final it operates in the pure-linear mode, the two p-channel MOS mechanical characteristics of the products at 28 days. transistors compose a source-follower circuit for the floating diffusion and the capacitor does its load. Thus the pixel acts just same as the 4-TR one and offers low dark current and low kTC noise. When it 6501-15, Session 3 operates in the linear-logarithmic mode, the transfer gate is always Shared transistor architecture with diagonally connected open and the reset transistor and one p-channel MOS transistor pixels for a CMOS image sensor operate in the sub-threshold regime. The former logarithmically converts the photocurrent to a voltage fed to the gate of the latter, Y. Kudoh, F. Koga, T. Abe, H. Taniguchi, M. Sato, H. Ishiwata, S. whose source is connected to the capacitor and outputs a signal Ooki, R. Suzuki, H. Mori, Sony Corp. (Japan) logarithmically proportional to the integrate amount of the photocurrent. Thus no flickers occur. We have developed a pixel unit for a CMOS image sensor (CIS) having shared transistor architecture with diagonally connected pixels. This pixel unit has four photodiodes and seven transistors. It has a pixel 6501-18, Session 3 size of 2.5 ?m square. The transistors were designed using 0.18- micron aluminum process technology. Shared diffusion for reading A wide-dynamic-range CMOS imager by hybrid use of signal electrons occurs between the corners of two photodiodes. In active and passive pixel sensors addition, the two diffusions from a pair of pixels and the reset transistor’s source diffusion are connected by a metal wire. In this K. Kagawa, Y. Adachi, Nara Institute of Science and Technology pixel unit, no pair of photodiodes are on the same line; thus, line (Japan); Y. Nose, H. Takashima, K. Tani, A. Wada, SANYO Electric access architecture enables fast data output. This layout has the Co., Ltd. (Japan); M. Nunoshita, J. Ohta, Nara Institute of Science advantages of a long amplifier gate length and a large photodiode and Technology (Japan) area. Floating diffusions for reading signal electrons which are

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 149 Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X We have proposed a wide dynamic range CMOS imager with a new 6501-21, Session 3 pixel readout scheme based on a hybrid use of active and passive pixel sensors (APS and PPS). Wide dynamic range active pixel sensor arrays for digital We utilize a pixel structure with an overflow transistor and a capacitor. x-ray imaging using a-Si:H technology Photo charges overflowed from a photodiode to the in-pixel capacitor J. Lai, A. Nathan, Univ. of Waterloo (Canada); J. A. Rowlands, at the high illumination are read out in a passive manner by a charge Sunnybrook and Women’s Health Sciences Ctr. (Canada) amplifier of a column. The advantage of the passive pixel sensor is that the charges under the threshold voltage of in-pixel readout Hydrogenated amorphous silicon (a-Si:H) active matrix flat panel transistor can be read out, while a source follower used in active pixel imagers (AMFPIs) have gained considerable significance in digital sensors cannot read the photodiode voltage under the threshold diagnostic medical imaging applications, in view of their large area voltage. Therefore, the total amount of the detectable photo charges, readout capability. Current interests lie in a multi-faceted a-Si:H array namely dynamic range of the imager, increases. We have fabricated a which is compatible with multiple x-ray imaging modalities that 128x96-pixel CMOS imager with the proposed scheme in a 0.35-um demands sufficiently wide dynamic range and variable signal gain. CMOS technology. This work proposes an APS circuit with a global shutter TFT that From the experiments, basic operations of our pixel readout scheme, decouples the storage and sensing capacitances. Unlike previously dynamic range extension, automatic active and passive mode reported APS designs, this design performs charge sensing and switching according to the intensity of the incident light were storage with separate dedicated capacitors and the analog operation confirmed. of the SHUTTER TFT. The capacitance ratio between storage and sensing capacitors enhances the in-pixel charge gain and yield at least a seven-fold increase to previously reported APS circuits for 6501-19, Session 3 small input level with similar pixel size. The pixel transfer characteristic demonstrates piece-wise linearity with sensitivity from Small distance detective camera module sub-pico-ampere to nano-ampere photocurrent at 33 ms integration I. Oyama, T. Iijima, K. Imada, S. Tamaki, T. Suenaga, N. Imamura, time. This transfer characteristic widens the pixel dynamic range and T. Ota, M. Suzuki, M. Nagashima, Matsushita Electric Industrial proves amenable for both low dosage dynamic imaging and high Co., Ltd. (Japan) input static imaging. Together with external amplification, this APS circuit delivers the performance needed in terms of signal gain, Camera modules have been used for various applications, and there dynamic range, and readout rate entailed by fluoroscopic and is continuing demand for smaller and multifunctional camera modules. radiographic imaging applications. Our new type of camera module is small, thin and can detect the distances of objects as well as those images. Our camera module has a four-lens-array, one imaging sensor and optical filters. The imaging 6501-22, Session 3 sensor is divided to four regions. Two of those regions are covered with infrared filters, and the other two regions are covered with green Demonstration of a low-voltage three-transistor-per-pixel filters. A prototype with a focal length of 2.63mm and a baseline CMOS imager based on a pulse-width-modulation length of 2.59mm was fabricated. The two images of same optical readout scheme employed with a one-transistor in-pixel filters have parallaxes, and can detect the distancees of objects. We comparator use infrared images with infrared LED at night, and use green images during the daytime. With sufficient calibrations of images, we S. Shishido, I. Nagahata, T. Sasaki, K. Kagawa, M. Nunoshita, J. achieved the accuracy of distance detection less than +/-2.5% at 1m Ohta, Nara Institute of Science and Technology (Japan) in spite of a small baseline length. Consequently, our new distance- A pulse-width-modulation(PWM) pixel reading scheme is effective in detective camera module is small and thin, and generates an accurate implementing a low-voltage single-power-supply CMOS image depth-map as well as an object image. Hence our camera module can sensor. In our scheme, a ramp-shaped voltage signal is applied to the be applicable for automobile vehicles, security systems and three- source node of the in-pixel gate-common amplifier transistor working dimensional images. as a comparator to perform single-slope ramp-ADC. The features of our sensor are low-voltage operation by a PWM scheme, small input referred noise and reduced column-to-column fixed-pattern- 6501-20, Session 3 noise(FPN) due to a large voltage gain of the in-pixel amplifier A digital high-dynamic-range CMOS image sensor with transistor, and suppressed pixel-to-pixel FPN by a feedback reset. Because pixel value (light intensity) is encoded to the temporal multi-integration and pixel readout request information, namely, width of the digital pulses, reduction of the A. Guilvard, École Nationale Supérieure de l’Aéronautique et de power-supply voltage less affects the dynamic range. We have l’Espace (France) and STMicroelectronics (France); J. Segura, fabricated a 128x96-pixel prototype sensor in a 0.35um CMOS STMicroelectronics (France); P. Magnan, P. Martin-Gonthier, École technology, in which a ramp generator for pixel-level ramp-ADC, Nationale Supérieure de l’Aéronautique et de l’Espace (France) bootstrap circuits for pixel select and reset signals were integrated. The number of the transistors in a pixel was three, which was same as The natural scenes can have very different lightening conditions, from that of the ordinary 3-Tr active pixels sensors. We have successfully 10-3 lux for night vision to 105 lux for bright sunlight and even more demonstrated the operations of the prototype image sensor with a for direct lightening source. In several applications such as 1.5-V single power supply voltage. automobile or automotive, the roughly 60dB dynamic of a standard CMOS imager does not allow to keep all the relevant information content of the captured scene. 6501-23, Session 3 In the sensor presented several exposure are used to extend the A 960-fps sub-sampling object extraction CMOS image sensor dynamic range and to build a unique HDR frame. The different integration times are none linearly spread. Entirely digital, the sensor sensor with 12-bit column parallel ADCs and ALUs uses photodiode based pixel with an in pixel ADC, that allows a built Y. Motohashi, T. Kubo, H. Kanto, T. Tate, S. Sugawa, Tohoku Univ. in dynamic compression function. The sensor implements an (Japan) innovative and fast event-driven readout chain, to avoid the readout of the entire pixel array after each integration. Pixels directly request for 1A CMOS image sensor with highly accurate object extraction pre- readout, and the captured frame is rebuilt along the different processing functions by 960-fps sub-sampling operation, 12-bit integrations, with no pixel loss or measurement errors thanks to a column parallel successive approximation ADCs and column parallel acknowledged readout principle. The pixel values are coded using the ALUs has been developed. The pixel is composed of four transistors mantissa exponent principle which allows the use of numerous built in type pixel which shares the source follower transistor and the pixel dynamic compression functions, making this sensor very versatile. select transistor. The each ADC is composed of the noise and signal

150 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X holding capacitance, the noise reduction circuit, the comparator and above four light to frequency conversion pixels configured in a the small DAC that combined both the reference voltage ratios and common centroid arrangement. The sensor was fabricated in 0.18µm capacitance ratios. In the ALU, the object categorization pre- CMOS technology as part of a test chip. Polarization sensitivity was processing is performed by the each macro block of 3 Å~ 3 pixels observed which agreed with theoretical analysis. A modulation depth which has a reference pixel and its neighboring eight pixels. The three of 10% was also observed, as was a transmittance efficiency of 40%. image features which are the edge of object, the direction of edge- This transmittance efficiency is comparable to similar sensors vector and the average of light-intensity of 3 Å~ 3 pixels implemented using dichroic film based polarizers. However this corresponded to each pixel are extracted by the ALUs. The image and sensor presents the advantage of requiring no additional materials or the results of the object extraction pre-processing are outputted by manufacturing stages to that of a standard CMOS process and every 60-fps. The image sensor was fabricated by 0.35-µm 2P3M therefore can be produced at a reduced cost. It was found that the technology. The pixel pitch is 5.3-?m, the number of pixels is 640H Å~ sensor could measure angular rotation to within 0.7 degrees and it is 360V and the chip size is 4.9-mm square. This image sensor can be expected that with improved pixel performance this value will be utilized in various image-processing applications. reduced.

6501-24, Session 3 6501-27, Session 4 A temperature-resistant wide dynamic range CMOS A diffractive multispectral image sensor with on- and off- image sensor die signal processing and on-Die optics in 0.18-micron K. Mizobuchi, S. Adachi, T. Yamashita, S. Okamura, H. Oshikubo, CMOS Texas Instruments, Inc. (Japan); N. Akahane, S. Sugawa, Tohoku C. J. Thomas, R. I. Hornsey, York Univ. (Canada) Univ. (Japan) On-die optics have been proposed for imaging, spectral analysis, and A temperature resistant wide dynamic range (WDR) CMOS image communications applications. These systems typically require extra sensor has been developed using the very low dark current front-end process steps to fabricate on-die optics. Fabrication of diffractive of line (VLDC FEOL) and the metal hermetic-sealed package optics using the metal layers in commercial CMOS processes suppressing the degradation of the spectra response of the on-chip circumvents this requirement, but produces optical elements with micro lens and color filters (OCML/OCCF). A 1/4-inch VGA 5.6-?m poor imaging behavior. This paper discusses the application of Wiener pixel WDR CMOS image sensor has been fabricated through 0.18-?m filtering to reconstruction of images suffering from blurring and 2P3M process with VLDC FEOL which contains the pinned chromatic abberation, and to identification of the position and photodiode with less electrical fields, the less plasma etching wavelength of point sources. Adaptation of this approach to analog damages, the transfer gate with the suppressed current at Si-SiO2 and digital FIR implementations are discussed, and the design of a interface and the furnace temperature process for the re- multispectral imaging sensor using analog FIR filtering is presented. crystallization. The sensor chips with the conventional OCML/OCCF Simulations indicate that off-die post-processing can determine point assembled into the metal hermetic-sealed package by the high source wavelength to within 5% and position to within 0.05 radian, vacuum welding machine without residual oxygen has finally received and resolve features 0.3 radian in size in images illuminated by white the thermal stress test (150-deg.C/500-hours). The dark current is light. The analog hardware implementation is simulated to resolve 350-pA/cm2 at 85-deg.C (50-pA/cm2 at 60-deg.C). No degradation of features 0.3 radian in size illuminated by monochromatic light, and 0.6 the spectra response in any of R/G/B pixels is observed after the radian with white light. thermal stress test. It is found that the thermal decomposition of the OCML/OCCF (phenol resin) is not caused easily in nitrogen with the low residual oxygen concentration. The sample images captured by 6501-28, Session 4 the WDR CMOS image sensor assembled camera keep good quality up to 85-deg.C. Diffraction grating lens array T. Korenaga, T. Ando, S. Moriguchi, F. Takami, Y. Takasu, S. Nishiwaki, M. Suzuki, M. Nagashima, Matsushita Electric 6501-42, Session 4 Industrial Co., Ltd. (Japan) Characterization and system modeling of a 5 Mpixel We have proposed a new type of camera module with thin structure CMOS Array and distance-detection capability. This camera module has a four- D. Dorn, K. Cabanas-Holmen, Pelco, Inc. lens-array with diffraction gratings (one for blue, one for red, and two for green). The diffraction gratings on a mold are formed mechanically, We present characterization results for a 5 million pixel CMOS image and the plastic lens-array is fabricated by injection molding process. sensor designed for high speed applications. This sensor is capable The two green images are compared to detect parallaxes, and of outputting 12 frames per second and incorporates on-chip 12-bit parallax-corrected blue, red and green images are composed to digitization. We present measurements of system gain, read noise, generate a color image. We have developed a design software and dark current, charge capacity, linearity, photo response non- molding technologies for the grating lenses. The height and period of uniformity, defects, color reproduction, quantum efficiency, and blazed gratings and shapes of aspheric lenses are optimized, and modulation transfer function. The image sensor incorporates blue, red and two green aspheric lenses with gratings are molded as exposure control functionality, windowing, on-chip binning, anti- one unified four-lens-array. The diffraction gratings on both surfaces blooming capability and a on-rolling shutter architecture to implement of lens improve field curvature especially, and realize wide-angle image capture mode. The results show a favorable aspect of the images. However, blazed gratings sometimes cause unnecessary ability to achieve high speed, high resolution, and very good diffraction gratings to prevent high-resolution images. We have sensitivity in a monolithic CMOS sensor. Architecture trades for high developed a new method to measure necessary first order diffraction speed imaging systems utilizing CCDs and CMOS sensors are also lights and unnecessary diffraction lights separately. By use of this presented. method, we can show relations between molding conditions and necessary/unnecessary diffraction lights. The unnecessary diffraction lights can be diminished with proper molding processes, and our 6501-26, Session 4 grating lenses can be used for image capture. Low-cost optical polarization sensor P. Baxter, J. M. Raynor, STMicroelectronics (United Kingdom); D. Renshaw, The Univ. of Edinburgh (United Kingdom) A low cost polarization sensor has been presented. This sensor implements metal grid polarizers formed from a standard metal layer

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 151 Conference 6501: Sensors, Cameras, and Systems for Scientific/Industrial Applications X

6501-29, Session 5 6501-40, Session 5 Critical assessment of the integrity of thin organic films Development of wavelength changeable multiband by shearography mixing visual device and its application K. J. Habib, Kuwait Institute for Scientific Research (Kuwait) F. Ding, Y. Chen, K. Chao, USDA Agricultural Research Service In the present work, the temperature versus thermal deformation Previously, we showed that two- and three-band color-mixing (strain) with respect to time, of different coating films were studied by techniques could be used to achieve results optically equivalent to a non-destructive technique (NDT) known as shearography. An two- and three-band ratios that are normally implemented using organic coating, i.e., ACE Premium Enamel, on a metallic alloy, i.e., a multispectral imaging systems, for enhancing identification of single carbon steel, was investigated at a temperature range simulating the target types against a background and for separation of multiple severe weather temperatures in Kuwait especially between the targets by color or contrast. In this paper, a prototype of a daylight and the night time temperatures, 20-60 0C. The investigation wavelength-changeable two- and three-band color-mixing device is focused on determining the in-plane displacement of the coating, presented and its application is demonstrated. The wavelength- which amounts to the thermal deformation (strain) with respect to the changeable device uses changeable central wavelength bandpass applied temperature range. Furthermore, the investigation focused on filters and various filter arrangements. The experiments showed that a determining the thermal expansion coefficients of coatings, the slope color-mixing technique implemented in a pair of binoculars coupled of the plot of the thermal deformation (strain) versus the applied with an imager could greatly enhance target identification of color- temperature range. In other words, one could determine, from the blindness test cards with hidden numbers and figures as the targets. decreasing value of the thermal expansion coefficients of coatings, a Target identification of color blindness cards was greatly improved by critical (steady state) value of the thermal expansion coefficients of using two-band color-mixing with filters at 620 nm and 650 nm, which coatings, in which the integrity of the coatings can be assessed with were selected based on the criterion of uniform background. Target respect to time. In fact, determination of critical (steady state) value of identification of a different set of color blindness test cards was also the thermal expansion coefficients of coatings could be accomplished improved using three-band color-mixing with filters at 450 nm, 520 independent of parameters, i.e., UV exposure, Humidity, exposure to nm, and 632 nm, which were selected based on the criterion of chemical species, and so on, normally are considered in conventional maximum chromaticness difference. These experiments show that methods of the assessment of the integrity of coatings. In other color-mixing techniques can significantly enhance electronic imaging words, with the technique of shearography, one would need only to and visual inspection. determine the critical (steady state) value of the thermal expansion coefficients of coatings, regardless of the history of the coating, in order to assess the integrity of coatings. Furthermore, results of shearography indicate that the technique is very useful NDT method not only for determining the critical value of the thermal expansion coefficients of different coatings, but also the technique can be used as a 2D- microscope for monitoring the deformation of the coatings in real-time at a submicroscopic scale.

6501-30, Session 5 Deformation determination of aircraft parts by photogrammetry M. Varshosaz, K.N. Toosi Univ. of Technology (Iran) This paper investigates the precision and capability of low cost off-line industrial photogrammetry for determination of deformations in aircraft parts. The study concentrates on deformation analysis of an airplane propeller through standard photogrammetric procedures. In this regard, around 100 targets were fixed on the blades of the propeller and imaged from several stations. The left and right models of the blades were developed using the measured points. The two models were then compared with each other to reveal any deformations in the propeller. The results indicated a maximum of 2.5 mm deformation in the blades. The investigations carried out in this research, suggest that with an ordinary non-metric digital camera, low cost targets, and scale bars, an accuracy of around 1:20,000 (up to 50?m in object measurements) can be achieved. In addition, compared to current quality control techniques used in aviation, photogrammetry offers more flexibility, convenience, and a reasonable accuracy suitable for different measurement applications in aviation.

152 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6502: Digital Photography III Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6502 Digital Photography III

6502-01, Session 1 other ergonomic factors (such as single-hand-hold vs. both-hand- hold, shutter button position, capture delay and camera-mass) affect Mobile imaging camera-motion pattern. The results indicate that camera-motion K. Johnson, F. Xiao, J. Pincenti, Motorola, Inc. roughly follows a straight line for exposure duration shorter than 1/8 second. Based on this finding, we proposed a generalized rule-of- The booming mobile imaging industry is closing the gap with digital thumb (for maximum exposure duration) that links the rule-of-thumb still cameras (DSCs) in terms of image quality. In this talk, we will used in 35mm film camera world with digital camera world (mobile discuss the challenges that the mobile imaging industry is facing. Due imaging included). We also made a few recommendations to limit the to the size constraints of cell phone camera modules (especially the impact of camera-motion on captured images. thin trend set by the RAZR), one dilemma that the mobile imaging industry is facing is whether we want to pack more pixels into the same die size. Everybody knows that higher pixel count doesn’t 6502-04, Session 2 necessarily equal higher image quality because you lose sensitivity with smaller pixel size. To make matters worse market reports find Modeling of pixel reset associated wide dynamic range that camera phones are typical used in darker environment than CMOS log image sensor DSCs. The real question becomes at what point we stop the current M. Hashimoto, Omron Automotive Electronics Technology trend for imagers with smaller and smaller pixels. In the Japanese (Germany); Y. Tani, Omron Corp. (Japan); J. Sochos, Omron market prints are shown to demonstrate quality rather than relying on Automotive Electronics Technology (Germany) a pixel count. This further indicates that the mobile imaging industry needs to focus more seriously on factors other than pure pixel count A logarithmic response pixel (often referred as 3T pixel) captures wide and understand the trade off between cost, size and reliability. dynamic range images at pixel level. Since the 3T pixel captures image without need of photo charge integration in principle, a continuous image capturing operation mode was widely used. Image 6502-02, Session 1 lag was an issue of the operation. Parasitic capacitance at pixel photo current sensing node is the cause of the lag. The lag can be removed Ultracompact optical zoom lens for mobile phone by initializing the sensing node (that is, by applying a reset signal to K. Matsusaka, S. Ozawa, R. Yoshida, T. Yuasa, Y. Souma, Konica that node) since it occurs when a part of photo charge captured in the Minolta Opto, Inc. (Japan) previous cycle be reminded in the pixel parasitic capacitance. Once the pixel initialization step is introduced, the pixel response becomes Several key technologies which are essential to mount optical zoom time variant. It however does not mean a loss of pixel sensitivity at lens units onto mobile phones maintaining the compactness of them low lighting condition. A high sensitivity pixel is possible to realize by will be described. Those are optical design method, precise lens minimizing capacitance of pixel sensing node. In this paper, the reset alignment system, and compact zooming actuators. By integrating operation associated pixel output response is discussed. Its analytical those technologies, we have developed very small zoom lens unit and easy to handle approximation expression will be derived. The suitable for 3 to 5 mega-pixel cameras. Those technologies have analytical expression is contributing to a fixed pattern noise correction shown the new orientation in the future of high-end mobile camera algorithm development that is suitable to our reset associated 3T pixel phones. First, we have designed new optical zoom lens system which array. can achieve ultra compact size and mass productivity by utilizing plural aspherical glass-mold lenses and controlling decentering sensitivities. Second, we have developed precise lens alignment 6502-05, Session 2 system to compensate the field tilt caused by errors in the fabrication processes such as lens assembling and glass lens molding. Finally, Leakage characteristics for the buried photodiode we have also developed a long-stroke compact lens actuator based structure on CMOS image sensors on our unique SIDM (Smooth Impact Drive Mechanisms) technology. S. Lee, S. Lim, C. Lee, S. Bang, S. Kang, J. Park, J. Park, Dongbu This newly-developed actuator has made an important contribution to the miniaturization of the zoom lens units. Recently, the demand of Electronics Co., Ltd. (South Korea); R. A. Martin, S. Bae, Foveon, the camera phones having higher mega pixel cameras is getting Inc. bigger. We are developing new zoom lens units such as folding-type The leakage characteristics of the buried photodiode structure have and straight-type to meet this requirement by utilizing the foregoing been investigated in direct color CMOS image sensor with a stacked technologies. We will introduce those optical systems. photodiode (PD) structure tailored for detecting red, green and blue light. Image quality was investigated showing that the blue photodiode has surface related effects while the red and green PDs 6502-03, Session 1 do not. From these experiments, it is found that the activation energy Camera-motion and mobile imaging of PDs display dependence on area, periphery, and corners and the corner component dominants. Leakage characteristic of PDs show F. Xiao, J. Pincenti, G. John, K. Johnson, Motorola, Inc. similar behavior to normal n+pwell diode of similar structure. Also the separate contribution from the area, periphery and corners, and their For the booming mobile imaging industry, there are several challenges relationship to STI was analyzed by TCAD. that need to be addressed in order to achieve DSC-like image quality. Minimizing camera-motion introduced image blur is one of them. Film For the first time, we have analyzed the vertical buried photodiode photographers have long used a rule-of-thumb that a hand held structure and found that corner components on red and green PD can 35mm camera should have an exposure in seconds that is not longer be source of leakage current. We also found that surface contact of than the inverse of the focal length in millimeters. Due to the lack of blue PD can be a noise source, reducing image quality. Therefore, to scientific studies on camera-motion, it is still an open question how to maintain high image quality, the blue photo diode of a CIS has to be generalize this rule-of-thumb to mobile imaging. In a previous study, designed as a buried structure and the connections to the buried red we used captured pictures of a point light source to measure how and green PDs has to be free from STI sidewall contact. camera-mass, user’s skill level and exposure duration affect camera- motion. However, this method couldn’t provide reliable measurement of camera-motion for exposure duration shorter than 1/8 second (which is more relevant for mobile imaging). In this study, we used a gyroscope-based system to measure much more accurately how

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 153 Conference 6502: Digital Photography III

6502-06, Session 2 We apply this to our main goal, the understanding of “good” spectral sensitivities. One can expect that color fidelity favors spectral CMOS color image sensors with silicon photodiode and sensitivities matching those of the human eye, SNR the widest overlaid organic photoconductive layer having narrow spectra and color sensitivity wide but non-overlapping spectra. We absorption band test these hypothesis by consider a hypothetical, simple family of spectral sensitivities and evaluate the “best” filters thus obtained also S. Takada, M. Ihama, Fuji Photo Film Co.,Ltd. (Japan); T. from the point of view of color fidelity and perceptual performance. Komatsu, T. Saito, Kanagawa Univ. (Japan) At EI2006, we proposed the CMOS image sensor, which was overlaid with organic photoconductive layers in order to incorporate in it large 6502-09, Session 3 light-capturing ability of color films characterized by their multiple- Joint demosaicing and super-resolution imaging from a layer structure, and showed the pictures taken by a trial product of the set of unregistered aliased images proposed CMOS image sensor overlaid with an organic layer having green sensitivity. In this study, we have tried to get the optimized P. Vandewalle, École Polytechnique Fédérale de Lausanne spectral sensitivity for the proposed CMOS image sensor to meet the (Switzerland); D. Alleysson, Univ. Pierre-Mendes-France (France); compatibility between light-capturing efficiency and color S. E. Süsstrunk, École Polytechnique Fédérale de Lausanne reproduction by means of the simulation to get the change in S/N (Switzerland) ratio in terms of luminance owing to the operation to minimize the color difference between the original Macbeth chart and the one The goal in super-resolution imaging is to reconstruct a high- reproduced from the spectral sensitivity of the sensor. As a result, it resolution image from a set of lower resolution input images. Typically, has been confirmed that the proposed CMOS image sensor with such super-resolution algorithms are applied to RGB images after multiple-layer structure possesses high potential capability in terms of demosaicing. Such an approach is sub-optimal, however, as image-capturing efficiency when it is provided with the optimized demosaicing can introduce interpolation artifacts: color aliasing spectral sensitivity. appears in high frequency regions due to undersampling of the color image with the Bayer CFA. Those artifacts are then considered as part of the signal in the super-resolution image reconstruction. In this 6502-07, Session 3 article, we thus present a joint demosaicing and super-resolution algorithm from unregistered input images. Matrixing, number of color channels, and color gamut of First, we extract the luminance information from each of the mosaiced camera devices CFA input images. Then, the images are aligned using a frequency F. Gasparini, R. Schettini, S. Bianco, Univ. degli Studi di Milano- domain approach on the luminance channels. As the high frequency Bicocca (Italy) part of the spectrum is most affected by aliasing, we only use the low frequencies for the image registration. Once the input images are For digital camera devices the gamut can be defined as the range of aligned, they can be combined on a common grid. The luminance and color stimuli across which their responses show differences. Most of chrominance information can then be separately interpolated at a the methods proposed in the literature, assuming available the whole higher resolution from the nonuniform grid of known pixel values. range of inputs that the device can capture, try to determine the Finally, the luminance and chrominance information is combined to response of the device to each sample, and then estimate the gamut result in a high-resolution color image. boundaries from these data. We propose an alternative method based on the inversion of the processing pipeline of an imaging device. Its greatest advantage is that it does not have to process anymore all the 6502-10, Session 4 possible input color stimuli, but instead, starting for the digital outputs and inverting the processing pipeline, it estimates all the possible Autofocus survey: a comparison of algorithms color stimuli that belong to the camera color gamut. This color gamut L. C. Shih, Cypress Semiconductor Corp. it then compared with the gamut of all physically possible color stimuli, to have an idea of its extension. Numerous autofocus algorithms are well documented in literature, but tradeoffs between them have not been widely researched. Hence, In this paper we investigate how different matrixing methods influence choosing a suitable autofocus algorithm for a given application may the estimation of the device color gamut. We compare the Least often not be a straightforward process. This study explores the Squares, the White Point Preserving Least Squares procedure, and relative merits of a set of autofocus algorithms via examining their another method, developed by the authors. performance against a variety of scene conditions. Furthermore we investigate how the input gamut changes by adding We create a statistics engine that considers a scene taken through a an optimal forth filter to the three original camera color channels. range of focal values and then computes the best focal position using each autofocus algorithm. The process is repeated across a survey of 6502-08, Session 3 test scenes containing different representative conditions. The results are assessed against ideal solutions which are determined by From spectral sensitivities to noise characteristics manually focusing the scenes. Through examining these results, we then derive conclusions about the relative merits of each autofocus F. Guichard, DxO Labs. (France); J. Buzzi, École Polytechnique algorithm with respect to various scene conditions. (France); H. Hornung, DxO Labs. (France) For a given noise at the photosite level and output color space, spectral sensitivities of an imaging chain (determined by color and 6502-11, Session 4 infrared filters and photoelectric properties) impact its output level of Monte Carlo evaluation of ten focus measures noise as they determine, e.g., the color matrix which expands and/or mixes the individual channel noises. Hence the final noise may be Y. Tian, Univ. of California/Berkeley sometimes very different from the usually documented sensor level A high-performance focus measure is one of the key components in noise and present strong correlations between channels making any auto-focusing systems based on digital image processing. More individual channel SNR misleading. than a dozen of focus measures have been proposed and evaluated in We first evaluate existing chains (reflex, bridge and phone cameras) the literature, yet there have be no comprehensive evaluations that both by classical signal-to-noise ratios and more complete indicators include most of them. The purpose of the current study is to evaluate like “color sensitivity”. We compare these indicators among and compare the performance of ten focus measures using Monte themselves and to perceptual evaluations. We also show to what Carlo simulations, run on a local computer cluster with distributed extent actual performance can be predicted given a simple output computing capacity. From the perspective of a general framework for color space.

154 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6502: Digital Photography III focus measure evaluations, we calculate the true point spread 6502-14, Session 5 functions (PSFs) from defocus and aberrations represented by OSA standard Zernike polynomials using FFT. For each run, a range of Adaptive color bleeding removal for video and still DCT defocus levels are generated, the PSF for each defocus level is compressed sequences convoluted with an original image, and a certain amount of noise is added to the resulting defocused image. Each focus measure is A. Castorina, G. Spampinato, A. Bruna, A. Capra, applied to all the blurred images, and result in a focus measure curve. STMicroelectronics (Italy) The procedure is repeated on a few representative images for different DCT based compression engines, in particular for low bit rates, are types and levels of noise (Gaussian, Poisson, salt & pepper, and well known to introduce different kind of distortions. Among them, speckle). The performance of the ten focus measures is compared in blocking and ringing distortions have been deeply investigated, while terms of monotonicity, unimodality, defocus sensitivity, noise the chromatic color bleeding effect, due to both chromatic sub- sensitivity, effective range and computational efficiency. sampling and quantization in very high compression rates, has only recently received proper attention. The scope of this paper is to propose an innovative and effective algorithm to overcome this kind of 6502-12, Session 4 color artifacts. This technique can be applied on both still and video Combining strategies for white balance DCT based compression algorithms and is well suited for real time applications, due to its lightweight nature. The algorithm, in order to F. Gasparini, R. Schettini, S. Bianco, Univ. degli Studi di Milano- identify the macro blocks possibly affected by bleeding artifacts, Bicocca (Italy) performs an edge detection based on the mean magnitude of Sobel filters on luma and chroma channels. Afterwards, on these blocks, the White balance is the process of removing from digital images de-bleeding filter, adaptively generated by means of simple fuzzy sets unrealistic color casts mostly due to the acquisition condition. Digital fired by the previously computed measures, is applied and the results cameras often have great difficulty with auto-white balance (AWB), is weighted with the unfiltered data, to remove distortions while that, in the worst cases, can introduce unrealistic damaging colors. preventing edge corruption. Experimental results, on differently coded Several strategies for AWB are proposed in the literature. These, in still pictures and video sequences have shown valuable improvement general, require some information about the camera being used, or both through objective quality metrics (PSNR) and subjective are based on assumptions about the statistical properties of the inspection. expected illuminants and surface reflectances. Computationally, white balance is a two-stage process: the illuminant is estimated, and the image colors are then corrected on the basis of this estimate. The 6502-15, Session 5 correction generates a new image of the scene as if taken under a known standard illuminant. Super-resolution total-variation decoding of JPEG- In this work six methods of AWB were considered: Gray World, White compressed image data Point, Color by Correlation, 2D Gamut Mapping, an Iterative White T. Saito, T. Komatsu, Kanagawa Univ. (Japan) Balance procedure and our Self Tunable Color Balancing. The idea investigated in this work is to not rely on a single method, In digital cameras, output images are often compressed with the but instead to consider a consensus decision that takes into account JPEG encoder. When the coding rate is not high enough, in JPEG- the compendium of the responses of all the considered algorithms. decoded images there appear noticeable artifacts such as the blocky We analyze different combining schemes to improve the results in distortions and the ringing distortions. In this paper, we present a new terms of RMS error in the estimated illuminant chromaticity. decoding approach that recovers quality-improved images from JPEG-compressed image data while suppressing the occurrence of those particular coding-distortions. The new decoding approach is a 6502-13, Session 5 kind of super-resolution image restoration approach based on the total-variation image regularization; to reduce artificial ringing signal- Removal of signal-dependent noise for a digital camera variations near image edges it restores the DCT coefficients truncated by the JPEG compression, whereas in originally smooth image T. Saito, Y. Ishii, R. Nosaka, T. Komatsu, Kanagawa Univ. (Japan) regions it flattens unnecessary signal-variations to eliminate the In a digital camera, several factors cause signal-dependency of blocky distortions. additive noise. Many denoising methods have been proposed, but L.I. Rudin, S. Osher and E. Fetami first introduced the total-variation most of them do not work well for the actual signal-dependent noise. minimization as an image regularization criterion for solving inverse To solve the problem of removing the signal-dependent noise of a image-restoration problems in 1992. It has proved to be very efficient digital camera, this paper presents a new denoising approach via the for regularizing images without smoothing the boundaries of objects nonlinear image-decomposition. In the nonlinear decomposition-and- in images. In 2001, F. Malgouyres and F. Guichard applied the total- denoising approach, at the first nonlinear image-decomposition stage, variation regularization to the joint problem of de-blurring and over- multiplicative decomposition is performed, and a noisy image is sampling where images are magnified while image blurs caused by represented as a product of its two components so that its structural the 0-th order-hold are removed. They defined the energy functional to component corresponding to a cartoon approximation of the noisy be minimized as the weighted sum of the total-variation energy and image may not corrupted by the noise and its texture component may the data-fidelity energy in which the sampling and the hold-blur collect almost all the noise. At the successive nonlinear denoising operators are used as the image-degradation operator. They proved stage, intensity of the separated structural component is utilized that the total-variation approach achieves super-resolution; it zooms instead of the unknown true signal value, to adapt the soft- and simultaneously sharpens images effectively, without producing thresholding-type denoising manipulation of the texture component to ringing artifacts near edges. Our super-resolution decoding approach the signal dependency of the noise. At the final image-synthesis of JPEG-compressed image data rests on the above-mentioned stage, the separated structure component is combined with the de- theoretical findings. noised texture component, and thus a sharpness-improved denoised image is reproduced. Experimental simulations show that the The JPEG encoder quantizes only significant DCT coefficients, and nonlinear decomposition-and-denoising approach selectively removes discards insignificant DCT coefficients. The JPEG encoder is regarded the signal-dependent noise of a digital camera without blurring sharp as the image-degradation operator corresponding to band-limitation edges and destroying visually important textures. and quantization in the DCT domain. Instead of the sampling and the hold-blur operators, we introduce the JPEG encoder as the image- degradation operator used in the data-fidelity energy, and thus we define the total-variation energy functional whose minimization gives the super-resolution JPEG decoding. However, as side effects, the total-variation regularization tends to flatten original signal-variations, and hence fine textures in images may be degraded. To remedy the

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 155 Conference 6502: Digital Photography III drawback, in this paper we introduce a spatially adaptive technique 6502-18, Session 6 that changes the weighting parameter of the total-variation energy dependently on the local image-structure around a pixel. Kernel-size selection for defect pixel identification and To minimize the energy functional, we employ an iterative descent correction algorithm. The usual JPEG-decoded image is used as an initial image, E. T. Chang, Cypress Semiconductor Corp. and then iteratively the decoded image is refined. The update terms are composed of the divergence of the gradient of the total-variation In a digital camera, the Image Signal Processing (ISP) pipeline may be energy and the term determined by the difference between the JPEG- called to identify and hide defective pixels in the image sensor. Often encoded data of the recovered image and the initial JPEG-encoded filters are designed and implemented to accomplish these tasks, data. Within about fifty iterations, the decoded image converges to a without considering the cost in memory or the effect on actual steady restoration image. images. Experimental simulations demonstrate that our super-resolution We create a simulation system which uses an inverse ISP to model decoding approach with the total-variation regularization efficiently raw sensor data with defect pixels. The simulation includes lens blur, removes the coding distortions specific to the JPEG compression. inverse gamma, additive white noise, and mosaic. Defect pixels are Near sharp edges, especially, to compensate for the ringing added to the simulated raw data, which is then run through various distortions our decoding approach successfully recovers the DCT defect pixel correction algorithms. The end result is compared against components that have been truncated by the JPEG compression. the original simulated raw data to measure the effect of the added defects and defect pixel correction. We implement a bounding min-max filter as our defect pixel correction 6502-16, Session 6 algorithm. We conclude that the choice of kernel size and other A bad pixel location algorithm for cell phone cameras parameters depends not only on memory requirements, but also expected defect pixel rate. With a large defect pixel rate, we must M. Aleksic, S. R. Goma, ATI Technology (Canada) choose algorithms which are more aggressive in the defect correction but also result in higher accidental degradation. As CMOS imaging technology advances, sensor to sensor differences increase, suggesting an increasing need for individual, per sensor, calibration steps. Traditionally, the cell-phone market has a low 6502-36, Session 6 tolerance for complex per unit calibrations. This paper is proposing an algorithm able to do automatic calibration on a per phone basis, MEMS digital camera without a complex testing environment. The algorithm tries to locate pixels with characteristics out of the mean range of the values R. C. Gutierrez, T. K. Tang, R. J. Calvet, E. R. Fossum, Siimpel specified by the manufacturer in terms of light response, usually Corp. referred as ‘bad pixels’. It uses several images captured from a sensor MEMS technology uses photolithography and etching of silicon without using a mechanical shutter or predefined scenes. The wafers to enable mechanical structures with less than 1 um tolerance, implementation that follows uses two blocks: a dynamic detection important for the miniaturization of imaging systems. In this paper, we block (local area based) and a static correction block (location table present the first silicon MEMS digital auto-focus camera for use in cell based). The dynamic block fills the location table of the static block phones with a focus range of 10 cm to infinity. At the heart of the new using clustering techniques. The result of the algorithm is a list of silicon MEMS digital camera, a simple and low-cost electromagnetic coordinates containing the location of the found ‘bad pixels’. An actuator impels a silicon MEMS motion control stage on which a lens example is given of how this method can be applied to several is mounted. The silicon stage ensures precise alignment of the lens different cell-phone CMOS sensors. with respect to the imager, and enables precision motion of the lens over a range of 300 um with < 5 um hysteresis and < 2 um repeatability. Settling time is < 15 ms for 200 um step, and < 5ms for 6502-17, Session 6 20 um step enabling AF within 0.36 sec at 30 fps. The precise motion An optic model of the appearance of blemishes in digital allows COTS optics to maintain MTF \> 0.8 at 20 cy/mm up to 80% field over the full range of motion. Accelerated lifetime testing has photographs shown that the alignment and precision of motion is maintained after A. Zamfir, FotoNation Romania SRL (Romania); A. Drimbarean, 8,000 g shocks, thermal cycling from -40 C to 85 C, and operation FotoNation Ireland Ltd. (Ireland); V. V. Buzuloiu, M. Zamfir, Univ. over 20 million cycles. Politehnica Bucuresti (Romania); E. Steinberg, FotoNation Inc.; D. V. Ursu, ProOptica (Romania) 6502-19, Session 7 Changing the lens of a DSLR camera has the drawback of allowing small dust particles from the environment to be attracted onto the Establishing imaging sensor specifications for digital still sensors’ surface. As a result, unwanted blemishes may compromise cameras the normally high quality of photographs. The particles can be M. A. Kriss, Consultant removed by physically cleaning the sensor. A second, more general approach is to locate and remove the blemishes from digital photos Digital Still Cameras, DSCs) have just about replaced conventional by employing image processing algorithms. This paper presents a film based cameras in the consumer market if not in the advanced model that allows computing the physical appearance (actual size, and professional imaging markets. The main selling feature over the shape, position and transparency) of blemishes in a photograph as a last 12 years has been the number of pixels supported by a given function of camera settings. In order to remove these blemishes with camera. Almost all new cameras sport 6-mega pixels, be they thin, sufficient accuracy, an initial algorithm calibration must be performed compact cameras or rather large SLR type cameras. This paper for any given pair camera-lens. The purpose of this step is to estimate provides a systematic way to evaluate the quality of the resulting some parameters of the model that are not readily available. To images as a function of pixel count, pixel size, sensor architecture, achieve this, a set of “calibration images” must be carefully taken lens quality and other imaging system component characteristics. The under conditions that will allow the blemishes to become easily metrics used in the quality evaluation are system sharpness, the identifiable. Then, in a second step, based on the photo’s header potential to alias (artifact generation), ISO speed, dynamic range and content, the actual appearance of the blemishes in the given exposure latitude. Two previously developed models for DSCs are photograph is computed and used in the automatic removing used to make the evaluations. The first is a “linear” model of the algorithm. Computing formulas and results of our experiments are imaging elements that control sharpness and the potential for aliasing also included. artifacts. The second is a physics based model for the ISO speed, full well capacity, dynamic range and exposure latitude for a sensor when it is treated in a manner similar to conventional film. The results clearly

156 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6502: Digital Photography III show that not all pixels are equal and that in general a SLR type the handshake about two axes simultaneously but it is possible to use camera with a given number of pixels will produce better images than only one single motion direction as well. Controlling of the vibration the more compact camera with the same number of pixels, but where unit and parameter input are both effected by user via connected the pixels are smaller and the lens systems slightly poorer. computer. We also adopted the amplitude measurement method of human physiological tremor by Bradley J. Davis and John O’Conell with regard to holding photographic cameras and used the 6502-20, Session 7 ascertained values for our tests. The paper describes in detail the basic approaches and considerations of our measurement method Quality criterion for digital still camera followed by results from exemplary stabilizing systems tests. S. N. Bezryadin, KWE International, Inc. The main quality requirements for a digital still camera are color 6502-23, Session 8 capturing accuracy, low noise level, and quantum efficiency. Different consumers assign different priorities to the listed parameters, and New generation of image editing algorithms camera designers need clearly formulated methods for their S. N. Bezryadin, KWE International, Inc. evaluation. While there are procedures providing noise level and quantum efficiency estimation, there are no effective means for color Any color image editing software has Brightness, Contrast, and capturing accuracy estimation. Introduced in this paper criterion Saturation controls. However, because it usually imitates allows to fill this gap. Luther-Ives condition for correct color corresponding adjusting knobs of a Color TV, and thus, corresponds reproduction system became known in the beginning of the last to mid 20th century scope of engineering, adjusting one of the century. However, since no detector system satisfies Luther-Ives parameters affects all three parameters, and modification of condition, there are always stimuli that are distinctly different for an Brightness or Contrast does not preserve chromatic coordinates. A observer, but which detectors are unable to distinguish. To estimate person should be very experienced with the sequential control conformity of a detector set with Luther-Ives condition and calculate a operations in order to get a result equivalent to a simple measure of discrepancy, an angle between detector sensor sensitivity expocorrection. and Cohen’s Fundamental Color Space may be used. In this paper, A set of new generation algorithms described in this paper is free from the divergence angle is calculated for some typical CCD sensors and the above-mentioned defects and includes: Brightness and Contrast a demonstration provided on how this angle might be reduced with a editing which does not affect chromatic coordinates; Local Contrast corrective filter. In addition, it is shown that with a specific corrective editing that causes only minor modification of Global Dynamic Range; filter Foveon sensors turn into a detector system with a good Luther- Global Dynamic Range modification which affects neither chromatic Ives condition compliance. coordinates, nor Local Dynamic Range; and Saturation modification which affects neither Brightness, nor Hue. 6502-21, Session 7 The efficiency of color image editing software depends on the choice of a basic CCS (Color Coordinate System). A CCS that is effective for Digital camera resolution measurements by sine one editing procedure might be less effective for another. This paper modulated Siemensstars presents a set of correlated CCSs with a specification of their preferable area of application. C. M. Loebich, Image Engineering (Germany); A. Jaeger, Canon Deutschland GmbH (Germany); B. Klingen, Fachhochschule Köln (Germany); D. Wueller, Image Engineering (Germany) 6502-24, Session 8 Measuring the Spatial Frequency Response using a Slanted Edge has Characteristics of experimental single-chip color HDTV proven to be a reliable method when applied to raw image data. As image acquisition system with 8M-pixel image sensor soon as sharpening or compression is applied to an image the edge spread function still provides useful information about the sharpening H. Shimamoto, R. Funatsu, T. Yamashita, K. Mitani, Y. Nojiri, NHK and its effect on edges but the SFR derived from it does not reflect Science & Technical Research Labs. (Japan) the resolution of that camera anymore. To measure the resolution of a digital camera from processed image data a different method had to We developed an experimental single chip color HDTV video image be found. The result of a study at the Cologne University of Applied acquisition system with 8M-pixel CMOS image sensor. The imager Science presented in this paper is a method, which uses a sinusoidal has 3840 (H) x 2160 (V) effective pixels and built-in analog-to-digital Siemens Star. An image of that star taken with the camera under test converters, and its frame rate is 60-fps with progressive scanning. The is analyzed by a software which basically takes the digital values of MTF characteristic we measured with this system on luminance signal pixels along a radius of the star - representing a specific frequency - in horizontal direction was about 45% on 800 TV lines. This MTF is and fits in a sine function with the least square error. From this better than conventional three-pickup broadcasting cameras, optimized sine function the contrast for this frequency is derived. therefore we are able to keep the enhancement gain low and hence Measured at different frequencies the result is a MTF. The method has the noise level after detail enhancement process low. We also been tested in the past three years and is considered to become part measured the color characteristics and corrected the color gamut of a future revision of ISO 12233. using matrix gain on primary colors. We set the color correction target similar to that of three-pickup color cameras in order to use multiple cameras to shoot for broadcasting, where all cameras are controlled 6502-22, Session 7 in the same manner. In this case, the difference between the single- chip system and three-pickup cameras becomes very small. The color Measurement method for image stabilizing systems correction gain of the matrix is also less than that for the ideal color space. B. Golik, D. Wueller, Image Engineering (Germany) In the photographic practice, whether digital or analogue, the handshake of the photographer often results in disturbing blur. In 6502-25, Session 8 medicine the phenomenon of rhythmic, involuntary muscle contractions occurring in all healthy individuals is known as Full spatial resolution RGB digital camera physiological tremor. The frequency of this tremor has its peak at 8-12 R. Kumontoy, J. P. Allebach, G. T. C. Chiu, J. Huang, D. Howard, Hz as described by Kevin T. Wyne or Sharon Smaga. Camera and lens N. Bajaj, Purdue Univ. manufacturers developed various stabilizing systems to compensate for handshake. In order to evaluate the quality of image stabilizing In a conventional CCD, CID, or CMOS sensor array, only one of three systems we have developed a method based on automated different types of filters: red, green, or blue, is placed on top of each reproducible mechanical simulation of human handshake followed by photoreceptor site in the sensor array. This results in reduced effective resolution measurement of captured images. We are able to simulate spatial resolution of the captured image. To yield a full-spatial-

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 157 Conference 6502: Digital Photography III resolution image in each of the three channels, demosaicing is Conventional RGB color filter arrays (CFA) in single-chip color needed to generate the two missing channels at each pixel through cameras greatly reduce light intensity, and therefore limits the ISO interpolation algorithms. The interpolation process is particularly speed of cameras. A novel CFA and a method of forming color images prone to errors in image areas that contain fine detail. As a result, the are proposed. 75% of the CFA are transparent elements, and the image may suffer from a loss of this detail. It may even contain remaining 25% are repeated color filter blocks. A compact artifacts, such as moiré patterns. In this paper, our approach in arrangement of color filter elements in each block helps to reduce obtaining a full spatial resolution RGB images is by moving a typical color artifacts. Such an arrangement features a higher sampling rate mosaiced sensor array with a high speed and high precision scanning of luminance and a lower sampling rate of chrominance. Black-and- stage. By appropriately shifting the sensor array using the scanning white images (BI) with high resolution can be acquired from the stage and taking three additional exposures of the image. A full- transparent elements, and color images (CI) with low resolution from resolution RGB image can be readily obtained from the three exposed the color filter blocks. To generate output color images (OI), the CI in images. Experimental data have verified the feasibility of the proposed RGB format is transformed into CIE Lab space, and the luminance approach. Image analysis results comparing the full spatial resolution components are replaced with the high-resolution BI. Based on such RGB image and that obtained from various interpolation approaches a principle compatible to JPEG format, the visual quality of OI is will be presented. satisfactory. Simulation was conducted using raw images acquired with a Canon 20D camera. Results show the potential of the CFA in making digital cameras with high ISO speed. Applications can be 6502-26, Poster Session security day/night cameras and cell phone cameras that are able to capture quality images under sufficient and low light levels. Fast nonlocal means for image denoising R. C. Bilcu, M. Vehvilainen, Nokia Research Ctr. (Finland) 6502-29, Poster Session The digital images captured by a multimedia device always suffer from several distortions (noise, blur, etc). An important degradation is the The usage of digital cameras as luminance meters noise that can be introduced by various sources such as: the D. Wueller, Image Engineering (Germany); H. Gabele, FRAMOS recording medium (film, digital sensor), the transmission medium, Electronic Vertriebs GmbH (Germany) measurement and quantization errors. Although, the manufacturing technology of the digital sensors is constantly improving, the level of Many luminance measuring tasks require a luminance distribution of the sensor noise is still high even in the latest camera modules. One the total viewing field. The approach of image-resolving luminance interesting de-noising algorithm, called Non-Local means (NL-means) measurement, which could benefit from the continual development of algorithm, was recently reported in the literature. This method is position-resolving radiation detectors, represents a simplification of based on the idea that images contain many similar patterns (like such measuring tasks. edges, etc) and averaging similar patterns will reduce the additive Luminance measure cameras, which are specially manufactured for noise level. The Gaussian weighted Euclidean distance between the measuring tasks with very high requirements already exist. Due to MxM windows centered at the two pixel positions is computed. The high-precision solutions these cameras are very expensive and for weighted euclidean distances computed between neighbouring pixels many image resolving measuring tasks are not commercially viable. are further used to compute a weighted average that will smooth out Therefore it is desirable to measure luminance with digital still the noise while retaining the image details. Although this algorithm cameras which are freely available at reasonable prices. produces excellent de-noising results, its computational complexity is high which makes it unpractical. In this paper we propose a novel The paper presents a method for the usage of digital still cameras as modified NL-means algorithm for image de-noising containing a pre- luminance meters independent of the exposure settings. A calibration classification step to decrease the computational load. of the camera is performed with the help of an OECF measurement and the luminance is calculated with the camera’s digital RGB output values. The test method and computation of the luminance value 6502-27, Poster Session irrespective of exposure variations is described. The error sources which influence the result of the luminance measurement are also CMOS image sensor noise reduction method for image discussed. signal processor in digital cameras and camera phones Y. Yoo, S. Lee, W. Choe, C. Kim, SAMSUNG Advanced Institute of 6502-30, Poster Session Technology (South Korea) Motion deblurring based on fusing differently exposed Digital images captured from CMOS image sensors suffer Gaussian noise and impulsive noise. To efficiently reduce the noise in Image images Signal Processor (ISP), we analyze noise feature for imaging pipeline M. Tico, M. Vehvilainen, Nokia Research Ctr. (Finland) of ISP where noise reduction algorithm is performed. The Gaussian noise reduction method and impulsive noise reduction method are We present an approach to motion deblurring based on exploiting the combined for proper ISP implementation in Bayer domain. The information available in two differently exposed images of the same proposed method takes advantage of the analyzed noise feature to scene. Besides the normal-exposed image of the scene, we assume calculate noise reduction filter coefficients. Thus, noise is adaptively that a short exposed image is also available. Due to their different reduced according to the scene environment. Since noise is amplified exposures the two images are degraded differently: the short exposed and characteristic of noise varies while the image sensor signal image is affected by noise, whereas the normal-exposed image could undergoes several image processing steps, it is better to remove be affected by motion blur. The method presented in this paper noise in earlier stage on imaging pipeline of ISP. Thus, noise reduction estimates the motion blur point spread function (PSF) that models the is carried out in Bayer domain on imaging pipeline of ISP. The method degradation of the normal-exposed image, following to recover the is tested on imaging pipeline of ISP and images captured from image of the scene by deconvolution. The main processing steps Samsung 2M CMOS image sensor test module. The experimental detailed in the paper are: image registration and motion blur PSF results show that the proposed method removes noise while estimation. Image registration operation includes a preprocessing effectively preserves edges. step meant to cancel the differences between the two images due to their different exposures. Next, the registration parameters are estimated by matching the preprocessed images based on an image 6502-28, Poster Session based registration approach. Motion blur PSF estimation is carried out by exploiting the difference between the degradation models of the A novel color filter array with 75% transparent elements two images, as well as certain prior assumptions about a typical G. Luo, Harvard Medical School motion blur PSF. Experiments and comparisons are presented in order to validate the proposed method.

158 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6502: Digital Photography III

6502-31, Poster Session defects, little information is available about the nature or quantity of defects that develop in the field. Anecdotal evidence from digital Image demosaic via Nevatia-Babu edge detection camera users indicates that the development over time of “dead,” C. C. Reinhart, J. Bodie, California Lutheran Univ. “stuck”, and “hot” pixels is commonplace. Traditional fault detection schemes require the camera be returned to the factory for evaluation An N-pixel digital camera must produce 3N bytes of image data to and may explain the lack of information. In this paper, we overcome produce a color image of the requisite quality expected by the end that limitation and extend our defect identification methodology that user. All but the high end sensors can produce only N bytes of image uses only normally-captured images from the sensor itself. With the data. The other 2N bytes must be interpolated via a demosaic aim of tracking defects in real-world sensors, models are developed algorithm. The sensors (pixels) are typically laid out in a sparse color for the behaviour of defects in cameras utilizing several imaging Bayer pattern but other configurations are certainly possible. In the technologies. The prevalence and response characteristics of hot and Bayer pattern, green sensors populate 1⁄2 of the image plane while partially-stuck pixels are extracted via laboratory calibration while blue and red populate 1⁄4 each. The end requirement is that each abnormal sensitivity defects are shown to be rare in the test devices. pixel contains all three colors. Combined results are also used to investigate long-term stability of Methods of interpolation of the missing data fall into three basic defects. Finally, in-camera image processing is shown to further categories: 1) bilinear interpolation; 2) multi-frame super resolution; compound uncompensated defects and confirms that in-field and 3) gradient-based interpolation. In our work we investigate a detection and correction are necessary. gradient-based method using a template matching approach to demosaic (reconstruct) a full color image from the sparse pixel data captured by a CMOS imager. The proposed method is based on the 6502-34, Poster Session Nevatia-Babu template-based linear feature extraction algorithm. This Adaptive nonlinear imaging characteristic for wide- approach provides the color accuracy of gradient based algorithms dynamic-range image sensor using variable integration yet reaps the benefit of regularity of processing of the bilinear approach. In this paper we give some background on the Nevatia- time Babu linear feature extractor, describe our algorithm (gradient F. Suzuki, T. Hamamoto, Science Univ. of Tokyo (Japan) estimation and color interpolation), compare results to other approaches, and present a hardware implementation. The As the issue of wide dynamic range imaging, some adjustments are consideration of hardware implementation is especially important as needed to display an image with wider dynamic range than that of CMOS imagers find their way into low cost devices such as cell displays. Because the dynamic range of existing displays is limited to phones and other novelty camera applications. Finally, we discuss 10 from 8 bits for each color, the image quality degrades when the future directions of the research. characteristic is not suitable for the objects. One of our smart image sensors can control the integration time in each pixel. In the sensor, the intermediate photodiode value is compared with arbitrary 6502-32, Poster Session threshold at arbitrary timing and reset depending on the result. We call this “judgment and reset process”. The characteristic for wide Automated digital camera sensor characterization dynamic range imaging is decided by thresholds and timings. So optimization of the parameters adapted to the brightness distribution C. M. Normand, P. R. Fornaro, R. Gschwind, Univ. of Basel of objects and the new image sensor which realizes the proposed (Switzerland) method are reported. Before the first judgment, the PD values are Color characterization of professional imaging devices typically read out and passed to the external circuit in order to estimate the involves the capture of a reference color target under the scene- histogram of the image at a frame end. In the histogram, successive specific lighting conditions and the use of dedicated profiling pixels are recognized as the detected block. Quota of output values in software. However, the limited set of color patches on the target may the characteristic are mainly decided by the block width. In not adequately represent the reflection spectra found in the scene. We consequence, the dynamic range is widened and the contrast is present a solution developed in collaboration with a camera dramatically improved. manufacturer for the automatic color characterization of the sensors without the need of a physical color target. The optimal color transforms are computed based on the individually measured sensor 6502-35, Poster Session sensitivities, computer generated sets of color spectra forming a Illumination flicker detection and frequency classification virtual characterization target and a mathematical model of the camera. The use of a virtual target enables the optimization of the methods color transform for specific image capturing situations by selective T. Tajbakhsh, R. Grigat, Technische Univ. Hamburg-Harburg generation of the reflection spectra. Both the use of matrices and (Germany) look-up-table are investigated, whereas only the matrix approach is suited for in-camera processing. The results are compared to In many modern CMOS imagers employing pixel arrays the optical traditional color profiling solutions for a variety of scenes and objects. integration time is controlled by the method known as rolling shutter. This integration technique, combined with fluorescent illuminators exhibiting an alternating light intensity, causes spacial flicker in 6502-33, Poster Session images through the sequence. This flicker can be avoided when the integration time of the imager corresponding to the shutter width is Identification of in-field defect development in digital adjusted to a multiple of the flicker period. Since the spatial frequency image sensors of the flicker can vary upon the local AC power frequency, a classification must be performed beforehand. This is either performed J. Dudas, L. M. Wu, C. G. Jung, D. C. Chen, G. H. Chapman, utilizing an additional illumination intensity detector or in the case we Simon Fraser Univ. (Canada); I. Koren, Z. Koren, Univ. of focus on, restricting to image information only. Massachusetts/Amherst In this paper we review the state of the art techniques of flicker Given the trend in digital image sensors toward shrinking pixel detection and frequency classification, and propose two robust dimensions, more pixels per sensor, and larger sensor areas, the classification methods based on a clear mathematical model of the likelihood of developing defective pixels is greatly increased. However, illumination flicker problem. Finally we present another approach for while various techniques are available to locate manufacture-time compensating for flicker in single images suffering from these artifacts by inversing the flicker model function. Thereby, the flicker phase, amplitude and frequency are to be adjusted adaptively from frame to frame. This however compensates for the fact, that the shutter width is not limited to a multiple of the flicker period.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 159 Conference 6503: Machine Vision Applications in Industrial Inspection XV Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6503 Machine Vision Applications in Industrial Inspection XV

6503-01, Session 1 then used as the test statistic to decide the verity of the null hypothesis. The stain is then segmented by a decision rule based on Automatic characterization of cross-sectional coated the probability map generated by the EM algorithm. The proposed particle nuclear fuel using greedy coupled Bayesian approach was tested on a dataset of 48 fabric images soiled with snakes stains of ketchup, corn oil, mustard, ragu sauce, revlon makeup and grape juice. The decision theoretic part of the algorithm produced a J. R. Price, D. Aykac, J. D. Hunn, A. K. Kercher, Oak Ridge correct detection rate (true positive) of 92 % and a false alarm rate of National Lab. 5.2% on these set of images. We describe new image analysis developments in support of the U.S. Department of Energy’s (DOE) Advanced Gas Reactor (AGR) Fuel Development and Qualification Program. We previously reported a 6503-04, Session 1 non-iterative, Bayesian approach for locating the boundaries of Compact multispectral imaging system for contaminant different particle layers in cross-sectional imagery. That method, detection on poultry carcasses however, had to be initialized by manual preprocessing where a user must select two points in each image, one indicating the particle M. Kise, B. Park, K. C. Lawrence, W. R. Windham, USDA center and the other indicating the first layer interface. Here, we Agricultural Research Service describe a technique to eliminate this manual preprocessing and provide full automation. With a low resolution image, we use The objective of this research is to design and fabricate a compact, “EdgeFlow” to approximate the layer boundaries with circular cost effective multi-spectral instrument and to collect and analyze templates. Multiple snakes are initialized to these circles and spectra for real-time contaminant detection for poultry processing deformed using a greedy Bayesian strategy that incorporates coupling plants. The prototype system developed in this research consists of a terms as well as a priori information on the layer thicknesses and multi-spectral imaging sensor, illumination system and handheld PC. relative contrast. Once our described convergence criterion is met, To develop the system cost-effectively, all components are selected these boundaries are propagated to the full resolution image and the from off-the-shelf products and manually assembled. The multi- previously published technique is applied. We show results indicating spectral imaging sensor developed in this research is a two-port the effectiveness of the proposed method. imaging system that consists of two identical monochrome cameras, optical system and two narrow band-pass filter whose center of the wavelength are 520 and 560 nm, respectively. A spectral reflectance 6503-02, Session 1 from a chicken carcass is collected and split in two directions by an optical system including a beam-splitter and lenses, and then two Basic scanner for parallepipedic manufactured pieces identical collimated lights are filtered by the narrow band-pass filters and delivered to the cameras. Lens distortions and geometric H. Hrimech, J. Pavéglio, F. S. Marzani, Y. Voisin, Univ. de misalignment of two cameras are mathematically compensated to Bourgogne (France) register two images. This paper describes a machine vision dedicated to the dimensional The prototype system is tested in the real environment and shows that control of power capacitors. The geometry of these pieces is it can effectively detect feces and ingesta on the surface of poultry parallepipedic. This system is in keeping with the category of the carcass. active vision systems. It is built with two cameras and a LCD projector. This one will be used to determine the deformation of the parallelepiped. The calibration of the system is done with the 6503-05, Session 1 Faugeras-Toscani’s method. The set of points for the calibration is obtained by detecting the corners of the black squares on the 3D Chinese videotext detection based on directional pattern. For this, we have used a technique based on the Harris gradient angular information corner detectors. The LCD projector is calibrated by using the last J. Peng, Xidian Univ. (China) calibration. It is used to determine the deformation of the parallelepiped. The capacitor is held on a rotating table and we Videotext refers to text superimposed on still images and video examine it from all sides. So we can find all the points of interest of frames, and plays an important role in the Content-Based Indexing the capacitor. Detected points are stored and will be used to give the and retrieval system. Detection of videotext in a frame is the result of dimensional measure. In the beginning of the process, an operator image analysis involving text character segmentation and location. enters the references of the capacitor and the dimensions that he Though well progress has been made in this research area in recent wants to control. All the dimensions of the capacitors produced are years, the technique improvement is needed still. Especially the stored in a data base. It’s easy and quick to check if these existing method does not take the attributes of Chinese characters measurements are in adequacy with the specifications. into consideration. In this work, we present an approach to Chinese text detection from digital video frames based on directional gradient angular histogram. : 6503-03, Session 1 In this work, it is analyzed that the angular information of image edges Statistical modeling, detection, and segmentation of is more important than the modular magnitude information to describe an object. Therefore, we use the angular distribution of image edges stains in digitized fabric images in forming a feature vector that should have strong discriminative A. Gururajan, H. Sari-Sarraf, E. F. Hequet, Texas Tech Univ. power to support image content detection. And then, we analyzed the Chinese character that the angular of edge to Chinese text is mostly This paper will describe a novel and automated system based on a about 0, 45, 90 and 135 degree (suppose that the angular is from 0 to computer vision approach, for objective evaluation of stain release on 180 degree). So we pick up n edge pixels that have the n largest cotton fabrics. Digitized color images of the stained fabrics are magnitude values in the edge map to construct an angular histogram obtained, and the image intensities in the saturation plane of these with K bins by counting the number of edge pixels for which the angle images are probabilistically modeled as a Gaussian Mixture Model falls into a certain interval. Then we can give a judgment whether the (GMM). Stain detection is posed as a decision theoretic problem, region exist Chinese text by detecting the peak of histogram in bins where the null hypothesis corresponds to absence of a stain. The null near 0, 45, 90 and 135 degree. : hypothesis and the alternate hypothesis mathematically translate into a first order GMM and a second order GMM respectively. The Our algorithm procedure can summarize following: It can start with parameters of the GMM are estimated using the Expectation- wavelet decomposition by an anti-symmetric biorthogonal wavelet, Maximization (EM) algorithm. Minimum Description Length (MDL) is then detect edge using wavelet decomposition coefficients, and then

160 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6503: Machine Vision Applications in Industrial Inspection XV compute the gradient angular histogram for a windowed area at the advantage of this system is that it delivers directly images of the lowest resolution level and classify it into text or non-text. After object in the chosen channel. Up to now mainly military applications procedure is complement for the whole picture at a given resolution were investigated, but there is a large potential for civil applications. level, we can move to the next higher level and test the blocks for which the corresponding areas in the lower level has been classified as text area, When detection is completed in the highest resolution 6503-08, Session 2 level, the test-rock will be refined further by using information contented in the successive frames. The final output is a binary image Infrared imaging and machine vision of detected text areas with text as foreground in black on a white F. Meriaudeau, Univ. de Bourgogne (France) background. This can be the input to an OCR which recognizes the text characters. : This article will be a review of various machine vision systems which rely on infrared imaging. In our method, we detect edge using the decomposition coefficients of anti-symmetric biorthogonal wavelet, so the directional gradients The article will present systems based on active infrared come from a smoothed version of the image, in which the noise thermography and passive infrared thermography. After a brief effects are greatly reduced. In addition, the proposed method can be overview of heat transfer theory, active systems such as pulse infrared carried out in the compressive data domain, provided the image thermography, step hearing and lockin thermography will be coding is also based on the same wavelet decomposition. : presented. Passive applications will also be presented. Most of these application which involved low temperature acquisitions use The experimental results proved the performance of our method in commercial infrared cameras which are working in the IR range [2 -14 terms of Chinese videotext detection accuracy. ?m]. We will also present systems which were developed for high 6503-07, Session 2 temperature measurements to control process in the industry. These systems are based on either CMOS or standard CCD cameras. For Spectroscopic imaging from 400 nm to 1800 nm with each application, the radiometric model which is use to calibrate the liquid crystal tunable filters camera is carefully presented. H. Rothe, C. F. Hahlweg, Helmut-Schmidt Univ. (Germany) Problem Statement & Motivation: 6503-09, Session 3 In spectroscopic imaging it is desi-rable to get real images of scenes Integrating advanced 3D vision technology in instead of pushbroom images. This (i)avoids the need to compute the manufacturing “real” specroscopic image from the spectrum of a slit, and (ii) the need for scan-ning or a relative movement between camera and A. Pichler, H. Bauer, C. Eberst, C. Heindl, J. Minichberger, object. Profactor Produktionsforschungs GmbH (Austria) Methods: Today’s markets and economies are becoming increasingly volatile, For fluorescence microscopy liquid crystal tunable filters (LCTFs) are unpredictable, they are changing radically and even the innovation available in the spectral ranges from 400nm - 700nm, 700nm - speed is accelerating. Manufacturing and production technology and 1100nm and 850nm - 1800nm (CRI Inc). Therefore the multi-spectral systems must keep pace with this trend. To counter these challenging camera system from 400nm - 1800nm has to consist of three prospects, technologies enabling autonomous production systems channels. From 400nm - 1100nm CMOS-cameras can be used. In the are needed. The paper gives an overview of a novel innovative vision SWIR region other sensors have to be em-ployed, in this case we system embedded in manufacturing applications. The system is able used an InGaAs camera (XEVA). Since the transmission of the LCTFs to handle (1) unknown products, (2) in-available data, (3) large is wavelength dependend, a radiometric calibration is necessary and disturbances and deviations (shape, position, or even part type) and a high dynamic range of the cameras is desirable. From 400nm - (4) even very small lot-sizes. The concept foresees a full integration 1100nm cameras with a dynamic range of 22 bits have been into production technology. Products to be processed are past employed, while in the SWIR only 14 bits are available. Because the through a flexible 3D sensor cell. In principle, the 3D imaging process software controlled switching of the LCTFs takes some tens of is based on subsequent modules: (1) 3D data retrieval, (2) 3D milliseconds, a whole series of 140 images (10nm FWHM of one reconstruction and (3) 3D data interpretation. The proposed system channel of the LCTF) takes some tens of seconds, including the adapts its sensor system to arbitrary product geometry and builds a transfer time of the images from the cameras to the laptop computer. 3D world model of the task space using fast and robust 3D object Therefore only quaisi static scenes can be observed. However, this is recognition system. Having a flexible 3D imaging system paired with not a disadvantage, since the system will be used to find out automatic process planners the approach has proven to be a interesting spectral ranges for a particular applications. Since the promising alternative to conventional teach in and OLP based three cameras are not on one optical axis we have to deal also with programming and has shown feasible especially for high variant, low the disparity of the images. A good method therefore is cross volume parts. correlation. To be able to use the system in the field we added a hydrogen fuel cell as supply of electrical energy. 6503-10, Session 3 Experimental results The system was used to look a the multispectral behaviour High-performance surface inspection method for thin- (camoflage) of German and French uniforms. An automatic film sensors calssification of the type of uniforms was possible. The camoflage V. V. W. Wieser, S. Larndorfer, B. Moser, F. Kossak, Software paints of the German main battle tank Leoprad II and the naval attack Competence Ctr. Hagenberg (Austria) plane Tornado was caracterised. It was possible to perform image segmentation for ship silhouettes. Thin-film sensors for use in automotive or aeronautic applications must conform to very high quality standards. Due to defects that It was also possible to classify different liquids that are transparent cannot be addressed by conventional electronic measurements, an and colorless in the VIS and NIR according to their spectral charac- accurate optical inspection is imperative to ensure long-term quality teristics in the SWIR. aspects of the produced thin-film sensor. In this paper, a new method The system was used to chracterise the appearance of surfaces to is proposed that solves the problem of handling local deformations the eye and find out spectral flaws and differen-ces. due to production variability without involving the compensation of Conclusions local image registration operations. The main idea of this method is based on a combination of efficient morphological preprocessing and By using LCTFs it was possible to establish an imaging spectroscopic a multi-step comparison strategy based on logical implication. The system from 400nm - 1800nm with a channel width of 10nm. The

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 161 Conference 6503: Machine Vision Applications in Industrial Inspection XV

main advantage of this approach is that the neighborhood operations 6503-13, Session 5 that care for the robustness of the image comparison can be computed in advance and stored in a modified master image. By Evaluation of two and a half-dimensional surface data virtue of this approach, no further neighborhood operations have to with form component and groove bands be carried out on the acquired test image during inspection time. A systematic, experimental study shows that this method is superior to B. Xin, Univ. Karlsruhe (Germany) existing approaches concerning reliability, robustness, and Machine work pieces with ground, broached or milled surfaces have computational efficiency. The work is substantiated with theoretical frequently micro textures consisting of stochastically placed straight arguments and a comprehensive analysis of the obtained tool marks. In this paper we’ll exploit the depth data acquired by performance and practical usability in the above-mentioned white-light interferometer for the surface analysis. We present a new challenging industrial environment. algorithm for efficiently extracting extensive groove bands from depth images. The images are treated as a composition of shape component, straight line structures and background. The cylindrical 6503-11, Session 4 shape component is extracted using robust least squares methods. Toward an automation of 3D acquisition and post- The outliers are removed by adaptive center weighted median filter. The undefined regions due to sensor failures are interpolated using processing of cultural heritage artefact successive over-relaxation algorithm. An algorithm for the separation B. Loriot, R. Seulin, P. Gorria, Univ. de Bourgogne (France) of groove bands is introduced. Straight line in three-dimensional space is parameterized and used as primitive for groove separation. Historians and museums are very interested in 3D digitizing and After having determined the orientation of the groove group, straight duplication of cultural heritage artefacts. The aim is to provide models line segment produced with Digital Differential Analyser can be used and duplicata for historians to ease the manipulation and observation by the scanning algorithm for estimating the straight lines. This of very fragile and unique objects. decomposition enables a separate evaluation of different components To scan objects half-automatically, some algorithms solve the “next of the surface data. The results of the pre-processings and the best view” (NBV) problem: find the next position for the range scanner separation turn out to be fast and robust, which is verified by real given its previous scans of the object. There are no known general depth data. automated geometric modeling environments or performance- oriented view planning systems in industrial usage. Most of these algorithm do not take into account of the quality of the final model, 6503-14, Session 5 except for the density of points. Application of edge field analysis to a blurred image Proposed system allows to perform half-automatically digitizing and post-processing. The system uses modified MVC (Mass Vector M. Matsumoto, S. Hashimoto, Waseda Univ. (Japan) Chain)algorithm to determine the NBV to create a coarse model and This paper introduces a method for quasi-motion extraction from the holes remaining in the mesh to complete it. Then, each acquired blurred image utilizing edge field analysis (EFA). Exposing a film for a view is segmented (object-environnement), and cleaned: all certain time, we can directly photograph the trajectory of the moving abnormals faces are removed (angle of view, non-manifold, ...). object as an edge in a blurred image. As the edge trajectories are not Proposed system goes from the 3D acquistion to the generation of a exactly the same but similar to the optical flows, which allow us to 3D high-quality model. treat the edge image as a pseudo-vector field. We define three line integrals in the edge image on closed curve similar to vector analysis. These integrals correspond to three flow primitives of the scene such 6503-12, Session 4 as the translation, rotation and divergence. As the images, we utilized Appliance of stereoscopy for bath level detection of steel some images such as the storm, the bottle rocket and a moving casting pans object with random patterns. In order to evaluate the proposed method, we conducted the experiments of estimating the eye of the C. Rittenschober, K. S. Niel, Fachhochschule Wels (Austria); R. storm, the center of the explosion in terms of bottle rocket, and the Rössler, voestalpine Stahl AG (Austria) centers of the rotation and divergence of the moving object. Bath level detection of steel casting pans is a main issue within the automated steel manufacturing process. Different measuring systems 6503-15, Session 5 like micro wave, laser scanning have been implemented but without success due to the harsh environment caused by hot steel, slag, Camera calibrationless vision calibration for transfer chaotic movement of the bath surface and smoke. A stereoscopic robot system image acquisition system can manage these conditions. N. Kimura, T. Moriya, K. Matsumoto, Hitachi, Ltd. (Japan) With correspondence analysis and geometrical calculations it is possible to avoid complex perspective equalisation and to measure This research is focused on a system in which a manipulator with the bath level by using the pictures of the stereo cameras. Matlab robot vision transfers objects to a mobile robot which moves on a flat simulations show the weak points of the standard form of floor. In this system, the end effector of the manipulator operates on a stereoscopic analysis and the boundaries of the system. Next to plane surface, so only single vision is required. In a robot vision distortions of the camera axes the angle of incidence of the cameras system, many processes are usually needed for vision calibration, and has a great influence on the measurement failure. But there is a way to one of them is to measure camera parameters. Here we describe a calculate this angle automatically. calibration technique that does not need camera parameters explicitly Generally the investigations of the stereo system have shown which to reduce calibration processes for the researched system. specifications including certain tolerances have to be fulfilled by the In this technique, we measure the relation between the coordinate designer of the mechanical construction to get accurate and reliable systems of the image and mobile robot in the moving plane using a results for the automation process. Furthermore it is important to projective transformation framework. And we also measure the optimise the image acquisition, the pattern-matching parameters and relation between the image and the manipulator in the arbitrary plane to adjust the stereo system within the given tolerances too. If these in the same way. By analyzing the results, we can obtain the relation operations are correct, image processing with stereoscopic distance between the mobile robot and the manipulator without calculating the measurement can be a working system to detect a bath level. camera parameters explicitly. It means the procedure of capturing the images of the calibration board can be skipped. Lastly, we tested the calibration technique using an object-transfer task. The test showed the technique has sufficient accuracy to achieve the task.

162 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6503: Machine Vision Applications in Industrial Inspection XV

6503-16, Session 5 for a variety of applications. We present our latest results of research and development on: Subpixel evaluation of distorted circles characterisitcs * The artificial apposition compound eye, directly integrated with the F. Mairesse, T. M. Sliwa, S. Binczak, Y. Voisin, Univ. de Bourgogne CMOS-imager, improved resolution and field of view and blocking of (France) ghost images by an additional internal aperture array layer. The subject of this paper is the improvement of measures on * Hyperacuity of an artificial apposition compound eye for position- imperfect circular forms. Indeed, Simple geometric forms have been detection of point sources or edges with an accuracy far beyond the resolution limit of the compound eye with respect to bar targets. well studied in image processing. Thus, articles describing circles on a discrete framework are numerous but the case of imperfect geometric * Applying more than one pixel per channel in the artificial apposition forms, in return, is hardly ever deepen. However, it is a classical compound eye for improved sensitivity (neural superposition) or problem in industrial vision control process to not have a perfect, or colour vision without loss of resolution. perfectly discretized, geometric object due to, notably, fabrication * The rotating artificial apposition compound eye column for process, industrial environment (dust, vibrations, objects distortion-free 360° panoramic vision by a highly compact stick-like displacement, etc.), interferences on acquisition chain (electronic device. noise, lenses imperfections, etc. ). The authors present a comparison of measurement methods of circles characteristics subpixel * The artificial apposition compound eye on a curved basis by a estimation (center’s coordinates and radius) for several distortions special type of laser beam writer. (geometric or not). The estimators proposed are classic least mean * The microoptical telescope compound eye, which is providing a squares, 3D Hough algorithms and a method combining a Radon regular image of higher resolution due to the upright transfer of parts transform based estimator and a FitzHugh-Nagumo partial differential of the overall field of view by each channel. equation based active region algorithm. The originality of the method * The ultra-short, large object-size microscope by a unity-magnifying is to furnish a set of geometric envelopes in a single pass from a array of microlens telescopes. roughest to a full detailed representation. Moreover, this multiple active region principle also offers interesting electronic implementation possibilities for real time image processing for 6503-19, Session 6 metrology on production chains. High-performance camera for fast quality inspection in industrial printing application 6503-17, Session 5 J. Fürtler, E. Bodenstorfer, K. J. Mayer, J. Brodersen, Austrian Face recognition by using ring rotation invariant Research Ctrs. Seibersdorf Research GmbH (Austria); C. Eckel, K. transform and coping with image rotation and image Gravogl, Oregano Systems Design & Consulting GmbH (Austria); shifting problems H. Nachtnebel, Technische Univ. Wien (Austria); P. Roessler, Fachhochschule Technikum Wien (Austria) C. Su, Da Yeh Univ. (Taiwan) Today, printed matter which must meet highest quality standards, e.g., Generally, for the face image recognition, we must cope with the banknotes, stamps, vouchers, is automatically checked by optical image shift and image rotation problem. To cope with the image- inspection systems. Typically, the examination of fine details of the shifting problem, this research uses one pixel inside the sample image print or security features demands images taken from various to compare with the around pixels that surrounding the corresponding perspectives, with different spectral sensitivity (visible, infrared, pixel that inside the unknown image. The “ring rotation invariant ultraviolet), and with high resolution. Consequently, the inspection transform” technique is used to transfer the geometry feature of the system, equipped with several cameras, has to cope with an face image to another more salient feature. By this approaching one enormous data rate to be processed in real-time. Hence, it is can obtain more salient geometry feature of the face image. By this desirable to move image processing tasks into the camera to reduce more salient geometry feature, one can judge whether or not the the amount of data which has to be transferred to the (central) image sample image and the unknown image are the same image. The “ring processing system. The idea is to transfer relevant information only, rotation invariant transform” technique can solve the image rotation i.e., features of the image instead of the image itself. These features problem. In this research, three different kinds of extracted ring are then further processed. This paper presents a line scan camera signals are generated. The extracted ring signals are generated by the which is based on a commercial CMOS area image sensor and a field following ring-circles - ring-radius-31-circle, ring-radius-22-circle, and programmable gate array (FPGA). The camera implements extraction ring-radius-13-circle. These extracted ring-signals are used to of image features, which are used to detect print flaws like blotches of generate the rotation invariant vector magnitude quantities. These ink, smears, and scratches. Several units implemented on the FPGA rotation invariant vector magnitude quantities are combined as one are described in detail. Image processing methods implemented in entity and this entity is saved inside one specific corresponding pixel the camera include flat field correction, color space conversion, as in the BMP file. By this approach, one pixel will possess more well as decimation and neighborhood operation. In addition, an geometry-features of the face images. The obtained entity of the application of the camera in multi-sensor image processing systems combined signals of one specific pixel inside the sample image will be is addressed. compared to the entities of the combined signals of the entire pixels located in the corresponding radius-6-cake in the unknown image. By this comparison, one can find the most-matching point of the 6503-20, Session 6 geometry-feature of the pixels between the sample image and the unknown image. Setting up task-optimal illumination automatically for inspection purposes 6503-18, Session 6 J. J. Uusitalo, R. O. Tuokko, Tampere Univ. of Technology (Finland) Latest developments in microoptical artificial compound eyes: a promising approach for next generation Illumination with correct intensity and direction can help solving even very complicated inspection and measurement problems by hiding ultracompact machine vision unnecessary information and enhancing the contrast where it is J. Duparré, P. Dannberg, A. H. Bräuer, Fraunhofer-Institut für needed. However, achieving the optimal illumination solution can be Angewandte Optik und Feinmechanik (Germany) complicated and the widely adopted ‘wave and look’ approach requires wide-ranging knowledge about illumination techniques and Recent micro-optical fabrication technologies allow for a highly the measurement problem. Worse yet, in small-batch production precise generation of ultra-thin monolithic compound eye objectives where the inspected end-product changes often, the traditional

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 163 Conference 6503: Machine Vision Applications in Industrial Inspection XV

illumination solution is too inflexible and expensive. In this paper we 6503-23, Poster Session describe approaches for producing flexible light sources for machine vision purposes. The described devices allow one to change the Real-time vehicle detection and tracking based on traffic scene illumination by only updating software parameters. Often only scene analysis one programmable light source is needed per one vision system despite very different end-products or features are to be inspected. Z. Zeng, S. Wang, X. Ding, Tsinghua Univ. (China) We also describe an automated tuning approach that controls In this paper, we propose a real-time vehicle detection and tracking programmable light sources to correct the image seen by the sensor algorithm based on traffic scene analysis for driving assistant use. The to produce a desired illumination on the image. The approach can contribution we make mainly includes three parts: at first we describe also be used for maximising the illumination uniformness, contrasts, a general traffic scene analysis framework for vehicle detection and or even the part measurement robustness. The devices and tracking based on roadside detection. Through the analysis of traffic approaches described in the paper ensure a better diffusion of vision scene, we cut down the computing cost greatly to make the whole sensors in the industry as they become easier to be taken into use. detection and tracking process real-time; secondly a new object detection algorithm via fusion of global classifier and part-based classifier is presented to detection vehicles in region of interest; finally, 6503-21, Poster Session we propose an algorithm integrating classifying confidence and local shadow to track vehicles detected from interesting region. The Imaging based logics for ornamental stones quality chart integration makes the tracking algorithm not only robust to definition environment condition, but also adaptive to view change of the G. Bonifazi, A. Gargiulo, S. Serranti, Univ. degli Studi di Roma/La vehicle. Sapienza (Italy) We have tested our algorithm on several video sequences captured Ornamental stone products are commercially classified on the market from highways and suburban roads under different background according to several factors related both to intrinsic lythologic environments, weather types and light conditions. It successfully characteristics and on the base of their visible pictorial attributes. detect and track all passing vehicles with no false alarm, while the Sometimes these latter aspects prevail in quality criteria definition and time cost is only 47ms per frame (640x480 sized). The test results assessment. Pictorial attributes are in any case also influenced by show good time efficiency and high robustness of the proposed working actions performed and utilized tools to realize the final stone algorithm when coping with environment change, illumination variation manufactured product. This latter spect is quite critical because stone and vehicle view transition. surface finishing can contribute to enhance, or not, certain aesthetic features of the stone itself. The study was addressed to develop an innovative set of methodologies and techniques able to quantify the 6503-24, Poster Session aesthetic quality level of stone products taking into account both the 2DPCA face recognition based on block scan physical and the aesthetical characteristics of the stones. C. Zhang, Sun Yat-Sen Univ. (China) A new technique called two-dimensional principal component analysis 6503-22, Poster Session (2DPCA) was proposed for human face representation and Design and implementation of confocal imaging systems recognition. 2DPCA is computationally more efficient than PCA because of avoiding high-dimensional vector space. Image with a generalized theoretical framework covariance matrix is constructed directly using the original image G. Chang, C. Liao, National Taiwan Normal Univ. (Taiwan) matrices, there is no need to transform 2D image into 1D vector. The essence of 2DPCA had been examined through analysis: 2DPCA is Confocal imaging is primarily based on the use of apertures in the equivalent to a special case of block-based PCA, i.e. each block is a detection path to provide the acquired three-dimensional images with row of the images. Therefore, the correlation information among the satisfactory contrast and resolution. For many years, it has become column vectors of the images is lost. To overcome this problem, we an important mode of imaging in microscopy. In biotechnology and proposed a new scheme called block-scanned 2DPCA. In the new related industries, this technique has powerful abilities of biomedical scheme, image block was introduced instead of image row, so the inspection and material detection with high spatial resolution, and correlation information in both rows and columns of the images are furthermore it can combine with fluorescence microscopy to get more considered. Thus, the image covariance matrix can be more useful information. The objective of this paper is first to present a accurately evaluated and the recognition rate will be better. To test generalized theoretical framework for confocal imaging systems, and and evaluate the performance of block-scanned 2DPCA, experiments then efficiently to design and implement such systems with were performed on the FERET database. The recognition rate was satisfactory imaging resolutions. In our approach, a theoretical review higher using block-scanned 2DPCA than standard 2DPCA. The for confocal imaging is given to investigate this technique from theory experimental results also indicated that less number of principal to practice. Also, computer simulations are performed to analyze the components were need using block-scanned 2DPCA than standard imaging performance with varying optomechanical conditions. For 2DPCA. instance, the effects of stray light on the microscopic systems are examined using the simulations. In this paper, a modified optomechanical structure for the imaging process is proposed to 6503-26, Poster Session reduce the undesired effects. From the simulation results, it appears that the modified structure highly improves the system signal-to-noise Machine vision system for the inspection of relective ratio. Furthermore, the imaging resolution is improved through the parts in the automotive industry investigation on the tolerance of fabrication and assembly of the optical components. In the experiments, it is found that the imaging F. Meriaudeau, O. Morel, G. Salis, Univ. de Bourgogne (France) resolution of the proposed system is less sensitive than that of Specular surfaces inspection remains a delicate task within the common confocal microscopes, to the position deviations arising automatic control of products made by plastic plating. These objects from installations of the optical components, such as those from the are of very varied shape and their surface is highly reflective acting pinhole and the objective lens. like a mirror. This paper presents steps to follow in order to detect geometric aspect surface defects on objects made by plastic plating. The projection of a binary fringes pattern is used and enables to reveal the defects near the transition between a dark fringe and a bright fringe. Indeed, the surface imperfections provoke important light ray deviations. By moving this dynamic lighting, and thanks to a saturated camera, the system brings an aspect image where the defects appear very contrasted on a dark background. A simple

164 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6503: Machine Vision Applications in Industrial Inspection XV image processing algorithm is then applied leading to a very efficient segmentation. To obtain such resulting images, the translation step, the duty cycle and also the number of images are constraint. This article finally shows how to adjust these parameters according to the various sizes of defect and to the objects shape.

6503-27, Poster Session Validation of a site-finding algorithm for NIF optic scattered-light images M. Reed, Harvey Mudd College; L. M. Kegelmeyer, J. A. Liebman, M. Bolourchi, J. N. McElroy, R. R. Prasad, Lawrence Livermore National Lab. The Conditioning, Initiation and Mitigation Facility at Lawrence Livermore National Laboratory’s National Ignition Facility processes switch crystals before they are installed on the beam lines. Optics are conditioned by raster-scanning them with an ultraviolet laser, which increases their resistance to laser-induced bulk damage. The few surface sites that occur during the process are counted and micrographed to verify that they do not exceed size criteria. Currently, operators manually examine each optic and pilot the microscope to each site of interest. Software has been developed to automatically locate sites from a full-optic image and calculate site statistics such as scattering intensity. Another tool was developed to calculate the effective diameter of sites from micrographs. The relationship between site intensity and effective diameter was investigated. Sites were manually classified by morphology and if necessary excluded from the analysis. A loose linear relationship was found by analyzing remaining sites. It will be used to automatically select sites that must be micrographed and produce statistics documenting site creation and growth. Issues with pixel saturation and site morphology currently limit the precision of this relationship but may be ameliorated in future. This automation promises to reduce optic processing time by up to 40%.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 165 Conference 6504: Multimedia Computing and Networking 2007 Wednesday-Thursday 31 January-1 February 2007 Part of Proceedings of SPIE Vol. 6504 Multimedia Computing and Networking 2007

6504-01, Session 1 networks with fluctuating bandwidths. The Base Layer of FGS is often encoded at very low bitrates, resulting in unacceptable deterioration Optimal bit allocation for fine-grained scalable video of visual quality. In this paper, we propose a novel technique, termed M. M. Hefeeda, C. Hsu, Simon Fraser Univ. (Canada) as FMOE-MR, which delivers significantly improved rate distortion performance compared to existing MPEG-4/H.264 Base Layer We present optimal schemes for allocating bits of fine-grained scalable encoding techniques. The video frames are re-encoded in multiple- video sequences among multiple senders streaming to a single resolutions, keeping semantically and visually important regions at a receiver. This allocation problem is critical in optimizing the perceived higher resolution, such that further MPEG-4/H.264 compression leads quality in peer-to-peer and distributed multi-server streaming to improved rate-distortion performance. FMOE-MR is an integrated environments. Senders in such environments are heterogeneous in their approach that requires only encoder-side modifications, and is outgoing bandwidth and they hold different portions of the video transparent to the decoder. Further, since the FMOE-MR scheme stream. We formulate the allocation problem as an optimization incorporates “smart” video preprocessing, it requires no change in problem, which is nonlinear in general. We use rate-distortion models in existing MPEG-4 codecs. As a result, it is straightforward to use the the formulation to achieve the minimum distortion in the rendered proposed FMOE-MR scheme with any existing MPEG codec, thus video, constrained by the outgoing bandwidth of senders, availability of allowing great flexibility in implementation. In this paper, we have video data at senders, and incoming bandwidth of receiver. We show described, and implemented, unsupervised and semi-supervised how the adopted rate-distortion models transform the nonlinear algorithms to create the FMO-Mask from a given video sequence, problem to an integer linear programming (ILP) problem. We then which is used to create the content aware multi-resolution re- design a simple rounding scheme that transforms the ILP problem to a encoding of the video frames, using state-of-the-art computer vision linear programming (LP) one, which can be solved efficiently using algorithms. common optimization techniques such as the Simplex method. We prove that our rounding scheme always produces a feasible solution, and the solution is within a negligible margin from the optimal solution. 6504-04, Session 1 We also propose a new algorithm (FGSAssign) for the allocation problem that runs in O(n log n) steps, where n is the number of senders. Impact of reference distance for motion compensation We prove that FGSAssign is optimal. Because of its short running time, prediction on video quality FGSAssign can be used in real time during the streaming session. Our Y. Wang, EMC Corp.; M. Claypool, R. Kinicki, Worcester experimental study validates our analytical analysis and shows the Polytechnic Institute effectiveness of our allocation algorithm in improving the video quality. Transmitting high-quality, real-time interactive video over lossy networks is challenging because data loss due to the network can 6504-02, Session 1 severely degrade video quality. A promising feedback technique for low-latency video repair is Reference Picture Selection (RPS), A framework for MPEG-21 DIA-based adaptation and whereby the encoder selects one of several previous frames as a perceptual encryption of H.264 video reference frame for predictive encoding of subsequent frames. RPS R. Iqbal, S. Shirmohammadi, A. El Saddik, Univ. of Ottawa can operate in two different modes: an optimistic policy that uses (Canada) negative acknowledgements (NACKs) and a more conservative policy that relies upon positive acknowledgements (ACKs). The choice A ubiquitous computing concept permits end users to have access to between RPS ACK mode and NACK mode to some extent depends multimedia and digital content anywhere, anytime and in any way they upon the effects of reference distance on the encoded video quality. want. As a consequence, the importance of resource customization This paper provides a systematic study of the effects of reference according to user preferences and device requirements set the primary distance on video quality for a range of video coding conditions. High- challenge towards seamless access. Moreover, once a suitable quality videos with a wide variety of scene complexity and motion customization approach has been decided (e.g. adaptation), deploying characteristics are selected and encoded using H.264 with a it in the existing network requires a generic and widely accepted bandwidth constraint and a range of reference distances. Two standard applied to the process. With the advancement of time, objective measures of video quality, PSNR and VQM, are analyzed to performing encryption in the compressed domain should also be taken show that scene complexity and motion characteristics of the video care of not only for serving sensitive digital contents but also for determine the amount of degradation in quality as reference distance offering security as an embedded feature of the adaptation practice to increases. In particular, videos with low motion degrade in quality ensure digital right management and confidentiality. In this paper, we more with an increase in reference distance since they cannot take present an architecture for temporal adaptation of ITU-T H.264 video advantage of the strong similarity between adjacent frames. Videos conforming to ISO/IEC MPEG-21 DIA. In addition, we present a with high motion do not suffer as much with higher reference distance perceptual encryption scheme that is integrated in the system for video since the similarity between adjacent frames is already low. The encryption. The framework enables video bitstreams to be adapted and motion characteristics also determine the initial quality under the encrypted in the compressed domain, eliminating cascaded adaptation bandwidth constraint. The data presented should be useful for (i.e. decoding - adaptation - encoding). The encryption framework is selecting ACK or NACK mode or for modeling video repair applied on the adapted video content, which reduces computational techniques. overhead compared to that on the original content. A prototype, based on the proposed architecture and experimental evaluations of the system as well as its performance supporting the architecture are also 6504-05, Session 1 presented. Perceptual multimedia compression based on the predictive Kalman filter eye movement modeling 6504-03, Session 1 O. V. Komogortsev, J. I. Khan, Kent State Univ. FMOE-MR: content-driven multiresolution MPEG-4 fine- In this paper we propose an algorithm for predicting a person’s grained scalable layered video encoding perceptual attention focus (PAtF) through the use of a Kalman Filter S. Chattopadhyay, X. Luo, S. M. Bhandarkar, K. Li, The Univ. of design of the human visual system. The concept of the PAtF allows significant reduction of the bandwidth of a video stream and Georgia computational burden reduction in the case of 3D media creation and The MPEG/H.264 Fine Grained Scalability (FGS) profile aims at scalable transmission. This is possible due to the fact that the human visual layered video encoding, in order to ensure efficient video streaming in system has limited perception capabilities and only 2 degrees out of

166 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6504: Multimedia Computing and Networking 2007 the total of 180 provide the highest quality of perception. The the hierarchy. We also characterize content access patterns and their peripheral image quality can be decreased without a viewer noticing influence on system performance, as well as quality experienced by image quality reduction. Multimedia transmission through a network users and user behavior. In our investigation, scalability is one of the introduces a delay. This delay reduces the benefits of using a PAtF most important aspects of the system performance evaluation. due to the fact that the person’s attention area can change drastically Although the range of the prototype operation is limited, as far as the during the delay period, thus increasing the probability of peripheral number of users and the content repository is concerned, we believe image quality reduction being detected. We have created a framework that data collected from such a system provides a valuable insight which uses a Kalman Filter to predict future PAtFs in order to into efficiency of a CDN-type of solution to large scale streaming compensate for the delay/lag and to reduce the bandwidth/creation services. We find that the systems exhibits good performance and low burden of any visual multimedia. resource usage.

6504-06, Session 2 6504-09, Session 3 Share with thy neighbors Olfactory enhanced multimedia applications: S. Chandra, X. Yu, Univ. of Notre Dame perspectives from an empirical study Peer to peer (P2P) systems are traditionally designed to scale to a O. Ademoye, G. Ghinea, Brunel Univ. (United Kingdom) large number of nodes. However, we focus on scenarios where the The increasing demand for multi-modal systems and applications that sharing is effected only among neighbors. Localized sharing is are highly interactive and multi-sensory in nature has led to the particularly attractive in scenarios where wide area network introduction of new media and new user interface devices in connectivity is undesirable, expensive or unavailable. On the other multimedia computing. Computer generated smell, also known as hand, local neighbors may not offer the wide variety of objects olfactory data, is one of such media objects currently generating a lot possible in a much larger system. The goal of this paper is to of interest in the multimedia industry. We are currently focusing our investigate a P2P system that shares contents with its neighbors. We attention on exploring user perception of computer generated smell analyze the sharing behavior of Apple iTunes users in an University when combined with other media to enrich their multimedia setting. iTunes restricts the sharing of audio and video objects to experience. In this paper, we present the results of an empirical study peers within the same LAN sub-network. We show that users are into users’ perception of olfactory enhanced multimedia displays. already making a significant amount of content available for local Results showed that users generally enjoy an enhanced multimedia sharing. We show that these systems are not appropriate for experience when augmented by olfactory stimuli, and that the applications that require access to a specific object. We argue that presence of such stimuli increases the sense of relevance. Whilst mechanisms that allow the user to specify classes of interesting there is a general positive bias towards olfactory enhanced objects are better suited for these systems. Mechanisms such as multimedia applications, specific properties of smell such as diffusion bloom filters can allow each peer to summarize the contents available and lingering mean that in practice specific attention needs to be in the neighborhood, reducing network search overhead. This given when a mix of smells is associated with visualised multimedia research can form the basis for future storage systems that utilize the content; moreover, it was found that whilst smell was incorrectly shared storage available in neighbors and build a probabilistic storage identified in some instances, the presence of smell per se is generally for local consumption. enough to create a richer user multimedia experience.

6504-07, Session 2 6504-10, Session 3 Video to Go: the effects of mobility on streaming media The effects of resolution on users playing first person in a CDMA2000 1xEV-DO network shooter games P. Sessini, M. Leventer, A. Mahanti, Univ. of Calgary (Canada) K. Claypool, Oracle USA; M. Claypool, Worcester Polytechnic This paper experimentally examines the performance of streaming Institute media applications over a CDMA2000 1xEV-DO network. The Computer games are often played on devices with varying display performance of streaming in a cellular network is tested across three resolutions. While higher resolutions generally provide more immersive different levels of mobility, two applications, and the two transport game play they can yield reduced frame rates and/or increased costs, layer protocols, TCP and UDP. Findings of this study are that making choosing the optimal resolution important. Despite this streaming applications are impacted more by sources of interference importance, to the best of our knowledge, there has been no such as high-rise buildings than by increased velocity. Also, when the extensive study of the effects of resolution on users playing computer mobile client is stationary, high data rates and high video quality are games. This paper presents results from extensive user studies consistently achieved. We also find that for the streaming applications measuring the impact of resolution on users playing First Person considered, UDP streams outperform TCP streams, consistently Shooter games. The studies focus on the effects of resolution in achieving higher bandwidth. conjunction with low and high contrast virtual environments, full screen and windowed modes and identification of long-range objects. Analysis indicates resolution has little impact on performance over the 6504-08, Session 2 range of conditions tested and only matters when the objects being iTVP: large-scale content distribution for live and on- identified are far away or small and are reduced to too few pixels to be distinguishable. demand video services E. Kusmierek, M. Czyrnek, C. Mazurek, M. Stroinski, Poznan Supercomputing and Networking Ctr. (Poland) 6504-11, Session 4 iTVP is a system built for IP-based delivery of live TV programming, Efficient cost-based scheduling for scalable media video-on-demand and audio-on-demand with interactive access over streaming IP networks. It has a country-wide range and is designed to provide service to a high number of concurrent users. iTVP prototype contains N. J. Sarhan, B. Qudah, Wayne State Univ. the backbone of a two-level hierarchical system designed for The number of media streams that can be supported concurrently is distribution of multimedia content from a content provider to end highly constrained by the stringent requirements of real-time playback users. In this paper we present experience gained during a few and high transfer rates. To address this problem, media delivery months of the prototype operation. We analyze efficiency of iTVP techniques, such as Batching and Stream Merging, utilize the content distribution system and resource usage at various levels of multicast facility to increase resource sharing. The achieved resource

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 167 Conference 6504: Multimedia Computing and Networking 2007

sharing depends greatly on how the waiting requests are scheduled 6504-14, Session 5 for service. Scheduling has been studied extensively when Batching is applied, but up to our knowledge, it has not been investigated in the On providing reliability guarantees in live video streaming context of stream merging techniques, which achieve much better with collaborative clients resource sharing. In this study, we analyze scheduling when stream merging is employed and propose a simple, yet highly effective A. Raghuveer, Univ. of Minnesota; Y. Dong, Univ. of Hawaii; D. H. scheduling policy, called Minimum Cost First (MCF). MCF exploits the Du, Univ. of Minnesota wide variation in stream lengths by favoring the requests that require Although many overlay and P2P approaches have been proposed to the least cost. We present two alternative implementations of MCF: assist large-scale live video streaming, how to ensure service quality MCF-T and MCF-P. We compare various scheduling policies through and reliability still remains a challenging issue. Peer dynamics, extensive simulation and show that MCF achieves significant especially unscheduled node departures, affect the perceived video performance benefits in terms of both the number of requests that quality at a peer node in two ways. In particular, the amplitude of can be serviced concurrently and the average waiting time for service. quality fluctuations and the duration for which stable quality video is available at a node heavily depend on the nature of peer departures in the system. In this paper, we first propose a service quality model to 6504-12, Session 4 quantify the quality and stability of a video stream in a P2P streaming Resource management in a shared infrastructure video environment. Based on this model, we further develop tree construction algorithms that ensure that every peer in the CDN collaborative network receives a video stream with a statistical A. J. Cahill, C. J. Sreenan, National Univ. of Ireland/Cork (Ireland) reliability guarantee on quality. A key advantage of the proposed approach is that we can now explicitly control the quality and stability In any large scale distribution architecture, considerable thought of the video stream supplied to every node. Also, the proposed tree needs to be given to resource management, particularly in the case of construction schemes decide the position of a node in the delivery high quality TV on-demand. This work presents a globally accessible tree based on both its estimated reliability and upstream bandwidth network storage architecture operating over a shared infrastructure, contribution while striving to minimize the overall load on the server. termed Video Content Distribution Network (VCDN). The goal of which is to store all TV content broadcast over a period of time within the network and make it available to clients in an on-demand fashion. 6504-15, Session 5 This paper evaluates a number of content placement approaches in terms of their ability to efficiently manage system resources. Due to Priority-progress CPU adaptation for elastic real-time the dynamic viewing patterns associated with TV viewing, the applications effectiveness of content placement is expected to change over time, and so too should the content placement. The placement of content C. C. Krasic, The Univ. of British Columbia (Canada) within such a system is the single most influential factor in resource As multimedia-capable, network-enabled devices become ever more usage. As well as manage system resources a good placement abundant, device heterogeneity and resource sharing dynamics algorithm should also be able to converge on a placement solution remain difficult challenges in networked continuous media quickly. Work examines a number of placement algorithms, each with applications. These challenges often cause the applications to exhibit different properties, such as minimizing delivery path. This work also very brittle real-time performance. Due to heterogeneity, minor outlines the large number of variables in such a system, and examines variations in encoding can mean a continuous media item performs their impact on each algorithms performance. well on some devices but very poorly on others. Resource sharing can mean that content can work for some of the time, but real-time delivery is frequently interrupted due to competition for resources. 6504-13, Session 4 Quality-adaptive approaches seek to preserve real-time performance, Flash on disk for low-power multimedia computing by evaluating and executing trade-offs between the quality of application results and the resources required and available to L. C. Singleton IV, R. Nathuji, K. Schwan, Georgia Institute of produce them. Since the approach requires the applications to adapt Technology the results they produce, we refer to them as elastic real-time applications. In this paper, we use video as a specific example of an Mobile multimedia computers require large amounts of data storage, elastic real-time application. yet must consume low power in order to prolong battery life. Solid- state storage offers low power consumption, but its capacity is an order of magnitude smaller than the hard disks needed for high- 6504-16, Session 6 resolution photos and digital video. In order to create a device with the space of a hard drive, yet the low power consumption of solid- Web video search and the emergence of socially state storage, hardware manufacturers have proposed using flash immersive media memory as a write buffer on mobile systems. This paper evaluates the power savings of such an approach and also considers other possible R. Sarukkai, Yahoo! Inc. flash allocation algorithms, using both hardware- and software-level With the rapid emergence of viral Web 2.0 Video sites, the prominence flash management. Its contributions also include a set of typical of video on the web has grown significantly over the last year. multimedia-rich workloads for mobile systems and power models based upon current disk and flash technology. Based on these By leveraging the burgeoning social networks for rapid, viral workloads, we demonstrate an average power savings of 267 mW distribution of video, the web has dramatically altered the way video is (53% of disk power) using hardware-only approaches. Next, we being published, viewed and shared. In this talk, key trends in social propose another algorithm, termed Energy-efficient Virtual Storage networks and online video search are covered in addition to using Application-Level Framing (EVS-ALF), which uses both highlighting critical technologies/tools that can continue accelerating hardware and software for power management. By collecting the growth of this area. information from the applications and using this metadata to perform The talk will also cover how this new world of video is shaping up, intelligent flash allocation and prefetching, EVS-ALF achieves an what emerging technologies are critical, and what hurdles need to be average power savings of 307 mW (61%), another 8% improvement overcome? over hardware-only techniques.

168 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6504: Multimedia Computing and Networking 2007

6504-17, Session 7 this paper examines the potential collaborations among nodes across multiple channels. Our investigation shows that collaboration among CompatPM: enabling energy efficient multimedia nodes across different channels can improve the overall performance workloads for distributed mobile platforms of the multi-channel P2P streaming system. However, the collaboration strategies need to be carefully selected. Simple R. Nathuji, K. J. O’Hara, K. Schwan, T. Balch, Georgia Institute of collaboration strategies, such as treating collaborative nodes (those Technology “borrowed” from other channels) the same as a channel’s native The computation and communication abilities of modern platforms nodes (those playing the channel), tend to have marginal or even are enabling increasingly capable cooperative distributed mobile negative effects on the whole system performance. This result is systems. An example is distributed multimedia processing of sensor contrary to common impression—the larger population the better data in robots deployed for search and rescue, where a system performance of P2P system—and we found that this is caused by the manager can exploit the application’s cooperative nature to differences between P2P streaming and traditional P2P file-sharing successfully accomplish the mission. systems. Furthermore, this paper proposes a set of simple strategies that controls the upload-download ratio of collaborative nodes. We Because of limited battery capacities, a critical task a manager must showed that this set of strategies produces a much better perform is online energy management. While support for power collaboration result for multi-channel P2P streaming systems. management has become common for the components that populate Although only a preliminary study, we believe the results will promote mobile platforms, what is lacking is integration and coordination further investigation on the topic of multi-channel P2P streaming. across the management actions performed in a variety of system layers. This papers develops an integration approach for distributed multimedia applications, where a manager specifies both an operating 6504-20, Session 7 point and a workload for a node to execute. When jointly considering power and QoS, experimental evaluations show that using a simple Resource estimation methodology for multimedia deadline-driven approach to assigning frequencies can be non- applications optimal. These trends are further affected by certain characteristics of underlying power management mechanisms, which in our research, H. Kalva, R. Shankar, T. Patel, C. Cruz, Florida Atlantic Univ. are identified as groupings that classify component power Reducing the product development cycle time is one of the most management as “compatible” (VFC) or “incompatible” (VFI) with important and challenging problems faced by the industry today. voltage and frequency scaling. We build on these findings to develop Developing products rapidly in the face of this increasing complexity CompatPM, a vertically integrated control strategy for power requires new methodologies and tools. This paper presents a management in distributed mobile systems. methodology for estimating the resources consumed by a video decoder. The proposed methodology enables resource estimation based on high level user requirements. Component architecture for a 6504-18, Session 7 H.264 video decoder is developed to enable design space Efficient MPEG-21-based adaptation decision taking for exploration. The resources required to decode H.264 video are estimated based on a measure of the complexity of the H.264 scalable multimedia content bitstreams and the target architecture. The proposed approach is I. Kofler, C. Timmerer, H. Hellwagner, Univ. Klagenfurt (Austria); A. based on the hypothesis that the complexity of a H.264 video Hutter, F. Sanahuja, Siemens AG (Germany) bitstream significantly influences resource consumption and the complexity of a bitstream can thus be used to determine resource The MPEG-21 standard defines a framework for the interoperable estimation. The bitstream complexity is characterized to capture the delivery and consumption of multimedia content. Within this data dependencies using a process called Bitstream Abstraction. The framework the adaptation of content plays a vital role in order to decoder is componentized and component level resource support a variety of terminals and heterogeneous access networks. In requirements determined in a process called Decoder Abstraction. most cases the multimedia content can be adapted by applying The proposed methodology enables high level resource estimation for different adaptation operations that result in certain characteristics of multimedia applications without a need for extensive and time the content. Therefore, a component has to decide which adaptation consuming simulations. operations have to be performed to achieve a satisfactory result. This process is known as adaptation decision-taking and makes extensive use of metadata describing the possible adaptation operations, the 6504-21, Session 7 usage environment of the consumer, and constraints concerning the adaptation. Based on this metadata an optimization problem can be Multimodal event streams for virtual reality formulated and its solution yields the optimal parameters for the J. von Spiczak, Brigham and Women’s Hospital and Univ. adaptation operations. The metadata is represented in XML resulting in a verbose and inefficient encoding. In this paper, an architecture for Karlsruhe (Germany); E. Samset, Brigham and Women’s Hospital an Adaptation Decision-Taking Engine (ADTE) is introduced, which and Univ. of Oslo (Norway) and Harvard Medical School; S. P. operates both on XML metadata and on metadata encoded with DiMaio, Brigham and Women’s Hospital and Harvard Medical MPEG’s Binary Format for Metadata (BiM). This enables an efficient School; G. Reitmayr, Univ. of Cambridge (United Kingdom); D. metadata processing by separating the problem extraction from the Schmalstieg, Technische Univ. Graz (Austria); C. R. Burghart, Univ. actual optimization step. Furthermore, several optimization algorithms Karlsruhe (Germany); R. Kikinis, Brigham and Women’s Hospital which are suitable for scalable multimedia formats are reviewed. and Harvard Medical School Applications in the fields of virtual and augmented reality as well as 6504-19, Session 7 image-guided medical applications make use of a wide variety of hardware devices. Existing frameworks for interconnecting low-level Cross-channel collaborations in peer-to-peer streaming devices and high-level application programs include high-frequency event streaming frameworks as well as dynamically typed message H. Li, K. Li, The Univ. of Georgia passing systems. Neither of these approaches exploits the full Peer-to-Peer (P2P) streaming has become a very popular technique to potential for processing events coming from arbitrary sources and realize live media broadcast over the Internet. Most previous research neither is easily generalizable. In this paper, we will introduce a new of P2P streaming focuses on the delivery of a single media stream multi-modal event processing methodology that uses dynamically- (called a channel). The widely deployed implementations, however, all typed event attributes for event passing between multiple devices and concurrently offer multiple channels through their P2P networks. This systems. The existing OpenTracker framework was modified to paper investigates the overlay organization for multi-channel P2P incorporate a highly flexible and extensible event model, which can streaming systems through modeling and simulations. In particular, store data that is dynamically created and arbitrarily typed at runtime. The software architecture that is introduced in this work provides

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 169 Conference 6504: Multimedia Computing and Networking 2007 multi-modal event streaming, network and file logging support, and through given application-specific service and server topologies, and the ability to merge separate event streams. The main factors that the data being captured and distributed is manipulated by third impacting the library’s throughput were determined experimentally party service codes. CameraCast is a simple, easily used system-level and the overall performance was shown to be sufficient for most solution to remote video access. It provides a logical device API so typical applications. Several sample applications were developed to that an application can identically operate on local vs. remote video take advantage of the new dynamic event model that is provided by sensor devices, using its own service and server topologies. In the library, thereby demonstrating its flexibility and expressive power. addition, the application can take advantage of API enhancements to protect remote video information, using a capability-based model for differential data protection that offers fine grain control over the 6504-22, Session 7 information made available to specific codes or machines, thereby limiting their ability to violate privacy or security constraints. A multichannel multi-encoding transmission scheme for Experimental evaluations of CameraCast show that the performance wireless video streaming of accessing remote video information approximates that of accesses A. Kolekar, W. Feng, Portland State Univ.; M. Venkatachalam, Intel to local devices, given sufficient networking resources. High Corp. performance is also attained when protection restrictions are enforced. The wireless industry has seen a surge of interest in upcoming broadband wireless access (BWA) networks, like WiMAX that are based on orthogonal frequency division multiplexing (OFDM). These 6504-25, Session 8 wireless access technologies have several key features such as centralized scheduling, fine-grained allocation of transmission slots, Plasma: a scripting language for processing media adapting the modulation and coding schemes (MCS) to the SNR streams variations of the wireless channel, flexible and connection oriented T. Zhu, P. Korshunov, B. Liu, W. Ooi, National Univ. of Singapore MAC layer as well as QoS awareness and differentiation for (Singapore) applications. As a result, such architectures provide new opportunities for cross-layer optimization, particularly for applications that can Media streaming has found applications in many domains such as tolerate some bit errors. In this paper, we describe a multi-channel education, entertainment, communication and video surveillance. video streaming protocol for video streaming over such networks. In Many of these applications require non-trivial manipulations of media addition, we propose a new combined channel coding and streams, beyond the usual capture/playback operations supported by proportional share allocation scheme for multicast video distribution typical multimedia software and tools. To support rapid development based upon a video’s popularity. Our results show that we can more of such applications, we have designed and implemented a scripting efficiently allocate network bandwidth while providing high quality language called Plasma. Plasma treats media streams as first-class video to the application. objects, and caters to the characteristic differences between stored media files and live media streams. In this paper, we illustrate the design and features of Plasma through several small examples, and 6504-23, Session 7 describe two example applications that we developed on top of Plasma. These two applications demonstrate that using Plasma, Fine granularity adaptive multireceiver video streaming complex applications that compose, mix, and filter multimedia V. S. W. Eide, Simula Research Lab. (Norway); F. Eliassen, J. A. streams can be written with relatively little effort. Michaelsen, F. Jensen, Univ. of Oslo (Norway) Efficient delivery of video data over computer networks has been 6504-26, Session 8 studied extensively for decades. Still, multi-receiver video delivery is challenging, due to heterogeneity and variability in network availability, SenseTK: a multimodal multimedia sensor networking end node capabilities, and receiver preferences. Our earlier work has toolkit shown that content-based networking is a viable technology for fine P. Sitbon, Portland State Univ. granularity multi-receiver video streaming. By exploiting this technology, we have demonstrated that each video receiver is This paper describes the design and implementation of a multi-modal, provided with fine grained and independent selectivity along the multimedia capable sensor networking framework called SenseTK. different video quality dimensions region of interest, signal to noise SenseTK allows application writers to easily construct multi-modal, ratio for the luminance and the chrominance planes, and temporal multimedia sensor networks that include both traditional scalar-based resolution. Here we propose a novel adaptation scheme combining sensors as well as sensors capable of recording sound and video. The such video streaming with state-of-the-art techniques from the field of distinguishing features of such systems include the need to push adaptation to provide receiver-driven multi-dimensional adaptive application processing deep within the sensor network, the need to video streaming. The scheme allows each client to individually adapt bridge extremely low power and low computation devices, and the the quality of the received video according to its currently available need to distribute and manage such systems. This paper describes resources and own preferences. The proposed adaptation scheme is the design and implementation of SenseTK and provides several validated experimentally. The results demonstrate adaptation to diverse examples to show the flexibility and unique aspects of variations in available bandwidth and CPU resources roughly over two SenseTK. Finally, we experimentally measure several aspects of orders of magnitude and that fine grained adaptation is feasible given SenseTK. radically different user preferences.

6504-24, Session 8 CameraCast: flexible access to remote video sensors J. Kong, I. Ganev, K. Schwan, Georgia Institute of Technology; P. Widener, The Univ. of New Mexico New applications like remote surveillance and online environmental or traffic monitoring are making it increasingly important to provide flexible and protected access to remote video sensor devices. Current systems use application-level codes like web-based solutions to provide such access. This requires adherence to user-level APIs provided by such services, access to remote video information

170 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX Monday-Thursday 29 January-1 February 2007 Part of Proceedings of SPIE Vol. 6505 Security, Steganography, and Watermarking of Multimedia Contents IX

6505-01, Session 1 methods. Experimental results show that some steganography and steganalysis methods’ empirical performance accord with this theory. Improving steganographic security by minimizing the embedding impact 6505-04, Session 1 J. Fridrich, Binghamton Univ. Mel-Cepstrum based steganalysis for VoIP In this paper, we propose a general framework and practical coding methods for constructing steganographic schemes that minimize the steganography statistical impact of embedding. By associating a cost of an C. Kraetzer, J. Dittmann Otto-von-Guericke-Univ. Magdeburg, embedding change with every element of the cover, we first derive (Germany) bounds on the minimum theoretically achievable embedding impact and then propose a framework to achieve it in practice. The method is Steganography and steganalysis in VoIP applications are important based on syndrome codes with low-density generator matrices research topics as speech data is an appropriate cover to hide (LDGM). The problem of optimally encoding a message (e.g., with the messages or comprehensive documents. In our paper we introduce a smallest embedding impact) requires a binary quantizer that performs Mel-cepstrum based analysis known from speaker and speech near the rate-distortion bound. We implement this quantizer using recognition to perform a detection of embedded hidden messages. In LDGM codes with a survey propagation message passing algorithm. particular we combine known and established audio steganalysis Since LDGM codes are guaranteed to achieve the rate-distortion features with the features derived from Mel-cepstrum based analysis bound, the proposed methods are guaranteed to achieve the minimal for an investigation on the improvement of the detection performance. embedding impact (maximal embedding efficiency). We provide Our main focus considers the application environment of VoIP- detailed technical description of the method for practitioners and steganography scenarios. demonstrate its performance on matrix embedding. The evaluation of the enhanced feature space is performed for classical steganographic as well as for watermarking algorithms. With this strategy we show how general forensic approaches can detect 6505-02, Session 1 information hiding techniques in the field of hidden communication as well as for DRM applications. For the later the detection of the Merging Markov and DCT features for multiclass JPEG presence of a potential watermark in a specific feature space can lead steganalysis to new attacks or to a better design of the watermarking pattern. T. Pevny, J. Fridrich, Binghamton Univ. Following that the usefulness of Mel-cepstrum domain based features for detection is discussed in detail. Blind steganalysis based on classifying feature vectors derived from images is becoming increasingly more powerful. For steganalysis of JPEG images, features derived directly in the embedding domain from 6505-07, Session 2 DCT coefficients appear to achieve the best performance (e.g., the DCT features and Markov features). The goal of this paper is to Robust message authentication code algorithm for construct a new multi-class JPEG steganalyzer with markedly digital audio recordings improved performance. We do so first by extending the 23 DCT S. Zmudzinski, M. Steinebach, Fraunhofer-Institut für Integrierte feature set, then applying calibration to the Markov features described in and reducing their dimension. The resulting feature sets are Publikations- und Informationssysteme (Germany) merged, producing a 274 dimensional feature vector. The new feature Current systems and protocols for integrity and authenticity set is then used to construct a Support Vector Machine multi-classifier verification of media data do not distinguish between legitimate signal capable of assigning stego images to six popular steganographic transformation and malicious tampering that manipulates the content. algorithms — F5, OutGuess, Model Based Steganography without, Furthermore, they usually provide no localization or assessment of the and with deblocking, JpHide&Seek, and Steghide. Comparing to our relevance of such manipulations with respect to human perception or previous work on multi-classification, the new feature set significantly semantics. We present an algorithm for a robust message improves the performance of the classifier. authentication code to verify the integrity of audio recordings by means of audio fingerprinting and robust perceptual hashing. Experimental results show that the proposed algorithm provides both 6505-03, Session 1 a high level of distinction between perceptually different audio data Batch steganography and the threshold game and a high robustness against signal transformations that do not change the perceived information. A. D. Ker, Univ. of Oxford (United Kingdom) In Batch Steganography we assume that a Steganographer has to 6505-08, Session 2 choose how to allocate a fixed amount of data between a large number of covers. Given the existence of a steganalysis method for Blind audio watermark synchronization by passive audio individual objects (satisfying certain assumptions) we assume that a fingerprinting Warden attempts to detect a payload by pooling the evidence from all the objects. This paper works out the details of a particular method M. Steinebach, Fraunhofer-Institut für Integrierte Publikations- for the Warden, which counts the number of objects of which the und Informationssysteme (Germany) detection statistic surpasses a certain threshold. This natural pooling Synchronization is still one of the most important issues in digital method leads to a game between the Warden and Steganographer, watermarking. Many attacks do not remove the watermark from the and there are different varieties depending on whether the moves are cover, but only disable the synchronization between the watermark sequential or simultaneous. The solutions are intriguing, suggesting and the detector. Most watermarking algorithms feature some that the Steganographer should always concentrate the payload in as synchronization strategy, but especially in audio watermarking this few covers as possible, or exactly the reverse, but never adopt an may not be sufficient to fight de-synchronization attacks: As the intermediate strategy. Furthermore, the Warden’s optimal strategies watermark is embedded over a given length of audio data, a good are instructive for the benchmarking of quantitative steganalysis synchronization at the starting point of the retrieval process may be

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 171 Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX lost during retrieval. An example for this is time stretching, where the the underlying watermark algorithm. In our attack, a random vector effects of playback speed modification sums up during retrieval. We uniform over the n-dimensional sphere is scaled to a fixed length and introduce a novel synchronization approach applying passive audio added to a watermarked image. This is repeated hundreds or fingerprinting to synchronize each watermarking bit individually. thousands of times, each time retaining the noise only if the resulting Storage of the fingerprint values is not necessary in our approach, image lies within the detector region. The resulting noisy image lies improving the usability compared to existing solutions in this area. We along the edge of the detection region, with power distributed provide test results showing a high robustness against time proportionately within and without the watermark feature space. stretching, where a success rate of 100% for watermark retrieval after There are many uses for such false alarm vectors. The correlation time stretching of 6% was achieved. Additionally, by providing between two of them can be used to estimate the threshold of a exemplary test results for mp3 32 kbps lossy compression, we show detector based on normalized correlation or correlation coefficient. In that the applied hash scheme is robust against strong attacks on the combination with other oracle-derived features, we can deduce the audio quality. feature space dimension as well. By sampling the detector’s failure rate at three distinct noise power levels, we can combine the false alarm vectors to deduce both the threshold and feature space size. 6505-09, Session 2 Theory tells us that as the noise power overwhelms that of the Audio watermarking robust to geometrical distortions watermark, the noise vector approximates a ray from the origin of the based on dyadic wavelet transform normalized correlation detection cone to a point somewhere on its surface. Such a vector consists of a component along the watermark Y. Wang, J. Huang, Sun Yat-Sen Univ. (China) axis plus a random vector on an n-1-dimensional sphere. Given two such watermarks, independently generated, the off-axis components Geometrical transforms such as time-scale modification (TSM), will have an expected dot-product of 0; this means that the random removal(RR), random duplication(RD), and cropping, are of correlation of two long-noise vectors predictably follow a cosine law common operations on audio signals while presents many challenges relating their dot-product to the angle subtended by the detection to robust audio watermarking. The existing algorithms aiming at cone. As an interesting example, we can deduce the asymptotic solving the geometrical distortions have various drawbacks e.g. high false alarm rate of a normalized correlation detector in less than 1000 false alarm probability, heavy computation load, small data hiding oracle queries—even though the false alarm rate is on the order of capacity, and low robustness performance. In this paper an audio 1.0e-33. watermarking algorithm based on dyadic wavelet transform robust to geometrical distortions is proposed. Watermark synchronization is In this paper we derive the optimal noise levels needed at each step achieved using the geometrical invariant properties of dyadic wavelet to maximize the false alarm magnitude in the fewest steps possible. transform. A well-designed coding scheme is proposed for lowering This optimal noise level varies with the number of features and the the bit error rate of the watermark. The experimental results show that detection threshold, both of which are unknown to the reverse- the watermark is robust to geometrical transforms and other common engineer; however, other data extractable from a watermark detector operations. Compared with other existing algorithms the proposed can be used to estimate the optimal noise value per step. algorithm has several advantages of high robustness, large data hiding capacity and low computation load. 6505-12, Session 3 6505-10, Session 3 On distortion measures associated to random desynchronization attacks Efficient noninteractive zero-knowledge watermark F. Balado, Univ. College Dublin (Ireland); F. Perez-Gonzalez, Univ. detector robust to sensitivity attacks de Vigo (Spain) J. R. Troncoso, F. Pérez-González, Univ. de Vigo (Spain) Desynchronization attacks based on fine resampling of a watermarked Zero-knowledge watermark detectors presented to date are based on signal can be very effective from the point of view of degrading a linear correlation between the asset features and a given secret decoding performance. Nevertheless, the actual perceptual impact sequence. This detection function is susceptible of being attacked by brought about by these attacks has not been considered in enough sensitivity attacks, for which zero-knowledge does not provide depth in previous research. In this work, we investigate geometric protection. In this paper, an efficient zero-knowledge version of the distortion measures which aim at being simultaneously general, Generalized Gaussian Maximum Likelihood (ML) detector is related to human perception, and easy to compute. Our approach is presented. The inherent robustness that this detector presents against based on combining the stochastic characterization of the sampling sensitivity attacks, together with the security provided by the zero- grid jitter applied by the attacker with empirically relevant perceptual knowledge protocol that conceals the keys that could be used to measures. Using this procedure, we show that the variance of the remove the watermark or to produce forged assets, results in a robust sampling grid, which is a customary geometric distortion measure, and secure protocol. has to be weighted in order to carry more accurate perceptual meaning. Indeed, the spectral characteristics of the geometric jitter Two versions of the zero-knowledge detector are presented; the first signal have to be relevant from a perceptual point of view, as makes use of two new zero-knowledge proofs for modulus and intuitively seen when comparing constant shift resampling and white square root calculation, that are also introduced here; the second is jitter resampling. Finally, as the geometric jitter signal does not an improved version applicable when the spreading sequence is describe in full the resampled signal, we investigate more accurate binary, and it has minimum communication complexity. approaches to producing a geometric distortion measure that takes Completeness, soundness and zero-knowledge properties of the into account the amplitude modifications due to resampling. developed protocols are proved, and they are compared with previous zero-knowledge watermark detection protocols in terms of receiver operating characteristic, resistance to sensitivity attacks and 6505-13, Session 3 communication complexity. Fast and automatic watermark resynchronization based 6505-11, Session 3 on Zernike moments X. Kang, Sun Yat-Sen Univ. (China); W. Zeng, Univ. of Missouri/ Reverse-engineering a detector with false alarms Columbia; J. Huang, Sun Yat-Sen Univ. (China) S. A. Craver, J. Yu, Binghamton Univ. In this paper, we propose a fast watermark resynchronization method Using a watermark detector as an oracle, we can quickly grow high- based on Zernike moments, which requires only search over scaling magnitude noise vectors which fail to damage an embedded factor to combat RST geometric distortion, thus significantly reducing watermark. These noise vectors can be used to deduce facts about the computation load. We apply the proposed method to circularly 172 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX symmetric watermarking. According to Plancherel’s Theorem and the arbitrary (but invertible) linear time-invariant (LTI) system. We then rotation invariance property of Zernike moments, the rotation come up with an encoding-decoding scheme for the special case of a estimation only requires performing DFT on Zernike moments cyclic, fractional pixel-shift attack for QIM watermarking, and derive correlation value once. Thus for RST attack, we can estimate both Cramer-Rao bounds on the estimation error for the desynchronization rotation angle and scaling factor by searching for the scaling factor to parameter at the decoder. The proposed ideas can be generalized find the overall maximum DFT magnitude mentioned above. With the and bounds derived for other types of LTI attacks as well. estimated rotation angle and scaling factor parameters, the watermark can be resynchronized. In watermark detection, the normalized correlation between the watermark and the DFT magnitude of the test 6505-18, Session 4 image is used. Our experimental results demonstrate the advantage of our proposed method. The watermarking scheme is robust to global A novel interpretation of content authentication RST distortion as well as JPEG compression. In particular, the P. Comesaña-Alfaro, Univ. de Vigo (Spain); F. Balado, Univ. watermark is robust to print-rescanning and randomization-bending College Dublin (Ireland); F. Pérez-González, Univ. de Vigo (Spain) local distortion in Stirmark 3.1. An application of data-hiding technologies that typically has not been paid as much attention as copyright protection, or the embedding of 6505-14, Session 4 additional information (as metadata) on a host signal is authentication. It was only recently that a true interest on this topic has been raised. Security of spread-spectrum-based data hiding In the last years editing tools have undergone an impressive evolution, L. L. Pérez-Freire, Univ. de Vigo (Spain); P. Moulin, Univ. of Illinois putting very powerful capabilities on the hands of average users at Urbana-Champaign; F. Pérez-González, Univ. de Vigo (Spain) without needing specific advanced skills. On the other hand, this might also constitute a risk, as the the reliability on the authenticity of This paper presents an information-theoretic analysis of security for the observed contents can be seriously compromised. data hiding methods based on spread spectrum. The security is quantified by means of the mutual information between the observed Authentication deals with the capability of detecting modifications on watermarked signals and the secret carrier (a.k.a. spreading vector) the received version of a watermarked signal. Given that some level of that conveys the watermark, and this information-theoretic measure robustness is required in the authentication system in order not to can be used to bound the number of observations needed to estimate reject a truly watermarked signal when it only suffers slight the carrier up to a certain accuracy. The main results of this paper modifications (such as transcoding, small-powered noise, and other permit to establish fundamental security limits for this kind of methods typical signal processing operations that do not change the semantic and to draw conclusions about the tradeoffs between robustness and meaning of the content), not all the modifications should be detected, security. but just those that modify the meaning of the content. In this paper we introduced a novel information theoretic interpretation of this problem. 6505-16, Session 4 Trellis coded modulation to improve dirty paper trellis 6505-19, Session 4 watermarking Capacity and optimal collusion attack channels for C. Wang, G. Doërr, I. J. Cox, Univ. College London (United Gaussian fingerprinting games Kingdom) Y. Wang, P. Moulin, Univ. of Illinois at Urbana-Champaign Lattice codes and quantization-based techniques have received considerable attention as a method of digital watermarking with side In content fingerprinting, a fingerprint that is unique to each user is information. However, these algorithms are also recognized to be embedded into each copy of the distributed covertext. In a collusion highly sensitive to common signal processing primitives such as attack, two or more users may combine their copies in an attempt to valuemetric scaling, e.g. changes in volume of a song, or re- remove their fingerprints and to forge a pirated copy. To trace the quantization, e.g. JPEG compression. Hence, it is reasonable to forgery back to members of the coalition, we need fingerprinting investigate alternative codes which may offer the potential for superior codes that can reliably identify the fingerprints of those members. The performance. In previous work, we designed a new class of codes, maximal number of users that can be reliably accommodated is called dirty paper trellis codes (DPTC), which are by construction exponential in the length N of the covertext, and the exponential rate immune to valuemetric scaling. is the fingerprinting capacity. Although DPTC have shown themselves to be resistant to certain In this paper, we derive fingerprinting capacity and the capacity- classes attacks, the trellises were generated randomly, thus leading to achieving optimal probability distributions (which guide the suboptimal codes. In this paper, Ungerboeck’s work on trellis coded development of optimal codes) under the assumptions of Gaussian modulation (TCM) is exploited to significantly improve the cover signals, mean square error distortion metric, and memoryless performance of the DPTC. Experimental results show that the collusion attacks. The optimal distributions are Gaussian, the optimal resulting trellises significantly outperform the original design. attacks involve linear averaging of the colluders’ copies, and capacity is the same no matter whether the original covertext is available to the decoder or not. 6505-17, Session 4 Cramer-Rao bound on watermark desynchronization 6505-20, Session 4 parameter estimation accuracy Sphere packing lower bound on fingerprinting error S. Sadasivam, P. Moulin, Univ. of Illinois at Urbana-Champaign probability Various decoding algorithms have been proposed in the literature to N. Kiyavash, P. Moulin, Univ. of Illinois at Urbana-Champaign combat desynchronization attacks on QIM watermarking schemes. Nevertheless, none of these seem promising from the perspective of Digital fingerprinting schemes are devised for traitor tracing. In this the decoder, which is seen to consistently perform poorly. The need to work, we study the theoretical performance of one such scheme, investigate fundamental limitations on the decoder’s performance namely additive spherical fingerprints for a focused detector. Focused under a desynchronization attack is thus clear. detector has access to the host signal (non-blind detection) and aims to verify whether a user of interest is colluding. The colluders create a In this paper, we look at the class of two-step decoders which noise-free forgery by applying a mapping f to their individual copies, estimate the desynchronization attack parameter(s) before performing and then add an independent identically distributed (i.i.d) Gaussian the decoding step. We model the desynchronization attack as an noise sequence E to form the actual forgery. The colluder’s noiseless

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 173 Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX forgery can be any mapping as long as it satisfies fairness and a bound on the maximum number of modified versions of a content separation properties. Fairness property requires that all members of that satisfy good perceptibility requirement on one hand and the coalition incur equal risk, while the separation property ensures destructive averaging property on the other hand. that the host signal can be separated from the forgery at the detector. A special subclass of such attacks are order statistic attacks. The cost function in this problem is the detector’s worst case probability of 6505-23, Session 5 error. We derive a sphere packing lower bound for the spherical fingerprints on the worst case probability of error under the above Experimental study on watermark interference in general class of attacks and show that there is a regime where this multiple re-watermarking lower bound coincides with the random coding upper bound derived A. Uhl, Univ. Salzburg (Austria) in our earlier work. Watermark interference is a threat to reliable detection in multiple re- watermarking scenarios. The impact of using disjoint frequency bands 6505-21, Session 5 and/or different embedding domains in limiting those interferences is evaluated and compared. Employing disjoint frequency bands for Nested object watermarking: comparison of block- embedding different watermarks turns out to be more effective as luminance and blue channel LSB wet paper code image compared to the use of different embedding domains and is capable watermarking of maintaining reasonable detection correlation in multiple embedding applications. C. Vielhauer, J. Dittmann, Otto-von-Guericke-Univ. Magdeburg (Germany) Annotation watermarking (sometimes also used as caption or 6505-24, Session 5 illustration watermarking) denotes a specific application of Evaluation of standard watermarking techniques watermarks, which embeds supplementary information directly in the media, so that additional information is intrinsically linked to media E. Marini, F. Autrusseau, P. LeCallet, Univ. de Nantes (France); P. content and does not get separated from the media by non-malicious Campisi, Univ. degli Studi Roma Tre (Italy) processing steps such as image cropping or compression. In the last decade digital watermarking techniques have been devised Recently, nested object annotation watermarking (NOAWM) has been to answer the ever-growing need to protect the intellectual property of introduced as a specialized annotation watermarking domain, digital still images, video sequences or audio from piracy attacks. whereby hierarchical object information is embedded in photographic Although copyright protection was the very first application of images. In earlier work, the Hierarchical Graph Concept (HGC) has watermarking, different uses have been recently proposed in literature been suggested as a first approach to model object relations, which such as data authentication, fingerprinting, broadcast monitoring, are defined by users during editing processes, into a hierarchical tree multimedia indexing, content based retrieval applications, medical structure. The original HGC method uses a code-book decomposition imaging, etc. Despite the proliferation of watermarking algorithms and of the annotation tree and a block-luminance algorithm for their applications some benchmarks have been created in order to embedding. help watermarkers comparing their algorithms in terms of robustness In this article, two new approaches for embedding nested object against various attacks (i.e. Stirmark, Checkmark). However, no equal annotations are presented and experimentally compared to the attention has been devoted to the proposition of benchmarks tailored original HGC approach. The first one adopts the code-book scheme to assess the watermark perceptual transparency. In this work, we of HGC using an alternative embedding based on Wet Paper Codes in study several well known watermarking techniques in terms of blue-channel LSB domain, whereas the second suggests a new invisibility through subjective experiments and we test the ability of method based on the concept of intrinsic signal inheritance by sub- several objective metrics to be correlated with subjective scores. band energy and phase modulation of image luminance blocks. A Besides the possibility to simply check the invisibility of an embedded comparative experimental evaluation based on more than 100 test watermark, such benchmarks can also be used to support images is presented in the paper, whereby aspects of transparency watermarkers tuning the strength of their embedding algorithm to the and robustness with respect to the most relevant image modifications visibility threshold. to annotations, cropping and JPEG compression, are discussed The conclusions drawn in the paper are supported by extensive comparatively for the two code-book schemes and the novel experimentations. inheritance approach. 6505-25, Session 6 6505-22, Session 5 Digital imaging sensor identification: further study Deterring watermark collusion attacks using signal processing techniques M. Goljan, J. Fridrich, M. Chen, Binghamton Univ. A. N. Lemma, M. van der Veen, Philips Research Labs. In this paper, we revisit the problem of digital camera sensor (Netherlands) identification using photo-response non-uniformity noise (PRNU). Considering the identification task as a joint estimation and detection Collusion attack is a malicious watermark removal attack in which the problem, we use a simplified model for the sensor output and then hacker has access to multiple copies of the same content with derive a Maximum Likelihood estimator of the PRNU. The model is different watermarks and tries to remove the watermark using also used to design optimal test statistics for detection of PRNU in a averaging. In the literature, several solutions to collusion attacks have specific image. To estimate unknown shaping factors and the been reported. The main stream solutions aim at designing watermark distribution of the test statistics for the image-camera match, we codes that are inherently resistant to collusion attacks. The other construct a predictor of the test statistics on small image blocks. This approaches propose signal processing based solutions that aim at enables us to obtain conservative estimates of false rejection rates for modifying the watermarked signals in such a way that averaging each image under Neyman-Pearson testing. We also point out a few multiple copies of the content leads to a significant degradation of the pitfalls in camera identification using PRNU and ways to overcome content quality. In this paper, we present signal processing based them by preprocessing the estimated PRNU before identification. technique that may be deployed for deterring collusion attacks. We formulate the problem in the context of electronic music distribution where the content is generally available in the compressed domain. Thus, we first extend the collusion resistance principles to bit stream signals and secondly present experimental based analysis to estimate

174 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX

6505-26, Session 6 image de-noising, wavelet analysis, and neighborhood prediction, and then obtain statistical features from each characterization. Joint forensics and watermarking approach for video Experimental results demonstrate that the proposed method can authentication effectively identify the correct scanner brands/models with high accuracy. S. Thiemert, H. Liu, M. Steinebach, L. Croce Ferri, Fraunhofer- Institut für Integrierte Publikations- und Informationssysteme (Germany) 6505-29, Session 7 In our paper we discuss and compare the possibilities and Detection of malevolent changes in digital video for shortcomings of both content-fragile watermarking and digital forensic applications forensics. We analyze if the combination of both techniques allows the identification of more than the sum of all manipulations identified by R. Caldelli, N. Mondaini, V. Cappellini, A. Piva, Univ. degli Studi di both techniques on their own due to synergetic effects. Firenze (Italy); M. Barni, Univ. degli Studi di Siena (Italy) In the first part of the paper we discus the theoretical possibilities The practice of forging digital contents has become easier, it is then offered by a combined approach, in which forensics and important to create reliable, inexpensive, and fast tools for watermarking are considered as complementary tools for data identification of digital video and image authenticity. A particular field authentication or deeply combined together, in order to reduce their of application for such instruments is digital forensic. error rate and to enhance the detection efficiency. After the conceptual discussion we propose in the second part a concrete We propose a new method for the identification of digital video origin framework in which the joint approach is applied to video and, moreover, for the detection of forgeries, by resorting to the authentication. Specific forensics techniques are analyzed and sensor pattern noise inserted by each specific camera. The reference expanded to handle efficiently video data. The examples show pattern noise is self-built by means of the initial video frames and possible extensions of passive-blind image forgery detection to video previous knowledge on the camera is not needed. By observing the data, where the motion and time related characteristics of video are correlation between the camera pattern noise and the noise brought efficiently exploited. by the checked frame, its authenticity can be asserted. The technique can determine if a video sequence has been altered in some parts, if its duration has been modified and if some frames have been added. 6505-27, Session 6 Intra-frame integrity verification is provided too and fake inserted and/ or cloned objects are detected. Furthermore robustness of the Image splicing detection using 2D phase congruency and methodology against MPEG compression and motion compensation statistical moments of characteristic function has been positively checked. W. Chen, Y. Shi, New Jersey Institute of Technology; W. Su, U.S. Army Research Lab. 6505-30, Session 7 Image splicing is the process of combining image fragments from the Forensic classification of imaging sensor types same or different images without further post-processing. It is a challenging task to detect image splicing. The proposed novel N. Khanna, A. K. Mikkilineni, G. T. C. Chiu, J. P. Allebach, E. J. scheme tackles this issue by extracting distinguishing features from Delp III, Purdue Univ. both magnitude and phase information of the test images. The Digital images can be captured or generated by a variety of sources splicing process may introduce a number of sharp transitions such as including digital cameras, scanners, and various image editing lines and edges. Phase congruency has been known as a sensitive software. In many cases it is important to be able to determine the measure of such sharp transition. The statistics (mean, variance, source of a digital image. Techniques exist to identify the source skewness and kurtosis) of 2-D image phase congruency array are camera or scanner of a digital image, but these techniques rely on hence proposed as features. In addition to the phase information, the having prior knowledge on whether the image was captured magnitude information has been examined to catch the difference between the authentic images and splice images, specifically, the by a scanner or camera. We propose a method to differentiate moments of characteristic function (defined as the Fourier transform between scanned and non-scanned images based on the difference of probability density function) are proposed as another part of in geometry of the sensors used in digital cameras and scanners. features. Consequently, there are totally 120 distinguishing features Digital cameras use a two dimensional grid of sensor array while most which include 96 moment-related features and 24 phase-related scanners use a linear sensor array. In the case of flatbed scanners, features. With the SVM (support vector machine) classifier, the the same linear array is translated across the images to capture the experimental results have demonstrated that the proposed method entire image. Methods are developed for image source classification outperforms the state-of-the-art. The dependence of classification and classification results are obtained for a set of scanners and digital performance on the training sample size is also investigated and it has cameras. This method is able to classify a digital image as scanned or been found that the proposed image features are effective and not non-scanned with a very high accuracy. sensitive to the training size. 6505-31, Session 8 6505-28, Session 7 Encoding with errors: a confusing approach Robust scanner identification based on noise features M. Topkara, U. Topkara, M. J. Atallah, Purdue Univ. H. Gou, A. Swaminathan, M. Wu, Univ. of Maryland/College Park A substantial portion of the text available online is of a kind that tends A large portion of digital image data available today is acquired using to contain many typos, e.g., emails, blogs, forums. It is therefore not digital cameras or scanners. While cameras allow digital reproduction surprising that, in such texts, one can carry out information-hiding by of natural scenes, scanners are often used to capture hardcopy art in the judicious injection of typos. What is surprising is that, as this more controlled scenarios. This paper proposes a new technique for paper demonstrates, this form of embedding can be made quite non-intrusive scanner model identification, which can be further resilient. The primary technique consists of introducing ambiguous extended to perform tampering detection on scanned images. Using typos that have many possible corrections, making them harder to only scanned image samples that contain arbitrary content, we automatically restore to their original form: When considering construct a robust scanner identifier to determine the brand/model of alternative typos, we prefer ones that are also close to other the scanner used to capture each scanned image. The proposed vocabulary words. Such encodings do not materially degrade the scanner identifier is based on statistical features of scanning noise. text’s meaning because, compared to machines, humans are very We first analyze scanning noise from several angles, including through good at disambiguation. We use typo confusion matrices and word

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 175 Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX level ambiguity to carry out this kind of encoding. Unlike robust syntactic transforms that can hide information in plain text without synonym substitution that also cleverly used ambiguity, the task here changing the meaning or style of a document. A large representative is harder because typos are very conspicuous and an obvious target collection of newspaper text is fed through a prototype system. In for the adversary (synonyms are stealthy, typos are not). Our resilience contrast to previous work, the output is subjected to human testing to does not depend on preventing the adversary from correcting without verify that the text has not been significantly compromised by the damage: It only depends on a multiplicity of alternative corrections. In information hiding procedure, yielding a success rate of 96% and fact, even an adversary who has boldly “corrected” all the typos by bandwidth of 0.3 bits per sentence. randomly choosing from the ambiguous alternatives has, on average, destroyed fewer than w/4 of our w-bit mark. 6505-35, Session 8 6505-32, Session 8 Statistically constrained shallow text marking: techniques, evaluation paradigm, and results A bibliography of linguistic steganography B. Murphy, C. Vogel, The Univ. of Dublin, Trinity College (Ireland) R. Bergmair, Univ. of Cambridge (United Kingdom) In this paper we present three natural language marking strategies Linguistic steganography is, as a research area, still in its infancy, with that are based on fast and reliable shallow parsing techniques, and on an average of only 4.5 publications appearing per year since 2001. widely available lexical resources: lexical substitution, adjective Nevertheless, the field seems to be gaining attention lately as conjunction swaps, and relativiser switching. We tests these research groups around the world independently discover linguistic techniques on a random sample of the British National Corpus. steganography as a relevant area of study. Individual candidate marks are checked for goodness of structural There is no single widely accepted publication venue for this highly and semantic fit, using both lexical resources, and the web as a interdisciplinary subject, which is why relevant contributions are corpus. A representative sample of marks are given to 25 human scattered throughout publication venues of information security, judges to evaluate for acceptability and preservation of meaning. This information hiding, imaging and watermarking, cryptology, and natural allows us to establish a correlation between corpus based felicity language processing. Another problem is that the field does not have measures and perceived quality, and to make qualified predictions. an agreed upon name which could serve as an index into the Grammatical acceptability correlates with our automatic measure literature. A Google search for “linguistic steganography”, for strongly (Pearson’s r=0.795, p=0.001), allowing us to account for example, will only return a very small fraction of the relevant work. about two thirds of variability in human judgements. A moderate but statistically insignificant (Pearson’s r=0.422, p=0.356) correlation is This makes literature search in linguistic steganography a tedious task found with judgements of meaning preservation, indicating that the and a comprehensive bibliography a valuable research aid. Therefore contextual window of five content words used for our automatic we have compiled what we hope to be a comprehensive account of measure may need to be extended. the literature in linguistic steganography indexed as of November 2006. The bibliography resides on http://semanticom.ucam.org/ biblingsteg/ and is available in BibTeX and PDF formats. 6505-36, Session 9 Adaptive and distributed cryptography for signature 6505-33, Session 8 biometrics protection Syntactic tools for natural language watermarking P. Campisi, E. Maiorana, M. Gonzalez-Prats, A. Neri, Univ. degli H. M. Meral, B. Sankur, S. Ozsoy, Bogaziçi Univ. (Turkey) Studi Roma Tre (Italy) Text watermarking aims at embedding additional information in the The most emerging technology for people identification and text itself with the goals of subliminal communication and hidden authentication is biometrics. In contrast with traditional recognition information transport, content/authorship authentication, and approaches, biometric authentication relies on who a person is or enriching the text with metadata. We study the characteristics of an what a person does, being based on strictly personal traits, much agglutinative language (Turkish), amenable to text watermarking and more difficult to be forgotten, lost, stolen, copied or forged than explore its relevant syntactic tools. Although Turkish possesses traditional data. alternative watermarking tools (swapping synonyms, morphological In this paper, we focus on two vulnerable points of biometric systems: manipulations and punctuation flexibility), it turns out that syntactic the database where the templates are stored and the communication approach, which includes syntactic manipulations that leave channel between the stored templates and the matcher. semantics intact, offers the most prolific set of text manipulation. With Specifically, we propose a method, based on user adaptive error free word order possibilities, and a number of variations on morpho- correction codes, to achieve securitization and cancelability of the syntactic structure, Turkish provides a good ground for the syntax- stored templates applied to dynamic signature features. More in based watermarking. The idea of syntactic manipulation is best detail, the employed error correction code is tailored to the intra-class illustrated via tree diagrams, one or more synonymous tree variety is variability of each user’s signature features. This leads to an mapped to one logical bit value, while the varieties in the remaining enhancement of the system performance expressed in terms of false group to the alternate bit value. Adverb placement, conjunction acceptance rate. structures, active-passive, simplex-complex structures and phrasal- clausal representation of lexical elements are used as syntactic Moreover, in order to avoid corruptions or interception of the stored watermarking tools. We will also consider the dependence of templates in the transmission channels, we propose a scheme based watermarking capacity of Turkish texts depending upon the document on threshold cryptography: the distribution of the certificate authority genre and their vulnerability to non-intentional or malicious attacks. functionality among a number of nodes provides distributed, fault- tolerant, and hierarchical key management services. Experimental results show the effectiveness of our approach, when 6505-34, Session 8 compared to traditional non-secure correlation-based classifiers. The syntax of concealment: reliable methods for plain text information hiding 6505-42, Session 10 B. Murphy, The Univ. of Dublin, Trinity College (Ireland) The first BOWS contest: break our watermarking system Many plain text information hiding techniques demand deep semantic A. Piva, Univ. degli Studi di Firenze (Italy); M. Barni, Univ. degli processing, and so suffer in reliability. In contrast, syntactic Studi di Siena (Italy) processing is a more mature and reliable technology. Assuming a perfect parser, this paper evaluates a set of automated and reversible 176 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX

In the framework of the activities carried out by the European Network directions according to a simplified model of contrast masking. In this of Excellence for Cryptology ECRYPT, a 4-year network of excellence way the attacker makes use of prior knowledge that robust funded within the Information Societies Technology (IST) Programme watermarking algorithms are subject to a visibility constraint. of the European Commission’s Sixth Framework Programme (FP6), it was proposed to launch the BOWS Contest. BOWS, acronym of Break Our Watermarking System, was designed to allow to investigate 6505-46, Session 10 how and when an image watermarking system can be broken though preserving the highest possible quality of the modified content, in Tackling BOWS with the sensitivity attack case that the watermarking system is subjected to a world-wide A. Westfeld, Technische Univ. Dresden (Germany) attack. BOWS contest was not intended as an attempt to prove how well-performing a watermarking system is, but it was expected by We present our results that in the first BOWS challenge (Break Our means of this test to better understand which are the disparate Watermarking System). There were three given digital photos possible attacks, perhaps unknown at the moment of the start of the containing an invisible watermark that was introduced using informed contest, the BOWS participants could carry out to perform their action coding and embedding. The goal was to remove the watermark from and comprehend in this way the degree of difficulty of breaking the the photos while keeping a minimum quality of at least 30 dB PSNR embedded watermark. This paper describes how the Contest has (peek signal to noise ratio). We focus on the method used to win the been designed and analyzes the general progress of the attacks second phase of the contest with about 58 dB PSNR (18 dB ahead of during the Contest. the best fellow contributions). This method will be viewed from three different perspectives: Phase one and two of the contest as well as with complete knowledge about the implementation and the secret 6505-43, Session 10 key. The good, the bad, and the ugly: three different approaches to break their watermarking system 6505-47, Session 10 G. Le Guelvouit, Capgemini (France); T. Furon, IRISA (France); F. Two different approaches for attacking BOWS Cayre, École Nationale Supérieure d’Ingénieurs Electriciens de P. Comesaña-Alfaro, F. Pérez-González, Univ. de Vigo (Spain) Grenoble (France) From December 15th, 2005 to June 15th, 2006 the watermarking The Good is Blondie, a wandering gunman with a strong personal community was challenged to remove the watermark from 3 different sense of honor. The Bad is Angel Eyes, a sadistic hitman who always 512x512 watermarked images while maximizing the Peak Signal to hits his mark. The Ugly is Tuco, a Mexican bandit who’s always only Noise Ratio (PSNR) obtained comparing the attacked signals with looking out for himself. Against the backdrop of the BOWS contest, their watermarked versions. This challenge, which was named Break they search for a watermark in gold buried in 3 images. Each knows Our Watermarking System (BOWS), was in the scope of the European only a portion of the gold’s exact location, so for the moment they’re Network of Excellence Ecrypt, and its main objective was the dependent on each other. However, none are particularly inclined to improvement of the comprehension of the degree share... of difficulty of breaking a watermarking system; in this sense, BOWS was not aimed at checking the vulnerability of the specific chosen 6505-44, Session 10 watermarking scheme against attacks, but to inquire in the different strategies the attackers would follow to achieve their target. Watermarking attack: BOWS contest In this paper the main results obtained by the authors when attacking J. Bennour, J. E. Dugelay, M. Federico, Institut Eurécom (France) the BOWS system are introduced. Mainly, the strategies followed can be divided into two different approaches: blind sensitivity attacks and In this paper, we describe some attack strategies we have applied to exhaustive search of the secret key. the three protected images proposed within the particular context of BOWS (Break Our Watermarking System). The technique we have first used exploits the notion self-similarities included in images. This 6505-48, Session 10 technique alone was powerful enough for the first image whereas for the two other images the watermark has been removed but the use of How we broke the BOWS watermark some additional processing steps was necessary to preserve the minimum required PSNR. Additional tools used in this context were S. A. Craver, I. Atakli, J. Yu, Binghamton Univ. composed of basic signal processing techniques (e.g. filtering). The Break Our Watermarking System (BOWS) contest provided an In our article, we also propose a possible use of BOWS as a teaching excellent opportunity to reverse-engineer a secret watermarking tool for master students. A homework (derived from BOWS) was then algorithm from an oracle. We credit this strategy with our victory in proposed. 30 students were asked to break the watermark of the the first phase of the contest—rather than attempt to damage the three images while keeping the highest possible PSNR. They had to mark, we spent much of our time submitting experimental images to return their results as well as a report within a delay of two weeks. the online watermark detector, to deduce the watermark domain, the This homework was an attractive training for students to better number of features, and the type of detection algorithm used. This understand the difficulty of breaking a watermark and to develop a allowed us to focus our attack, once we had identified a weakness in critical point of view about watermarking algorithm; in short, to the unknown detector structure. We learned several lessons from the increase their global understanding and knowledge on digital experience. In this paper we introduce the notion of “super- watermarking. robustness,” a property of watermarking systems which is actually undesirable, and perhaps unavoidable. A watermark is super-robust if it survives such an extreme amount of distortion that the successful 6505-45, Session 10 detection could properly be called a false alarm. The problem with super-robustness is that it can be used to reverse- Tangential sensitivity analysis with prior information engineer the watermarking algorithm. If we inflict noise so severe that J. Earl, Univ. of Cambridge (United Kingdom) a watermark should be destroyed, its detection informs us that our noise is concentrated outside the feature space, or is irrelevant to the A novel sensitivity analysis attack is introduced in the context of the detector. Since a watermark will be super-robust to very specific BOWS (Break Our Watermarking System) contest. The method operations, these operations can be used experimentally to classify a follows the boundary of the detection region, rather than directly watermark algorithm using a detection oracle. We exploited this estimating the normal to the boundary, so it is applicable even if the technique to determine the watermark feature space, and rule out boundary is not well approximated by a hyperplane. The efficiency of common detection algorithms like normalized correlation. A second the method is improved by scaling the (randomly selected) search

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 177 Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX lesson that we learned is that specific signal processing components one of biggest challenge of the video watermark algorithm. With a lot carry security flaws. We observed that the detection algorithm was of common video tools the modification at the video data can also the not a common simple correlation structure—it was later revealed to frame types and existent watermark solutions then would try to detect be a Viterbi decoder—and that it was particularly sensitive to noise the watermark in the wrong I-frame, not in the original marked I-frame. spikes. Our attacked images multiplied a few sensitive AC Our strategy enables the detection of the selectively marked frames. components by factors of at least three. We describe the techniques The idea is the calculation of robust hash values of the marked I- we used to both reverse-engineer and defeat the watermarks in the frames and storing of the results for later assisting the synchronization BOWS contest, and describe general techniques to exploit super- of the watermarking algorithm at the retrieval stage. robustness in watermarking systems. The presented system is a framework which consists of four independent modules. It is possible to exchange the modules to adjust the system as easily as possible. The first module extracts the 6505-49, Session 11 features of the video data for following hash calculation of the I- Adaptive MPEG-2 video data hiding scheme frames in the second module. The third module is the detection of the marked video frames based on the hash calculation. The last module A. Sarkar, U. Madhow, S. Chandrasekaran, B. S. Manjunath, Univ. is the detection of the watermarking information from the collected of California/Santa Barbara frames. Different test show the improvement of the framework for watermark detection against without the usage of it. We have investigated adaptive mechanisms for high-volume transform-domain data hiding in MPEG-2 video which can be tuned to sustain varying levels of compression attacks. The data is hidden in 6505-53, Session 12 the uncompressed domain by scalar quantization index modulation (QIM) on a selected set of low-frequency Discrete Cosine Transform Source digital camcorder identification using CCD photo (DCT) coefficients. We propose an adaptive hiding scheme where the response nonuniformity embedding rate is varied according to the type of frame and the reference quantization parameter (decided according to MPEG-2 rate M. Chen, J. Fridrich, M. Goljan, Binghamton Univ. control scheme) for that frame. For a 1.5 Mbps video and a frame-rate Photo-response non-uniformity (PRNU) of digital sensors was recently of 25 frames/sec, we are able to embed almost 7500 bits/sec. Also, proposed [1] as a unique identification fingerprint for digital cameras. the adaptive scheme hides 20% more data and incurs significantly The PRNU extracted from a specific image can be used to link it to less frame errors (frames for which the embedded data is not fully the digital camera that took the image. Because digital camcorders recovered) than the non-adaptive scheme. Our embedding scheme use the same imaging sensors, in this paper, we extend this technique incurs insertions and deletions at the decoder which may cause de- for identification of digital camcorders from video clips. We also synchronization and decoding failure. This problem is solved by the investigate the problem of determining whether two video clips came use of powerful turbo-like codes and erasures at the encoder. The from the same camcorder and the problem of whether two differently channel capacity estimate gives an idea of the minimum code transcoded versions of one movie came from the same camcorder. redundancy factor required for reliable decoding of hidden data The identification technique is a joint estimation and detection transmitted through the channel. To that end, we have modeled the procedure consisting of two steps: (1) estimation of PRNUs from MPEG-2 video channel using the transition probability matrices given video clips using the Maximum Likelihood Estimator and (2) detecting by the data hiding procedure, using which we compute the (hiding the presence of PRNU using normalized cross-correlation. We scheme dependent) channel capacity. anticipate this technology to be an essential tool for fighting piracy of motion pictures. Experimental results are shown to demonstrate the 6505-50, Session 11 reliability and generality of our approach. Digital video watermarking in P-frames 6505-54, Session 12 M. Noorkami, R. M. Mersereau, Georgia Institute of Technology Blind identification of cellular phone cameras Most video watermarking algorithms embed the watermark in I- frames, which are essential for the video signal, but refrain from B. Sankur, O. Celiktutan, Bogaziçi Univ. (Turkey); I. Avcibas, embedding anything in P- and B-frames that are highly compressed Uludag Üniv. (Turkey) by motion compensation. Furthermore, these algorithms do not take We propose a novel forensic method for cell-phone camera advantage of temporal masking in subjective perception of the video. identification, in order to determine the brand of the source camera In this paper, we explore the capacity of P-frames and the temporal with which a given image was captured. The main idea is based on masking effect in the video signal. The challenge in embedding the telltale effects inherent in the proprietary image formation pipeline. watermark bits in P-frames is that the video bit rate can increase The main differences between cameras originate from the color filter significantly. Thus, we choose to embed watermark bits only in array that is used to interpolate between color pixels. These nonzero AC residuals in P-frames. Changing zero-valued coefficients proprietary interpolation algorithms leave footprints in the correlations to nonzero values can significantly increase the video bit rate because between contiguous bit planes of an image. Therefore, we track down H.264 (and earlier coders as well) uses run length codes. We show these characteristics using various statistical features. One set of that avoiding zero-valued coefficients significantly reduces the features, called binary similarity measures, directly address correlation percentage increase in the compressed video bit rate and makes properties within and between planes. Another set, called image watermark embedding in P-frames practical. Since the nonzero quality measures, captures camera-specific noise using the difference residuals in P-frames correspond to non-flat areas that are in motion, between an image and its denoised version. Thirdly, we use the so- temporal and texture masking will be exploited at the same time. This called higher order wavelet statistics, which focus on the wavelet is confirmed by showing the resemblance of the plots of the number decomposition of the image. A subset of these features is then of nonzero residuals in each frame to the motion intensity plots. selected using the sequential forward feature selection methodology, and finally we build a multiclass support vector machine classifier. We 6505-51, Session 11 have obtained classification performance of 93% for threesomes of camera brands and of 85% when all nine camera types were involved. Framework for combined video frame synchronization We have also observed that decision fusion in the form of plurality and watermark detection voting improves the performance by two points. E. Hauer, T. Bölke, M. Steinebach, Fraunhofer-Institut für Integrierte Publikations- und Informationssysteme (Germany) At watermark retrieval, the detection of selectively marked I-frames is

178 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX

6505-55, Session 12 6505-58, Session 12 Forensics for flatbed scanners A generalized Benford’s law for JPEG coefficients and its T. Gloe, E. Franz, A. Winkler, Technische Univ. Dresden (Germany) applications in image forensics Based on a current method for the identification of digital cameras, D. Fu, Y. Shi, New Jersey Institute of Technology; W. Su, U.S. we investigated approaches for identifying flatbed scanner image Army Research Lab. origin. Generally, the success of identifying digital camera images It is well known that the distribution of the JPEG block-DCT (discrete origin is based on analyzing imaging device dependent noise cosine transform) coefficients can be modeled as generalized patterns. Since flatbed scanners and digital cameras use similar Gaussian distribution or Laplacian distribution. However, the technologies, the utilization of image sensor noise for identifying the distribution of the most significant digits of the JPEG block-DCT origin of scanned images seems to be possible. Within this article, we coefficients has not been reported in the literature. In this paper, a point out possible problems with directly applying the original novel statistical model based on Benford’s law for the probability identification method on flatbed scanner images. Tests investigating distribution of the first digits of the JPEG block-DCT coefficients is resolution and flatfielding as examples for influences confirmed our presented. Specifically, a parametric logarithmic model, called assumptions. Generally, our tests aimed at describing the distribution generalized Benford’s law, is formulated. Furthermore, some image of noise produced by one CCD element. While the first naïve tests forensic applications of this model are discussed in this paper, which only utilized the mean of the estimated noise per column, further tests include the detection of JPEG compression for images in bitmap considered additional statistical parameters. However, experiments format, the estimation of JPEG compression Q-factor for JPEG showed that these analyzes of noise per column are not suited for compressed bitmap image and the detection of double compressed identification. JPEG image. The results of our extensive experiments demonstrate To conclude, further research is required to develop a reliable the effectiveness of the proposed statistical model. identification method for flatbed scanners, e.g., by evaluating additional features like patterns of dust and scratches. Of course, resolution and flatfielding have also to be considered. Particularly, 6505-59, Session 13 flatfielding seems to be a possibility to avoid correct image identification in case of digital cameras. A semi-blind image watermarking scheme: embedding a pseudo-random number sequence in two-frequency bands in the DCT domain 6505-56, Session 12 A. Koc, A. M. Eskicioglu, The City Univ. of New York Image authentication via intrinsic fingerprints The proposed watermarking scheme was tested using a range of A. Swaminathan, M. Wu, K. J. R. Liu, Univ. of Maryland/College attacks performed using Matlab. The chosen attacks were low pass Park filtering, Gaussian noise, JPEG compression, resizing, cropping, intensity adjustment, histogram equalization, gamma correction, and With growing popularity of digital imaging devices and low-cost image salt and pepper noise. editing software, the integrity of image content can no longer be taken The experiments show that for one group of attacks (JPEG for granted. This paper introduces a methodology for forensic analysis compression, Gaussian noise, resizing, salt and pepper noise, and of digital camera images, based on the observation that many in- low pass filtering), the correlation with the real watermark exceeds the camera and post-camera processing operations leave distinct traces threshold in the low bands. For another group of attacks (gamma on digital images. We present methods to identify these intrinsic correction, intensity adjustment, histogram equalization, and fingerprint traces of the various processing operations and employ cropping), the correlation with the real watermark exceeds the them to verify the authenticity of digital data. An explicit ground-truth threshold in the high bands. The randomly generated watermarks do model is developed to characterize the properties that should be not exceed the threshold in either band. satisfied by a direct camera output. Any further processing applied to the camera captured image is modeled by a manipulation filter, and its coefficients are estimated using blind deconvolution. Utilizing 6505-60, Session 13 reference patterns obtained from untampered images in the training stage, the proposed methods are capable of detecting manipulations Distortion-free robust watermarking: a case study made by previously unseen operations and steganographic embedding. D. Coltuc, Univ. Valahia din Targoviste (Romania) and Institut National Polytechnique de Grenoble (France); J. Chassery, Institut National Polytechnique de Grenoble (France) 6505-57, Session 12 This paper investigates the distortion-free robust watermarking by Scanner identification using sensor pattern noise multiple marking. Robust watermarking is first performed. Then, by reversible watermarking, the information needed to invert both the N. Khanna, A. K. Mikkilineni, G. T. C. Chiu, J. P. Allebach, E. J. robust and the reversible watermarking is embedded. In case of no Delp III, Purdue Univ. attacks, the robust watermark is detected and the authorized party exactly recovers the original. In case of attacks, one can suppose that This paper presents methods for authenticating images that have the robust watermark can still be detected, but the reversibility is lost. been acquired using desktop scanners. The method is based on using The approach relies on the embedded capacity of the reversible the pattern noise of the imaging sensor as a fingerprint for the watermarking. The overall scheme inherits the robustness of the first scanner. Similar methods have been used for authenticating digital marking stage. The selection of the robust and of the reversible cameras. To identify the source scanner of an image from a given set watermarking schemes is discussed. In order to improve the of scanners, a reference pattern is first estimated for each scanner. robustness of the first marking stage against the second one a joint Since the noise in each image contains two components, a random marking procedure is proposed. A case study of reversible and fixed component, the reference pattern is found by averaging the watermarking providing very good robustness against JPEG noise obtained from multiple images. The reference pattern is treated compression is presented. as a unique fingerprint of the scanner. The noise pattern extracted from an unknown image is then compared with the known scanner reference patterns to determine the image origin. A recently 6505-61, Session 13 developed anisotropic local polynomial estimator for image restoration based on directional multiscale optimizations is used along Hidden context highlighting with JPEG2000 imagery with a wavelet based filter to reduce the number of images needed to estimate the sensor pattern noise. R. U. Rosenbaum, Sr., H. Schumann, Univ. Rostock (Germany) electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 179 Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX

Within single imagery some parts might be more interesting than that describes the TSH decomposition of the image. The proposed others. In order to point these regions out, they are usually method allows to significantly improve the robustness and secureness highlighted. To achieve this, many different ideas as spatial of Barni’s approach. transformations or color coding have been proposed. To avoid the unhandy signalization of these regions via a second channel, highlighting is often applied directly to the content and not removable. 6505-65, Session 15 However, it might not always be desired to indicate the respective contents to all persons having access to the image. This opens up for A method of text watermarking using presuppositions the completely new and rather uncommon idea of hiding the B. Macq, O. Vybornova, Univ. Catholique de Louvain (Belgium) highlighting. We propose a method for watermarking texts of arbitrary length using This contribution introduces a strategy to highlight pre-defined focus natural-language semantic structures. For the key of our approach we regions without inserting new contents into the image. This is possible use a linguistic semantic phenomenon of presupposition. by taking advantage of effects of the HVS for highlighting, and Presupposition is implicit information considered as well-known or features of the new image coding standard JPEG2000 for its which readers of the text are supposed to treat as well-known; this implementation. An integral novelty, however, are the options to hide information is a semantic component of certain linguistic expressions and control the access to the highlighting. As the proposed ideas can (lexical items and syntactical constructions called presupposition be implemented by un-complex transcoding operations, this allows triggers). The same sentence can be used with or without for a fast and flexible application to already encoded but also plain presupposition, or with a different presupposition trigger, provided imagery. Contrary to related strategies, the highlighting is fully that all the relations between subjects, objects and other discourse reversible and might also be removed if no longer required. referents are preserved - then such transformations will not change the meaning of the sentence. We define the distinct rules for presupposition identification for each trigger and regular 6505-62, Session 14 transformation rules for using/non-using the presupposition in a given Data hiding based wavelet on finite field sentence. A. Benjelloun Touimi, France Telecom R&D (France); L. Godard, Isolated sentences can carry the proposed watermarks. However, the École Nationale Supérieure d’Ingénieurs Electriciens de Grenoble longer is the text, the more efficient is the watermark. The proposed approach is resilient to main types of random transformations, like (France) passivization, topicalization, extraposition, preposing, etc. The web of This paper proposes a new data hiding methods for multimedia resolved presupposed information in the text will hold the integrity of signals (e.g. audio) using the wavelet on finite field representation. the text, introducing “secret ordering” into the text structure to make it Using such representation for the watermark embedding and resilient to “data loss” attacks and “data altering” attacks. detection improves the security level compared to the existing algorithms. The polynomial generator of the filter bank is used as a secure key of the watermarking algorithm. The paper gives a brief 6505-66, Session 15 review on the theory of wavelet on finite field and their main properties used for watermarking. Since the embedding and detection methods An evaluation of lightweight JPEG2000 encryption with have to be designed in respect to the constraints of this new anisotropic wavelet packets representation, the paper introduces new techniques for these D. Engel, A. Uhl, Univ. Salzburg (Austria) operations. The first one is based on the statistical properties of the signal and uses a substitution method for embedding. With respecting In this paper we evaluate a lightweight encryption scheme for the constraints of the new representation, the second technique is JPEG2000 which relies on a secret transform domain constructed issued from the spread spectrum method and uses the correlation for with anisotropic wavelet packets. The pseudo-random selection of the the detection. The resulting methods are sensitive to the modification bases used for transformation takes compression performance into of the watermarked signal and the additive noise. Hence fragile account, and discards a number of possible bases which lead to poor watermarking and steganography are the main applications of the compression performance. Our main focus in this paper is to answer proposed methods. the important question of how many bases remain to construct the keyspace. In order to determine the trade-off between compression performance and keyspace size, we compare the approach to a 6505-63, Session 14 method that selects bases from the whole set of anisotropic wavelet packet bases following a pseudo-random uniform distribution. The Perceptual data hiding in tree structured Haar transform compression performance of both approaches is compared to get an domain estimate of the range of compression quality in the set of all bases. We then analytically investigate the number of bases that are M. Carli, M. Cancellaro, Univ. degli Studi di Roma Tre (Italy); K. O. discarded for the sake of retaining compression performance in the Egiazarian, J. T. Astola, Tampere Univ. of Technology (Finland) compression-oriented approach as compared to selection by uniform In this paper a novel, fast, and robust data hiding technique based on distribution. Finally, the question of keyspace quality is addressed, i.e. key-dependent basis functions is presented. One of the main how much similarity between the basis used for analysis and the basis vulnerabilities of classical watermarking schemes is the public used for synthesis is tolerable from a security point of view and how knowledge of the domain used for the embedding. It can reduce the this affects the lightweight encryption scheme. effectiveness of the system if portions of the original image are available. To counteract with this problem and to improve the security of watermarking applications, the concept of key-dependent basis 6505-67, Session 15 has been introduced. By keeping a secret key used to generate the Tamper-proofing of electronic and printed text basis functions, it is impossible to guess the exact domain in which the mark is spread. Recently, a generalization of the Haar transform, documents called Tree-Structured Haar transforms (TSH), was developed. It is R. Villan, Sr., S. V. Voloshynovskiy, O. J. Koval, F. Deguillaume, T. based on a non uniform sampling of the time interval splitting. The Pun, Univ. de Genève (Switzerland) image watermarking method based on the characteristics of the Human Visual System (HVS) designed by Barni, Bartolini and Piva is In this paper, we deal with the problem of authentication and tamper- adopted here in TSH domain. The secret key is represented by the proofing of text documents that can be distributed in electronic or key-generator of the tree (so-called the Discontinuity Point Vector) printed forms. We advocate for the combination of robust text hashing and text data hiding technologies as an efficient solution to this problem. First, we consider the problem of text data hiding in the

180 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6505: Security, Steganography, and Watermarking of Multimedia Contents IX algorithm creates “exceptions” in H.264 codespace that only the scope of the Gel’fand-Pinsker text data hiding framework. For decoder understands. The codespace is defined by the Context illustration, the two modern text data hiding methods, namely color Adaptive Variable Length Codeded(CAVLC) of the stream. Each index modulation (CIM) and location index modulation (LIM), are CAVLC is defined by a 5-tuple. The watermarker securely maps explained. Second, we study two approaches of robust text hashing eligible 5-tuples to unused portions of the codespace. What makes that are well suited for the considered problem. In particular, both this algorithm possible is the discovery that most of H.264 codespace approaches are compatible with CIM and LIM. The first approach is in fact unused. We have come up with a code mapping that makes use of optical character recognition (OCR) and a classical embeds the watermark by altering eligible elements of each 5-tuple, cryptographic message authentication code (MAC). The second hence mapping it to the unused portion. Security is achieved through approach is new and can be used in some scenarios where OCR does a shared key between embedder and decoder. The watermarked not produce consistent results. The experimental work compares both stream retains its file size, remains visually transparent, is secure approaches and shows their robustness against typical intentional/ against forging and is reversible. Since the watermark is placed post unintentional document distortions including electronic format compression it remains fragile to re encoding and other tampering conversion, data hiding, printing, scanning, photocopying, and faxing. attempts.

6505-68, Session 15 Channel model and capacity analysis of printed text documents A. K. Mikkilineni, P. Chiang, G. T. C. Chiu, J. P. Allebach, E. J. Delp III, Purdue Univ. In today’s digital world securing different forms of content is very important in terms of protecting copyright and verifying authenticity. One example is watermarking of digital audio and images. We believe that a marking scheme analogous to digital watermarking but for documents is important and that average users need to be able to print secure documents, for example boarding passes and bank transactions. Our signature embedding process takes place inside the printer hardware which allows more control over the type of marks which can be added to the document and also deters hacking or modification of the encoded data before it is printed. Previously we have presented techniques which allow embedding up to 33 bits into a 12 point text document, or 2 bits for every 3 lines of text. More recently we have been able to reliably embed up to 3 bits per text line. In this paper we will present our recent work on modelling and capacity analysis of a text document as a communication channel. The approach we take is a band limited fast fading channel with signalling periods equal to the height of a single text line, and with multipath receiver diversity.

6505-69, Session 15 Watermarking geographic vectordata using a variable stripsize scheme M. Voigt, Technische Univ. Darmstadt (Germany) A blind watermarking system for geographic vector-data is proposed which is robust against the following attacks: polyline simplifications such as the Douglas-Peucker algorithm, cropping and shifting of data-points as well as small perturbations of random noise. The watermark information is embedded by changing the x/y-coordinates of the data-points within the tolerance range of the data. Therefore, the map is divided into horizontal and vertical stripes whose widths depend on a secret key. Each strip possesses a 0-line and a 1-line. Depending on the watermark information all points within a stripe are shifted on either one of the two lines. As a consequence of the variable strip-size structure no extra bits for synchronization are necessary, leading to an improved capacity compared to our former system with equal strip-sizes. The application of the proposed scheme is not restricted to the case of vector geo-data, other kinds of more general vector graphics can be treated in the same way.

6505-70, Session 15 Authentication of H.264 streams by direct watermarking of CAVLC blocks B. G. Mobasseri, Y. J. Naik Raikar, Villanova Univ. In this work we report on what we believe to be the first H.264 authentication watermarker that operates directly in the bitstream. The

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 181 Conference 6506: Multimedia Content Access: Algorithms and Systems Wednesday-Thursday 31 January-1 February 2007 Part of Proceedings of SPIE Vol. 6506 Multimedia Content Access: Algorithms and Systems

6506-01, Session 1 can be compared so that the medical expert can make a better diagnosis. To this effect, this paper introduces a search and retrieval A model-based conceptual clustering of moving objects system for brain MR databases. We also describe a fully automated in video surveillance and fast brain MR image analysis system to detect degenerative iron accumulation in brain, as it is the case in Alzheimer’s and Parkinson’s. J. Lee, Univ. of Bridgeport The system is composed of several novel image processing Data mining techniques have been applied in video databases to algorithms and has been extensively tested in Leiden University identify various patterns or groups. Clustering analysis is used to find Medical Center over so far more than 600 patients. the patterns and groups of moving objects in video surveillance systems. Most existing methods for the clustering focus on finding the optimum of overall partitioning. However, these approaches cannot 6506-04, Session 2 provide meaningful descriptions of the clusters. Also, they are not very Assessment of end-user response to sports highlights suitable for moving object databases since video data have spatial and temporal characteristics, and high-dimensional attributes. In this extraction for personal video recorders paper, we propose a model-based conceptual clustering (MCC) of A. Divakaran, K. Wittenburg, K. A. Peker, Mitsubishi Electric moving objects in video surveillance based on a formal concept Research Labs.; R. Radhakrishnan, Dolby Labs. Inc. analysis. Our proposed MCC consists of three steps: model formation, model-based concept analysis, and concept graph We tested our previously reported sports highlights playback for generation. The generated concept graph provides conceptual personal video recorders with a carefully chosen set of sports descriptions of moving objects. In order to assess the proposed aficionadoes. Each subject spent about an hour with the content, approach, we conduct comprehensive experiments with artificial and going through the same basic steps of introduction, trying out the real video surveillance data sets. The experimental results indicate system, and follow up questionnaire. The main conclusion was that that our MCC dominates two other methods, i.e., generality-based the users unanimously liked the functionality very much even when it and error-based conceptual clustering algorithms, in terms of quality made mistakes. Furthermore, the users felt that if the user interface of concepts. were made much more responsive so as to quickly compensate for false alarms and misses, the functionality would be vastly enhanced. The ability to choose summaries of any desired length turned out to 6506-02, Session 1 be the main attraction. Image watermarking based on a color quantization process 6506-05, Session 2 A. Trémeau, J. Thomas, G. Chareyron, Univ. Jean Monnet Saint- Examining user interactions with video retrieval systems Etienne (France) M. G. Christel, Carnegie Mellon Univ. The purpose of this paper is to propose a color image watermarking scheme based on an image dependent color gamut sampling of the The Informedia group at Carnegie Mellon University has since 1994 L*a*b* color space. The main motivation of this work is to control the been developing and evaluating surrogates, summary interfaces, and reproduction of color images on different output devices in order to visualizations for accessing digital video collections containing have the same color feeling, coupling intrinsic informations on the thousands of documents, millions of shots, and terabytes of data. image gamut and output device calibration. This paper is focused This paper reports on TRECVID 2005 and 2006 interactive search firstly on the research of an optimal LUT (Look Up Table) which both tasks conducted with the Informedia system by users having no circumscribes the color gamut of the studied image and samples the knowledge of Informedia or other video retrieval interfaces, but being color distribution of this image. This LUT is next embedded in the experts in analyst activities. The TRECVID studies confirm the utility of image as a secret message. The principle of the watermarking multiple access strategies to news video beyond just query-by-text. scheme is to modify the pixel value of the host image without causing Query-by-image-example, query-by-concept, and query-by-best-of- any change neither in image appearance nor on the shape of the topic collectively are used much more than query-by-text, despite the image gamut. analysts’ high level of expertise with text retrieval and inexperience with video retrieval. Think-aloud protocols, questionnaires, and interviews were also conducted with this user group to assess the 6506-03, Session 1 contributions of various video summarization and browsing techniques with respect to exploratory investigations of broadcast Search and retrieval of medical images for improved news corpora. Lessons learned from these user interactions are diagnosis of neurodegenerative diseases reported, with recommendations on both interface improvements for video retrieval systems and enhancing the ecological validity of video A. Ekin, R. S. Jasinschi, E. Turan, R. Engbers, Philips Research retrieval interface evaluations. Labs. (Netherlands); J. van der Grond, M. van Buchem, Leids Univ. Medisch Ctr. (Netherlands) 6506-06, Session 2 In the medical world, the accuracy of diagnosis is mainly affected by either lack of sufficient understanding of some diseases or the inter-, Video summary evaluation and visualization and/or intra-observer variability of the medical experts. The former requires understanding the progress of diseases at much earlier C. M. Taskiran, Z. Li, Motorola, Inc. stages, extraction of important information from ever growing Although many algorithms have been proposed to generate automatic amounts of data, and finally finding correlations with certain features summaries for video programs we believe there are two major open and complications that will illuminate the disease progression. The problems in the field: the evaluation of summaries generated and latter (inter-, and intra- observer variability) is caused by the deriving intuitive summary visualizations for different types of devices. differences in the experience levels of different medical experts (inter- In this paper we examine these two problems. Evaluating summaries observer variability) or by mental and physical tiredness of one expert for video documents is more complex than the corresponding (intra-). We believe that the use of large databases can help improve problem for text documents. Still, we believe that many of the the current status of disease understanding and decision making. By evaluation schemes developed to evaluate text summarization comparing large number of patients, some of the otherwise hidden systems are applicable to the video domain. We outline some of the relations can be revealed that results in better understanding, patients problems associated with video summary evaluation and discuss the with similar complications can be found, the diagnosis and treatment

182 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6506: Multimedia Content Access: Algorithms and Systems methodologies we have developed to address these problems. background subtraction. Semantics extracted from the remaining Another important issue if how to present the generated summaries in object trajectories are then used to segment the video. The results of an intuitive and compact manner to users. Obviously, the best manner the algorithm when applied to surveillance data are compared with of visualization will depend on the particular device being used to hand labelled segmentation to obtain precision recall curves and view the summary, e.g., viewing a program summary on the PC is a harmonic mean. Comparisons with traditional background subtraction very different experience than viewing it on a mobile device. We and video segmentation algorithms show a drastic improvement in present the visualization approaches we have developed to present performance. summaries on cell phones. 6506-10, Session 4 6506-07, Session 3 Knowledge discovery for better photographs Efficient re-indexing of automatically annotated image J. Yen, Toshiba America Business Solutions; P. Wu, D. R. Tretter, collections using keyword combination Hewlett-Packard Labs. A. G. Yavlinsky, S. Rueger, Imperial College London (United A photograph captured by a digital camera usually includes camera Kingdom) metadata in which sensor readings, camera settings and other This paper presents a framework for improving the image index capture pipeline information are recorded. The camera metadata, obtained by automated image annotation. Within this framework, the typically stored in an EXIF header, contains a rich set of information technique of keyword combination is used for fast image re-indexing reflecting the conditions under which the photograph was captured. based on initial automated annotations. It aims to tackle the This set of rich information can be potentially useful for improvement challenges of limited vocabulary size and low annotation accuracies in digital photography but its multi-dimensionality and heterogeneous resulting from differences between training and test collections. It is data structure make it difficult to be useful. Knowledge discovery, on useful for situations when these two problems are not anticipated at the other hand, is usually associated with data mining to extract the time of annotation. We show that based on example images from potentially useful information from complex data sets. In this paper we the automatically annotated collection, it is often possible to find use a knowledge discovery framework based on data mining to multiple keyword queries that can retrieve new image concepts which automatically associate combinations of high-dimensional, are not present in the training vocabulary, and improve retrieval results heterogeneous metadata with scene types. In this way, we can of those that are already present. We demonstrate that this can be perform very simple and efficient scene classification for certain types done at a very small computational cost and at an acceptable of photographs. We have also provided an interactive user interface in performance tradeoff, compared to traditional annotation models. We which a user can type in a query on metadata and the system will present a simple, robust, and computationally efficient approach for retrieve from our image database the images that satisfy the query finding an appropriate set of keywords for a given target concept. We and display them. We have used this approach to associate EXIF report results on TRECVID 2005, Getty Image Archive, and Web image metadata with specific scene types like back-lit scenes, night scenes datasets, the last two of which were specifically constructed to and snow scenes. To improve the classification results, we have support realistic retrieval scenarios. combined an initial classification based only on the metadata with a simple, histogram based analysis for quick verification of the discovered knowledge. The classification results, in turn, can be used 6506-08, Session 3 to better manage, assess, or enhance the photographs. Video to the rescue of audio: shot boundary assisted speaker change detection 6506-12, Session 4 A. Samour, M. Karaman, L. Goldmann, T. Sikora, Technische Univ. Organizing a daily visual diary using multifeature Berlin (Germany) clustering Speaker change detection (SCD) is a preliminary step for many audio C. E. Ó Conaire, N. E. O’Connor, A. F. Smeaton, Dublin City Univ. applications such as speaker segmentation and recognition. Thus, its (Ireland) and Adaptive Information Cluster (Ireland); G. Jones, robustness is crucial to achieve a good performance in the later steps. Dublin City Univ. (Ireland) Especially, misses (false negatives) affect the results. For some applications, domain-specific characteristics can be used to improve The SenseCam is a prototype device from Microsoft that facilitates the reliability of the SCD. In broadcast news and discussions, the automatic capture of images of a person’s life by integrating a colour cooccurrence of shot boundaries and change points provides a camera, storage media and multiple sensors into a small wearable robust clue for speaker changes. In this paper, two multimodal device. However, efficient search methods are required to reduce the approaches are presented that utilize the results of a shot boundary user’s burden of sifting through the thousands of images that are detection (SBD) step to improve the robustness of the SCD. Both captured per day. In this paper, we describe experiments using colour approaches clearly outperform the audio-only approach and are spatiogram and block-based cross-correlation image features in exclusively applicable for TV broadcast news and plenary conjunction with accelerometer sensor readings to cluster a day’s discussions. worth of data into meaningful events, allowing the user to quickly browse a day’s captured images. Two different low-complexity algorithms are detailed and evaluated for SenseCam image clustering. 6506-09, Session 3 A trajectory based video segmentation for surveillance 6506-13, Session 5 applications Recognizing persons in images by learning from videos N. M. Thomas, N. Canagarajah, Univ. of Bristol (United Kingdom) R. W. Lienhart, E. Hörster, Univ. Augsburg (Germany) Video segmentation for content based retrieval has traditionally been done using shot cut detection algorithms that search for abrupt In this paper, we propose an approach for automatically recognizing changes in scene content. Surveillance videos however, usually use persons in images based on their general outer appearance. Therefore still cameras, and do not contain any shots. Hence, a novel high level we build a statistical model for each person. Large amounts of semantic change detection algorithm is proposed in this paper that training data are collected and labeled automatically by using a visual uses object trajectory features to segment surveillance footage. These sensor array capturing image sequences containing the person to be trajectory features are extracted automatically, using background learnt. Foreground-background segementation is performed to subtraction and a multiple blob tracking algorithm. The trajectory seperate the person from background, thus enabeling to learn the features are first used to remove false object detections from persons appearance independent of the background. Color and

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 183 Conference 6506: Multimedia Content Access: Algorithms and Systems gradient features are extracted representing the segmented person. concepts in the ontology. Using ontology concepts and relationships Person recognition of incoming photos is carried out using (k)-Nearest allows building complex queries with true semantic content. Image Neighbor(s) classification and normalized histogram intersection metadata can be used as powerful criteria to query image data which match value is the distance measure. Reported experimental results are related to original data. Our results of image retrieval are show that the presented approach performs well. represented using a structural graph by exploiting ontology relationships rather than a listed table. Applying this to retrieve images from the same subject at different levels of resolution opens a new 6506-14, Session 5 field for the analysis of image content. Storage format for personalized broadcasting content consumption 6506-17, Session 6 S. Jin, Information and Communications Univ. (South Korea) Adaptation of video game UVW mapping to 3D In this paper, we propose a storage format which binds digital visualization of gene expression patterns broadcasts with related data such as TV-Anytime metadata, additional P. D. Vize, V. Gerth, Univ. Calgary (Canada) multimedia resources, and personal viewing history. The goal of the proposed format is to make it possible to offer personalized content Analysis of gene expression patterns within an organism plays a consumption after recording broadcasting contents to storage critical role in associating genes with biological processes in both devices, e.g., HD-DVD and Blu-ray Disc. To achieve that, we adopt health and disease. During embryonic development the analysis and MPEG-4 file format as a container and apply a binary format for comparison of different gene expression patterns allows biologists to scenes (BIFS) for representing and rendering personal viewing history. identify candidate genes that may regulate the formation of normal In addition, TV-Anytime metadata is used to describe broadcasts and tissues and organs and to search for genes associated with to refer to the additional multimedia resources, e.g, images, audio congenital diseases. No two individual embryos, or organs, are clips, and short video clips. To demonstrate the usefulness of the exactly the same shape or size so comparing spatial gene expression proposed format, we introduce an application scenario and test it on in one embryo to that in another is difficult. We will present our efforts that scenario. in comparing gene expression data collected using both volumetric and projection approaches. Volumetric data is highly accurate but difficult to process and compare. Projection methods use UV mapping 6506-15, Session 5 to align texture maps to standardized spatial frameworks. This approach is less accurate but is very rapid and requires very little A unified and efficient framework for court-net sports processing. We have built a database of over 180 3D models video analysis using 3D camera modeling depicting gene expression patterns mapped onto the surface of spline J. Han, Technische Univ. Eindhoven (Netherlands) based embryo models. Gene expression data in different models can easily be compared to determine common regions of activity. The extensive amount of video data stored on available media (hard Visualization software, both Java and OpenGL optimized for viewing and optical disks) necessitates video content analysis, which is a 3D gene expression data will also be demonstrated. cornerstone for different user-friendly applications, such as, smart video retrieval and intelligent video summarization. This paper aims at finding a unified and efficient framework for court-net sports video 6506-18, Session 6 analysis. We concentrate on techniques that are generally applicable for more than one sports type to come to a unified approach. To this Classification of yeast cells from image features to learn end, our framework employs the concept of multi-level analysis, pathogen conditions where a novel 3-D camera modeling is utilized to bridge the gap F. J. Verbeek, P. van der Putten, Leiden Univ. (Netherlands) between the object-level and the scene-level analysis. The new 3-D camera modeling is based on collecting features points from two Morphometrics may reveal differences between classes of objects planes, which are perpendicular to each other, so that a true 3-D present in images. We have performed an image-feature-based reference is obtained. Another important contribution is a new classification for the basidiomycetous yeast Cryptococcus tracking algorithm for the objects (i.e. players). The algorithm can neoformans. This yeast can cause infections such as meningitis, track up to four players simultaneously. The complete system meningoencephalitis or skin infections. The presence of an extra- contributes to summarization by various forms of information, of cellular capsule is known to be related to virulence. This paper reports which the most important are the moving trajectory and real-speed of on the approach towards developing classifiers for detecting each player, as well as 3-D height information of objects and the potentially more or less virulent cells in an image of a sample using a semantic event segments in a game. We illustrate the performance of range of features derived from the shape or density distribution. The the proposed system by evaluating it for a variety of court-net sports classifier can henceforth be used for automating screening and videos containing badminton, tennis and volleyball, and we show that annotating existing image collections. We will present a holistic the feature detection performance is above 92% and events detection overview of our methods for creating samples, collecting images, about 90% image preprocessing, identifying “yeast cells” and extracting features from the images. We contrast various expertise based and fully automated methods of feature selection and benchmark a range of 6506-16, Session 6 classification algorithms and discuss successful applications for this particular domain. Ontology driven image search engine in multimodal imaging databases F. J. Verbeek, Y. Bei, J. Dmitrieva, M. Belmamoune, Leiden Univ. 6506-20, Session 7 (Netherlands) Analysis of unstructured video based on camera motion Image collections are most often domain specific. We have developed G. Abdollahian, E. J. Delp III, Purdue Univ. a system for image retrieval of multimodal microscopy images. That is, the same object of study visualized with a range of microscope Although considerable work has been done in management of techniques and with a range of different resolutions. In microscopy, “structured” video such as movies, sports, and television programs image content is depending on the preparation method of the object that has known scene structures, “unstructured” video analysis is still under study as well as the microscope technique. Both are taken into a challenging problem due to its unrestricted nature. The purpose of account in the submission phase as metadata whilst at the same time this paper is to address issues in the analysis of unstructured video ontologies are employed to annotate the image. As a result image and in particular video shot by a typical unprofessional user (i.e. home data are interrelated through the relationships derived from annotated video). We describe how one can make use of camera motion

184 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6506: Multimedia Content Access: Algorithms and Systems information for unstructured video analysis. A new concept, “camera cycles in human motions. Our system can provide three different viewing direction,” is introduced as the building block of home video types of analysis results: (1) foreground person detection; (2) motion analysis. Motion displacement vectors are employed to temporally recognition in the sequence; (3) 3-D modeling of human motion based segment the video based on this concept. We then and the on generic human models. The proposed framework was evaluated correspondence between the camera behavior with respect to the and proved its effectiveness as it achieves the motion recognition and subjective importance of the information in each segment and body-orientation classification at the accuracy of 95% and 98%, describe how different patterns in the camera motion can indicate respectively. levels of interest in a particular object or scene. By extracting these patterns, the most representative frames, keyframes, for the scenes are determined and aggregated to summarize the video sequence. 6506-24, Session 8 A study on video viewing behavior 6506-21, Session 7 S. Mongy, C. Djeraba, Univ. des Sciences et Technologies de Lille Audiovisual shot-reverse-shot analysis (France) J. Nesvadba, Philips Research (Netherlands); B. Kroon, In this paper, we present a study on video viewing behavior. Based on Technische Univ. Delft (Netherlands) a well-suited Markovian model, we have developed a clustering algorithm called K-Models and inspired by the K-Means technique to The proliferation of captured personal and broadcast content in cluster and analyze behaviors. These models are constructed using personal consumer archives necessitates comfortable access to the different actions proposed to the user while he is viewing a video stored audiovisual content. Intuitive retrieval and navigation solutions sequence (play, pause, forward, rewind, jump, stop). We have applied require however a semantic level that cannot be reached by generic our algorithm with a movie trailer mining tool. This tool allows users to multimedia content analysis alone. A fusion with film grammar rules perform searches on basic attributes (cast, director, onscreen date...) can help to boost the reliability significantly. The current paper and to watch selected trailers. With an appropriate server, we log describes the fusion of low-level content analysis cues including face every action to analyze behaviors. First results obtained from a set of parameters and inter-shot similarities to segment commercial content beta users answering to a set of defined questions reveals interesting into film grammar rule-based entities and subsequently classify those typical behaviors. sequences into so-called shot reverse shots, i.e. dialog sequences. Moreover shot reverse shot specific mid-level cues are analyzed augmenting the shot reverse shot information with dialog specific 6506-25, Session 8 descriptions. ARGOS: French evaluation campaign for benchmarking of video content analysis methods 6506-22, Session 7 P. Joly, Univ. Paul Sabatier (France); J. Benois-Pineau, Univ. Edit while watching: home video editing made easy Bordeaux I (France); G. Quenot, Univ. Joseph Fourier (France) M. Campanella, H. Weda, M. Barbieri, Philips Research Labs. The paper presents the Argos evaluation campaign of video content (Netherlands) analysis tools supported by the French Techno-Vision program. This project aims at developing the resources of a benchmark of content In recent years, more and more people capture their experiences in analysis methods and algorithms. Evaluation tasks are related to basic home videos. However, home video editing still is a difficult and time- topics gathering a large number of scientific contributions. They are consuming task. We present the Edit While Watching system that separated in two groups corresponding to two successive evaluation allows users to automatically create and change a summary of a steps in the campaign. The first one lead to the evaluation of low-level home video in an easy, intuitive and lean-back way. Based on content feature extracting tools such as shot segmentation and transition analysis, video is indexed, segmented, and combined with proper effect identification, camera motion identification, shooting location music and editing effects. The result is an automatically generated identification (inside / outside), person presence detection, and text home video summary that is shown to the user. While watching it, presence detection. Tasks of the second group are partially or fully users can indicate whether they like certain content, so that the based on the aggregation or the fusion of the first group results. The system will adapt the summary to contain more content that is similar content set has been developed for various application domains such or related to the displayed content. Edit While Watching does not as video surveillance, broadcast video, video archiving. It includes require a complex user interface: a TV and a few keys of a remote about 40 hours of heterogeneous video contents. This content set control are sufficient. A preliminary user study has shown that it is includes TV news programs, scientific documentaries, and a rich set simple to learn and allows users to achieve predictable and satisfying of video surveillance recordings. The paper describes the type of the editing results with little effort. evaluated tasks, the way the content set has been produced, metrics and tools developed for the evaluations and results obtained at the end of the first phase. 6506-23, Session 8 Multimodule human motion analysis from a monocular video W. Lao, Technische Univ. Eindhoven (Netherlands) In this paper, we propose an effective framework for semantic analysis of human motion from a monocular video. As it is difficult to find a good motion description for humans, we focus on a reliable recognition of the motion type and estimate the body orientation involved in the video sequence. Our framework analyzes the body motion in three modules: a pre-processing module, matching module and semantic module. The proposed framework includes novel object-level processing algorithms, such as a local descriptor and a global descriptor to detect body parts and analyze the shape of the whole body as well. Both descriptors jointly contribute to the matching process by incorporating them into a new weighted linear combination for matching. We also introduce a simple cost function based on time-index differences to distinguish motion types and

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 185 Conference 6507: Multimedia on Mobile Devices 2007 Monday-Tuesday 29-30 January 2007 Part of Proceedings of SPIE Vol. 6507 Multimedia on Mobile Devices 2007

6507-01, Session 1 with an emphasis on the advances and potential in developing direct- view displays further to meet the goal of enabling multimedia in the Color adaptation of videos for mobile devices mobile domain. S. Kopf, W. Effelsberg, Univ. Mannheim (Germany) A large number of videos cannot be visualized on mobile devices, 6507-05, Session 2 such as mobile phones or PDA’s, due to an inappropriate color depth of the displays. Important details would be lost if the color depth was A neighborhood analysis based technique for real-time reduced. Although much work was done on video transcoding only error concealment in H.264 intra pictures few approaches focus on preserving the semantic content of videos. S. T. C. Beesley, C. Grecos, A. J. Armstrong, E. A. Edirisinghe, In the following, we present a novel adaptation algorithm to enable the Loughborough Univ. (United Kingdom) playback of videos on color-limited mobile devices. Dithering algorithms diffuse the error to neighbor pixels and do not work very H.264s extensive use of context-based adaptive binary arithmetic or well for videos. We propose a non-linear transformation of luminance variable length coding makes streams highly susceptible to channel values and the use of textures to reduce the color depth in videos. errors, a common occurrence over networks such as those used by Additionally, we enhance the videos by highlighting significant edges. mobile devices. Even a single bit error will cause a decoder to discard all stream data up to the next fixed length resynchronisation point, the worst scenario is that an entire slice is lost. In cases where 6507-02, Session 1 retransmission and forward error concealment are not possible, a decoder should conceal any erroneous data in order to minimise the Complexity-constrained rated-distortion optimization for impact on the viewer. Wyner-Ziv video coding Stream errors can often be spotted early in the decode cycle of a L. Liu, Purdue Univ.; Z. Li, Thomson Corporate Research; E. J. macroblock which if aborted can provide unused processor cycles, Delp III, Purdue Univ. these can instead be used to conceal errors at minimal cost, even as part of a real time system. This paper demonstrates a technique that A new coding scheme, referred to Wyner-Ziv video coding, was utilises Sobel convolution kernels to quickly analyse the proposed in recent years. The information theoretic principle behind neighbourhood surrounding erroneous macroblocks before the problem dates back to 1970s. The main characteristic of Wyner- performing a weighted multi-directional interpolation. This generates Ziv video coding is that side information is available only to the significantly improved statistical (PSNR) and visual (IEEE structural decoder. Many current Wyner-Ziv video coding schemes encode the similarity) results when compared to the commonly used weighted sequences by two ways, which leads to the division of key frames and pixel value averaging. Furthermore it is also computationally scalable, Wyner-Ziv frames. The key frames are encoded by the conventional both during analysis and concealment, achieving maximum video coding schemes and the Wyner-Ziv frames are encoded by the performance from the spare processing power available. channel coding techniques. At the decoder, the reconstructed key frames are served as the side information to reconstruct the Wyner- Ziv frames. Previously we presented a backward-channel-aware 6507-06, Session 2 motion estimation method to encode the key frames, where motion estimation was performed at the decoder and motion information was Reducing computational complexity of three-dimensional transmitted back to the encoder to save the computational complexity discrete cosine transform in video coding process at the encoder. We refer to these backward predictively coded frames as BP frames. Then the key frames can be I frames, P frames or BP J. J. Koivusaari, J. H. Takala, M. Gabbouj, Tampere Univ. of frames. Rate-distortion (R-D) analysis has been extensively studied to Technology (Finland) address the tradeoff between the rates used to code a video Low complexity video coding schemes are aimed to provide video sequence and the fidelity of the decoded video sequence. In this encoding services also for devices with restricted computational paper, we present a framework for modeling the computational power. Video coding process based on the three-dimensional discrete complexity of the key frames. Based on the complexity model, we cosine transform (3D DCT) can offer a low complexity video encoder propose a rate-distortion optimization of the key frames with by omitting the computationally demanding motion estimation complexity constraint for Wyner-Ziv video coding. operation. In this coding scheme, extended fast transform is also used, instead of the motion estimation, to decorrelate the temporal dimension of video data. Typically, the most complex part of the 3D 6507-04, Session 2 DCT based coding process is the three-dimensional transform. In this Displays enabling mobile multimedia paper, we demonstrate methods that can be used in lossy coding process to reduce the number of one-dimensional transforms required J. S. Kimmel, Nokia Research Ctr. (Finland) to complete the full 3D DCT or its inverse operation. Because unnecessary computations can be omitted, fewer operations are With the rapid advances in telecommunications networks, mobile required to complete the transform. Results include the obtained multimedia delivery to handsets is now a reality. While a truly computational savings for standard video test sequences. The immersive multimedia experience is still far ahead in the mobile world, savings are reported in terms of computational operations and significant advances have been made in the constituent audio-visual computation time. Generally, the reduced number of computational technologies to make this become possible. One of the critical operations also implies longer battery lifetime for portable devices. components in multimedia delivery is the mobile handset display. While such alternatives as headset-style near-to-eye displays, autostereoscopic displays, mini-projectors, and roll-out flexible 6507-07, Session 3 displays can deliver either a larger virtual screen size than the pocketable dimensions of the mobile device can offer, or an added Restructuring a software based MPEG-4 video decoder degree of immersion by adding the illusion of the third dimension in for short latency hardware acceleration the viewing experience, there are still challenges in the full deployment of such displays in real-life mobile communication O. J. Silvén, J. J. Boutellier, T. Erdelyi, Univ. of Oulu (Finland) terminals. Meanwhile, direct-view display technologies have The multimedia capabilities of emerging high-end battery powered developed stadily, and can provide a development platform for an mobile devices rely on monolithic hardware accelerators that have even better viewing experience for multimedia in the near future. The long latencies to minimize the interrupt and software overheads. paper presents an overview of the mobile display technology space When compared to pure software implementations the monolithic

186 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6507: Multimedia on Mobile Devices 2007 hardware accelerator solutions need an order of magnitude less implemented because of the computational limitations of these power, however, they are rather inflexible and difficult to modify to devices. provide support for multiple coding standards and bit streams. A more In this paper, we present an efficient implementation of a state-of-the- flexible alternative is to employ multiple short latency accelerators that art video denoising algorithm based on three-dimensional discrete implement individual functions needed in the codecs. Unfortunately, cosine transform (3D DCT) and Block Matching (BM). Noise the software overheads become prohibitively high, if interrupts are suppression is locally performed in 3D arrays size of 8x8x8 sliding used for software/hardware synchronization. Instead, means to over the video data in three directions (two spatial and one temporal). interface hardware accelerators to software preferably at the same For every chosen spatial position in each frame, we use a block- cost as respective software functions are needed. We present a matching algorithm to estimate similar 2D blocks of size 8x8 from restructuring of an MPEG-4 decoder task in a manner that enables the neighboring pictures and stack the matched blocks to form a 3D simultaneous decoding of multiple bit streams using short latency array. The forward 3D-DCT transform is applied to the selected 3D hardware accelerators. The approach takes multiple streams as input array, and then, locally adaptive hard-thresholding is used on its and produces a multiplexed stream that in turn controls the coefficients to attenuate noise. The inverse 3D-DCT reconstructs a accelerators without interrupts. In principle, the decoding of each local estimate of the noise-free signal in the array. After processing of stream can be considered as a separate thread that shares the same all spatial positions of the current frame, we achieve a set of local hardware resources with the others. The results based on software estimates which constitute an over-complete representation of the simulations show that the energy efficiency falls in between a noise-free estimated signal. This is due to overlaps between sliding monolithic hardware accelerator and pure software implementation. 3D window. The final denoised video picture is constructed as a weighted average of the overlapping local estimates. This algorithm, apart from the architectural simplicity, also allows for computational 6507-08, Session 3 scalability because of the sliding-window based processing. Main Hand-held analog television over WiMAX executed in SW components of the algorithm are 8-point DCT and block matching, both of which can be efficiently computed with TI C55x DSP hardware D. S. Iancu, H. Ye, M. Senthilvelan, V. Kotlyar, J. C. Glossner, M. extension. Moudgill, S. Agrawal, S. Jinturkar, A. Iancu, G. Nacer, S. Stanley, Our target platform is the OMAP Innovator development kit, a dual Sandbridge Technologies, Inc.; J. H. Takala, Tampere Univ. of processor environment including ARM 925 RISC general purpose Technology (Finland) processor (GPP) and TMS320C55x digital signal processor (DSP). The This paper describes a device capable of performing the following C55x DSP offers hardware acceleration support to compute the tasks: it samples and decodes the composite video Analog TV signal, transform as well as the motion estimation (block matching) required it encodes the resulting RGB data into an MPEG-4 stream and sends by the denoising algorithm. Hardware acceleration can offer notable it over a WiMAX link. On the other end of the link a similar device performance increase in comparison to assembler optimization of receives the WiMAX signal, in either TDD or FDD mode, decodes the source codes. MPEG data and displays it on the LCD display. The device can be a The result will demonstrate that the video denoising algorithm can be hand held device, such as a mobile phone or a PDA. The algorithms efficiently realized also on mobile devices with limited computational for the Analog TV, WiMAX physical layer, WiMAX MAC and the MPEG resources. encoder/decoder are executed entirely in software in real time, using the Sandbridge Technologies low power SB3011 digital signal processor. The SB3011 multithreaded digital signal processor 6507-10, Session 3 includes four DSP cores with eight threads each, and one ARM processor. The execution of the algorithms requires the following Multimedia support for J2ME on high-end PDAs processing resources: 16 threads for the WiMAX receiver, 8 threads T. Preuss, M. Rambow, M. Lehmann, Brandenburg Univ. of for the WiMAX transmitter, 4 threads for the MPEG encoder, one Applied Sciences (Germany) thread for the MPEG decoder and 3 threads for the Analog TV. The WiMAX MAC is executed on the ARM processor. The total power Platform independence is the major advantage of the Java consumption for receive mode is less than 200mW. programming language. Whereas Java is widespread on servers, desktop computers and mobile phones, PDA and Pocket PC applications are usually based on C or C# applications based on the 6507-09, Session 3 .NET framework. The paper will focus on the J2ME standard and its suitability for Pocket PCs and PDA. Further an overview of existing Efficient implementation of DCT-based video denoising Java Virtual Machines (JVMs) will be given. In detail we will focus on algorithm on OMAP(tm) platform Esmertecs Jbed CDC, IBM Websphere Everyplace, Creme and D. Rusanovskyy, J. J. Koivusaari, K. O. Egiazarian, J. H. Takala, MySaifu. These different JVMs will be compared according functional criterias as well as standard conformance and performance. Further Tampere Univ. of Technology (Finland) we will run a set of tests to benchmark these different JVMs. Finally, The recent progress in the development of embedded systems has we will implement an example application as part of the Bosporus enabled hardware devices to be small with increasing computational project to evaluate the JVMs from the programers perspective. power. However, even with these developments, there exists limited number of video applications for mobile devices. Due to the high computational requirements of video processing algorithms, they 6507-11, Session 4 need to be optimized for the target platform enabling them to meet the resource constraints. In general, optimization of an algorithm is Underflow prevention for AV streaming media under quite a time consuming task and is highly dependent on the target varying channel conditions platform. In many cases, development resources are not spent to fine- S. G. Deshpande, Sharp Labs. of America, Inc. tune the algorithms. Instead, people tend to wait for new hardware with more processing power to implement their algorithms. On the We propose a method to prevent receiver buffer underflow for AV other hand, Texas Instrument Innovator Development Kit for OMAP streaming media under varying channel conditions using a variable processing platform offers a versatile development environment to scale factor adaptive media playout algorithm. Our proposed implement sophisticated video processing algorithms also for algorithm dynamically calculates a slow and/or fast playout factor embedded systems. based on the current buffer state, target buffer level, past history of media data reception, estimate of future data arrival, content One example of a challenging video processing problem is video characteristics and the estimated current network conditions. As a noise reduction. Due to image sensor imperfections, digital video are result our algorithm results in prevention of underflow in situations often contaminated with noise. This is especially true in the case of where prior art approach can still result in underflow. Our algorithm cameras in low-cost mobile terminals. Despite the need for camera can also avoid oscillations between slow and fast playout. Also noise reduction in such devices, the noise reduction is rarely variants of our algorithm can result in more smooth transition to

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 187 Conference 6507: Multimedia on Mobile Devices 2007

normal playout from adaptive playout stage thus improving the and interactive data in mobile environments. The T-DMB standard is perceptual user experience. The proposed algorithm can also be used based on the Eureka-147 Digital Audio Broadcasting (DAB) System. to reduce initial buffering latency at the start of a media stream Since the DAB system is designed for digital audio and data services, playback while achieving same robustness by reaching the desired various MPEG-4 technologies such as MPEG-4 AVC, MPEG-4 BSAC target buffer level. We present extensive network simulation results for and MPEG-4 Systems are adopted for multimedia services. our proposed approach. Especially, MPEG-4 Systems technology enables various types of interactive data services for T-DMB. In MPEG-4 Systems, broadcasting contents are described by the combination of a large 6507-12, Session 4 number of nodes, routes and descriptors. In order to provide interactive data services through the T-DMB network, it is essential to Seamless content consumption experience across media have an interactive contents authoring system which allows contents delivery channels and devices authors to compose interactive contents easily and conveniently even C. M. Taskiran, B. Gandhi, W. Wang, D. Li, Motorola, Inc. if they lack any background on MPEG-4 Systems technology. This paper introduces an interactive contents authoring system which can The digital availability of content is accelerating the separation easily and conveniently produce interactive contents for T-DMB. The between content and its delivery medium. However, current content introduced authoring system provides powerful graphical user delivery and consumption paradigms artificially bind the digital interface and produces interactive broadcasting contents in the forms content to specific delivery mechanisms (e.g., broadcast, cable, etc.) of binary and textual format which are called mp4 (MPEG-4 file specifically targeting specific devices (e.g., set-top-box / TV, Personal format) and XMT (eXtensible MPEG-4 Textual format), respectively. Computer, etc.). Moreover, a new paradigm of content delivery is emerging where the traditional service providers are delivering a variety of content on 6507-15, Session 4 existing channels. For example, cable operators have been moving Musical slideshow MAF with protection and governance from a pure broadcast mode to allow a pull mechanism via on demand technologies. Furthermore, they are providing music content using MPEG-21 IPMP components and REL in addition to traditional video content. There is also an emergence of M. S. H. Sabirin, H. Tan, M. Kim, Information and new service providers who are typically web-based that are Communications Univ. (South Korea) supporting additional services such as video, images, and music in addition to text-based web-pages. The Musical Slide Show Multimedia Application Format (MAF) which currently being standardized by the Moving Picture Expert Group There is also a trend on the part of the user to want time-shifted [1] (MPEG) conveys the concept of combining several established and place-shifted [2, 3] consumption experiences. standard technologies in a single file format. It defines the format of Observing this trend, we are developing media association technology packing up MP3 audio data, along with MPEG-7 Simple Metadata to enable a seamless consumption experience that blurs the Profile and MPEG-21 Digital Item Declaration metadata, with JPEG boundary between delivery channels and across devices. We use the images and optional text, and synchronizes them all together to flow diagram shown in Figure 1 to illustrate the process of seamless create a slideshow of JPEG image data associated to MP3 audio data content consumption. during the audio playback. The implementation of MPEG Musical Slide Show MAF can be applied in a multimedia rich-contents file such as a music karaoke file or a story-telling file. By standardizing 6507-13, Session 4 the format for combining already standardized and verified resources, we can expect to reduce the effort and time consumed in producing a MDC image coding using cascaded correlating new file format. Moreover, to avoid violation in obtaining the content transforms from legal distribution chain, we need to have the content protected. R. Knoerig, T. Sikora, Technische Univ. Berlin (Germany) In this paper, we are presenting the implementation of our proposal in defining the file format of Musical Slide Show MAF with an We describe a joint source-channel coding framework combining enhancement in adding other established standards: MPEG-21 cascaded correlating transforms as proposed by Goyal with an Intellectual Property Management and Protection (IPMP) and MPEG- optimal estimation algorithm in the MSE sense. 21 Rights Expression Language (REL) metadata to create a protected The cascaded correlating transform, an extension of the well-known content. pairwise correlating transform to transforms of higher order, can be seen as a detunable decorrelating transform. By reducing the transforms ability to decorrelate, a higher amount of source 6507-16, Session 5 correlation “survives” in the signal. The digital aura, a framework for a sensible mobile This increased redundancy will be used for concealing channel errors. phone entertainment experience: a future-oriented Since the detuning can be performed stepless an arbitrary amount of evaluation of possibilities of applying sensor technology redundancy can be selected, allowing ?ne-tuned trade-offs between on mobile phones toward increasing the consumer coding efficiency and robustness to channel errors. This is an entertainment experience advantage over the classic approach by combining source- and channel coders since even shortened convolution coders offer only a A. R. Lugmayr, Jr., Tampere Univ. of Technology (Finland) discrete and therefore not stepless set of coding rates. Moreover, our Ambient media are media, which are manifesting in the natural approach affects only the transform and the inverse transform stages environment of the consumer. The perceivable borders between the and will be transparent to other stages of the coding system (e.g. media and the context, where the media is used are getting more and quantization or entropy coding). more blurred. The consumer is moving through a digital space of services throughout his daily life. As we are developing towards an experience society, the central point in the development of services is 6507-14, Session 4 the creation of a consumer experience. This paper reviews The interactive contents authoring system for terrestrial possibilities and potentials of the creation of entertainment digital multimedia broadcasting experiences with mobile phone platforms. It reviews sensor network capable of acquiring consumer behaviour data, interactivity strategies, W. Cheong, S. Ahn, J. Cha, K. Moon, Electronics and Telecommunications Research Institute (South Korea) Terrestrial Digital Multimedia Broadcasting (T-DMB) is an upcoming mobile multimedia broadcasting service which provides audio, video

188 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6507: Multimedia on Mobile Devices 2007 psychological models for emotional computation on mobile phones, 6507-20, Session 6 and lays the foundations of a nomadic experience society. The paper rounds up with a presentation of several different possible service Smooth transitions for mobile imagery browsing scenarios in the field of entertainment and leisure computation on R. U. Rosenbaum, Sr., H. Schumann, Univ. Rostock (Germany) mobiles. The goal of this paper is to present a framework and evaluation of possibilities of applying sensor technology on mobile Modern mobile devices allow taking and handling photos, and thus, platforms to create an increasing consumer entertainment experience. must provide options to browse image galleries. To lower the cognitive efforts imposed by the switch between two images, smooth transitions are mandatory. A common approach to do so is Alpha 6507-17, Session 5 blending applied in pixel domain. However, this technique is not appropriate in mobile environments. To enable the application of Sensometrics: identifying pen digitizers by statistical transitions, a completely new approach founded on the new image multimedia signal processing compression standard JPEG2000 is proposed. As within a A. Oermann, C. Vielhauer, Otto-von-Guericke-Univ. Magdeburg JPEG2000-data-stream encoded image parts can be independently (Germany) accessed, the blending is accomplished by successively replacing data within the source image by the associated data of the destination In this paper a new approach will be introduced to identify pen-based image. By introducing a certain criterion, the new data perfectly fits at digitizer devices based on handwritten samples used for biometric the respective position within the old data-stream. A frequent user authentication. This new method of digitizer identification based decoding of the modified image during the replacement procedure on their signal properties can also be seen as an influencing part in leads to the aspired blending effect. There are different variances of the new research area of so-called Sensometrics. The goal of the the exchange procedure and options to steer the blending. The work presented in this paper is to identify statistical features, derived introduced approach reduces the need for the critical system from signals provided by pen-based digitizer tablets during the writing resources by blending the contents in JPEG2000 domain (computing process, which allow identification, or at least group discrimination of power) and a sophisticated transmission strategy (bandwidth). different tablet types. Based on a database of a total of 5754 writing samples taken on 23 different digitizer tablets, specific features for class discrimination will be chosen and a novel feature vector based 6507-21, Session 6 classification system will be implemented and experimentally validated. The goal of our experimental validation is to study the class Experienced quality factors: qualitative evaluation space that can be obtained, given a specific feature set, i.e. to which approach to audiovisual quality degree single tablets and/or groups of tablets can be identified using S. Jumisko-Pyykkö, Tampere Univ. of Technology (Finland); J. P. the suggested scheme under the constraint of zero false Häkkinen, Nokia Research Ctr. (Finland) classifications. The results confirm that a group discrimination of devices can be achieved. By applying this new approach, the 23 Subjective evaluation is used to identify impairment factors of different tablets from our database can be discriminated in 19 output multimedia quality. The final quality is often formulated via quantitative groups. experiments, but this approach has its constraints, as subject’s quality interpretations, experiences and quality evaluation criteria are disregarded. To identify these quality evaluation factors, this study 6507-18, Session 5 examined qualitatively the criteria participants used to evaluate audiovisual video quality. A semi-structured interview was conducted Semantic consumption of photos on mobile devices with 60 participants after a subjective audiovisual quality evaluation S. Lee, Information and Communications Univ. (South Korea) experiment. The assessment compared several, relatively low audio- video bitrate ratios with five different television contents on mobile Recently, digital camera becomes a powerful, popular means to device. In the analysis, methodological triangulation (grounded theory, capture one’s daily life. It is also possible and getting popular to neural networks and correspondence analysis) was applied to access to photos on portable multimedia devices such as PDA and approach the qualitative quality. The results showed that the most mobile phone. However, while most users want to consume their important evaluation criteria were the factors of visual quality, photos in semantically meaningful order, most of the existing photo contents, factors of audio quality, usefulness - followability and consumption tools produce low quality of service in a sense of audiovisual interaction. Some relations between the quality factors semantic photo browsing. A few album tools support useful semantic and the similarities between the contents were identified. As the metadata, e.g., ITPC in Exif, for effective photo browsing. However, research methodological recommendation, the focus on content and the metadata is unstructured and hard to be interpreted since it is usage related factors need to be further examined to improve the generated by free annotation. Furthermore, any application-oriented quality evaluation experiments. metadata is often dysfunctional in different applications due to lack of compatibility. In this paper, we propose a promising photo consumption for mobile application, which enables augmented use of 6507-22, Session 7 digital photos over a wide range of mobile applications as well as semantic photo consumption with minimal user’s manual tasks. The Tablet PC interaction with digital micromirror device proposed photo album consists of interactively coupled two solo (DMD) applications: collection engine and photo player. Collection engine is responsible for generation of collection of photos and associated H. H. Refai, J. J. Sluss, Jr., Univ. of Oklahoma metadata while photo player is used to consume the photo collection Digital light processing (DLP) is a revolutionary display technology that files. This sort of packaging could give an effective way to synchronize uses an optical switch array, known as a digital micromirror device photos and associated metadata. (DMD), which allows digital control of light. To date, the main successes of the DMD have been as high-speed spatial light modulators for projector applications. Tablet PC is a notebook or 6507-19, Session 6 slate-shaped mobile PC. Its touch screen or digitizing tablet Wireless video sensor networks: convergence of video, technology allows the user to operate the notebook with a stylus or digital pen instead of using a keyboard or mouse. In this paper, we wireless, and vision developed an interface solution that translates any sketch in the tablet Z. He, Univ. of Missouri/Columbia screen to an identical copy over the DMD micromirrors to be No abstract available

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 189 Conference 6507: Multimedia on Mobile Devices 2007

projected on a special screen. An algorithm has been created to 6507-25, Session 7 control each single micromirror of the hundreds of thousands of micromirrors that cover the DMD surface. We demonstrate the Interfaces for mobile image browsing successful application of a DMD to a high-speed 2D scanning R. U. Rosenbaum, Sr., H. Schumann, Univ. Rostock (Germany) environment, acquiring the data from the tablet screen and launching its contents to the projection screen; with very high accuracy up to The available screen space of mobile hardware is usually rather small 13.68 um x 13.68 um of mirror pitch. The authors are not aware of any leading to the problem that not the whole image fits on screen. This reports in the literature where the DMD has been evaluated for this often leads to a lost in space situation of the user. This can be application. We present data from measurements of optical power overcome by an appropriate user interface. Such an interface consists and image sketching which are very promising and indicates that the of the two main parts content representation and means for DMD is suitable for interaction with tablet PC and rapidly controls the interaction. The goal of content representation is the communication micromirrors according to the acquired data from the motion tracking of relevant image information and is usually implemented by of the user’s pen. displaying regions associated to the current user interests. This also requires means for their interactive determination. To consider the fact that there is no approache appropriate for all kind 6507-23, Session 7 of data, device, task or user preference, 3 different interfaces for TiDi browser: a novel photo browsing technique for mobile image browsing - Grid-based Zoom&Pan, the largeDetail-View, mobile devices and the localLens-View are introduced. Beside statements to their function, each technique is evaluated regarding their properties to C. Tominski, Univ. Rostock (Germany); G. Bieber, B. Urban, reduce the most limiting factors in mobile environments -screen Fraunhofer-Institut für Graphische Datenverarbeitung (Germany) space, computing power, and bandwidth. To decrease the required bandwidth, for each technique a strategy for dynamic image Today’s digital photos can be tagged with information about when and streaming based on JPEG2000 is proposed. Overall, a significant where they were taken. On stationary computers, this information is enhancement regarding representation, interaction and use of often used to drive photo browsing. This is not the case for mobile bandwidth compared to traditional strategies could be achieved. devices. We describe preliminary results of our work on a novel photo browsing technique called the TiDi Browser. The TiDi Browser exploits time and location information available in digital photos. Along with a 6507-26, Session 8 main view and thumbnail previews, our browser application provides two timelines. One timeline visualizes the number of photos taken per Novel layered scalable video coding transmission over temporal unit (e.g., day, week, etc.). This allows users to easily detect MIMO wireless systems with partial CSI and adaptive personal events in time. The second timeline visualizes location channel selection information. Since two- or three-dimensional locations are difficult to represent on small displays, we reduce the location information to D. Song, C. W. Chen, Florida Institute of Technology one-dimensional distance information. The distance is shown in the We described in this paper a practical layered scalable video coding second timeline. Both timelines serve a second purpose as graphical (SVC) transmission scheme over multi-input multi-output (MIMO) user interface, meaning that they can be used to browse in time. We wireless system with considering both channel estimation error and implemented our ideas in an interactive prototype that uses a client- information feedback delay. This scheme makes proper use of the server-architecture. To save bandwidth, we transmit appropriately feedback of the estimated partial channel state information in terms of scaled photos that fit the display dimensions of the client (mobile the ordering of each sub-channel’s SNR strength. Based on the device). To enhance the user’s browsing experience, we apply caching feedback of the estimated partial channel state information (CSI), and pre-fetching strategies. unequal error protection (UEP) for layered SVC can be achieved by the proposed adaptive channel selection (ACS) algorithm: base-layer 6507-24, Session 7 is launched to the best sub-channel and the highest enhancement- layer to the worst sub-channel among multiple transmitters. This Of MOS and men: bridging the gap between objective proposed scheme enables us to both overcome the challenge of the and subjective quality measurements in mobile TV perfect CSI that is assumed in the existing approaches and utilize the estimated CSI from the receiver so as to achieve UEP for layered SVC T. C. M. de Koning, P. Veldhoven, TNO TPD (Netherlands); H. O. transmission over MIMO system. In addition, the automatically Knoche, Univ. College London (United Kingdom); R. E. Kooij, TNO obtained UEP in this scheme does not require any power control at TPD (Netherlands) the transmitter. It was shown by various simulation results that the proposed ACS system provides a better PSNR performance and In this paper we explore the relation between the binary video quality shows the robustness against incorrect partial CSI in terms of channel measure Acceptability and the objective quality measure Mean estimation error and time delay. Opinion Score (MOS). We obtained objective MOS values from previously used video (football, news, music, and animation) clips using the video quality measuring tool VQM. Using the ITU defined 6507-27, Session 8 mapping (M2G) from MOS to binary Good or Better (GoB) values, we compared the M2G translated values to the clips’ subjective An efficient client-based JPEG2000 image transmission Acceptability scores at various encoding bitrates (32-224kbps) and protocol sizes (120x90, 168x126, 208x156 and 240x180). The results show that in the domain of mobile TV the ITU mapping results in a serious J. P. García Ortiz, Univ. de Almería (Spain) overestimation of Acceptability. The mapping M2A, between MOS and JPEG2000 is a new still image coding standard that allows to Acceptability, that we suggest provides a significant improvement of implement efficient applications for remote image browsing. In this 76% in the root mean square error (RMSE) over M2G. We show that kind of developments, the clients retreives from a server only specific Acceptability depended on more than just the visual quality and that regions of the remote images. In the Part 9 of this standard is defined both content type and size are essential to provide accurate estimates the JPIP protocol, offering a complete set of definitions and methods of Acceptability in the field of mobile TV. We illustrate this gain in for the remote browsing of JPEG 2000 images. Using the JPIP Acceptability predictions for the popular content type football protocol, the server side is where all the hard processes are done, so (soccer). In terms of RMSE our content dependent mapping (M2Af) the clients have only to request the desired region and wait for the yielded an improvement of 39% over M2A. Future research will associated information. This little advantage has many other validate the predictive power of our suggested mapping on other disadvantages like, for example, to not support the partial proxy video material. caching, to difficult the implementation of improvement techniques like data prefetching or to offer a poor behaviour in wireless

190 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6507: Multimedia on Mobile Devices 2007 environments. An image transmission protocol with a more complex intuitive. In order to overcome these problems, we are developing client side avoids all these disadvantages, only requiring a bit more of novel, intuitive interaction methods for mobile devices. They are process load in clients. In this paper it is analyzed the current JPIP based on the detection of 3D self-motion of the phone using the protocol, showing how affect its philosophy in the mentioned integrated camera. Possible movements involve changes in position, situations. A new JPEG2000 image transmission protocol is pitch, roll and yaw. The motion tracking has to be performed in real- presented, J2KP. Its performance is compared with JPIP and it is time under the constraints of the limited memory and computing demostrated that all the features rejected by this protocol offer a resources available on a handheld device. The movements can then considerable increment of performance when are applied to the be used to control various applications on the device. Specifically, we proposed protocol. Moreover, the process overload in the clients is have developed a mobile car racing game that is controlled by moving nearly negligible. the device. We also conducted a user study to test our system for usability, ease of interaction, responsiveness and sensitivity of the system. 6507-28, Session 8 Error resilient image transmission based on virtual SPIHT 6507-33, Poster Session R. Liu, J. He, X. Zhang, Beijing Univ. of Aeronautics and Bidirectional traffic status information service based on Astronautics (China) T-DMB In this paper, we integrate new error resilient tools into wavelet coding Y. Jeong, S. Kim, S. Cho, G. Kim, C. Ahn, S. Lee, Electronics and algorithm and present an error-resilient image transmission scheme based on virtual set partitioning in hierarchical trees (SPIHT), EREC Telecommunications Research Institute (South Korea) and self truncation mechanism. The virtual spatial-orientation trees in T-DMB (Terrestrial-Digital Multimedia Broadcasting) can provide a the wavelet domain are individually encoded using virtual SPIHT. variety of data services such as traffic and travel information (TTI), Since the self-similarity across sub bands is preserved, a high source EPG and BWS. Among these data services, TTI service has been coding efficiency can be achieved. The scheme is essentially a tree- spotlighted in aspects of economic influence and information usability. based coding, thus error propagation is limited within each virtual In order to providing TTI services, TPEG protocol was developed. tree. EREC is also adopted to enhance the error resilience capability However, it has been applied merely to two application areas such as of the compressed bit streams. At the receiving side, the self- RTM and PTI. One is to transfer the road traffic status message (RTM) truncation based on self constraint mechanism of set partition trees is in case of an accident, out-of-door gathering, and so forth. The other introduced. The decoding of any sub-tree halts in case the violation of is to convey the public transport information message (PTI) such as self-constraint relationship occurs in the tree. So the bits impacted by schedules and routes of buses, trains, flights, ships, and so forth. the error propagation are limited and more likely located in the low bit- Recently, the traffic status image information (TSI) service based on layers. In additional, inter-trees interpolation method is applied, thus still or moving pictures becomes more important because public some errors are compensated. Preliminary experimental results confidence about the accuracy of currently used coded-traffic demonstrate that the proposed scheme can achieve much more information is low. In this paper, we propose the novel bidirectional benefits on error resilience. TSI service, which is satisfied to be interoperable with TPEG, and verify the stability and effectiveness of that. The TSI application protocol is designed to overcome the drawbacks of conventional RTM 6507-29, Session 8 service, which cannot provide any subjective decision chance to users. The TSI enables us to easily access the close circuit TV (CCTV) Complexity analysis and control in joint channel information of the desired location, and it can be linked to the traffic protection system for wireless video communication information provided by RTM service. X. Jin, G. Zhu, Huazhong Univ. of Science and Technology (China) In the wireless communication, channel coding and error control are 6507-34, Poster Session essential to protect the video data from wireless interference. The power it consumed, which is determined by the protection method it Codesign toolset for application-specific instruction set used, will directly affect the system performance especially in processors decoding side. In this paper, a channel coding and error control system, called joint channel protection (JCP) system here, is proposed P. Jääskeläinen, V. Guzma, A. Cilio, J. H. Takala, T. Pitkänen, as an improvement of the hybrid automatic repeat request (HARQ) Tampere Univ. of Technology (Finland) system to integrate the complexity controllability. The complexity Application-specific programmable processors tailored for the models of the encoder and decoder are established based on requirements at hand are often at the center of today’s embedded theoretical analysis and statistical data retrieval using the time systems. Therefore, it is not surprising that considerable effort has complexity concept, and the relative variation in the computational been spent on constructing tools that assist in co-designing complexity is carefully studied to provide a proportional variation application-specific processors for embedded systems. It is desirable reference for complexity control. Based on the models, strategies are that such design toolsets support an automated design flow from designed to control the system complexity by adjusting the packet application source code down to synthesizable processor description length, iterative decoding times and retransmission ratio according to and optimized machine code. TTA Co-design Environment (TCE) the decoding quality and complexity level. developed at Tampere University of Technology (TUT) is one such toolset. The customizable processor architecture template that forms the basis of TCE is a VLIW-derived architecture paradigm called 6507-30, Poster Session Transport Triggered Architecture (TTA). Intuitive user interface for mobile devices based on TCE addresses some of the pressing shortcomings found in existing visual motion detection toolsets, such as lack of automated exploration of the “design space”, limited run time retargetability of the design tools or restrictions in the S. Winkler, K. Rangaswamy, S. Zhou, National Univ. of Singapore customization of the target processors. On the other hand, existing (Singapore) toolsets can often be used for wider range of applications than TCE. Mobile phones and PDA’s are becoming increasingly powerful and This is due to the limitations placed by the use of TTA as the provide more and more entertainment functionality to the user. processor template in the toolset. The presented poster gives an However, the user interface is severely limited due to the small form overview of the TCE toolset, and outlines future research around TTA factor of these devices. Using the small, un-ergonomic keys for and automated processor design toolsets at TUT. control in applications other than calling is tedious and not always

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 191 Conference 6507: Multimedia on Mobile Devices 2007

6507-35, Poster Session Dynamic power management for UML modeled applications on multiprocessor SoC P. Kukkala, T. Arpinen, M. Setälä, M. Hännikäinen, T. D. Hämäläinen, Tampere Univ. of Technology (Finland) The paper presents a novel scheme of dynamic power management for UML modeled applications that are executed on a multiprocessor System-on-Chip (SoC) in a distributed manner. The UML models for both application and architecture are designed according to a well- defined UML profile for embedded system design, called TUT-Profile. Application processes are considered as elementary units of distributed execution, and their mapping on a multiprocessor SoC can be dynamically changed at run-time. Our approach on the dynamic power management balances utilized processor resources against current workload at runtime by (1) observing the processor and workload statistics, (2) re-evaluating the amount of required resources (i.e. the number of active processors), and (3) re-mapping the application processes to the minimum set of active processors. The inactive processors are set to a power-save state by using clock- gating. The approach integrates the well-known power management techniques with the UML based design of embedded systems in a novel way. We evaluated the dynamic power management with a WLAN terminal implemented on a multiprocessor SoC on Altera Stratix II FPGA containing up to five Nios II processors and dedicated hardware accelerators. Measurements proved up to 21% savings in the power consumption of the whole FPGA board.

6507-36, Poster Session The evolution of ambient learning R. Creutzburg, Fachhochschule Brandenburg (Germany) During the last years a lot of new technologies for distance and e- learning have been developed. Several new terms andn notions have been created: d-learning, e-learning, m-learning, ubiquitous learning, ambient learning, .... The aim of this paper is to give an overview of these developments and analyse the terms carefully.

6507-37, Poster Session Ambient learning in a mobile world R. Creutzburg, Fachhochschule Brandenburg (Germany) The aim of this paper is to give an overview of the new developments and the new paradigms in e-learning that result in future ambient learning in a mobile world.

192 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007 Tuesday-Thursday 30 January-1 February 2007 Part of Proceedings of SPIE Vol. 6508 Visual Communications and Image Processing 2007

6508-01, Session 1 6508-04, Session 1 Real-time video coding under power constraint based on Comparison of standard-based H.264 error-resilience H.264 CODEC techniques and multiple-description coding for robust L. Su, Institute of Computing Technology (China); Y. Lu, F. Wu, S. Li, MIMO-enabled video transmission Microsoft Research Asia (China); W. Gao, Institute of Computing M. Tesanovic, D. R. Bull, D. Agrafiotis, A. Doufexi, Univ. of Bristol Technology (China) (United Kingdom) In this paper, we propose a joint rate-distortion-complexity optimization Significant enhancements in spectral efficiency and error-resilience scheme for real-time H.264 video encoding under the power constraint. must be obtained from emerging wireless video systems if high-quality Firstly, the power constraint is translated to the complexity constraint multimedia transmission is to become cost-effective. MIMO (multiple- with quantitative analysis. Secondly, a computation allocation model input-multiple-output) systems offer potential for throughput increase (CAM) with virtual computation buffers (VCB) is proposed to facilitate and enhanced quality of service for multimedia transmission. Previous the optimal allocation of constrained computational resource for each research into MIMO algorithms has not focused on its optimisation frame. Thirdly, the complexity adjustable encoder based on adaptive specifically for video transmission. MIMO is a generic technology motion estimation and mode decision is developed to meet the which, depending on the actual architecture deployed and the channel allocated resource. The proposed scheme takes the advantage of the conditions encountered, produces variable error conditions. New error- advanced video coding tools such as early termination strategy in fast resilience techniques are required if the obtained benefits are to be fully motion estimation. Moreover, it can avoid suffering from the high exploited. This paper proposes the use of multiple-description coding overhead of the parametric power control algorithms and achieve fine (MDC) as an approach that outperforms the standard-based error- complexity scalability in a wide range with stable rate-distortion resilience techniques in the majority of these cases. performance. The proposed scheme also shows the potential of a Results obtained from the random packet-error generator are furthered further reduction of computation and power consumption in the through the use of realistic MIMO channel scenarios and argue in decoding without any change on the existing decoders. favour of the deployment of an MDC-based video transmission system. Singular value decomposition (SVD) is used to create orthogonal sub- channels within a MIMO system which provide an efficient means of 6508-02, Session 1 mapping video content. In addition, resource allocation in MIMO-SVD A low bit-rate video coding approach using modified systems (transmit power allocation and assignment of descriptions to adaptive warping and long-term spatial memory singular values) is examined in the context of transmitted video quality. Results indicate improvements in average PSNR of decoded test- Y. Chen, M. J. T. Smith, E. J. Delp III, Purdue Univ. sequences of up to 3 dB (5dB in the region of high PERs) compared to standard, single-description video transmission. This is also supported In this paper, we introduce an approach to further exploit long-term by significant subjective quality enhancements. These results provide a dependencies in a way that is compatible with the existing H.264/AVC framework for future wireless video transport. framework. Using a hybrid spatio-temporal subband representation, we attempt to capture high frequency spatial information at the receiver side in coordination with the encoder to improve the subjective 6508-05, Session 1 reconstruction quality. A key aspect of the proposed coding scheme is an adaptive warping method, which is used to perform motion Spatial and temporal models for texture-based video estimation and reconstruction at the receiver side. The approach is coding shown to yield improved subjective quality for sequences with low to moderate motion with many high frequency details. F. Zhu, Purdue Univ. In the past two decades, conventional hybrid video codecs have succeeded in increasing the video quality while reducing the data rate. 6508-03, Session 1 One way to increase the coding efficiency beyond the data rates Rate-prediction structure complexity analysis for achievable by modern codecs, such as H.264, is to not encode all the multiview video coding using hybrid genetic algorithms pixels in the sequence. In early coding systems this was achieved by either reducing the size of the frame and/or a combination of frame Y. Liu, Q. Dai, Z. You, W. Xu, Tsinghua Univ. (China) skipping. The quality of the reconstructed sequence was, of course, reduced. In this paper, we investigate spatial and temporal models for Efficient exploitation of the temporal and inter-view correlation is critical texture analysis and synthesis. The goal is to use these models to to multi-view video coding (MVC),and the key to it relies on the design increase the coding efficiency for video sequences containing textures. of prediction chain structure according to the various pattern of The models are used to segment texture regions in a frame at the correlations. In this paper, we propose a novel prediction structure encoder and synthesize the textures at the decoder. These methods model to design optimal MVC coding schemes along with tradeoff can be incorporated into a conventional video coder (e.g. H.264) where analysis in depth between compression efficiency and prediction the regions to be modeled by the textures are not coded in a usual structure complexity for certain standard functionalities. Focusing on manner but texture model parameters are sent to the decoder as side the representation of the entire set of possible chain structures rather information. We showed that this approach can reduce the data rate by than certain typical ones, the proposed model can give efficient MVC as much as 15%. schemes that adaptively vary with the requirements of structure complexity and video source characteristics (the number of views, the degrees of temporal and interview correlations). Our contributions 6508-06, Session 1 include an evaluation model for chain efficiency, a rate-structure complexity optimization model, and a hybrid genetic algorithm. These Content adaptive motion estimation for efficient video three form a rate-structure complexity analysis framework which compression establishes a theoretical basis and provides a practical guideline in coding design and performance optimization for multi-view video L. Liu, Purdue Univ.; Y. Liu, Sun Microsystems, Inc.; E. J. Delp III, systems. Also, this framework permits further investigation on other Purdue Univ. functionalities of MVC by simple alteration of the objective. Simulation Motion estimation is the most important step in the video compression. results demonstrate feasibility of our model. We believe our work will Most of the current video compression systems use forward motion provide some inspiration for the future researches. estimation, where motion information is derived at the encoder and sent to the decoder over the channel. Backward motion estimation does not

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 193 Conference 6508: Visual Communications and Image Processing 2007

derive an explicit representation of motion at the encoder. Instead, the with subspace representation encoder implicitly embeds the motion information in an alternative J. Gai, R. L. Stevenson, Univ. of Notre Dame subspace. Most recently, an algorithm that adopts least-square prediction (LSP) for backward motion estimation has shown great This paper describes a new, robustified Hidden Markov Model for potential to further improve coding efficiency. Forward motion target tracking using a subspace representation. The Hidden Markov estimation and backward motion estimation have both their Model (HMM) provides a powerful framework for the probabilistic advantages and disadvantages. Each is suitable for handling some modelling of observations and states. Visual tracking problems are specific category of patterns. In this paper, we propose a novel often cast as an inference problem within the HMM framework. approach that combines both forward motion estimation and Probabilistic Principal Component Analysis (PPCA), a classic backward motion estimation in one framework to adaptively exploit the subspace representation method, is a popular tool for appearance local motion characteristics in an arbitrary video sequence, thus modelling because it provides a compact representation for high- achieving better coding efficiency. We refer to this as Content- dimensional data. Previous subspace based tracking algorithms Adaptive Motion Estimation (CoME). The encoder in the proposed assume the image observations were generated from a Gaussian system is able to adjust the motion estimation method in a rate- distribution parameterized by principal components. One drawback of distortion optimized manner. According to the experimental results, using Gaussian density model is that atypical observations cannot be CoME reduces the data rate in both lossless and lossy compression. modelled well. Hence, they are very sensitive to outliers. To address this problem, we propose to augment the HMM by adding a set of latent variables to adjust the shape of the observation distribution. By 6508-07, Session 2 carefully choosing the distribution of , we obtain a more robust observation distribution with heavier tails than a Gaussian. Numerical Feature point tracking combining the interacting multiple experiments demonstrate the effectiveness of this new framework in model filter and an efficient assignment algorithm cases where the target objects are corrupted by noise or occlusion. D. Marimon, Y. Abdeljaoued, B. Palacios, T. Ebrahimi, École Polytechnique Fédérale de Lausanne (Switzerland) 6508-10, Session 2 An algorithm for feature point tracking is proposed. The Interacting Multiple Model (IMM) filter is used to estimate the state of a feature Optimal multiple sprite generation based on physical point. The problem of data association, i.e. establishing which feature camera parameter estimation point to use in the state estimator, is solved by an assignment M. Kunter, A. Krutz, Technische Univ. Berlin (Germany); M. K. algorithm, the solution proposed by Jonker and Volgenant (JV). Feature points in real scenarios describe abrupt motion changes and Mandal, Univ. of Alberta (Canada); T. Sikora, Technische Univ. hence are best tracked with different motion models filtering at the Berlin (Germany) same time. Moreover, feature points are extracted in textured regions. We present a robust and computational low complex method to Consequently, it is likely that these feature points are located close to estimate the physical camera parameters, intrinsic and extrinsic, for each other. In such environments, the assignment algorithms such as scene shots captured by cameras applying pan, tilt, rotation, and JV are best suited. Therefore, our approach shows robustness in front zoom. These parameters are then used to split a sequence of frames of different manoeuvres and dense feature point environments. A track into several subsequences in an optimal way to generate multiple management method is also developed. In particular a track sprites. Hereby, optimal means a minimal usage of memory while continuation method and a track quality indicator are presented. The keeping or even improving the reconstruction quality of the scene evaluation of the tracking system on real sequences shows that the background. Since wide angles between two frames of a scene shot IMM filter combined with the assignment algorithm outperforms the cause geometrical distortions using a perspective mapping it is Kalman filter, used with the Nearest Neighbour (NN) filter, in terms of necessary to part the shot into several subsequences. In our data association performance and robustness to sudden feature point approach it is not mandatory that all frames of a subsequence are manoeuvres. adjacent frames in the original scene. Furthermore the angle-based classification allows frame reordering and makes our approach very 6508-08, Session 2 powerful. Trajectory-based ball detection and tracking with aid of 6508-11, Session 2 homograph transform in broadcast tennis video X. Yu, Institute for Infocomm Research (Singapore); N. Jiang, J. Li, Geometrical image filtering with connected operators National Univ. of Singapore (Singapore) and image inpainting Ball-detection-and-tracking in broadcast tennis video (BTV) is a crucial M. Dimiccoli, P. Salembier, Univ. Politècnica de Catalunya (Spain) but challenging task in tennis video semantics analysis. Informally, the This paper deals with the joint use of connected operators and image challenges are due to camera motion and the other causes such as inpainting for image filtering. Connected operators filter the image by the presence of many ball-like objects and the small size of the tennis merging its flat zones while preserving the contour information. Image ball. The trajectory-based approach proposed by us in our previous inpainting restores the values of an image for a destroyed or papers mainly counteracted the challenges imposed by the causes consciously masked area of the image domain. The present paper other than camera motion and achieves a good performance. This shows the usefulness of using image inpainting to completely remove paper proposes an improved trajectory-based ball detection and the perceptual presence of the areas filtered by a connected operator. tracking algorithm in BTV with the aid of homography, which The proposed filtering scheme involves two main steps: first, the counteracts the challenges caused by camera motion and bring us image is simplified using a connected operator. Second, the multiple new merits. First, it acquires an accurate homography, which perceptually most important filtered regions are estimated using transforms each frame into the “standard” frame. Second, it achieves inpainting. The mask marking the regions to be inpainted is higher accuracy of ball identification. Third, it obtains the ball automatically computed and no user interaction is required. projection position in the real world, instead of ball location in the Comparative experiments have shown that the proposed scheme image. Lastly, it also identifies landing frames and positions of the ball. outperforms early filtering strategies in term of structural perceptual The experimental results show that the improved algorithm can obtain quality. not only higher accuracy in ball identification and higher accuracy in ball position but also ball landing frames and positions. The approach is general in the sense that any connected operator with any simplification criterion can be used. As a result, it is suitable for a large set of advanced filtering applications such as objects, 6508-09, Session 2 writing or defects removal. The extension of the morphological framework through inpainting methodology for image filtering seems A robustified hidden Markov model for visual tracking to be a very interesting field of research.

194 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007

6508-12, Session 2 processing. We discuss these challenges and our solutions via several examples. In conclusion, the solutions have enabled the market Maximum-entropy expectation-maximization algorithm introduction of several professional 3D products, and progress is for image processing made rapidly towards consumer 3DTV. H. Hong, D. Schonfeld, Univ. of Illinois/Chicago Estimation of an unknown probability density function (pdf) has been 6508-15, Session 3 intensively studied for various image processing applications. Finite mixture assumption is reasonable to estimate underlying density and A survey of polynomial fitting and segmentation methods has been widely used in that it reduces the problem into parameter for generalized principal component analysis estimation. However, the maximum entropy solution is required to S. R. Rao, Univ. of Illinois at Urbana-Champaign; H. Derksen, avoid degenerate solutions and overfitted density estimations. In this Univ. of Michigan; R. M. Fossum, Y. Ma, A. W. Wagner, A. Y. Yang, paper, we propose a maximum-entropy expectation-maximization Univ. of Illinois at Urbana-Champaign algorithm. The maximum-entropy constraint is imposed in order to ensure smoothness of the estimated density function. The exact We consider the problem of simultaneously segmenting data samples derivation of the maximum-entropy expectation-maximization drawn from multiple linear subspaces and estimating model algorithm requires determination of the covariance matrix combined parameters for those subspaces. This “subspace segmentation” with the maximum entropy likelihood function, which is difficult to problem naturally arises in many computer vision applications such as solve directly. We therefore introduce a new lower-bound for the EM motion and video segmentation, and in the recognition of human algorithm derived by using the Cauchy-Schwartz inequality to obtain a faces, textures, and range data. Generalized Principal Component suboptimal solution. We use the proposed algorithm for function Analysis (GPCA) has provided an effective way to resolve the strong interpolation and image segmentation. We propose the use of the EM coupling between data segmentation and model estimation inherent algorithm for image recovery from randomly sampled data and signal in subspace segmentation. Essentially, GPCA works by first finding a reconstruction from randomly scattered sensors, since conventional global algebraic representation of the unsegmented data set, and then EM algorithm is typically used for maximum likelihood estimation with decomposing the model into irreducible components, each incomplete samples. The proposed algorithm can estimate missing corresponding to exactly one subspace. Over the years there have data from given incomplete samples. The performance of the been many proposed methods for finding fitting polynomials and for proposed algorithm and proposed application is shown to provide segmenting the linear subspaces. However, it is not clear to what good results. extent these methods will make GPCA a stable algorithm in the presence of noise. We provide a survey of many of the polynomial fitting and segmentation methods that have been used for GPCA and 6508-13, Session 3 using a large testbed of synthetic and real experiments compare their efficiency and effectiveness. Mathematical aspects of shape analysis for object recognition P. F. Stiller, Texas A&M Univ. 6508-16, Session 3 We survey some of the mathematical techniques that have led to A lossy coding heuristic for mixed data segmentation useful new results in shape analysis and their application to a variety H. Derksen, Univ. of Michigan; Y. Ma, W. Hong, J. Wright, Univ. of of object recognition tasks. In particular, we will show how these Illinois at Urbana-Champaign techniques allow one to solve a number of fundamental problems related to object recognition for configurations of point features under Various real-life applications lead to data sets in a finite dimensional a generalized weak perspective model of image formation. Our real vector space. Often this data is mixed, in which case it is better to approximate the data by a subspace arrangement (union of linear or approach makes use of progress in shape theory and includes the affine subspaces) rather than a single subspace. We will discuss one development of object-image equations for shape matching and the heuristic to find a “good” segmentation of the data into subsets such exploitation of shape space metrices (especially object-image metrics) that each subset can be well-approximated by an affine subspace. to measure matching up to certain transformations. This theory is built With “good” we mean that the segmentation gives us an way to store on advanced mathematical techniques from algebraic and differential the data (we allow small errors) using as few bits as possible. We will geometry which are used to construct generalized shape spaces for derive nice formulas which estimate the “entropy” of the data. There various projection and sensor models. That construction in turn is are various nice interpretations of these formulas. This heuristic works used to find natural metrics that express the distance (geometric well in practice and can be used for image compression and image difference) between two configurations of object features, two segmentation for example. configurations of image features, or an object and an image pair. Such metrics are believed to produce the most robust tests for object identification; at least as far as the object’s geometry is concerned. 6508-17, Session 3 The most important feature of a shape theoretic approach is that all of the Integral invariants for curves in 3D: inductive approach matching tests and metrics are independent of the choice of I. A. Kogan, K. Hamid, S. Feng, North Carolina State Univ. coordinates used to express the feature locations on the object or in the image. In addition, the approach is independent of the camera/ Invariants under the actions of the Euclidean, affine and projective sensor position and any camera/sensor parameters. Finally, the groups are widely used in image processing and computer vision. method is also independent of object pose or image orientation. This Differential invariants, such as Euclidean curvature and torsion for is what makes the results so powerful. space curves, are the most classical. The practical utilization of differential invariants is limited, however, due to their high sensitivity to noise. This motivates interest in other types of invariants such as 6508-14, Session 3 semi-differential, or joint, invariants and various types of integral invariants. Since integration diminishes the effects of noise, integral Challenges in 3DTV invariants have advantage in such applications. The explicit A. Redert, R. M. Berretty, Philips Research Labs. (Netherlands) expressions for integral invariants, however, appear to be known only for curves in 2D, as computations become challenging in 3D. Philips provides autostereoscopic three-dimensional display systems In this paper we obtain, for the first time, explicit formulae for integral that will bring the next leap in visual experience, adding true depth to invariants for curves in 3D with respect to the special and the full video systems. We identified three challenges specifically for 3D affine groups. Using an inductive variation of the moving frame image processing: 1) bandwidth and complexity of 3D images, 2) method, we first compute Euclidean integral invariants and then use conversion of 2D to 3D content, and 3) object-based image/depth

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 195 Conference 6508: Visual Communications and Image Processing 2007

them to build the affine invariants. We use integral invariants to introduce directional wavelets and compare their directional basis construct signatures that characterize curves up to the special affine functions and image compression performances. Our experiments transformations. show that adaptive directional wavelets without side information has the worst performance. This is because its directional selectivity is significantly worse than with side information, the selectivity is further 6508-18, Session 3 affected by quantization. On the whole, directional wavelets have good performances. Modification of context-based arithmetic coding based Variable separation for 3D from 2D on the direction features of directional wavelets may improve their J. Zhang, M. Boutin, D. G. Aliaga, Purdue Univ. compression performances. Accurately reconstructing the 3D geometry of a scene or object observed on 2D images is a difficult problem: there are many 6508-62, Poster Session unknowns involved (camera pose, scene structure, depth factors) and solving for all these unknowns simultaneously is computationally Locally adaptive reconstruction of lost low-frequency intensive and suffers from numerical instability. In this paper, we coefficients in wavelet coded images algebraically decouple some of the unknowns so that they can be solved for independently. Decoupling the pose from the other variables J. Rombaut, A. Pizurica, W. R. Philips, Univ. Gent (Belgium) has been previously discussed in the literature. Unfortunately, pose In packet switched networks such as the Internet, packets may get estimation is an ill-conditioned problem. In this paper, we algebraically lost during transmission due to, e.g., network congestion. This leads to eliminate all the camera pose parameters (i.e., position and orientation) a quality degradation of the original signal. As video communication is from the structure-from-motion equations for an internally calibrated a bandwidth consuming application, the original data are first camera. We then also fully eliminate the structure coordinates from the compressed. This compression step increases the impact of equations. This yields a very simple set of homogeneous polynomial information loss even more. In wavelet based image and video coding, equations of low degree involving only the depths of the observed the low frequency data is the most important. Loss of low frequency points. When considering a small number of tracked points and coefficients results in annoying black holes in the received images and pictures (e.g., five points on two pictures), these equations can be video. This effect can be countered by post processing error solved using the sparse resultant method. concealment: a lost coefficient is estimated from its neighboring coefficients. 6508-60, Poster Session In this paper we present a locally adaptive interpolation method for the lost low frequency coefficients. For each lost low frequency coefficient, An improved fast interpolation approach for H.264 we estimate the optimal interpolation direction (horizontal or vertical) using novel error measures. In this way, we preserve the edges in the D. Wu, Institute for Infocomm Research (Singapore) reconstructed image much better. Compared to older techniques of As one of important components in H.264 video encoder, interpolation similar complexity, our scheme reconstructs images with the same or of half and quarter pixels is also computational intensive. Compared to better quality. This is reflected in the visual as well as in the numerical integer pixel motion estimation, “finer” interpolation provides lower results: there is an increase of up to 4.4 dB compared to bilinear residue because of better block match. Nevertheless, this good motion concealment. The proposed scheme is fast and simple, which makes it compensation performance is obtained at the expense of increased suitable for real-time applications. complexity. Thus, we aim to reduce repetitive interpolation operations. Based on our previous work, this paper presents an improved fast and adaptive interpolation method that further reduces the complexity of 6508-63, Poster Session H.264 video encoding process. Memory reorganization of interpolated Pose estimation from video sequences based on reconstructed image is also utilized to facilitate the overall algorithm. Experiments have been carried out for the typical test sequences. They Sylvester’s equation range from not too much motion such as ‘akiyo’ to sequences with C. Chen, J. Yang, D. Schonfeld, Univ. of Illinois/Chicago; M. A. much higher motion such as ‘Mobile’ and ‘Stefan’. Experimental results Mohamed, Motorola, Inc. on these sequences show that the proposed method is able to increase encoder speed ranging from 10% to 22% compared with our In this paper, we introduce a method to jointly track the object motion previous work without any PSNR loss or bit rate increase. Compared to and estimate pose within the framework of particle filtering. We focus JM software, our new method is able to increase encoder speed on direct estimation of the 3D pose from a 2D image sequence. Scale- ranging from 10% to 94%. Furthermore, our experimental results prove Invariant Feature Transform (SIFT) is used to extract feature points in that adoption of this new algorithm does not degrade image quality the images. We show that pose estimation from the corresponding (PSNR or bit rate) at all. feature points can be formed as a solution to Sylvester’s equation. We rely on a solution to Sylvester’s equation based on the Kronecker product method to solve the equation and determine the pose state. 6508-61, Poster Session We demonstrate that the classical Singular Value Decomposition (SVD) approach to pose estimation provides a solution to Sylvester’s A comparative study of image compression based on equation in 3D-3D pose estimation. The proposed approach to the directional wavelets solution of Sylvester’s equation is therefore equivalent to the classical SVD method for 3D-3D pose estimation, yet it can also be used for K. Li, W. Xu, Q. Dai, Tsinghua Univ. (China); Y. Wang, Polytechnic pose estimation from 2D image sequences. Finally, we rely on Univ. computer simulation experiments to demonstrate the performance of Discrete wavelet transform is an effective tool to generate scalable our algorithm on video sequences. stream, but it cannot efficiently represent edges which are not aligned in horizontal or vertical directions, while natural images often contain rich edges and textures of this kind. Hence, recently, intensive research 6508-64, Poster Session has been focused particularly on the directional wavelets which can A fast and efficient method to protect color images effectively represent directional attributes of images. Specifically, there are two categories of directional wavelets: redundant wavelets (RW) M. Chaumont, W. Puech, Lab. d’Informatique de Robotique et de and adaptive directional wavelets (ADW). One representative redundant Microelectronique de Montpellier (France) wavelet is the dual-tree discrete wavelet transform (DDWT), while adaptive directional wavelets can be further categorized into two types: In this paper, we propose a method to embed the color information of with or without side information. What’s more, adaptive directional an image in a corresponding grey-level image. The objective of this wavelets with side information mainly have two sampling modes: work is to allow free access to the grey-level image and give color orthogonal sampling and quincunx sampling. In this paper, we briefly image access only if you own a secret key. This method is made of

196 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007 three major steps which are a fast color quantization, an optimized 6508-68, Poster Session ordering and an adapted data hiding. The principle is to build an index image which is, in the same time, a semantically intelligible grey-level Pet fur color and texture classification image. In order to obtain this particular index image, which should be J. Yen, Toshiba America Business Solutions; D. Mukherjee, S. robust to data hiding, a layer running algorithm is proceeded to sort the K colors of the palette. The major contributions of this paper are Lim, D. R. Tretter, Hewlett Packard Labs. the fast color quantization, the optimized layer running algorithm, the Object segmentation is important in image analysis for imaging tasks color palette compression and the adapted data hiding. such as image rendering and image retrieval. Pet owners have been known to be quite vocal about how important it is to render their pets perfectly. We present here an algorithm for pet (mammal) fur color 6508-65, Poster Session classification and an algorithm for pet (animal) fur texture classification. Per fur color classification can be applied as a 3D face model tracking based on a multiresolution active necessary condition for identifying the regions in an image that may search contain pets much like the skin tone classification for human flesh M. Chaumont, W. Puech, Lab. d’Informatique de Robotique et de detection. As a result of the evolution, fur coloration of all mammals is Microelectronique de Montpellier (France) caused by a natural organic pigment called Melanin and Melanin has only very limited color ranges. We have conducted a statistical This paper deals with face tracking in difficult conditions of non analysis and concluded that mammal fur colors can be only in levels calibrated camera, strong head motions, thanks to a deformable 3D of gray or in two colors after the proper color quantization. This pet fur model. In those conditions, the proposed approach is able to detect color classification algorithm has been applied for peteye detection. and track a face. The novelty is mainly due to a multi-resolution Active We also present here an algorithm for animal fur texture classification Model search which allows to catch strong head motions. Results using the recently developed multi-resolution directional sub-band show an improvement between the single and the multi-resolution Contourlet transform. The experimental results are very promising as technique. Near real-time results are also provided. these transforms can identify regions of an image that may contain fur of mammals, scale of reptiles and feather of birds, etc. Combining the color and texture classification, one can have a set of strong 6508-66, Poster Session classifiers for identifying possible animals in an image. Region-based hidden Markov models for image categorization and retrieval 6508-69, Poster Session F. Li, Q. Dai, W. Xu, Tsinghua Univ. (China) A simple reversed-complexity Wyner-Ziv video coding Hidden Markov models (HMMs) have been widely used in various mode based on a spatial reduction framework fields, including image categorization and retrieval. Most of the existing methods train HMMs by low-level features of image blocks; D. Mukherjee, Hewlett-Packard Labs.; B. Macchiavello, R. L. de however, the block-based features can not reflect high-level semantic Queiroz, Univ. de Brasília (Brazil) concepts well. This paper proposes a new method to train HMMs by A spatial-resolution reduction based framework for incorporation of a region-based features, which can be obtained after image Wyner-Ziv frame coding mode in existing video codecs is presented, segmentation. Our work can be characterized by two key properties: to enable a mode of operation with low encoding complexity. The core (1) Region-based HMM is adopted to achieve better categorization Wyner-Ziv frame coder works on the Laplacian residual of a lower- performance, for the region-based features accord with the human resolution frame encoded by a regular codec at reduced resolution. perception better. (2) Multi-layer semantic representation (MSR) is The quantized transform coefficients of the residual frame are mapped introduced to couple with region-based HMM in a long-term relevance to cosets in a manner that enables reuse of the same entropy coder feedback framework for image retrieval. The experimental results that already exists in the regular codec, with minimal loss in efficiency. demonstrate the effectiveness of our proposal in both aspects of The decoder iteratively conducts motion-based side-information categorization and retrieval. generation and coset decoding, to gradually refine the estimate of the frame. Preliminary results are presented for application to the H.263+ 6508-67, Poster Session video codec. A new support tool for machine learning and pattern 6508-71, Poster Session recognition using tracking and motion segmentation E. Bichot, O. Masset, L. Mascarilla, P. Courtellemont, Univ. de La Motion refined medium granular scalability Rochelle (France) Z. Li, W. Yao, S. Rahardja, Institute for Infocomm Research (Singapore) A new support tool using object tracking and motion based segmentation is developed for machine learning and pattern In this paper, we propose an interesting scheme to obtain a good recognition. In the learning step, an object of interest is tracked while tradeoff between motion information and residual information for learning is performed from segmented frames. In the recognition step, medium granular scalability (MGS). The Scheme refines both motion target is tracked until favorable conditions allow identification. This information and residual information at enhancement layers when the tool is used in the context of the Aqu\@thèque project which includes scalable bitrate range is wide, whereas only residual information is an automatic fish recognition system. Motion based segmentation is refined when the range is narrow. In other words, for the case of wide performed by estimating optical flow and grouping similar motion bitrate range, there can be more than one motion vector fields (MVFs) vectors together. It results in coherent moving regions corresponding where one is generated at base layer and others are generated at to video objects. Tracking is a difficult task especially in case of real enhancement layers. For narrow bitrate range, only one MVF is world images. We use particle filtering approaches which are robust in necessary. The layers can either share one MVF or have its own, presence of background clutter. We incorporate motion based depending on the bitrate range cross layers. Unlike Coarse Granular segmentation measurements in importance sampling step. It results in Scalability (CGS), the correlation between two adjacent MVFs in MGS a new proposal density composed of a dynamical model and a is very strong. Hence MGS can be provided in the most important mixture of Gaussians centered on motion regions having similar bitrate range to achieve a better tradeoff between motion and residual motion to target. This proposal density drives target search toward information and a finer granularity in that range. CGS can be applied regions of interest of state space. Consequently, it improves tracking in less important bitrate ranges to give a coarse granularity. performance compared with Bootstrap filter in difficult situations Experimental results show that the coding efficiency can be improved (presence of objects with similar appearance to target). by up to 1dB compared with existing SNR scalability scheme at high bitrate.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 197 Conference 6508: Visual Communications and Image Processing 2007

6508-72, Poster Session algorithm. The algorithm is designed to optimize the Rate-Distortion- Complexity (R-D-C) performance. A complexity prediction model is An efficient multiframe dynamic search range motion designed and the prediction is based on past. When the complexity estimation for H.264 needs to be reduced to meet the target, the controller will decide configurations of inter motion search based on a simple analysis of the Q. Sun, J. Wang, X. Chen, L. Yu, Zhejiang Univ. (China) past statistics. When the complexity may be increased, the controller H.264/AVC achieves higher compression efficiency than previous will choose the configurations based on the complexity prediction video coding standards. However, this comes at the cost of increased model. The algorithm also works well across scene changes. Tests are complexity due to the use of variable block size motion estimation conducted for coding interlaced video (720x576 PAL). The complexity and long-term memory motion compensated prediction (LTMCP). In of motion estimation is accurately controlled within a fixed number of this paper, an efficient multi-frame dynamic search range motion frames. Average improvements are obtained in R-D performance estimation algorithm is proposed. This algorithm can adjust the spatial compared to the previously found off-line optimized settings. These search range and temporal search range according to the video improvements are considerable in some cases, up to 6.3% compared content dynamically. This algorithm can be on the top of many other with these off-line solutions. The improvements are up to 18% fast motion estimation (Fast ME) algorithms. Compared with the compared to a conservative solution. constant search range scheme used by multi-frame UMHexagonS algorithm, the proposed algorithm can be 4.86 time faster, with negligible degradation of video quality. 6508-76, Poster Session Wyner-Ziv residual coding for wireless multiview system 6508-73, Poster Session Z. Jin, M. Yu, G. Jiang, Ningbo Univ. (China) An area-efficient VLSI architecture for AVS intraframe For wireless multi-view video sensor array, whose abilities of storage encoder and computation are all very weak, it is essential to have a low-power and low-complexity encoder device. In this paper, a DCT-domain K. Zhang, L. Yu, Zhejiang Univ. (China) Wyner-Ziv residual coding scheme with low encoding complexity is In this paper, we propose a VLSI architecture for AVS intra frame proposed for wireless multi-view video coding. The scheme is designed encoder. Reconstruction loop hinders the parallelism exploration and to encode each view independently without motion and disparity becomes the critical path in an intra frame encoder. A First Selection estimation, shift the large encoding complexity to the decoder. At the Then Prediction (FSTP) method is proposed to break the loop and decoder, the proposed scheme performs joint decoding with side enable the parallel process of intra mode selection and reconstruction information interpolated from key frames and adjacent views. The on neighboring blocks. In addition, area-efficient modules were experimental results show that, the proposed WZRC-MVC scheme developed. Configurable intra predictor can support all the intra outperforms the H.264 intra coding about 1.1 dB in rate-distortion prediction modes. A CA-2D-VLC engine with an area-efficient Exp- performance. Golomb encoder was developed to meet the encoding speed demand with comparably low hardware cost. Synthesized with 0.18 µm CMOS standard-cell library, the overall hardware cost of the proposed intra 6508-77, Poster Session frame encoder is 89k logic gates at the clock frequency constraint of Progressive image transmission with RCPT protection 125MHz. Proposed encoder can satisfy real time encoding of 720Å~576 4:2:0 25fps video at the working frequency of 54MHz. L. Cao, The Univ. of Mississippi In this paper, a joint source-channel coding scheme is proposed for progressive image transmission over channels with both random bit 6508-74, Poster Session errors and packet loss by using rate-compatible punctured Turbo Fast luminance and chrominance correction based on codes (RCPT) protection only. Two technical components which are motion compensated linear regression for multiview different from existing methods are presented. First, a data frame is divided into multiple CRC blocks before being coded by a turbo code. video coding This is to secure a high turbo coding gain which is proportional to the W. Chen, L. Ding, L. Chen, National Taiwan Univ. (Taiwan) data frame size. In the mean time, the beginning blocks in a frame may still be usable although the decoding of the entire frame fails. Second, Luminance and chrominance correction (LCC) is important in multi- instead of employing product codes, we only use RCPT, along with an view video coding (MVC) because it provides better rate-distortion interleaver, to protect images over channels with combined distortion performance when encoding video sequences captured by ill- including BER and packet loss. With this setting, the effect of packet calibrated multi-view cameras. This paper presents a robust and fast loss is equivalent to randomly puncturing turbo codes. As a result, the LCC algorithm based on motion compensated linear regression which optimial allocation of channel code rates is needed for the random reuses the motion information from the encoder. We adopt the linear errors only, which largely reduces the complexity of the optimization weighted prediction model in H.264/AVC as our LCC model. In our process. experiments, the proposed LCC algorithm outperforms basic histogram matching method up to 0.4dB with only few computational overhead and zero external memory bandwidth. So, the dataflow of 6508-19, Session 4 this method is suitable for low bandwidth/low power VLSI design for future multi-view applications. Bayesian distributed articulated object tracking using multiple collaborative trackers 6508-75, Poster Session W. Qu, D. Schonfeld, Univ. of Illinois/Chicago Complexity control of fast motion estimation in H.264/ In this paper, we propose two novel articulated object tracking approaches. The Decentralized Articulated Object Tracking approach MPEG-4 AVC with rate-distortion-complexity avoids the common practice of using a high-dimensional joint state optimization representation for articulated object tracking. Instead, it introduces a M. Wu, S. Forchhammer, S. M. Aghito, Danmarks Tekniske Univ. decentralized scheme and models the inter-part interaction within an (Denmark) innovative Bayesian framework. To handle severe self-occlusions, we further extend the first approach by modeling high-level inter-unit A complexity control algorithm for H.264 advanced video coding is interactions and develop the Hierarchical Articulated Object Tracking proposed. The algorithm can adaptively control the complexity of algorithm within a consistent hierarchical framework. Preliminary integer inter motion estimation for a given target complexity based on experimental results have demonstrated the superior performance of Enhanced Predictive Zonal Search (EPZS) fast motion estimation the proposed approaches for real-world videos sequences.

198 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007

6508-20, Session 4 Since the location of the cropped image is specified by the end user, there exits a small but non-negligible time-lag between the intended Tracking people in mixed modality systems and transmitted cropped images. This imposes additional Y. Ivanov, Mitsubishi Electric Research Labs. computational constraints and requires collaborative tracking of targets both in LRT and asynchronous HRC images. We developed a In traditional surveillance systems tracking of objects is achieved by joint tracking method for static EPTZ cameras that automatically means of image and video processing. The disadvantage of such detects targets by background subtraction and tracks them using surveillance systems is that if an object needs to be tracked - it has to iterative refinement while keeping the computational costs minimum. be observed by a video camera. However, geometries of indoor spaces typically require a large number of video cameras to provide the coverage necessary for robust operation of video-based tracking 6508-23, Session 4 algorithms. Increased number of video streams increases the computational burden on the surveillance system in order to obtain Particle filter-based camera tracker fusing marker and robust tracking results. In this paper we present an approach to feature point-based cues tracking in mixed modality systems, with a variety of sensors. The D. Marimon, Y. Maret, Y. Abdeljaoued, T. Ebrahimi, École system described here includes over 200 motion sensors as well as 6 Polytechnique Fédérale de Lausanne (Switzerland) moving cameras. We track individuals in the entire space and across cameras using contextual information available from the motion This paper presents a video-based camera tracker that combines sensors. Motion sensors allow us to almost instantaneously find marker-based and feature point-based cues within a particle filter plausible tracks in a very large volume of data, ranging in months, framework. The framework relies on their complementary which for traditional video search approaches could be virtually performances. On the one hand, marker-based trackers can robustly impossible. We describe a method that allows us to evaluate when recover camera position and orientation when a reference (marker) is the tracking system is unreliable and present the data to a human available but fail once the reference becomes unavailable. On the operator for disambiguation. other hand, filter-based camera trackers using feature point cues can still provide predicted estimates given the previous state. However, the trackers tend to drift and usually fail to recover when the reference 6508-21, Session 4 reappears. Therefore, we propose a fusion where the estimate of the filter is updated from the individual measurements of each cue. The Multiple hypothesis shape tracking using particle filtering particularity of the fusion filter is to manipulate different sorts of cues and Hough-based observation models in a single framework. The framework keeps a single motion model A. Dore, M. Asadi, C. S. Regazzoni, Univ. degli Studi di Genova and its prediction is corrected by one cue at a time. More precisely, (Italy) the marker-based cue is selected when the reference is available whereas the feature point-based cue is selected otherwise. The filter’s In the last years, the Particle Filter algorithm has been extensively state is updated by switching between two different likelihood proposed and employed for handling the problem of visual tracking of distributions. Each likelihood distribution is adapted to the type of multiple moving objects under different assumptions. This wide usage measurement (cue). Evaluations on real cases show that the fusion of is due to the capability of performing a recursive multiple hypothesis these two approaches outperforms the individual tracking results. state estimation for non-linear non-Gaussian motion and observation models. In this paper a method, based on the Particle Filter framework, is proposed for multiple objects tracking, exploiting a 6508-24, Session 5 target representation consisting of position and shape described as a fixed dimensionality vector composed by a fixed number of grouped Robust distributed multiview video compression for target corners. However, usually, application domains of visual wireless camera networks tracking algorithms are characterized by non-rigid objects and high C. Yeo, K. Ramchandran, Univ. of California/Berkeley occlusions rate entailing new corners to appear and others to disappear at each frame. In order to cope with this problem, a voting We propose a novel method of exploiting inter-view correlation among method (i.e. the Generalized Hough Transform) is employed to cameras that have overlapping views in order to deliver error-resilient estimate the likelihood function to weight different propagated video in a distributed multi-camera system. The main focus in this particles (i.e. multiple corners configurations describing shapes) by work is on robustness which is imminently needed in a wireless means of the corners extracted from the currently observed image. setting. Our system has low encoding complexity, is robust while This method, in addition to the high dimensionality of the state satisfying tight latency constraints, and requires no inter-sensor representation, depicts the two main particularities of the presented communication. In this work, we build on and generalize PRISM Particle Filter. The proposed algorithm has been tested in a real-world [Puri2002], an earlier proposed single-camera distributed video domain and experiments indicate good results in tracking both rigid compression system. Specifically, decoder motion search, a key and non-rigid objects. attribute of single-camera PRISM, is extended to the multi-view setting to include decoder disparity search based on two-view camera geometry. Our proposed system, dubbed PRISM-MC (PRISM 6508-22, Session 4 multi-camera), achieved PSNR gains of up to 1.7 dB over a PRISM based simulcast solution in experiments over a wireless channel Collaborative tracking of objects in EPTZ cameras simulator. F. Bashir, F. M. Porikli, Mitsubishi Electric Research Labs. Recently, very high resolution cameras that are able to accommodate 6508-25, Session 5 full frame rate video become more feasible. These cameras can provide more than 1M pixels and deliver exceptional details of the Hybrid key/Wyner-Ziv frames with flexible macroblock depicted scene. However, data transmission bandwidth and ordering for improved low delay distributed video coding computational bottlenecks often limit the amount of video data to be D. Agrafiotis, P. Ferré, D. R. Bull, Univ. of Bristol (United Kingdom) analyzed at the user end for most existing system. To furnish a complete view of the scene and detailed information of a target region This paper proposes a concealment based approach to generating at the same time while obliged with bandwidth requirements, such the side information and estimating the correlation noise for low-delay, cameras offer a mode of transmission that supports two video pixel-based, distributed video coding. The proposed method employs streams. One stream corresponds to a low-resolution thumbnail (LRT) a macroblock pattern similar to the one used in the dispersed type of the overall field-of-view, and the second stream shows a high FMO of H.264 for grouping the macroblocks of each frame into intra resolution view of a cropped region (HRC). In other words, the coded (key) and Wyner-Ziv groups. Temporal concealment is then cropped image acts as an electronic pan-tilt-zoom (EPTZ) camera. used at the decoder for “concealing” the missing macroblocks (estimating the side information - predicting the Wyner-Ziv

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 199 Conference 6508: Visual Communications and Image Processing 2007

macroblocks) using information from the adjacent key macroblocks. MPEG-4 SVC. In this framework, for a video frame at a high resolution The actual intra coded/decoded (key) macroblocks are used for layer, the lowpass content is predicted from the information already estimating the correlation noise. The results indicate significant coded in lower resolution layer, but the highpass content is predicted by performance improvements relative to existing motion extrapolation exploiting the neighboring frames at current resolution. In this way, both based approaches (up to 25% bit rate reduction) . the cross-resolution correlation and temporal correlation are exploited simultaneously, which leads to much higher efficiency in prediction. Preliminary experimental results demonstrate that the proposed 6508-26, Session 5 framework improves the spatial scalability performance of current H.264/ MPEG-4 SVC. The improvement is significant especially for high-fidelity Distributed video coding based on constrained rate video coding. In addition, another advantage over wavelet-based in- adaptive LDPC codes scale scheme is achieved that the proposed framework can support R. Liu, Beijing Univ. of Aeronautics and Astronautics (China) and arbitrary down-sampling and up-sampling filters. Florida Institute of Technology; G. Hua, C. W. Chen, Florida Institute of Technology 6508-29, Session 5 In this paper, we present a distributed video coding scheme based on zero motion identification at the decoder and constrained rate Fast prediction model of adaptive GOP structure for SVC adaptive low density parity check (LDPC) codes. Zero-motion-block Y. Chen, C. Lin, C. Chen, L. Chen, National Taiwan Univ. (Taiwan) identification mechanism is introduced at the decoder, which takes the characters of video sequence into account. The constrained error Adaptive group-of-picture (GOP) structure is an important encoding tool control decoder can use the bits in the zero motion blocks as a in multi-level motion-compensated temporal filtering coding scheme. constraint to achieve a better decoding performance and further Compared to conventional fixed-GOP scheme, it can dynamically adapt improve the overall video compression efficiency. It is only at the the GOP size to enhance the coding performance based on each decoder side that the proposed scheme exploits temporal and spatial sequence’s characteristics. But the existing adaptive GOP structure redundancy without introducing any additional processing at the (AGS) algorithm proposed in JSVM requires huge computation encoder side, which keeps the complexity of the encoding as low as complexity. In this paper, a fast AGS prediction algorithm is proposed. At possible with certain compression efficiency. As a powerful alternative first, based on the relationship among coding performance, GOP size to Turbo codes, LDPC codes have been applied to our scheme. Since and corresponding intra block ratio, a sub-GOP size prediction model for video data are highly non-ergodic, we use rate-adaptive LDPC codes different decomposition levels is developed based on the encoded intra to fit this variation of the achievable compression rate in our scheme. block ratio. Then, a prediction scheme is proposed to implement AGS by We propose a constrained LDPC decoder not only to improve the the sub-GOP size prediction model. It can predict the following sub- decoder efficiency but also to speed the convergence of the iterative GOP size by current sub-GOP’s information instead of searching all decoding. Simulation demonstrates that the scheme has significant possible sub-GOP composition. The experimental results show that the improvement in the performances. In addition, the proposed proposed algorithm with linear threshold has almost equivalent coding constrained LDPC decoder may benefit other application. performance as AGS in JSVM but only one-fourth computation complexity for 4-level interframe coding scheme is required. 6508-27, Session 5 6508-30, Session 6 Unequal error protection using Wyner-Ziv coding Scalable direction representation for image compression L. Liang, Purdue Univ.; P. Salama, Indiana Univ.-Purdue Univ. at with direction-adaptive discrete wavelet transform Indianapolis; E. J. Delp III, Purdue Univ. T. Xu, C. Chang, B. Girod, Stanford Univ. Compressed video is very sensitive to channel errors. A few bit losses can derail the entire decoding process. Thus, protecting compressed The 2-D discrete wavelet transform (DWT) is the most important new video is imperative to enable visual communications. Since different image compression technique of the last decade. Conventionally, the 2- elements in a compressed video stream vary in their impact on the D DWT is performed as a separate transform by cascading two 1-D quality of the decoded video, unequal error protection can be used to transforms in the vertical and horizontal direction. However, such a provide efficient protection. This paper describes an unequal error separate transform cannot efficiently represent image features with the protection method for protecting data elements in a video stream, via edges not aligned in these two directions since it distributes the energy a Wyner—Ziv encoder that consists of a coarse quantizer and a Turbo of these edges into several sub-bands. The direction-adaptive discrete coder based lossless Slepian—wolf encoder. Data elements that wavelet transform (DA-DWT) locally adapts the filtering direction to the significantly impact the visual quality of decoded video, such as geometric flow in the image. DA-DWT image coders have been shown to modes and motion vectors as used by H.264, are provided more achieve a rate-distortion performance superior to non-adaptive wavelet parity bits than coarsely quantized transform coefficients. This results coders. However, since the direction information must always be in an improvement in the quality of the decoded video when the signalled regardless of total bit-rate, performance at very low bit-rates transmitted sequence is corrupted by transmission errors, than might be worse. In this paper, we propose two scalable direction obtained by the use of equal error protection. representations: the layered scheme which is similar to the scalable motion vector representation in scalable video coding and the level-unit scheme which provides finer granularity upon the layered scheme. In 6508-28, Session 5 addition, we observed that the distortion caused by using a coarser direction representation at the decoder is almost independent of the rate Generalized in-scale motion compensation framework of the wavelet coefficients in a wide range. Experimental results indicate for spatial scalable video coding that we can achieve the desirable performance at both low and high bit rates with our proposed level-unit scheme. Significant improvement in R. Xiong, Institute of Computing Technology (China); J. Xu, F. Wu, image quality (about 3-5 dB) is observed at very low bit rate, relative to S. Li, Microsoft Research Asia (China) non-scalable coding of the direction information. In existing video coding schemes with spatial scalability based on pyramid frame representation, such as the ongoing H.264/MPEG-4 SVC (scalable video coding) standard, video frame at a high resolution 6508-31, Session 6 is mainly predicted either from the lower-resolution image of the same Video coding with fully separable wavelet and wavelet frame or from the temporal neighboring frames at the same resolution. Most of these prediction techniques fail to exploit the two correlations packet transforms simultaneously and efficiently. This paper extends the in-scale M. S. Trocan, B. Pesquet-Popescu, École Nationale Supérieure des prediction technique developed for wavelet video coding to a Télécommunications (France) generalized in-scale motion compensation framework for H.264/

200 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007

Three-dimensional (t+2D) wavelet coding schemes have been investigated. In particular, we relate rate-distortion performance as demonstrated to be efficient techniques for video compression measured in PSNR to the matching scores as obtained by the applications. However, the separable wavelet transform used for recognition systems. JPEG2000 and SPIHT are correctly predicted by removing the spatial redundancy allows a limited representation of the PSNR to be the most suited compression algorithms to be used in 2D texture because of spatial isotropy of the wavelet basis functions. In fingerprint and face recognition systems. Fractal compression is this case, anisotropic transforms, such as fully separable wavelet identified to be least suited for the use in the investigated recognition transforms (FSWT), can represent a solution for spatial decorrelation. systems, although PSNR suggests JPEG to deliver worse recognition FSWT inherits the separability, the computational simplicity and the filter results in the case of face imagery. JPEG compression performs bank characteristics of the standard 2D wavelet transform, but it surprisingly well at high bitrates in face recognition systems, given the improves the representation of directional textures, as the ones which low PSNR performance observed. In general it may be stated that can be found in temporal detail frames of t + 2D decompositions. The PSNR provides a fairly well matching prediction at the high quality extension of both classical wavelet and wavelet-packet transforms to end, whereas for lower quality PSNR does not reliably predict fully separable decompositions preserve at the same time the matching scores in biometric recognition systems. low-complexity and best-bases selection algorithms of these ones. We apply these transforms in t + 2D video coding schemes and compare them with classical decompositions. 6508-108, Session 7 Automatic target segmentation in color dental images 6508-32, Session 6 J. Luo, M. R. Bolin, Eastman Kodak Co. Anisotropic multidirectional representation with oriented Automatic target segmentation is critical to computerized dental lifting scheme imaging systems, which are designed to reduce human effort and error. We have developed an automatic algorithm that is capable of G. Jeannic, V. Ricordel, D. Barba, Univ. de Nantes (France) outlining an intra-oral reference bar and the tooth of interest. In particular, the algorithm first locates the reference bar using unique In spite of its success, the standard 2-D discrete wavelet transform (2D- color and shape cues. The located reference bar provides an estimate DWT) is not completely adapted to represent image entities like edges for the tooth of interest in terms of both its scale and location. Next, or oriented textures. the estimate is used to initialize a trained active shape model (ASM) Indeed the DWT is limited by the spatial isotropy of its basis functions consisting of the bar and the tooth. Finally, a search process is that can not take advantage of edges regularity and moreover, direction performed to find a match between the ASM and the local image edge that is neither vertical or horizontal is represented using many of structures. Experimental results have shown that our fully automatic these wavelet basis functions which does mean that DWT does not algorithm provides accurate segmentation of both the reference bar provide a sparse representation for such discontinuities. and the tooth of interest, and it is insensitive to lighting, tooth color, Several representations have been proposed to overcome this lack. and tooth-shape variations. Some of them deal with more orientations while introducing redundancy (e.g. ridgelets, curvelets, contourlets) and their implementations are not trivial or require 2-D non separable filtering. 6508-36, Session 8 We present two oriented lifting-based schemes using separable Film grain noise modeling in advanced video coding filtering, lead by edge extraction, and inspired from bandelets and curved wavelets. B. T. Oh, Univ. of Southern California; S. Sun, L. Shawmin, Sharp Labs. of America, Inc.; C. J. Kuo, Univ. of Southern California An image is decomposed into a quadtree according to the edge elements orientation. For each leaf, a wavelet transform is performed A new technique for film grain noise extraction, modeling and along the most regular orientation, and then along its orthogonal synthesis is proposed and applied to the coding of high definition direction. Different adapted filters may be used for these two directions video in this work. The film grain noise is viewed as a part of artistic in order to achieve anisotropic filtering. presentation by people in the movie industry. On one hand, since the Our method permits also a perfect reconstruction and a critical film grain noise can boost the natural appearance of pictures in high sampling. definition video, it should be preserved in high-fidelity video processing systems. On the other hand, video coding with film grain noise is expensive. It is desirable to extract film grain noise from the 6508-33, Session 7 input video as a pre-processing step at the encoder and re-synthesize the film grain noise and add it back to the decoded video as a post- Multimodal image registration based on edges and processing step at the decoder. Under this framework, the coding junctions gain of the denoised video is higher while the quality of the final reconstructed video can still be well preserved. Following this idea, Y. Li, R. L. Stevenson, Univ. of Notre Dame we present a method to remove film grain noise from image/video This paper proposes an edge based multimodal image registration without distorting its original content. Besides, we describe a approach which registers two images taken by visible light and infrared parametric model containing a small set of parameters to represent light. We aim to address image registration as a whole rather than tackle the extracted film grain noise. The proposed model generates the film each of its elements separately. Edges are firstly extracted from images grain noise that is close to the real one in terms of power spectral as matching primitives, and then analyzed to look for those edges that density and cross-channel spectral correlation. Experimental results help design a suitable similarity metric without relying too much on are shown to demonstrate the efficiency of the proposed scheme. optimization algorithms. The selected edges provide meaningful initial conditions for an exhaustive search procedure which converges in a few evaluations. Experimental results show that the presented approach is a 6508-37, Session 8 robust and effective tool for multimodal image registration. Advances in hybrid video coding T. Wedi, S. Wittmann, T. Palfner, B. Schuur, F. Knicker, Panasonic 6508-35, Session 7 R&D Ctr. Germany GmbH (Germany) Comparison of compression algorithms’ impact on Hybrid video coding is known for a long time and is applied in all fingerprint and face recognition accuracy video coding standards such as MPEG-x or H.26x. This paper shows that there is still enough potential for further coding efficiency A. Uhl, Univ. Salzburg (Austria) improvements. The paper starts with an overview of state of the art The impact of using different lossy compression algorithms on the hybrid video coding schemes such as H.264/AVC. Based on this matching accuracy of fingerprint and face recognition systems is review, our advances on main building blocks of H.264/AVC are

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 201 Conference 6508: Visual Communications and Image Processing 2007

presented that significantly improve the coding efficiency. For instance, H.264/MPEG4-AVC is the latest video codec provided by the Joint Video adaptive filtering and high precision motion compensation improves the Team, gathering ITU-T and ISO/IEC experts. Technically there are no motion compensated prediction. The combination of transformation, drastic changes compared to its predecessors H.263 and MPEG-4 part quantization, and entropy coding of the prediction error is improved 2. It however significantly reduces the bitrate and seems to be using an advanced frequency selective coding technique. Intra progressively adopted by the market. The gain mainly results from the prediction is improved by changing scan directions and thus providing addition of efficient motion compensation tools, variable block sizes, better reference pixels for specific prediction directions. Finally, texture multiple reference frames, 1/4-pel motion accuracy and powerful Skip synthesis techniques are used to improve the subjective quality for and Direct modes. A close study of the bits repartition in the bitstream specific textures with homogenous statistical characteristics. This paper reveals that motion information can represent up to 40% of the total presents our above-mentioned techniques in detail. Depending on the bitstream. As a consequence reduction of motion cost is a priority for input sequence and the bit rate, the objective and/or subjective gains future enhancements. compared to H.264/AVC are quite significant. This paper proposes a competition-based scheme for the prediction of the motion. It impacts the selection of the motion vectors, based on a modified rate-distortion criterion, for the Inter modes and for the Skip 6508-38, Session 8 mode. Combined spatial and temporal predictors take benefit of Next-generation video coding for mobile applications: temporal redundancies, where the spatial median usually fails. An market trends, requirements, and technologies average 7% bitrate saving compared to a standard H.264/MPEG4-AVC codec is reported. In addition, on the fly adaptation of the set of M. Budagavi, M. Zhou, Texas Instruments Inc. predictors is proposed and preliminary results are provided. Handheld battery-operated consumer electronics devices such as camera phones, digital still cameras, digital camcorders, and personal 6508-41, Session 8 media players have become very popular in recent years. Video codecs are extensively used in these devices for video capture and/or playback. High-definition video coding with super-macroblocks The annual shipment of such devices already exceeds a hundred million units and is growing, which makes mobile battery-operated video S. Ma, C. J. Kuo, Univ. of Southern California device requirements very important to focus in video coding research A high definition video coding technique using super-macroblocks is and development. This paper highlights the following unique set of investigated in this work. Our research is motivated by the observation requirements for video coding for these applications: low power that the macroblock-based partition in H.264/AVC may not be efficient consumption, high video quality at low complexity, and low cost, and for high definition video since the maximum macroblock size of 16 Å~ 16 motivates the need for a new video coding standard that enables better is relatively small against the whole image size. In the proposed super- trade-offs of power consumption, complexity, and coding efficiency to macboblock based video coding scheme, the original block size MÅ~N meet the challenging requirements of portable video devices. This in H.264 is scaled to 2MÅ~2N. Along with the super-macroblock paper also provides a brief overview of some of the video coding prediction framework, a low-complexity 16 Å~ 16 discrete cosine technologies being presented in the ITU-T Video Coding Experts Group transform (DCT) is proposed. As compared with the 1D 8 Å~ 8 DCT, only (VCEG) standardization body for computational complexity reduction 16 additions are added for a 1D 16 points 16 Å~ 16 DCT. Furthermore, and for coding efficiency improvement in a future video coding an adaptive scheme is proposed for the selection the best coding mode standard. and best transform size. It is shown by experimental results that the super-macroblock coding scheme can achieve a higher coding gain. 6508-39, Session 8 6508-42, Session 9 Adaptive filtering for cross-view prediction in multiview video coding Cross-layer optimization for wireless video communication P. Lai, Univ. of Southern California; Y. Su, P. Yin, C. Gomila, D. Wu, Univ. of Florida; Z. He, Univ. of Missouri/Columbia THOMSON; A. Ortega, Univ. of Southern California With the rapid growth of wireless networks and increasing popularity of We consider the problem of coding multi-view video that exhibits portable video devices, wireless video communication is poised to mismatches in frames from different views. Such mismatches could be become the enabling technology for many multimedia applications over caused by heterogeneous cameras and/or different shooting positions wireless networks. Real-time wireless video transmission typically has of the cameras. In particular, we consider focus mismatches across requirements on quality of service (QoS). However, wireless channels are views, i.e., such that different portions of a video frame can undergo unreliable and the channel capacities are time-varying, which may cause different blurriness/sharpness changes with respect to the severe degradation to video presentation quality. In addition, for portable corresponding areas in frames from the other views. We propose an devices, video compression and wireless transmission are tightly adaptive filtering approach for cross-view prediction in multi-view video coupled through the constraints on data rate, power, and delay. These coding. The disparity fields are exploited as an estimation of scene issues make it particularly challenging to design an efficient real-time depth. An Expectation-maximization (EM) algorithm is applied to video compression and wireless transmission system on a portable classify the disparity vectors into groups. Based on the classification device. In this paper, we take a cross-layer approach to this problem; our result, a video frame is partitioned into regions with different scene- objective is to maximize the video quality under the constraints of depth levels. Finally, for each scene-depth level, a two-dimensional filter resource and delay. Specifically, we minimize the end-to-end video is designed to minimize the average residual energy of cross-view distortion under the constraints of resource and delay, over the prediction for all blocks in the class. The resulting filters are applied to parameters in physical, link, and application (video) layers. This the reference frames to generate better matches for cross-view formulation is general and capable of capturing the fundamental aspects prediction. Simulation results show that, when encoding across views, of the design of wireless video communication systems. Based on this the proposed method achieves up to 0.8dB gain over current H.264 formulation, we study how the resources could be intelligently allocated video coding. to maximize the video quality and analyze the performance limits of the wireless video communication system under resource constraints. 6508-40, Session 8 6508-43, Session 9 RD-optimized competition scheme for efficient motion prediction A more aggressive prefetching scheme for streaming J. Jung, G. Laroche, France Télécom R&D (France); B. Pesquet- media delivery over the Internet Popescu, École Nationale Supérieure des Télécommunications J. Yuan, Q. Sun, S. Rahardja, Institute for Infocomm Research (France) (Singapore)

202 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007

As the applications of media streaming has grown rapidly in recent 6508-46, Session 9 years, efficient delivery of streaming media content over the Internet becomes an important area of research. Many research works studied Constant quality JPEG2000 rate control for digital this problem based on the client-proxy-server structure and proposed cinema various mechanisms to address this problem such as proxy caching and prefetching. While the existing techniques can improve the M. Smith, Consultant; J. D. Villasenor, Univ. of California/Los performance of accesses to reused media objects, they are not so Angeles effective in reducing the startup delay for first-time accessed media A Digital Cinema image typically has dimension 4096x2160 (4K) or objects. In this paper, we try to address this issue by proposing a 2048x1080 (2K) with 12 bit pixel data for each of the 3 color planes. At more aggressive prefetching scheme to reduce the startup delay of a frame rate of 24 frames per second, this gives uncompressed data first-time accesses. In our proposed scheme, proxy servers rates of 7.6 and 1.9 Gbps for 4K and 2K respectively. Even after aggressively prefetch media objects before they are requested. We compression, average data rates in the hundreds of Mbits/sec. are make use of servers’ knowledge about access patterns to ensure the encountered. Recently, the Society of Motion Picture and Television accuracy of prefetching, and we try to minimize the prefetched data Engineers’ size by prefetching only the initial segments of media objects. Results of trace-driven simulations show that our proposed prefetching (SMPTE) has chosen JPEG2000 as the standard to be used for scheme can effectively reduce the ratio of delayed requests by up to cinema compression. Thus, methods to appropriately trade off rate 38% with very marginal increase in traffic. and quality for JPEG2000-compressed movies will have high importance in the next several years as systems are designed and deployed. 6508-44, Session 9 In this paper we introduce a new framework for rate control that enables a JPEG2000 encoder to achieve a user-specified quality, and Rate-smoothed encoding for real-time video streaming therefore makes it possible to produce constant applications quality from frame-to-frame in an image sequence. The new method J. Wu, J. Cai, Nanyang Technological Univ. (Singapore) makes direct use of the same JPEG2000 coding pass data as the traditional approaches, and thus can easily be adopted at the back For real-time video streaming applications over the constant bit rate end of JPEG2000 encoding engines. We compare the proposed channels, it is highly desired that video signals can be encoded in not method with two other common rate control techniques only good average quality but also smooth video quality. However, in the case that the network resource is sufficiently large and the video for JPEG2000. quality has reached the target quality, the quality smoothing is not necessary and the rate smoothing is desired to avoid overusing the unnecessary network resource but also achieve a smoothed traffic 6508-47, Session 9 rate. In this paper, we propose a novel real-time rate-smoothed On preserving robustness-false alarm tradeoff in media encoding scheme by applying the low pass filtering idea, which is originally proposed for single-pass quality-smoothed video encoding. hashing Both theoretical analysis and experimental results show that the S. Roy, X. Zhu, J. Yuan, Institute for Infocomm Research proposed rate-smoothed encoding scheme can achieve a target (Singapore); E. Chang, National Univ. of Singapore (Singapore) average quality while significantly reducing the peak rate and the rate variance. We have further proposed a joint quality and rate smoothed Robustness of media hashing algorithm is primarily determined by encoding scheme, which can provide adaptive smoothing according three factors, (1) robustness false alarm tradeoff achieved by the to different network situations. Experimental results show that the chosen feature representation, (2) accuracy of the bit extraction step proposed joint smoothing scheme can make an optimal balance which maps the features to a sequence of bits and (3) the distance between the quality fluctuation and the rate fluctuation, and hence measure used to measure the similarity between two hashes. The improve the overall system performance. robustness false alarm tradeoff at the feature representation stage defines a limit on the tradeoff achievable by the hashing method. And the distance measure chosen measures how far this tradeoff is 6508-45, Session 9 preserved through the bit extraction step. This makes the bit extraction step extremely crucial in determining the performance of Joint source-channel rate allocation in parallel channels the hashing algorithm. Although this is a known fact, to our L. Pu, M. W. Marcellin, I. B. Djordjevic, B. V. Vasic, A. Bilgin, The knowledge there are no work that highlight its importance and elucidate the performance of their hashing method in terms of how Univ. of Arizona well it preserve this tradeoff. This paper presents a bit extraction step Parallel channels are often used in multimedia transmission for more based on the idea of locality preserving projections that closely reliable communication or more efficient bandwidth usage. They have preserves the robustness false alarm tradeoff achieved at the feature attracted significant research interest, and yielded many modern representation stage and hence achieves a good performance communications techniques, such as multicarrier modulation (MCM) tradeoff at the hash level. Moreover this work also highlights the or discrete multitone (DMT). In multi-channel systems, subchannels importance of an appropriate choice of similarity measure in can have differing characteristics, such as differing signal-to-noise measuring the similarity between hashes that enhances this ratios (SNR) or error rates. In the literature, there are allocation performance tradeoff further. algorithms that allocate power and bit rate such that the symbol error rate (SER) or bit error rate (BER) is the same among subchannels. In joint designs for multimedia transmission, many algorithms aim to 6508-48, Session 10 optimize the system performance for a given average SNR over subchannels, entirely ignoring specific subchannel variations. In some Edge-based automatic white balancing with linear cases, varying subchannel SNR is considered, but not fully utilized for illuminant constraint a joint design. In this paper, the potential of varying SNR over all H. H. Chen, C. Shen, National Taiwan Univ. (Taiwan) subchannels is exploited to design a joint source-channel coding scheme for robust transmission of scalable sources in multi-channel Automatic white balancing is an important function for digital systems. Scalable sources are transmitted via fixed-length packets. cameras. It adjusts the color of an image and makes the image look For each packet, the proposed algorithm selects a subchannel among as if it is taken under canonical light. White balance is usually all available subchannels, and chooses a channel code rate, such that achieved by estimating the chromaticity of the illuminant and then the expected length of the correctly received source data is using the resulting estimate to compensate the image. The grey world maximized. The resulting scheme offers unequal error protection method is the base of most automatic white balance algorithms. It (UEP) and applications to JPEG2000 transmission show that generally works well but fails when the image contains a large object significant UEP gains are achieved. or background with a uniform color. The algorithm proposed in this

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 203 Conference 6508: Visual Communications and Image Processing 2007

paper solves the problem by considering only pixels along edges and information as well as resolution and quality to achieve such effective by imposing an illuminant constraint that confines the possible colors streaming methods. Experimental results confirm the effectiveness of of the light source to a small range during the estimation of the the proposed system in terms of objective measures and subjective illuminant. By considering only edge points, we reduce the impact of evaluation. the dominant color on the illuminant estimation and obtain a better estimate. By imposing the illuminant constraint, we further minimize the estimation error. The effectiveness of the proposed algorithm is 6508-52, Session 10 tested thoroughly. Both objective and subjective evaluations show that the algorithm is superior to other methods. A 2D gel electrophoresis DNA image analysis algorithm with automatic thresholding 6508-49, Session 10 N. Kaabouch, R. R. Schultz, Univ. of North Dakota Polymerase chain reaction (PCR) and gel electrophoresis are two Highly automated image recomposition: picture you wish widely used techniques for genetic studies that require the bench you had taken scientist to perform many tedious manual steps. Advances in J. Luo, P. Lei, Eastman Kodak Co. automation are making these techniques more accessible, but detection and image analysis still remain labor-intensive. Although Have you ever lamented, “I wish I had taken a different picture of that several commercial software packages are now available, DNA image ‘Kodak Moment’ when everyone was smiling and no one blinked”? analysis still requires some intervention by the user, and thus a certain With image recomposition, we strive to deliver the “one-click fix” level of image processing expertise. To allow researchers to speed up experience to customers so they can easily create the perfect pictures their analyses and obtain more repeatable results, we present a fully that they never actually took. To accomplish this, a graphic user automated image analysis system for DNA or protein studies with high interface was created that integrates existing and new algorithms, accuracy. The proposed system is based mainly on four steps: including face detection, facial feature location, face recognition, automatic thresholding, shifting, filtering, and processing. The expression recognition, face appearance and pose matching, and automatic thresholding that is used to equalize the gray values of the seamless blending. Advanced modes include face relighting. This gel electrophoreses image background is one of the key and novel system is capable of performing image recomposition from a mixture operations in this algorithm. An enhancement is also used to improve of videos and still photos, with ease of use and a high degree of poor quality images that have faint DNA bands. Experimental results automation. show that the proposed method eliminates defects due to noise for good and average quality gel electrophoresis images, while it also improves the appearance of poor quality images. 6508-50, Session 10 Symmetric trinocular dense disparity estimation for car 6508-53, Session 10 surrounding camera array Automatic estimation and compensation of geometric Y. Chang, Y. Tsai, L. Chen, National Taiwan Univ. (Taiwan) distortions in copy videos This paper presented a novel dense disparity estimation method B. Chupeau, A. Massoudi, F. Lefebvre, THOMSON (France) which is called as symmetric trinocular dense disparity estimation. Also a car surrounding camera array application is proposed to The proliferation of pirate copies of feature films on peer-to-peer improve the driving safety by the proposed symmetric trinocular networks arouses a great interest to countermeasures such as the dense disparity estimation algorithm. The symmetric trinocular insertion of (invisible) forensic marks in projected movies, to deter their property is conducted to show the benefit of doing disparity illegal capture. The registration of pirate copies with the original estimation with three cameras. A 1D fast search algorithm is content is however a prerequisite to the recovery of such embedded described to speed up the slowness of the original full search messages, as severe geometric distortions often occur in illegally algorithms. And the 1D fast search algorithm utilizes the horizontal camcorded contents. After a brief state-of-the-art in image registration, displacement property of the cameras to further check the the paper details an algorithm for video registration, focusing on the correctness of the disparity vector. The experimental results show that compensation of geometric distortions. Control points are the symmetric trinocular property improves the quality and automatically extracted in original and copy pictures, followed by pre- smoothness of the disparity vector. and post-matching filtering steps to discard not relevant control points and erroneous matched pairs of control points respectively. This enables the accurate numerical estimation of an 8-parameter 6508-51, Session 10 homographic distortion model, used to register the copy frames with the original reference grid. Such an image registration algorithm is Surveillance system with mega-pixel scalable transcoder inserted into a general video registration scheme. Results are T. Hata, N. Kuwahara, Mitsubishi Electric Corp. (Japan); D. L. presented on both natural and synthetic test material. Schwenke, A. Vetro, Mitsubishi Electric Research Labs. This paper presents a video surveillance system that displays mega- 6508-54, Session 11 pixel streams effectively, while transmitting and processing the streams efficiently with limited resources such as bandwidth, A novel statistical learning-based rate distortion analysis computing power and display resolution. The system stores high- approach for multiscale binary shape coding resolution and high-quality video data encoded in JPEG2000 and Z. Chen, K. N. Ngan, The Chinese Univ. of Hong Kong (Hong Kong associated object metadata, which includes ROI (Region-of-Interest) information. Two streaming methods are proposed. In MCPTZ, China) multiple cameras are displayed with lower resolution and lower quality In this paper, we propose a statistical learning based approach to at the same time and the resolution and quality of the selected analyze the rate-distortion characteristics of multiscale binary shape cameras can be increased dynamically. This is done to decrease the coding. We employ the polynomial kernel function and incorporate total bit rate and facilitate the display of mega-pixel streams under rate-distortion related features for our support vector regression. restricted system resources. OLRH displays ROIs in high-resolution $\epsilon$-Insensitive loss function is chosen to improve the and high-quality and an overall view of the scene with low-resolution estimation robustness. The parameter tuning is also studied. Moreover, and low-quality simultaneously to observe each object in detail we discuss the feature selection which helps to improve the estimation without missing the overall scene context. We also propose accuracy. Comparing to the traditional method, our proposed JPEG2000 transcoder that uses a combination of three types of framework provides better rate distortion estimation not only on simple scalability: quality, resolution and position, and transcodes mega-pixel shapes but also on complex shapes. images efficiently in the compressed-domain based on the ROI

204 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Conference 6508: Visual Communications and Image Processing 2007

6508-55, Session 11 6508-58, Session 12 Drift-compensated coding optimization for fast bit-rate Motion estimation performance models with application reduction transcoding to hardware error tolerance P. Zhang, X. Ji, W. Gao, Institute of Computing Technology (China) H. Cheong, A. Ortega, Univ. of Southern California Bit rate adaptation is one of the basic problems in universal multimedia The progress of VLSI technology towards deep sub-micron feature access. Therefore, in many cases, there exists a strong requirement for sizes, e.g., sub-100 nanometer technology, has created a growing a very fast transcoding implementation. Usually, open-loop transcoding impact of hardware defects and fabrication process variability, which architecture is the desired solution with the lowest computational lead to reductions in yield rate. To address these problems, a new complexity. However, severe drift error makes it infeasible to most approach, named the System-Level Error Tolerance (ET) technique, applications. In this paper, we propose a drift-compensated coding has been recently introduced. Considering that a significant optimization scheme, by which generated bitstreams can be effectively percentage of the entire chip production is discarded due to minor transcoded into lower bitrate by open-loop transcoder with little drift imperfections, this approach is based on additionally accepting propagation. The encoder integrates a virtual open-loop transcoder, in imperfect chips that introduce imperceptible/acceptable degradation, which drift error is effectively compensated by suitably adjusting the RD so that the overall effective yield can be improved. The critical factor based mode selection and motion estimation for each macroblock. of the error tolerance approach is to accurately predict if a defective Simulation results show that compared with traditional coding, the chip will provide acceptable performance without having to plug it into proposed mode selection scheme can yield better coding efficiency the application. In this paper, we investigate the impact of hardware when rate reduction transcoding to low bitrate occurs. And meanwhile, faults on the video compression performance with a focus on the it does not degrade the coding efficiency in comparison with the normal motion estimation (ME) process. More specifically, we provide an single layer coding in H.264/AVC. analytical formulation of the impact of single and multiple stuck-at- faults within ME computation. We further present a model for estimating the system level performance degradation due to such 6508-56, Session 11 faults, which can be used for the error tolerance based decision strategy of accepting a given faulty chip. Furthermore, we show how Fast intermode decision algorithm in H.264 for IPTV different faults and ME search algorithms are related in terms of error broadcasting server tolerance and define the characteristics of search algorithm that lead G. Y. Kim, B. Y. Yoon, Electronics and Telecommunications to increased error tolerance. Finally, we show that different hardware Research Institute (South Korea); Y. Ho, Gwangju Institute of architectures performing the same metric computation introduce Science and Technology (South Korea) different error tolerance level and present the optimal ME hardware architecture in terms of error tolerance. The new video coding standard H.264/AVC employs the rate-distortion optimization (RDO) method for choosing the best coding mode. However, since it increases the encoder complexity tremendously, it is 6508-59, Session 12 not suitable for real-time applications, such as internet protocol television (IPTV) broadcasting services. Therefore we need a fast mode Super-resolution based on region-matching motion decision algorithm to reduce its encoding time. In this paper, we have estimation proposed a fast inter mode decision algorithm considering quantization O. A. Omer, T. Tanaka, Tokyo Univ. of Agriculture and Technology parameter (QP). We have focused on the frequency of the best mode (Japan) according to QP. In order to consider these characteristics, we check coded block pattern (CBP) which reflects the number of quantized We consider the problem of recovering a high-resolution (HR) frame discrete cosine transform (DCT) coefficients. We have developed the from a sequence of low-resolution (LR) frames. It is challenging to early 16x16 mode decision method using reference frame and the sub- design a super-resolution (SR) algorithm for arbitrary video optimal best mode as well as motion vectors of 16x16 and SKIP modes. sequences. Video frames in general cannot be related through global We also consider the early SKIP mode decision. Experimental results parametric transformation due to the arbitrary individual pixel show that the proposed algorithm reduces encoding time by 74.6% for movement between frame pairs. Hence a local motion model needs to the Baseline profile and 72.8% for the Main profile compared to H.264/ be used for frame alignment. An accurate alignment is the key to AVC reference software. success of reconstruction-based super-resolution algorithms. Motivated by this challenge we propose to employ region-matching technique for image registration in this paper. The proposed algorithm 6508-57, Session 12 consists of the alignment step to produce a blurred version of the HR frame and the restoration step to restore the HR frame. The Enhanced SAD reuse fast motion estimation experimental results of the proposed algorithm are compared with the K. L. Tang, K. N. Ngan, The Chinese Univ. of Hong Kong (Hong results of using affine, block matching, and optical flow motion Kong China) models. It is shown that the use of region matching for SR is very promising in producing higher quality images. H.264 is a new video coding standard which outperforms the previous video coding standards such as H.261 and H.263. It employs many advanced video coding techniques to improve the coding performance. Variable block size motion estimation (ME) and compensation is one of the techniques that contributes to the excellent performance of H.264 but it is computational intensive. In this paper, a fast SAD reuse ME algorithm is proposed which reuses SAD within the same macroblock and uses pattern-based ME and refinement search to reduce the computational complexity of the variable block size ME with a little degradation of coding performance in terms of PSNR and bitrate. Experimental results show that the proposed algorithm reduces the ME time by more than 90% on average with only a little degradation of coding performance when comparing with that of the Fast Full Search (FFS). Additionally, the proposed algorithm is faster than the fast ME algorithms adopted by the H.264 reference software JM.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 205 Participants

A An, Chang [6500-33]S11 Bang, Yousun [6493-12]S4, Bezryadin, Sergey N. Bouwer, James C. [6501-08]S2 Anchini, Rosario [6491-07]S2, [6493-19]S5 [6502-20]S7, [6502-23]S8 Bovey, John D. [6495-25]S9 Aach, Til 6497 ProgComm [6491-40]S9 Baraniuk, Richard G. Bhandarkar, Suchendra M. Bovik, Alan C. [6492-68]S1, Abbott, Kevin [6499-11]S3 Andersen, Tim L. 6500 [6498-15]S4 [6504-03]S1 [6499-12]S3, 6508 Abdeljaoued, Yousri ProgComm Barba, Dominique Bianchini, Carlo [6491-08]S2 ProgComm [6508-07]S2, [6508-23]S4 Anderson, David V. [6492-61]SA, [6508-32]S6 Bianco, Simone [6502-07]S3, Bowen, Francis R. [6497-18]S4 Abdelzaher, Tarek F. 6504 [6508-100]S13 Barbieri, Mauro [6506-22]S7 [6502-12]S4 Bowen, Jayson J. ProgComm Anderson, James A. D. W. Bares, Jan 6493 ProgComm Bicego, Manuele [6492-42]S9 [6493-45]S12 Abdollahian, Golnaz [6499-01]S1, [6499-18]S5, Barney Smith, Elisa H. 6500 Bichot, Elodie [6498-29]S7, Bozdagi Akar, Gözde [6506-20]S7 [6499-19]S5 ProgComm [6508-67]S13 [6508-90]S13 Abe, Takashi [6501-15]S3 Andersson, Mattias Barni, Mauro [6492-17]S3, Bici, Mehmet O. [6508-90]S13 Brambilla, Carla [6494-06]S2 Abufadel, Amer [6498-11]S3 [6493-26]S7, [6493-29]S8 6505 ProgComm, 6505 Bieber, Gerald [6507-23]S7 Braudaway, Gordon W. 6505 Achkar, Firas [6508-91]S13 Ando, Shigeru [6491-10]S3 S10 SessChr, [6505-29]S7, Bilcu, Radu C. [6502-26]S9 ProgComm Adachi, Satoru [6501-24]S3 Ando, Takamasa [6501-28]S4 [6505-42]S10 Bilgin, Ali [6508-45]S9 Bräuer, Andreas H. Adachi, Yudai [6501-18]S3 Andra, Srinivas [6500-03]S2 Barrera, Junior 6497 Binczak, Stéphane [6501-07]S2, [6501-11]S2, Adams, Andrew A. Andrews, Keith Review ProgComm [6503-16]S5 [6503-18]S6 [6499-01]S1, [6499-18]S5 Andrienko, Gennady Review Barrett, William A. [6500-13]S5 Bitelli, Gabriele [6491-28]S7 Bredthauer, Gregory R. Adar, Eytan Review Animali, Mauro [6501-14]S2 Barth, Erhardt Review, 6492 Bitlis, Burak [6498-27]S6 [6501-01]S1 Ademoye, Oluwakemi Ansari, Rashid 6508 ProgComm, 6492 S2 Black, John A. [6492-43]S9, Bredthauer, Richard A. [6504-09]S3 ProgComm SessChr, [6492-09]S2 [6492-66]SA, Review [6501-01]S1 Agaian, Sos S. 6497 Antonacopoulos, Apostolos Bashir, Faisal [6508-22]S4 Blagi, Vantier [6497-06]S2 Bressan, Marco [6493-57]S9 ProgComm, 6497 S2 6500 ProgComm Baskurt, Atilla M. [6508-86]S13 Blais, Francois 6491 S3 Brichon, Mathieu [6493-35]S9 SessChr, [6497-16]S4, Apostolopoulos, John G. 6508 Basu, Samit K. 6498 SessChr, [6491-05]S2, Brill, Michael H. 6492 [6497-40]SB ProgComm, 6508 S1 ProgComm, 6498 S5 [6491-11]S3 ProgComm Agam, Gady 6499 SessChr SessChr Blakeley, Nicholas D. Brinkschulte, Uwe 6495 ProgComm, 6499 S4 Aragon, Cecilia R. [6496-06]S2 Battiato, Sebastiano [6498-53]S12 ProgComm SessChr, [6499-06]S2, Aragon, David B. [6496-06]S2 [6506-26]S8 Bloom, Jeffrey A. 6505 S2 Brodersen, Jörg [6496-17]S4, [6500-11]S4, [6500-28]S10 Arai, Jun [6490A-43]S12 Bauer, Harald [6503-09]S3 SessChr, 6505 ProgComm, [6503-19]S6 Agar, A. U. 6493 ProgComm Argamon, Shlomo Baxter, Patrick [6501-26]S4 [6505-52]S11 Brooks, Dana H. [6498-01]S1 Aggarwal, Jake K. [6499-12]S3 [6500-28]S10 Bazhyna, Andriy V. Bloomberg, Dan S. Brooks, Frederick P. Aghito, Shankar M. Arivazhagan, Manivannan [6497-21]S5, [6498-18]S4 [6500-16]S6 [6490B-47]S13 [6508-75]S13 [6500-29]S11 Beal, Matthew J. [6500-30]S11 Blouke, Morley 6501 Chr, Brooks, Stephen [6492-13]S3 Agrafiotis, Dimitris Armstrong, Andrew J. Beaulieu, Mario [6492-59]SA 6501 S5 SessChr, 6501 S4 Broussard, Randy P. [6508-04]S1, [6508-25]S5 [6507-05]S2 Beermann, Markus SessChr, 6501 S3 SessChr, [6496-03]S1, [6497-25]S6 Agrawal, S. [6507-08]S3 Arnabat, Jordi [6493-28]S8 [6508-104]S13 6501 S1 SessChr, Brueckner, Andreas Ahmed, Mohamed N. Arney, Jonathan S. Beesley, Steven T. C. [6501-02]S1 [6501-11]S2 [6497-47]SB [6494-21]S8 [6507-05]S2 Bober, Miroslaw [6498-07]S2 Bruna, Arcangelo [6502-14]S5 Ahn, Chunghyun [6507-33]S9 Arpinen, Tero [6507-35]S9 Bei, Yun [6506-16]S6 Bockenbach, Olivier Brunette, Kynthia [6495-26]S9 Ahn, Sangwoo [6507-14]S4 Arregui, Damian [6493-57]S9 Belcher, Craig S. [6494-13]S5 [6498-22]S5, [6498-51]S5 Brunnström, Kjell E. Ahrens, James P. [6495-16]S6 Asada, Hideki [6490A-20]S5 Belmamoune, Mounia Bodegom, Erik 6501 S2 [6492-15]S3 Ahumada, Albert J. 6492 Asadi, Majid [6508-21]S4 [6506-16]S6 SessChr, 6501 ProgComm, Bruyelle, Jean-Luc ProgComm, 6492 S5 Asai, Toshihiro [6491-17]S4 Ben Chouikha, Mohamed [6501-02]S1 [6490A-70]SPS SessChr Astola, Jaakko T. 6497 Chr, [6493-49]S12 Bodenstorfer, Ernst Budagavi, Madhukar Aida, Tahito [6491-33]S8 [6497-43]SB, Bender, Walter R. 6492 [6496-17]S4, [6503-19]S6 [6508-38]S8 Aizawa, Kiyoharu 6506 [6505-63]S14, ProgComm Bodie, James [6502-31]S9 Bugnon, Thomas [6493-35]S9 ProgComm, 6508 [6508-90]S13 Benjelloun Touimi, Abdellatif Boggs, Kasey L. [6501-01]S1, Bull, David R. [6508-04]S1, ProgComm Atakli, Idris [6505-48]S10 [6505-62]S14 [6501-05]S1 [6508-25]S5 Akahane, Nana [6501-24]S3 Atallah, Mikhail J. [6505-31]S8 Bennour, Jihane [6505-44]S10 Bolas, Mark T. 6490B Chr, Bur, Alexandre [6492-47]S9 Akar, Gozde B. [6508-83]S13 Atsumi, Eiji 6502 ProgComm Benois-Pineau, Jenny 6490B S14 SessChr Burghart, Catherina R. Akar, Nail [6508-83]S13 Atzeni, Carlo [6491-29]S7 [6506-25]S8 Boliek, Martin P. [6500-17]S6 [6504-21]S7 Akarun, Lale [6492-42]S9, Au, Oscar C. L. 6505 Benson, Thomas M. Bolin, Mark R. [6508-108]S7 Burke, Barry E. [6501-41]S1 [6499-13]S3 ProgComm [6498-23]S5 Bölke, Torsten [6505-51]S11 Burns, Peter D. SC807 Inst, Akca, Devrim [6491-31]S8 Autrusseau, Florent Ben-Zvi, Ofer [6496-11]S3 Bolme, David S. [6495-03]S1 6494 S2 SessChr, 6494 S3 Akhriev, Albert [6497-37]SB [6505-24]S5 Beraldin, J.-Angelo 6491 Chr, Bolourchi, Masoud SessChr, 6494 ProgComm, Akil, Mohamed 6496 Avcibas, Ismail [6505-54]S12 6491 S2 SessChr, [6503-27]S7 [6494-16]S6, [6494-20]S7 ProgComm Avila, Lisa [6495-13]S5 [6491-05]S2, [6491-11]S3, Bonifazi, Giuseppe Burton, Robert P. [6495-23]S9 Akopian, David 6507 Aykac, Deniz [6503-01]S1 [6491-15]S4, [6491-29]S7 [6501-14]S2, [6501-38]S6, Butkiewicz, Thomas ProgComm Beraldin, J.-Angelo [6503-21]S7 [6495-21]S8 Alattar, Adnan M. 6505 [6491-40]S9 Borgeat, Louis [6491-05]S2 Butler, Darren [6498-52]S4 ProgComm, 6505 S11 B Berche, Stéphane J. Börner, Katy 6495 Chr Buxbaum, Bernd [6491-14]S3 SessChr [6493-49]S12 Borovikov, Anna [6500-04]S2 Buyukbayrak, Hakan Aleksic, Milivoje [6502-16]S6 Babaguchi, Noboru 6506 Beretta, Giordano B. 6493 S7 Borovikov, Eugene [6500-14]S5 Alenius, Sakari [6497-20]S5 ProgComm SessChr, [6493-20]S6, [6500-04]S2 Buzuloiu, Vasile V. Aliaga, Daniel G. [6508-18]S3 Badea, Cristian T. [6491-19]S4 [6494-06]S2 Borsack, Julie [6500-27]S10 [6502-17]S6 Allebach, Jan P. 6492 Bae, Sanghoon [6502-05]S2 Bergeron, Dan Review Boughorbel, Faysal Buzzi, Jérôme [6502-08]S3 ProgComm, 6493 Bae, Tea-Meon [6492-65]SA, Bergeron, R. Daniel [6490A-22]S6 ProgComm, 6493 S11 [6507-03]S1 [6495-14]S6 Bouguet, Jean-Yves SessChr, [6493-44]S12, Baik, Aron [6493-12]S4 Bergmair, Richard [6505-32]S8 [6506-27]S8 C [6498-27]S6, 6500 Baird, Henry S. [6500-33]S11 Bergman, Ruth [6497-09]S2 Boujemaa, Nozha 6506 ProgComm, [6502-25]S8, Bajaj, Nikhil [6502-25]S8 Berkner, Kathrin 6500 ProgComm Cahill, Adrian J. [6504-12]S4 [6505-30]S7, [6505-57]S12, Bajcsy, Peter Review ProgComm, [6500-17]S6, Bouman, Charles A. Cai, Jianfei 6507 S5 SessChr, [6505-68]S15, EI115X Chr Bajorski, Peter SC804 Inst, EI115X ProgComm SympComm, [6493-03]S2, 6507 S6 SessChr, 6507 S3 Alleysson, David [6502-09]S3 SC805 Inst, SC806 Inst Berman, Ari [6498-51]S5 [6493-04]S2, 6498 S8 SessChr, 6507 Chr, Allili, Madjid [6499-15]S4, Bal, Gulsher [6500-11]S4 Berns, Roy S. [6493-08]S3 SessChr, 6498 S9 SessChr, [6508-44]S9, [6508-88]S13 [6499-16]S4 Bala, Raja [6493-40]S11 Berretty, Robert-Paul M. 6498 Chr, [6498-41]S12, Caldelli, Roberto [6505-29]S7 Alquié, Georges [6493-49]S12 Balado, Félix [6505-12]S3, [6490A-39]S10, [6498-55]S13 Calvet, Robert J. [6502-36]S6 Altunbasak, Yucel 6508 [6505-18]S4 [6508-14]S3 Bourgeat, Pierrick T. 6503 Campanella, Marco ProgComm Balasubramanian, R. Berry, Richard L. [6501-02]S1 ProgComm [6506-22]S7 Alvino, Christopher V. [6499-03]S1 Bertrand, Gilles 6499 Boutellier, Jani J. [6498-05]S2, Campisi, Patrizio 6497 S4 [6498-13]S3 Balasubramanian, Vineeth ProgComm [6507-07]S3 SessChr, [6497-01]S1, Alyüz, Neşe [6499-13]S3 [6492-66]SA Bethge, Matthias [6492-08]S2, Boutin, Mireille 6508 [6497-19]S5, [6497-34]SB, Amer, Aishy [6508-87]S13, Balch, Tucker [6504-17]S7 [6492-10]S2 ProgComm, 6508 S3 [6505-24]S5, [6505-36]S9 [6508-91]S13 Bang, Sun-Kyong [6502-05]S2 SessChr, [6508-18]S3

206 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Participants

Canagarajah, Nishan Chen, Chang Wen Collon, Maximilien D’Apuzzo, Nicola 6491 S8 Divakaran, Ajay 6506 [6506-09]S3 [6507-26]S8, 6508 Chr, [6490B-46]S13 SessChr, [6491-24]S6 ProgComm, 6506 S2 Cancellaro, Michela [6508-26]S5, [6508-99]S13 Colonna, Brian R. [6497-15]S4 Darbon, Jerome [6498-25]S6 SessChr, 6506 SB [6505-63]S14 Chen, Chaomei Review Coltuc, Daniela [6497-41]SB Das, Pankaj K. [6492-45]S9 SessChr, [6506-04]S2 Candes, Emmanuel J. Chen, Ching-Yeh [6508-29]S5 Coltuc, Dinu [6505-60]S13 Das, Sukhendu [6499-03]S1 Djeraba, Chabane [6506-24]S8 [6498-17]S4 Chen, Chong [6508-63]S13 Comer, Mary L. 6498 S2 Davis, James W. [6491-04]S1 Djordjevic, Ivan B. [6508-45]S9 Canosa, Roxanne L. Chen, David C. [6502-33]S9 SessChr, [6498-12]S3, Davison, Brian D. 6500 Dmitrieva, Julia [6506-16]S6 [6500-07]S3 Chen, Homer H. [6508-48]S10 [6508-103]S13 ProgComm Do, Minh N. 6498 S10 Cao, Lanying [6497-24]S6 Chen, Hong Review Comesaña-Alfaro, Pedro Dayal, Umeshwar [6495-04]S1 SessChr, [6498-08]S2, Cao, Lei [6508-77]S13 Chen, Liang-Gee [6508-29]S5, [6505-18]S4, [6505-47]S10 de Haan, Gerard 6508 [6498-36]S10 Cappellini, Vito [6505-29]S7 [6508-50]S10, Conroy, Richard M. ProgComm Dodgson, Neil A. 6490A S9 Capra, Alessandro [6508-74]S13 [6491-13]S3 de Koning, Tim C. M. SessChr, 6490A Chr, [6502-14]S5 Chen, Mo [6505-25]S6, Cook, Jamie [6498-52]S4 [6507-24]S7 [6492-13]S3 Carli, Marco [6505-63]S14 [6505-53]S12 Cooper, Brian E. [6494-11]S4 de Queiroz, Ricardo L. Doermann, David S. SC808 Carlsohn, Matthias F. 6496 Chr Chen, Philip C. 6495 Cooper, Michael J. [6508-69]S13 Inst, 6500 ProgComm, Carlson, Jeffrey J. [6497-38]SB ProgComm [6501-41]S1 de Ridder, Huib 6492 6507 ProgComm Carnegie, Dale A. [6491-13]S3 Chen, Tsuhan 6506 Cooper, Ted J. 6502 S6 ProgComm Doërr, Gwenaël [6505-16]S4 Carpendale, M. Sheelagh T. ProgComm SessChr, 6502 ProgComm De Vleeschouwer, Christophe Doerschuk, Peter C. 6498 Review, [6495-07]S2 Chen, Wei-Yin [6508-74]S13 Cooprider, Nathan D. [6508-82]S13 ProgComm Carrato, Sergio [6497-22]S5 Chen, Wen [6505-27]S6 [6495-23]S9 de With, Peter H. N. Doherty, Peter [6501-41]S1 Castañeda, Román Chen, Xin-Hao [6508-72]S13 Costello, Kathleen [6496-10]S3, [6496-23]S3 Doi, Motonori [6499-07]S2 [6493-52]SA Chen, Yi-Hau [6508-29]S5 [6492-63]SA De Zwart, Siebe T. Dolinsky, Margaret 6490B Castorina, Alfio [6502-14]S5 Chen, Ying [6508-02]S1 Cournoyer, Luc [6491-05]S2, [6490A-26]S7 CoChr Castro, Ana Paula A. d. Chen, Ying-Ruei Y. [6491-11]S3 Dedic, Renato [6499-16]S4 Dolmiere, Thierry [6492-16]S3 [6497-27]S7 [6490A-60]SPS Courtellemont, Pierre Deepti, D. R. [6497-44]SB Domiciano, Marco A. P. Castro-Pareja, Carlos R. 6496 Chen, Yud-Ren [6501-40]S5 [6498-29]S7, [6508-67]S13 DeFanti, Thomas A. [6497-28]S7 ProgComm Chen, Zhenzhong Cox, Ingemar J. 6505 [6490A-03]S1 Dong, Yingfei [6504-14]S5 Cayre, François [6505-43]S10 [6508-54]S11 ProgComm, [6505-16]S4 Degerman, Johan H. Dorai, Chitra 6506 ProgComm Celenk, Mehmet Review Chen, ZhiQiang [6498-45]SA Coyle, Edward J. [6498-33]S8 [6498-44]SA Dore, Alessio [6508-21]S4 Celentano, Augusto 6506 Cheng, Hui [6498-52]S4 Craen, Piere [6501-07]S2 Deguillaume, Frédéric Dorrington, Adrian A. ProgComm Cheng, Kennedy [6492-22]S4 Craver, Scott A. [6505-11]S3, [6505-67]S15 [6491-13]S3 Celiktutan, Oya [6505-54]S12 Cheng, Wei-Chung [6505-48]S10 DeJong, Eric S. MeetingVIP Doufexi, Angela [6508-04]S1 Cha, Jihun [6507-14]S4 [6492-39]S8 Crawfis, Roger Review Del Bimbo, Alberto 6506 Dougherty, Edward R. 6496 Cha, Kyung-Hoon Cheok, David A. 6504 Cree, Michael J. [6491-13]S3 ProgComm ProgComm, 6497 Chr [6490A-53]SPS ProgComm Crete, Frederique [6492-16]S3 Del Mastio, Andrea Drimbarean, Alexandru Chakareski, Jacob Cheong, Hye-Yeon Creusere, Charles D. 6508 [6491-27]S7 [6502-17]S6 [6508-101]S13 [6508-58]S12 ProgComm Delp, Edward J. 6505 Chr, Du, David H. 6504 ProgComm, Chalmers, Alan 6507 Cheong, Won-Sik Creutzburg, Reiner 6497 6505 S8 SessChr, [6504-14]S5 ProgComm [6494-29]S10, [6507-14]S4 ProgComm, 6497 S5 [6505-30]S7, [6505-57]S12, Du, Yan-Jun [6493-55]SA Chambah, Majed [6494-10]S4 Chi, Ed H. Review SessChr, 6507 S1 SessChr, [6505-68]S15, [6506-20]S7, Du, Yingzi [6491-43]S9, Chan, Din-Yuen [6508-95]S13 Chiang, Pei-Ju [6505-68]S15 6507 S2 SessChr, 6507 [6507-02]S1, 6508 [6494-13]S5, [6497-17]S4, Chandra, Surendar 6504 Chien, Wei-Ta [6508-95]S13 Chr, [6507-36]S9, ProgComm, 6508 S5 [6497-18]S4 ProgComm, [6504-06]S2, Chikatsu, Hirofumi 6491 [6507-37]S9 SessChr, [6508-02]S1, Duarte, Marco F. [6499-02]S1 6507 ProgComm ProgComm Croce Ferri, Lucilla [6508-06]S1, [6508-27]S5 Dubois, Eric 6508 ProgComm Chandramouli, Rajarathnam Chiricota, Yves [6495-22]S9 [6505-26]S6 Denney, Thomas S. 6498 Dudas, Jozsef [6502-33]S9 [6505-06]S1 Chiu, George T. C. Cruz, Camilo [6504-20]S7 ProgComm Dufaux, Frederic SC766 Inst, Chandrasekaran, Shivkumar [6502-25]S8, [6505-30]S7, Cucchiara, Rita 6506 Depalov, Dejan [6492-11]S2 6508 ProgComm [6505-49]S11 [6505-57]S12, ProgComm Derdar, Salah [6499-15]S4 Dugelay, Jean-Luc E. Chang, Chun-Yuan [6505-68]S15 Cui, Luke C. 6494 Chr, Derefeldt, Gunilla A. M. 6492 [6505-44]S10 [6508-95]S13 Cho, Min-ki [6493-19]S5 [6494-27]S9 ProgComm Dulong, Carole [6506-27]S8 Chang, Chuo-Ling [6508-30]S6 Cho, Sammo [6507-33]S9 Cutu, Florin Review DeRego, Paul J. [6492-50]S10 Dumke, Joel [6498-12]S3 Chang, Edward T. [6494-09]S4, Cho, Seongyoon [6495-01]S1 Cvek, Urska Review Derksen, Harm [6508-15]S3, Dummann, Uwe 6507 [6502-18]S6 Choe, Wonhee [6502-27]S9 Czyrnek, Miroslaw [6504-08]S2 [6508-16]S3 ProgComm Chang, Edward Y. 6506 CoChr Choh, Heui-Keun [6493-19]S5 Desai, Nehal [6495-16]S6 Dunham, Joshua [6498-01]S1 Chang, Ee-Chien 6505 Choi, Donchul [6493-18]S5 Deshpande, Sachin G. Duparré, Jacques [6501-11]S2, ProgComm, [6508-47]S9 Choi, Jiyoung [6498-46]SA D [6507-11]S4 [6503-18]S6 Chang, Gao-Wei [6493-16]S4, Choi, KwangNam [6496-22]S5 Desurmont, Xavier Dupont, Florent [6508-86]S13 [6493-36]S10, Choi, Pyung [6501-35]S6 da Fontoura Costa, Luciano F. [6490A-70]SPS, Durdle, Nelson G. [6497-12]S3 [6494-28]S10, [6501-36]S6, Choi, Seo Young [6493-07]S3 6496 ProgComm, 6503 [6491-03]S1, 6496 Duttweiler, Fred [6501-08]S2 [6503-22]S7 Chou, Cheng-Fu [6508-95]S13 ProgComm ProgComm, [6496-14]S4, Dykstra-Erickson, Elizabeth Chang, Remco [6495-21]S8 Choudhary, Ramesh K. Dai, Qionghai [6508-03]S1, [6496-26]S5 6507 ProgComm Chang, Samuel H. [6496-22]S5 [6497-32]S8 [6508-61]S13, Dias, Paulo [6490A-63]SPS Chang, Yu-Lin [6508-50]S10 Christel, Michael G. 6506 S5 [6508-66]S13, Díaz, José A. [6497-42]SB Chao, Hui 6500 ProgComm, SessChr, [6506-05]S2 [6508-78]S13 DiCarlo, Jeffrey M. 6502 Chr, E EI115X Chr Christodoulou, Lakis Dalal, Edul N. [6494-19]S7 6502 S4 SessChr Chao, Kaunglin [6501-40]S5 [6508-102]S13 Dalton, John C. 6492 Dill, Joe Review Earl, John [6505-45]S10 Chapdelaine, Claude Christopher, Galbraith ProgComm DiMaio, Simon P. [6504-21]S7 Easton, Roger L. [6500-07]S3, [6492-59]SA [6492-54]S10 Daly, Scott J. 6492 Chr, 6492 Dimiccoli, Mariella [6508-11]S2 [6500-10]S4 Chapman, Glenn H. Chua, TatSeng 6506 S8 SessChr, [6492-38]S8 Dimitrov, Rouslan [6496-21]S5 Eberst, Christof [6503-09]S3 [6502-33]S9 ProgComm Damiani, Luca [6501-14]S2 Ding, Fujian [6501-40]S5 Ebert, David S. Review Chareyron, Gaël [6506-02]S1 Chung, In Jae [6490A-27]S7 Dammann, John F. Ding, Li-Fu [6508-74]S13 Ebisu, Hiromitsu [6490A-33]S8 Chassery, Jean-Marc Chupeau, Bertrand [6490A-68]S1 Ding, Wenpeng [6508-98]S13 Ebrahimi, Touradj SC766 Inst, [6505-60]S13 [6508-53]S10 Danahy, Ethan E. [6497-40]SB Ding, Xiaoqing 6500 6508 S2 SessChr, 6508 Chatapuram-Krishnan, Ciaramello, Francis Dance, Christopher R. ProgComm, [6500-08]S3, ProgComm, [6508-07]S2, Narayanan [6492-43]S9, [6492-21]S4 [6493-57]S9 [6500-25]S8, [6503-23]S7 [6508-23]S4 [6492-66]SA Cilio, Andrea [6507-34]S9 Dandekar, Omkar DiPaola, Steve [6492-01]S1 Eck, Ralf [6495-12]S5 Chattopadhyay, Siddhartha Cittadini, Ferruccio [6498-21]S5 Dittmann, Jana SC816 Inst, Eckel, Christian [6496-17]S4, [6504-03]S1 [6493-49]S12 Daneu, John L. [6501-41]S1 6505 S5 SessChr, 6505 [6503-19]S6 Chaumont, Marc [6495-20]S8, Claypool, Kajal [6504-10]S3 Dang, Philip P. 6496 CoChr, [6505-04]S1, Edirisinghe, Eran A. [6508-64]S13, Claypool, Mark 6504 ProgComm [6505-21]S5, [6505-39]S9 [6496-01]S1, [6507-05]S2 [6508-65]S13 ProgComm, [6504-04]S1, D’Angelo, Angela [6492-17]S3 Edmundson, Kenneth [6504-10]S3 Dannberg, Peter [6503-18]S6 [6491-26]S6

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 207 Participants

Effat, Hala M. A. [6493-53]SA, Fishbain, Barak Ganev, Ivan [6504-24]S8 Greenwade, L. Eric 6495 Han, Jun-Gong [6506-15]S5 [6493-54]SA [6490A-55]SPS, Gao, Wen [6508-01]S1, ProgComm Han, Seong Wook Effelsberg, Wolfgang [6496-11]S3 [6508-55]S11, Gregor, Jens [6498-23]S5 [6493-47]S12 [6507-01]S1 Fisk, Lee Ann [6495-13]S5 [6508-106]S13 Greig, Darryl [6497-09]S2 Han, Zhun [6508-78]S13 Egami, Norifumi [6501-06]S2 Fitzpatrick, Darren M. García Ortiz, Juan Pablo Greiner, Guenther Review Handley, John C. 6497 Egawa, Yoshitaka [6492-57]SA [6499-20]S5 [6507-27]S8 Griffin, Lewis D. [6492-07]S2 ProgComm Egiazarian, Karen O. 6497 Chr, Flack, Julien C. 6490A S5 Garcia-Lamont, Jair Grigat, Rolf-Rainer Hands, David S. [6492-22]S4 6497 S1 SessChr, SessChr, [6490A-21]S5 [6496-28]S5 [6502-35]S9 Haneda, Eri [6493-03]S2 [6497-21]S5, [6497-43]SB, Flierl, Markus H. [6508-89]S13, Gargiulo, Aldo [6503-21]S7 Grinstein, Georges G. Review Hanjalic, Alan 6506 Chr [6498-18]S4, [6505-63]S14, [6508-96]S13 Gasparini, Francesca Griwodz, Carsten 6504 Chr Hanks, Kristin [6495-26]S9 [6507-09]S3 Floeder, Steven P. 6503 [6502-07]S3, [6502-12]S4 Gröhn, Matti T. 6495 Chr, 6495 Hännikäinen, Marko Eglin, Veronique [6500-12]S5 ProgComm Gaubatz, Matthew D. S4 SessChr, 6495 S5 [6507-35]S9 Eick, Steve Review Fofi, David 6503 ProgComm, [6492-12]S3 SessChr Hansen, D. Michael Eid, Ahmed [6494-11]S4 6503 S4 SessChr, 6503 S3 Gaykema, Frans 6494 Groller, Eduard Review [6500-10]S4 Eide, Viktor S. Wold SessChr ProgComm Gross, Markus Review Hansen, Lone [6491-39]S9 [6504-23]S7 Foley, John M. [6492-18]S4 Ge, Jinghua [6490A-03]S1 Grossman, David Hanusch, T. [6491-31]S8 Ekin, Ahmet 6506 S3 SessChr, Fonseca, Leila M. G. Gee, Christelle [6497-14]S4 [6500-28]S10 Hanzo, Lajos 6507 ProgComm [6506-03]S1 [6497-06]S2 Geisler, Jürgen [6495-12]S5 Grosso, Enrico [6492-42]S9 Hao, Ming C. 6495 CoChr, El Saddik, Abdulmotaleb Forchhammer, Søren Geisler, Wilson [6492-03]S2 Groth, Dennis Review 6495 S2 SessChr, [6504-02]S1 [6508-75]S13 Gerhardt, Jeremie Gruen, Armin 6491 [6495-04]S1 El-Hakim, Sabry F. 6491 Ford, Ralph M. 6503 [6493-41]S11 ProgComm, [6491-31]S8 Hardeberg, Jon Y. ProgComm, [6491-30]S7 ProgComm Gerth, Victor [6506-17]S6 Gschwind, Rudolf [6493-02]S1, [6493-41]S11 Eliassen, Frank [6504-23]S7 Fornaro, Peter R. [6502-32]S9 Gerwinn, Sebastian [6493-06]S2, [6502-32]S9 Harrington, Steven J. 6500 Elion, Chris [6498-28]S6 Foroud, Tatiana [6491-18]S4 [6492-10]S2 Guarnieri, Gabriele ProgComm, [6500-18]S6, Ellenrieder, Marc M. 6503 Forsell, Camilla [6495-10]S4 Gesquière, Gilles [6495-20]S8 [6497-22]S5 EI115X ProgComm ProgComm Fossum, Eric R. [6502-36]S6 Gevers, Theo 6506 CoChr Guest, Clark C. [6492-45]S9 Harrold, Jonathan Ellisman, Mark H. [6501-08]S2 Fossum, Robert M. Ghazal, Mohammed Guichard, Frédéric [6490A-19]S5 Emoto, Masaki [6492-53]S10 [6508-15]S3 [6508-87]S13 [6502-08]S3 Harvey, Euan S. [6491-01]S1 Emptoz, Hubert [6500-12]S5 Franceschetti, Andrea Ghinea, Gheorghita Guidi, Gabriele 6491 Hasegawa, Jun [6494-33]S5 Engbers, Rene [6506-03]S1 [6492-49]S10 [6504-09]S3 ProgComm, 6491 S6 Hashimoto, Masashi Engel, Dominik [6505-66]S15 Frank, Alan M. 6491 Giannetti, Fabio [6500-20]S7 SessChr, [6491-06]S2, [6502-04]S2 Erbacher, Robert F. 6495 Chr, ProgComm Gilboa, Guy [6498-26]S6 [6491-08]S2, [6491-29]S7 Hashimoto, Shuji [6497-10]S3, 6495 S1 SessChr, 6495 S8 Franz, Elke [6505-55]S12 Gille, Jennifer 6492 Guilvard, Alexandre [6503-14]S5 SessChr Fridrich, Jessica 6505 ProgComm [6501-20]S3 Hata, Toshihiko [6508-51]S10 Ercil, Aytul [6500-14]S5 ProgComm, 6505 S7 Gilson, Kevin J. [6490A-01]S1 Guleryuz, Onur G. 6508 Hatori, Yoshinori [6508-92]S13 Erdelyi, Tamas [6507-07]S3 SessChr, [6505-01]S1, Girado, Javier I. [6490A-03]S1 ProgComm Hauer, Enrico [6505-51]S11 Ertl, Thomas Review [6505-02]S1, [6505-25]S6, Girardi, Stefano [6491-30]S7 Gupta, Maya R. [6493-45]S12 Hauptmann, Alexander G. Eschbach, Reiner 6493 Chr, [6505-53]S12 Girelli, Valentina A. Gupta, Shalini [6499-12]S3 6506 ProgComm 6493 S1 SessChr, 6493 Frieder, Gideon [6500-11]S4 [6491-28]S7 Gurram, Prudhvi K. Hauser, Robert [6501-33]S6 S12 SessChr, [6493-05]S2 Frieder, Ophir [6500-11]S4, Girod, Bernd [6508-30]S6, [6490A-10]S3 Havre, Susan L. Review Eskicioglu, Ahmet M. [6500-28]S10 [6508-89]S13 Gurses, Eren [6508-83]S13 Hayat, Khizar [6495-20]S8 [6494-07]S3, 6505 Fritsch, Dieter 6491 Giuffrida, Giovanni Gururajan, Arunkumar He, Feng [6500-25]S8 ProgComm, 6505 S15 ProgComm [6506-26]S8 [6503-03]S1 He, Jie [6507-28]S8 SessChr, [6505-59]S13 Froner, Barbara [6490A-16]S4 Giusti, Christian [6492-46]S9 Gusev, Dmitri [6494-18]S7 He, Zhihai 6507 ProgComm, Etoh, Takeharu G. Frossard, Pascal 6504 Gloe, Thomas [6505-55]S12 Gustavson, Stefan [6507-19]S6, 6508 S9 [6491-34]S8 ProgComm, [6508-101]S13 Glossner, John C. [6507-08]S3 [6501-12]S2 SessChr, [6508-42]S9, Fruchterman, James R. Godard, Ludovic [6505-62]S14 Gustavsson, Tomas [6508-97]S13 [6500-26]S9 Goddard, Iain [6498-51]S5 [6498-44]SA Healey, Christopher G. Review, F Fu, Dongdong [6505-58]S12 Godding, Robert [6491-23]S6 Gutierrez, Roman C. [6495-06]S2 Fujii, Toshiaki [6490A-58]SPS, Godin, Guy [6491-05]S2 [6502-36]S6 Hefeeda, Mohamed M. Fageth, Reiner [6493-33]S8 [6490A-59]SPS Gokturk, Salih B. [6500-06]S3 Guzma, Vladimír [6507-34]S9 [6504-01]S1 Fairchild, Mark D. 6494 Fujimoto, Katsuhito Goldmann, Lutz [6506-08]S3 Hegazy, Mohamed N. ProgComm [6500-09]S4 Golik, Borys [6502-22]S7 [6493-53]SA, [6493-54]SA Fan, Zhigang Z. [6493-05]S2 Fujioka, Takeshi [6501-03]S1 Goljan, Miroslav [6505-25]S6, H Heindl, Christoph [6503-09]S3 Farasat, Yousef [6492-48]S10 Fujishima, Makoto [6501-03]S1 [6505-53]S12 Hellwagner, Hermann Farin, Dirk [6496-10]S3 Fukuda, Tsuyoshi Goma, Sergio R. [6502-16]S6 Ha, Yeong-Ho [6493-09]S3, [6504-18]S7 Farinella, Giovanni M. [6493-38]S10 Gomila, Cristina [6508-39]S8 [6493-34]S9 Helmbold, David P. [6506-26]S8 Fukushima, Norishige Goncalves, Joao G. M. 6491 Haakma, Reinder [6492-19]S4 [6499-20]S5 Farnand, Susan P. 6494 S9 [6490A-59]SPS ProgComm Habib, Ahsan 6504 ProgComm Hemami, Sheila S. SC812 Inst, SessChr, 6494 ProgComm Fuller, Andrew [6490B-47]S13 Gonzalez-Prats, Miguel Habib, Khaled J. [6501-29]S5 6492 ProgComm, Farrell, Joyce E. SC762 Inst Funatsu, Ryohei [6502-24]S8 [6505-36]S9 Haddleton, Graham P. 6491 [6492-12]S3, [6492-21]S4, Farsiu, Sina [6498-24]S6 Funt, Brian V. [6492-28]S6, Gonzo, Lorenzo [6491-30]S7 ProgComm, EI104X Chr [6492-64]SA Favalora, Gregg E. 6490A S6 [6492-67]SA Gooch, Amy [6492-35]S7 Hagan, Hans Review Henry, Matthew [6495-26]S9 SessChr, 6490A Furon, Teddy 6505 Gopal, Sucharita [6498-09]S3 Hagebeuker, Bianca Hepting, Daryl H. [6495-24]S9 ProgComm ProgComm, [6505-43]S10 Gormish, Michael J. [6491-14]S3 Hequet, Eric F. [6503-03]S1 Federico, Matta [6505-44]S10 Fürtler, Johannes [6500-17]S6 Haggren, Henrik G. A. 6491 Hernandez, Luis [6500-04]S2 Fedorovskaya, Elena A. 6492 [6496-17]S4, [6503-19]S6 Gorria, Patrick [6503-11]S4 ProgComm Hersch, Roger D. ProgComm Gotchev, Atanas P. 6497 Hahlweg, Cornelius F. [6493-35]S9, [6493-39]S10, Feghali, Rosario [6497-39]SB ProgComm, 6497 S3 [6503-07]S2 6493 ProgComm, 6506 Feiszli, Matt [6498-14]S4 G SessChr, [6508-90]S13 Hains, Charles M. ProgComm Feng, Guotong [6500-17]S6 Gou, Hongmei [6505-28]S7 [6493-42]S11 Hertel, Dirk W. 6494 Feng, Shuo [6508-17]S3 Gabbouj, Moncef [6507-06]S2 Govindaraju, Venu Häkkinen, Jukka P. ProgComm, 6494 S7 Feng, Wuchi [6504-22]S7 Gabele, Helke [6502-29]S9 [6500-32]S11 [6494-01]S1, [6507-21]S6 SessChr, [6494-09]S4 Feng, Wu-chi 6504 ProgComm Gader, Paul D. 6497 Goyal, Abhishek [6498-53]S12 Halvorsen, Pål 6504 Herzog, Patrick G. 6493 Fergason, James L. ProgComm Graebling, Pierre 6496 ProgComm ProgComm [6490A-72]S11 Gagnon, Langis [6492-59]SA ProgComm Hämäläinen, Timo D. Heynderickx, Ingrid E. J. Fernandez-Pineda, Cristobal Gai, Jiading [6497-04]S1, Gravier, Erwan [6498-03]S1 [6507-35]S9 [6490A-17]S4 [6491-25]S6 [6508-09]S2 Gravogl, Klaus [6496-17]S4, Hamamoto, Takayuki Hiddink, Martin [6490A-26]S7 Ferré, Pierre [6508-25]S5 Gallegos-Funes, Francisco J. [6503-19]S6 [6501-16]S3, [6502-34]S9 Hirai, Tadaaki [6501-06]S2 Ferreira, Carlos M. [6496-20]S5, [6497-46]SB Grecos, Christos [6507-05]S2 Hamann, Bernd [6495-05]S2 Hirayama, Teruo [6501-04]S1 [6492-60]SA Galli, Raffaello [6490A-05]S1 Green, Phil J. 6493 Hamid, Krim [6508-17]S3 Hirayama, Yuzo [6490A-42]S12 Figuera, Maribel [6493-04]S2 Gamadia, Mark N. SC809 Inst ProgComm, 6493 S5 Hampapur, Arun 6506 Fiorio, Christophe [6499-14]S4 Gandhi, Bhavan [6507-12]S4 SessChr, [6493-27]S8 ProgComm

208 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Participants

Hisatake, Shintaro Imiya, Atsushi 6499 Joutel, Guillaume [6500-12]S5 Kercher, Andrew K. Koike, Takafumi [6490A-28]S7 [6490A-61]SPS, ProgComm Ju, Junchang [6498-09]S3 [6503-01]S1 Koivusaari, Jari J. [6507-06]S2, [6490A-64]SPS Ince, Serdar [6490A-14]S3, Judd, Jamison E. [6495-26]S9 Khan, Javed I. [6504-05]S1 [6507-09]S3 Hix, Deborah Review [6490A-23]S6 Jumisko-Pyykkö, Satu Khanna, Nitin [6505-30]S7, Kokaram, Anil C. 6506 Ho, Yo-Sung [6508-56]S11 Inselberg, Alfred Review [6507-21]S6 [6505-57]S12 ProgComm Hochheiser, Harry Review Ip, Horace H. S. [6496-27]S5 Jun, Jiang [6496-27]S5 Khater, Hatem A. [6497-03]S1 Kolaczyk, Eric D. [6498-09]S3 Hollander, Ari J. [6490A-04]S1 Iqbal, Razib [6504-02]S1 Jung, Cory G. [6502-33]S9 Kikinis, Ron [6504-21]S7 Kolekar, Abhijeet [6504-22]S7 Holliman, Nicolas S. 6490A Isaacson, David [6498-50]S1 Jung, Joel [6508-40]S8 Kikuchi, Koji [6501-04]S1 Kollin, Joel S. [6490A-04]S1 ProgComm, 6490A S10 Ishii, Yuki [6502-13]S5 Kikuta, Kengo [6490A-29]S7 Komatsu, Takashi SessChr, [6490A-16]S4 Ishiwata, Hiroaki [6501-15]S3 Kim, Chang-Yeong [6497-45]SB, [6502-06]S2, Honda, Hiroto [6492-57]SA Isklandar, Danny D. K [6492-29]S6, [6494-03]S1, [6502-13]S5, [6502-15]S5 Honda, Takeshi [6497-17]S4 [6502-27]S9 Komogortsev, Oleg V. [6490A-56]SPS Itoh, Goh [6492-57]SA Kaabouch, Naima Kim, Cheon-Seog [6492-65]SA [6504-05]S1 Hong, Hunsop [6508-12]S2 Itti, Laurent 6492 ProgComm [6508-52]S10 Kim, Choon-Woo 6493 S4 Komulainen, Jeppe Hong, Hyungki [6490A-27]S7 Ivanov, Yuri [6508-20]S4 Kachelriess, Marc [6498-22]S5 SessChr, 6493 ProgComm, [6494-01]S1 Hong, Wei [6508-16]S3 Ives, Robert W. [6496-03]S1, Kacker, Dhiraj EI115X [6493-48]S12 Kong, Jae-Sung [6501-34]S6, Hopper, Darrel G. [6497-25]S6 ProgComm Kim, Dae-sik [6490A-25]S7, [6501-35]S6 [6490A-34]S8 Iwamoto, Tsuyoshi Kagawa, Keiichiro [6490A-53]SPS Kong, Jiantao [6504-24]S8 Horesh, Nadav [6490A-08]S2 [6501-17]S3 [6501-18]S3, [6501-22]S3 Kim, Eun Soo [6490A-65]SPS, Konrad, Janusz 6490A Horita, Yuukou EI115X Izquierdo, Ebroul 6506 Kahlmann, Timo [6491-12]S3 [6490A-67]SPS ProgComm, 6490A S3 ProgComm ProgComm Kaistinen, Jyrki [6494-01]S1 Kim, Geon [6507-33]S9 SessChr, [6490A-14]S3, Hornsey, Richard I. Izquierdo-Gil, Maria Amparo Kakeya, Hideki [6490A-18]S4, Kim, Geun Yong [6508-56]S11 [6490A-23]S6, [6501-27]S4 [6491-25]S6 [6490A-33]S8 Kim, Hak Soo [6508-99]S13 [6490A-37]S9, 6508 Hornung, Hervé [6502-08]S3 Kalker, Ton 6505 ProgComm, Kim, In Cheol [6500-24]S8 ProgComm Hörster, Eva [6506-13]S5 6505 S3 SessChr Kim, Inji [6492-27]S6 Koo, Jae-phil [6490A-53]SPS Hotta, Aira [6490B-48]S13 J Kalva, Hari [6504-20]S7, Kim, Jae-Hyun [6501-31]S6 Kooij, Robert E. [6507-24]S7 Hotz, Ingrid [6495-05]S2 [6508-102]S13 Kim, Kiduk [6493-13]S4 Kooima, Robert L. Howard, Dana [6502-25]S8 Jääskeläinen, Pekka Kalyanasundaram, P. Kim, Kyeong-Man [6493-18]S5 [6490A-03]S1, Hrimech, Hamid [6503-02]S1 [6507-34]S9 [6496-08]S2 Kim, Kyung-Do [6501-35]S6 [6490B-52]S14 Hsieh, Jiang [6498-41]S12 Jacobs, Richard [6492-22]S4 Kamata, Hiroe [6501-04]S1 Kim, Minwoo [6498-46]SA Kopf, Stephan [6507-01]S1 Hsu, ChengHsin [6504-01]S1 Jaeger, Anke [6502-21]S7 Kamon, Koichi [6501-17]S3 Kim, Munchurl [6507-15]S4 Koplowitz, Jack 6499 Hu, Chia-Lun J. [6497-30]S8 Jaimes, Alejandro 6506 Kanbara, Masayuki Kim, Sang Ho [6493-18]S5, ProgComm Hu, Jianying 6500 ProgComm ProgComm [6491-17]S4 [6493-48]S12 Koren, Israel [6502-33]S9 Hu, Junn-Yen [6508-95]S13 Jain, Ashish [6490A-37]S9 Kanbe, Hideo [6501-03]S1 Kim, Sang-Heon [6501-34]S6 Koren, Zahava [6502-33]S9 Hu, Yanming [6496-25]S5 Janesick, James SC504 Inst Kang, Byoung-Ho [6492-29]S6 Kim, Se-Eun [6493-19]S5 Korenaga, Tsuguhiro Hua, Guogang [6508-26]S5 Jang, Mijung [6493-18]S5 Kang, Sung-Hyun [6502-05]S2 Kim, Seon H. 6504 ProgComm [6501-28]S4 Huang, Jiwu [6505-09]S2, Janoos, Firdaus [6491-04]S1 Kang, Xiangui [6505-13]S3 Kim, Soonchoul [6507-33]S9 Korshunov, Pavel [6504-25]S8 [6505-13]S3 Jansson, Peter A. Kani, Yuya [6501-03]S1 Kim, Sung-Su [6492-29]S6 Kosara, Robert Review Huang, Jun [6502-25]S8 [6498-27]S6 Kankanhalli, Mohan S. 6506 Kim, Youn Jin [6494-03]S1 Kossak, Felix [6503-10]S3 Huang, Kuo-Chung Jasinschi, Radu S. ProgComm Kim, Yun-Tae [6493-19]S5 Kostis, Helen-Nicole G. [6490A-31]S8 [6506-03]S1 Kannenberg, John Kimachi, Akira [6491-10]S3 [6490B-52]S14 Huang, Wan-Jian Jasinski, Jacek [6498-34]S9 [6490B-52]S14 Kimmel, Jyrki S. [6507-04]S2 Kot, Alex C. 6508 ProgComm [6490A-31]S8 Jeannic, Guillaume Kanto, Hiroaki [6501-23]S3 Kimura, Masao [6501-03]S1 Kotlyar, V. [6507-08]S3 Huang, Yong [6493-10]S3 [6508-32]S6 Kanungo, Tapas 6500 Kimura, Nobutaka [6503-15]S5 Koval, Oleksiy J. [6505-41]S9, Hügli, Heinz [6492-47]S9 Jenkin, Robin B. 6494 ProgComm Kimura, Tetsuya [6490A-33]S8 [6505-67]S15 Hung, Yi-Ping [6490A-60]SPS ProgComm, 6494 S4 Karam, Lina J. 6508 Kinicki, Robert [6504-04]S1 Kozintsev, Igor V. [6506-27]S8 Hunn, John D. [6503-01]S1 SessChr, 6494 S5 SessChr, ProgComm Kise, Michio [6503-04]S1 Kpalma, Kidiyo [6500-02]S2 Hurst, Matthew F. 6500 [6494-23]S8 Karaman, Mustafa [6506-08]S3 Kiyavash, Negar [6505-20]S4 Kraetzer, Christian [6505-04]S1 ProgComm Jenkins, Colleen A. Karimov, Pavel [6491-34]S8 Klein, Stanley A. 6492 Kramm, Matthias [6492-14]S3 Hutchinson, Tara C. [6497-38]SB Karl, W. C. 6508 ProgComm ProgComm Krasic, Charles C. [6504-15]S5 [6498-45]SA Jensen, Frank [6504-23]S7 Karras, George I. 6491 Kleinfelder, Stuart [6501-08]S2 Kraus, Martin R. H. Review Hutter, Andreas [6504-18]S7 Jeong, Euijune [6501-31]S6 ProgComm Kleinmann, Johanna Krawczyk, Grzegorz Hwang, Dong-Choon Jeong, Se-Yoon [6494-29]S10 Karsten, Martin 6504 [6494-22]S8 [6492-40]S8 [6490A-67]SPS Jeong, Youngho [6507-33]S9 ProgComm Klingen, Bruno [6502-21]S7 Kress, William C. [6494-20]S7 Hyde, Damon E. [6498-01]S1 Jerald, Jason J. Kassim, Ashraf A. B. M. Klinger, Thomas [6501-33]S6 Krishna, Sreekar [6492-66]SA [6490B-47]S13 [6491-02]S1 Knaup, Michael [6498-22]S5 Kriss, Michael A. SympChair, Ji, Xiangyang [6508-55]S11, Katkovnik, Vladimir Knicker, Florian [6508-37]S8 SC810 Inst, 6493 S8 I [6508-106]S13 [6497-07]S2, [6497-20]S5 Knieser, Michael J. SessChr, 6493 ProgComm, Jiang, Gangyi [6508-76]S13 Kawai, Hiroshi [6490A-43]S12 [6497-18]S4 [6493-23]S7, [6493-24]S7, Iancu, A. [6507-08]S3 Jiang, Nianjuan [6508-08]S2 Kawai, Norihiko [6498-35]S10 Kniss, Joe M. Review 6502 ProgComm, Iancu, Daniel S. [6507-08]S3 Jiang, Xiaoyun [6493-14]S4 Kawai, Toshiaki [6501-06]S2 Knoche, Hendrik O. [6502-19]S7 Ichiyanagi, Toshitaka Jiang, Yi [6497-18]S4 Kawakita, Masahiro [6507-24]S7 Kroon, Bart [6506-21]S7 [6490A-58]SPS Jin, Liang [6501-08]S2 [6490A-43]S12, Knoerig, Ruediger [6507-13]S4 Krutz, Andreas [6508-10]S2 Ideses, Ianir A. Jin, Mingwu [6498-03]S1 [6491-33]S8 Ko, Kyung-Woo [6493-34]S9 Kubo, Takashi [6501-23]S3 [6490A-55]SPS, Jin, Sungho [6506-14]S5 Kegelmeyer, Laura M. Kobatake, Makito [6501-37]S6 Kubota, Akira [6508-92]S13 [6496-11]S3 Jin, Xin [6507-29]S8 [6503-27]S7 Kobayashi, Akira [6501-06]S2 Kubota, Misao [6501-06]S2 Ihama, Mikio [6502-06]S2 Jin, Zhipeng [6508-76]S13 Kehtarnavaz, Nasser SC809 Kobayashi, Miho [6490A-28]S7 Kuchi, Prem [6492-66]SA Ihara, Toru [6496-24]S1 Jinturkar, S. [6507-08]S3 Inst, 6496 S1 SessChr, Kobayashi, Tetsuro Kudoh, Yoshiharu [6501-15]S3 Iida, Yoshinori [6492-57]SA Johansson, Jimmy 6496 Chr, [6496-05]S2 [6490A-61]SPS, Kukkala, Petri [6507-35]S9 Iijima, Tomokuni [6501-19]S3 [6495-10]S4 Keif, Malcolm G. [6500-21]S7 [6490A-64]SPS Kumontoy, Roy [6502-25]S8 Iizuka, Masayuki John, George [6502-03]S1 Keim, Daniel C. Review, Koc, Ali [6505-59]S13 Kunter, Matthias [6508-10]S2 [6490A-66]SPS Johnson, Allen [6491-19]S4 [6495-04]S1 Kocak, Tekin [6496-07]S2 Kuo, C.-C. J. 6508 IJsselsteijn, Wijnand A. Johnson, Garrett M. Kelly, Kevin F. [6498-15]S4 Kodama, Kazuya ProgComm, 6508 S8 [6490A-17]S4 [6494-15]S6 Kempken, Sebastian [6508-92]S13 SessChr, [6508-36]S8, Ikeda, Hiroshi [6497-10]S3 Johnson, Kevin [6502-01]S1, [6495-19]S8 Koenderink, Jan J. 6492 [6508-41]S8 Ikeda, Hisashi 6500 [6502-03]S1 Kender, John R. 6506 ProgComm Kuo, Chunghui [6494-17]S6 ProgComm Johnson, Matthew A. ProgComm Kofler, Ingo [6504-18]S7 Kupiec, Stephen A. Ilgner, Justus F. R. [6492-34]S7 Kennard, Douglas J. Koga, Fumihiko [6501-15]S3 [6490A-34]S8 [6490A-07]S2 Joly, Philippe [6506-25]S8 [6500-13]S5 Kogan, Irina A. [6508-17]S3 Kurceren, Ragip [6508-94]S13 Imada, Katsumi [6501-19]S3 Jones, Gareth [6506-12]S4 Kennell, Lauren R. [6497-25]S6 Koh, Chin Chye [6492-18]S4 Kurihara, Toru [6497-13]S3 Imai, Shigeki [6501-37]S6 Jones, Gawain [6497-14]S4 Kenny, Peter [6495-08]S2 Kohli, Luv [6490B-47]S13 Kurita, Takaaki [6494-24]S8 Imamura, Norihiro [6501-19]S3 Jordan, Jay B. [6497-38]SB Ker, Andrew D. [6505-03]S1, Koike, Atsushi [6494-08]S3 Kusmierek, Ewa [6504-08]S2 Joshi, Anand A. [6498-49]S11 [6505-05]S1

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 209 Participants

Kusuda, Masayuki [6501-17]S3 Lee, Seong-Deok [6492-29]S6, Liu, K. J. Ray [6505-56]S12 Madhow, Upamanyu Mayer-Patel, Ketan D. 6504 Kuwahara, Naoki [6508-51]S10 [6502-27]S9 Liu, Liang [6498-37]S11 [6505-49]S11 ProgComm Kwan, Stephanie [6492-12]S3 Lee, Seungsin [6494-03]S1 Liu, Limin [6507-02]S1, Madritsch, Christian Mazurek, Cezary [6504-08]S2 Kwon, Eun-Young [6493-34]S9 Lee, Sihyoung [6507-18]S5 [6508-06]S1 [6501-33]S6 McAllister, David F. Kwon, Oh-Seol [6493-34]S9 Lee, Soo-In [6507-33]S9 Liu, Rongke [6507-28]S8, Maggard, Eric [6494-17]S6 [6490A-71]SPS Lee, Tae-Hyung [6493-09]S3 [6508-26]S5 Magnan, Pierre [6501-20]S3 McCann, John J. 6492 Lee, Teahyung [6508-100]S13 Liu, Yebin [6508-03]S1, Mahanti, Anirban [6504-07]S2 ProgComm, [6492-26]S6, L Lefebvre, Frédéric [6508-78]S13 Maiorana, Emanuele [6492-41]S8, [6493-01]S1 [6508-53]S10 Liu, Yuxin [6508-06]S1, [6497-34]SB, [6505-36]S9 McCarthy, Ann L. [6494-19]S7 Labbe, D. [6490A-07]S2 Lehmann, Mike [6507-10]S3 [6508-94]S13 Mairesse, Fabrice [6503-16]S5 McCormick, Patrick S. Labusch, Kai [6492-09]S2 Lei, Phoury [6508-49]S10 Liversedge, Simon Malik, Amal [6493-05]S2 [6495-16]S6 Lacroute, Phil Review Lei, Shawmin 6508 S8 [6490A-16]S4 Mandal, Mrinal K. McDowall, Ian E. 6490B Chr, Ladret, Patricia [6492-16]S3 SessChr Loce, Robert P. [6493-46]S12 [6497-12]S3, [6508-10]S2 6490B S13 SessChr Lagendijk, Reginald L. 6505 Lemma, Aweke N. [6505-22]S5 Lodha, Suresh K. [6499-20]S5 Mane, Ketan K. Review McElroy, James N. ProgComm, 6508 Leppert, Valerie [6498-34]S9 Loebich, Christian M. Manjunath, Bangalore S. 6505 [6503-27]S7 ProgComm Lesser, Michael P. [6501-05]S1 [6502-21]S7 ProgComm, [6505-49]S11, McElvain, Jon S. [6493-42]S11 Lahanier, Christian Leung, Clement H. C. 6506 Lomheim, Terrence S. 6501 6508 ProgComm McLauchlan, Lifford [6491-05]S2 ProgComm ProgComm Mannheim, Volker [6501-33]S6 [6496-02]S1 Lai, Jackson [6501-21]S3 Leventer, Matei [6504-07]S2 Loomis, Andrew H. Manoharan, N. [6496-08]S2 Mecocci, Alessandro Lai, PoLin [6508-39]S8 Levoy, Marc S. Review [6501-41]S1 Mantiuk, Radoslaw [6492-49]S10 Laidlaw, David Review, Review Lew, Michael S. 6506 López Granado, Otoniel M. [6492-40]S8 Medeiros, Felipe L. L. Lam, Edmund Y. 6503 ProgComm [6508-81]S13 Mantiuk, Rafal [6492-40]S8 [6497-28]S7 ProgComm Lewis, David [6500-28]S10 Lopez Silva, Brenda Maple, Carsten Review Meesters, Lydia [6492-19]S4 Lam, Eric P. [6497-23]S5 Li, Baochun 6504 ProgComm [6490B-51]S14 Marcellin, Michael W. 6508 Mehrubeoglu, Mehrube 6496 Lambooij, Marc [6490A-17]S4 Li, Dongge [6507-12]S4 Lopresti, Daniel P. 6500 S9 SessChr, [6508-45]S9 ProgComm, 6496 S2 Landers, Deborah J. Li, Fei [6508-66]S13 ProgComm Marcu, Gabriel G. SessChr, [6496-02]S1 [6501-41]S1 Li, Heng [6498-02]S1 Lorenz, Michael G. SympComm, SC516 Inst, Melancon, Guy Review Lang, Andreas [6505-39]S9 Li, Hongchao [6504-19]S7 [6491-25]S6 SC075 Inst, 6493 S3 Melnyk, Pavlo [6499-04]S1 Langer, Michael S. Li, Jingqiang 6502 ProgComm Lorigo, Liana M. SessChr, 6493 Chr Memon, Nasir D. 6505 CoChr, [6492-06]S2, [6492-48]S10 Li, Jinzhi [6508-08]S2 [6500-32]S11 Maret, Yannick [6508-23]S4 6505 S6 SessChr, 6505 S1 Langley, Adam [6500-16]S6 Li, Kang 6504 ProgComm, Loriot, Benjamin [6503-11]S4 Marimon, David [6508-07]S2, SessChr, [6505-40]S9 Langton, John T. [6495-15]S6 [6504-03]S1, [6504-19]S7 Love, Shaun T. 6493 [6508-23]S4 Menegaz, Gloria [6492-17]S3, Lao, Weilun [6506-23]S8 Li, Kun [6508-61]S13 ProgComm Marini, Enrico [6505-24]S5 [6492-49]S10 Laplante, Phillip A. 6496 Li, Qiming [6505-40]S9 Lu, Jian [6497-29]S7 Markey, Mia K. [6499-12]S3 Menesatti, Paolo [6501-38]S6 ProgComm, [6496-09]S2 Li, Shengdong [6501-08]S2 Lu, Yan [6508-01]S1, Markov, Vladimr B. Meral, Hasan M. [6505-33]S8 Larndorfer, Stefan [6503-10]S3 Li, Shipeng 6508 ProgComm, [6508-98]S13 [6490A-34]S8 Meriaudeau, Fabrice 6503 S6 Laroche, G. [6508-40]S8 6508 S1 SessChr, Lu, Yue [6498-08]S2 Marosi, Istvan [6500-01]S1 SessChr, 6503 S1 SessChr, Larsen, Peter K. [6491-22]S5, [6508-01]S1, [6508-28]S5, Lugmayr, Artur R. [6507-16]S5 Marshall, Stephen 6497 6503 Chr, [6503-08]S2, [6491-39]S9 [6508-98]S13 Lukin, Vladimir V. 6497 ProgComm [6503-26]S7 Laska, Jason N. [6498-15]S4 Li, Shuo [6497-18]S4 ProgComm, [6497-43]SB, Martin, Ken [6495-16]S6 Merritt, John O. SC060 Inst, Lasser, Tobias [6498-01]S1 Li, Xin 6507 ProgComm [6498-18]S4 Martin, Russel A. 6502 Chr, 6490A S4 SessChr, 6490A Lastra, Anselmo A. Li, Yong [6497-04]S1, Luo, Gang [6502-28]S9 6502 S1 SessChr, Chr [6490B-47]S13 [6508-33]S7 Luo, Huitao [6493-21]S6 [6502-05]S2 Mersereau, Russell M. Latecki, Longin Jan SC754 Li, Yunfeng [6499-10]S3 Luo, Jiebo [6492-63]SA, 6508 Martinetz, Thomas [6505-50]S11 Inst, 6499 S3 SessChr, Li, Zhen [6507-02]S1 Chr, [6508-49]S10, [6492-09]S2 Messner, Richard A. 6499 Chr Li, Zhengguo [6508-71]S13 [6508-85]S13, Martínez Rach, Miguel O. [6499-04]S1 Lau, Ryan [6490A-23]S6 Li, Zhu [6506-06]S2 [6508-108]S7 [6508-81]S13 Meylan, Laurence [6492-38]S8, Lauginiger, Francis P. Liang, Liang [6508-27]S5 Luo, Ronnier M. [6493-07]S3, Martin-Gonthier, Philippe [6498-06]S2 [6498-51]S5 Liao, Chia-Cheng [6493-16]S4, [6494-03]S1 [6501-20]S3 Miao, Jingqi [6495-08]S2 Lawrence, Kurt C. [6503-04]S1 [6493-36]S10, [6501-36]S6, Luo, Xingzhi [6504-03]S1 Martins, Mauricio P. Michaelsen, Jørgen Andreas Lazarescu, Mihai Review [6503-22]S7 Luther, Wolfram [6495-19]S8 [6497-28]S7 [6504-23]S7 Lazarus, Edward M. Liao, Hongen [6490A-45]S12 Lynnerup, Niels [6491-22]S5, Marzani, Franck S. Micheli, Mario [6498-10]S3 [6490A-05]S1 Liebman, Judith A. [6491-39]S9 [6503-02]S1 Mikkilineni, Aravind K. Le, Daniel X. [6500-24]S8, [6503-27]S7 Masaoka, Kenichiro [6505-30]S7, [6505-57]S12, [6500-31]S11 Lienard, Bruno [6496-26]S5 [6492-53]S10 [6505-68]S15 Le Callet, Patrick [6492-61]SA Lienhart, Rainer W. 6506 M Mascarilla, Laurent Milanfar, Peyman 6498 Le Guelvouit, Gaëtan ProgComm, [6506-13]S5 [6508-67]S13 ProgComm, 6498 S6 [6505-43]S10 Liguori, Consolatina Ma, Kwan-Liu Review Mashtare, Dale [6494-17]S6 SessChr, [6498-24]S6, Leahy, Richard M. [6491-07]S2, [6491-40]S9 Ma, Siwei [6508-41]S8 Mason, Steve [6492-56]SA 6508 ProgComm [6498-49]S11 Likforman, Laurence Ma, Yi [6508-15]S3, Masset, Olivier [6508-67]S13 Milani, Simone [6508-107]S13 Lebart, Katia 6503 ProgComm [6500-15]S5 [6508-16]S3 Massoudi, Ayoub Milazzo, Anna-Clare Leblanc, Philippe C. Likova, Lora T. [6492-52]S10 Maari, Koichi [6501-03]S1 [6508-53]S10 [6501-08]S2 [6501-08]S2 Lillholm, Martin [6492-07]S2 Maas, Hans-Gerd 6491 Masurel, Paul [6497-13]S3 Millane, Rick P. [6498-53]S12 Lebowsky, Fritz [6493-10]S3 Lim, Su [6502-05]S2 ProgComm, 6491 S1 Matherson, Kevin J. SC753 Miller, Eric L. 6498 S1 SessChr, LeCallet, Patrick [6505-24]S5 Lim, SukHwan [6508-68]S13 SessChr, [6491-20]S5, Inst, 6501 ProgComm, 6498 Chr, [6498-01]S1 Lee, Byoungho Lin, Chia-Hua [6508-29]S5 [6491-21]S5 6502 ProgComm Milosevic, Nenad [6501-39]S6 [6490A-65]SPS Lin, Ching-Ju [6508-95]S13 Macaire, Ludovic [6492-67]SA Mathes, Tom [6508-82]S13 Min, Sung-Wook Lee, Byung-Joo [6490A-27]S7 Lin, Hoang Yan [6490A-31]S8 Macchiavello, Bruno Mathieu, Hébert [6493-39]S10 [6490A-54]S12 Lee, Chang-Eun [6502-05]S2 Lin, Jim C. [6495-27]S9 [6508-69]S13 Matsubara, Tomoki Minghim, Rosane Review Lee, Changhyung Lin, Qiwei [6493-56]SA MacDonald, Lindsay W. 6494 [6501-06]S2 Minichberger, Jürgen [6493-44]S12 Lin, Xiaofan 6500 Chr ProgComm Matsumoto, Kohsei [6503-09]S3 Lee, Changmin [6501-31]S6 Lin, Yu-Min [6493-16]S4 MacEachren, Alan Review [6503-15]S5 Minkovich, Vladimir P. Lee, Dong Su [6507-03]S1 Lind, Thomas A. [6501-41]S1 Machado, Patrick Matsumoto, Mitsuharu [6501-32]S6 Lee, Jeongkyu [6506-01]S1 Linsen, Lars [6495-05]S2 [6490A-39]S10 [6497-10]S3, [6503-14]S5 Mishra, Ajay K. [6491-02]S1 Lee, Jong Ok [6493-48]S12 Lippens, Stefaan [6497-33]SB Machiraju, Raghu [6491-04]S1 Matsusaka, Keiji [6502-02]S1 Mistry, Nilesh [6491-19]S4 Lee, Kilsung [6501-31]S6 Lipsa, Dan R. [6495-14]S6 Macq, Benoît [6490A-70]SPS, Mauthe, Andreas U. 6504 Mitani, Kohji [6502-24]S8 Lee, KuenJin [6490A-31]S8 Lipton, Lenny [6490A-15]S4 [6491-16]S4, 6505 ProgComm Mitra, Sanjit K. [6492-18]S4 Lee, Kun [6499-05]S2 Liu, Bing [6504-25]S8 ProgComm, 6505 S13 Mavlankar, Aditya A. Miyakawa, Kazunori Lee, Myong-Young Liu, Changsong [6500-08]S3 SessChr, 6505 S14 [6508-89]S13 [6501-06]S2 [6493-09]S3 Liu, Cheng-Lin [6500-03]S2 SessChr, [6505-65]S15, Mayer, Konrad J. [6496-17]S4, Miyake, Yoichi 6494 Chr Lee, Sang-Gi [6502-05]S2 Liu, Huajian [6505-26]S6 [6508-82]S13 [6503-19]S6 Miyatake, Shigehiro Lee, Sang-Yong 6496 Liu, Jim [6492-39]S8 Madeira, Joaquim [6501-17]S3 ProgComm [6492-60]SA, [6495-11]S4

210 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Participants

Miyazaki, Daisuke Nagata, Shojiro 6490A O Papin, Christophe Pimentel, Angela E. [6490A-56]SPS ProgComm, 6490A S12 [6508-85]S13 [6490A-63]SPS Miyazawa, Shinji [6501-03]S1, SessChr Ó Conaire, Ciarán E. Papon, Jeremie [6496-03]S1 Pincenti, John [6502-01]S1, [6501-04]S1 Nagy, George [6500-03]S2, [6506-12]S4 Pappas, Thrasyvoulos N. [6502-03]S1 Mizobuchi, Koichi [6501-24]S3 [6500-05]S3 Obrador, Pere [6500-19]S6 SC812 Inst, 6492 S3 Pinkus, Alan R. [6497-02]S1, Mobasseri, Bijan G. Naik, Sagar S. [6499-09]S3 O’Connor, Noel E. [6506-12]S4 SessChr, 6492 S4 SessChr, [6498-31]S7 [6505-70]S15 Naik Raikar, Yatish J. Oermann, Andrea [6507-17]S5 6492 Chr, [6492-11]S2, Pitkänen, Teemu [6507-34]S9 Mohamed, Magdi A. [6505-70]S15 Ogusu, Kouichi [6501-06]S2 6508 ProgComm Piva, Alessandro [6505-29]S7, [6508-63]S13 Nair, Dinesh 6503 ProgComm Oh, Byung Tae [6508-36]S8 Paquet, Eric [6491-15]S4 [6505-42]S10 Moll, Michael A. [6500-33]S11 Nandhitha, Nathakattuvalasu Oh, Hyun Soo [6493-18]S5 Parent, Richard [6491-04]S1 Pizlo, Zygmunt SC754 Inst, Möller, Torsten Review M. [6496-08]S2 O’Hara, Keith J. [6504-17]S7 Park, Bosoon [6503-04]S1 6498 S7 SessChr, 6498 Momonoi, Yoshiharu Nantung, Tommy [6497-18]S4 Ohkawa, Yuji [6501-06]S2 Park, Du-Sik [6492-27]S6, ProgComm, [6498-30]S7, [6490A-42]S12 Naoi, Satoshi [6500-09]S4 Ohm, Jens-Rainer [6492-29]S6, [6493-12]S4, 6499 ProgComm, Mondaini, Nicola [6505-29]S7 Narabu, Tadakuni [6501-09]S2 [6508-104]S13 [6494-03]S1 [6499-10]S3 Monga, Vishal [6493-40]S11 Nartker, Thomas A. 6500 Ohno, Akira [6501-16]S3 Park, Frederick [6498-17]S4 Pizurica, Aleksandra Mongy, Sylvain [6506-24]S8 ProgComm Ohta, Jun [6501-18]S3, Park, HyungJu [6490A-27]S7 [6508-62]S13 Montag, Ethan D. [6493-08]S3 Nasrabadi, Nasser M. SC491 [6501-22]S3 Park, J. J. H. [6490A-07]S2 Plaisant, Catherine Review Monteiro, Marcus V. T. Inst, 6497 S6 SessChr, Ohtsuka, Nao [6491-42]S9 Park, Jae-Sung Pointer, Michael R. [6497-28]S7 6497 S8 SessChr Oikawa, Michio [6490A-28]S7 [6490A-67]SPS [6493-07]S3 Moon, Kyung-Ae Nathan, Arokia [6501-21]S3 Ok, Hyunwook [6492-27]S6 Park, Jeong-Su [6502-05]S2 Poirier, Herve [6493-57]S9 [6494-29]S10, [6507-14]S4 Nathuji, Ripal [6504-13]S4, Okamoto, Jun [6494-24]S8 Park, Jinah Review, Pole, Joe [6498-02]S1 Moore, Elizabeth [6491-18]S4 [6504-17]S7 Okamura, Seiichiroh [6495-01]S1 Polito, Marzia [6506-27]S8 Morel, Jean-Michel Navarrete Michelini, Pablo [6501-24]S3 Park, Jin-Won [6502-05]S2 Politte, David G. [6498-04]S1 [6498-28]S6 [6498-33]S8 Okano, Fumio [6490A-43]S12 Park, Juun [6490A-27]S7 Pollak, Ilya 6498 S3 SessChr, Morel, Olivier [6503-26]S7 Nave, Demian [6491-19]S4 Okinaka, Tomoo [6491-34]S8 Park, Kee-Hyon [6493-09]S3 6498 Chr, [6498-54]S10 Moreland, Kenneth Review, Nefian, Ara V. [6506-27]S8 Okumura, Haruhiko Park, Kyoung Shin Pomerantz, Stuart [6491-19]S4 [6495-13]S5 Neophytou, Neophytos [6490B-48]S13 [6490A-54]S12 Ponomarenko, Nikolay N. Moreno-Escobar, José A. [6498-20]S5 Okuyama, Fumio Parraman, Carinna E. [6497-21]S5, [6497-43]SB, [6497-46]SB Neri, Alessandro [6497-34]SB, [6490A-06]S2 [6493-30]S8 [6498-18]S4 Mori, Hiroyuki [6501-15]S3 [6505-36]S9 O’Leary, Paul L. 6503 Partridge, Caroline Ponomaryov, Volodymyr I. Moriguchi, Seiji [6501-28]S4 Nesvadba, Jan 6506 S1 ProgComm [6492-22]S4 6496 S3 SessChr, 6496 Moriya, Toshio [6503-15]S5 SessChr, [6506-21]S7 Oliver, José [6508-81]S13 Pastrnak, Milan [6496-23]S3 ProgComm, [6496-16]S4, Moroney, Nathan SC811 Inst, Netanyahu, Nathan S. 6499 Olives, Jean-Luc [6494-02]S1 Patel, Tuhina [6504-20]S7 [6496-20]S5, [6497-46]SB 6494 S1 SessChr, 6494 ProgComm Olshausen, Bruno A. Pavéglio, Jean [6503-02]S1 Pontin, Marco [6491-30]S7 ProgComm, [6494-05]S2 Newell, Jonathan C. [6492-04]S2 Payne, Andrew D. [6491-13]S3 Poon, Ting-Chung Morris, Daniel D. [6497-15]S4 [6498-50]S1 Olsson, Roger [6508-84]S13 Pearlman, William A. 6508 [6490A-54]S12 Morris, Oeter [6494-17]S6 Newman, Todd [6494-14]S6 Olszewska, Joanna I. ProgComm Porikli, Fatih M. 6496 Morvan, Yannick [6490A-24]S6 Newsam, Shawn D. [6508-82]S13 Peinsipp-Byma, Elisabeth ProgComm, 6496 S4 Moser, Bernhard [6503-10]S3 [6497-08]S2 Omer, Osama A. [6508-59]S12 Review, [6495-12]S5 SessChr, [6496-07]S2, Mosny, Milan [6492-28]S6 Newsam, Shawn D. Onaka, Peter M. [6501-41]S1 Pekelsky, Jim R. [6491-11]S3 6508 ProgComm, 6508 S4 Motohashi, Yuichi [6501-23]S3 [6498-34]S9 Ono, Fumie [6501-16]S3 Peker, Kadir A. 6506 SessChr, [6508-22]S4 Motta, Ricardo J. 6502 Nezamabadi, Mahdi Ooi, Wei-Tsang 6504 ProgComm, [6506-04]S2 Portnoy, Vlad [6498-54]S10 ProgComm [6493-08]S3 ProgComm, [6504-25]S8 Pelagotti, Anna [6491-27]S7 Pozdin, Vladimir A. Mottin, Bruno [6491-05]S2 Ng, Yee S. [6494-17]S6 Ooki, Susumu [6501-15]S3 Pelah, Adar 6492 ProgComm [6491-43]S9 Moudgill, M. [6507-08]S3 Ngan, King Ngi [6508-54]S11, Ortega, Antonio 6508 S12 Peltier, Steve T. [6501-08]S2 Prabhakaran, Balakrishnan Moulin, Pierre 6505 [6508-57]S12 SessChr, [6508-39]S8, Peng, Jinye [6503-05]S1 [6499-09]S3 ProgComm, [6505-14]S4, Nguyen, Ha T. [6498-36]S10 [6508-58]S12 Peng, Qingzhong [6496-05]S2 Prasad, Rahul R. [6503-27]S7 [6505-17]S4, [6505-19]S4, Nicolas, Marina [6492-16]S3 Osher, Stanley J. [6498-26]S6 Pereira, Fernando 6508 Prêteux, Françoise J. 6497 [6505-20]S4, 6508 Niel, Kurt S. SC767 Inst, 6503 Oshikubo, Hiromichi ProgComm ProgComm ProgComm S2 SessChr, 6503 S5 [6501-24]S3 Perez Malumbres, Manuel J. Preuss, Thomas [6507-10]S3 Mount, David M. 6499 Chr, SessChr, 6503 Chr, O’Sullivan, Joseph A. 6498 [6508-81]S13 Price, Jeffery R. 6503 6499 S1 SessChr [6503-12]S4 ProgComm, [6498-04]S1 Pérez-Freire, Luis L. ProgComm, [6503-01]S1 Mueller, Chris Review Nillozza, Mario [6501-38]S6 Ota, Takashi [6501-19]S3 [6505-14]S4 Pu, Lingling [6508-45]S9 Mueller, Klaus D. Review, Nishiwaki, Seiji [6501-28]S4 Oyama, Ichiro [6501-19]S3 Pérez-González, Fernando Puech, William [6495-20]S8, [6498-20]S5 Noble, Sam [6499-19]S5 Ozawa, Shinji [6496-24]S1 6505 ProgComm, 6505 [6508-64]S13, Mueller, Nickolaus [6498-08]S2 Nojiri, Yuji [6492-53]S10, Ozawa, Sou [6502-02]S1 S10 SessChr, [6505-10]S3, [6508-65]S13 Mukai, Takaaki [6490A-56]SPS [6502-24]S8 Ozsoy, Sumru [6505-33]S8 [6505-12]S3, [6505-14]S4, Pulli, Kari 6507 ProgComm Mukherjee, Debargha Noorkami, Maneli Oztan, Basak [6493-05]S2 [6505-18]S4, [6505-47]S10 Pun, Thierry [6505-41]S9, [6508-68]S13, [6505-50]S11 Pernice, Eric [6498-34]S9 [6505-67]S15 [6508-69]S13 Norberg, Ole [6492-31]S6, Pesquet-Popescu, Béatrice Putnam, Gloria G. 6501 Mulligan, Jeffrey B. 6492 [6493-26]S7 P 6508 ProgComm, ProgComm, 6502 ProgComm Norkin, Andrey [6508-90]S13 Pagendarm, Hans-Georg 6495 [6508-31]S6, [6508-40]S8 ProgComm Murphy, Brian [6505-34]S8, Normand, Cedric M. ProgComm Péteri, Renaud [6498-29]S7 Putze, Torsten [6491-21]S5 [6505-35]S8 [6493-06]S2, [6502-32]S9 Paik, Jung Min [6499-05]S2 Peterka, Tom [6490A-03]S1 Murray, Alan [6491-04]S1 North, Chris Review Pala, Sunil [6490A-36]S9 Petkov, Nicolai [6497-01]S1, Murshed, Manzur M. 6507 North, Steve Review Palacios, Bruno [6508-07]S2 [6497-19]S5 Q ProgComm Nosaka, Reina [6502-13]S5 Palfner, Torsten [6508-37]S8 Pevny, Tomas [6505-02]S1 Muselet, Damien [6492-67]SA Nose, Yugo [6501-18]S3 Pfeiffer, Silvia 6506 ProgComm Qi, Jinyi [6498-39]S12 Paliy, Dmitriy V. [6497-07]S2, Qi, Xiaojun Review Myszkowski, Karol 6492 Noureldin, Samy [6497-18]S4 [6497-20]S5 Pfister, Hanspeter 6495 ProgComm Novak, David [6491-31]S8 ProgComm Qu, Huamin Review Pan, Shih-Yao [6494-28]S10 Qu, Wei [6508-19]S4 Ntziachristos, Vasilis Panchanathan, Sethuraman Philips, Wilfried R. [6498-01]S1 [6497-33]SB, [6508-62]S13 Quan, Shuxue [6493-14]S4 [6492-43]S9, [6492-66]SA, Qudah, Bashar [6504-11]S4 N Nuding, Ulrich [6492-02]S1 6507 ProgComm, 6508 Piater, Justus [6508-82]S13 Nunoshita, Masahiro Picard, Michel [6491-05]S2, Quenot, Georges [6506-25]S8 Nacer, G. [6507-08]S3 ProgComm Quigley, Aaron Review [6501-18]S3, [6501-22]S3 Panetta, Karen A. [6497-16]S4, [6491-11]S3 Nachlieli, Hila [6497-09]S2 Nyland, Lars S. 6491 Pichler, Andreas [6503-09]S3 Nachtnebel, Herbert [6497-40]SB ProgComm Pang, Alex T. 6495 ProgComm Pieroni, Goffredo G. [6496-17]S4, [6503-19]S6 Nyman, Göte S. [6494-01]S1, [6492-46]S9 R Nagahata, Itsuki [6501-22]S3 Panse, Christian Review, [6494-02]S1 [6495-18]S8 Pieroni, Laura [6492-46]S9 Rabbani, Majid SC468 Inst, Nagashima, Michiyoshi Nystrom, Marcus [6492-44]S9 Pietrosanto, Antonio [6501-19]S3, [6501-28]S4 Papari, Giuseppe [6497-01]S1, SC813 Inst, 6508 [6497-19]S5 [6491-07]S2 ProgComm Pilz, Thomas [6495-19]S8 Radhakrishnan, Regunathan [6506-04]S2

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 211 Participants

Radu, Bilcu C. [6497-07]S2 Rivait, Derek [6492-06]S2 Saito, Takahiro [6497-45]SB, Schreiber, Peter [6501-07]S2 Shirmohammadi, Shervin Radun, Jenni E. [6494-02]S1 Rizvi, Syed A. 6497 S7 6502 ProgComm, Schrider, Christina D. [6504-02]S1 Raghuveer, Aravindan SessChr [6502-06]S2, [6502-13]S5, [6497-11]S3 Shishido, Sanshiro [6504-14]S5 Rizzi, Alessandro [6491-30]S7 [6502-15]S5 Schrotter, Gerhard [6501-22]S3 Raguse, Karsten [6491-21]S5 Rizzi, Alessandro [6492-41]S8, Sakai, Masayoshi [6501-16]S3 [6491-31]S8 Shortis, Mark R. 6491 Chr, Rahardja, Susanto 6493 ProgComm, 6493 S9 Sakazawa, Shigeyuki Schuberth, Sebastian 6491 S7 SessChr, [6508-43]S9, [6508-71]S13 SessChr, 6493 S10 [6494-08]S3 [6498-51]S5 [6491-01]S1 Raj, Baldev [6496-08]S2 SessChr, [6493-01]S1, Sakoda, Gen [6501-03]S1 Schuchter, Arthur [6496-15]S4 Shtern, Alon [6496-11]S3 Rajapandi, Iniavan [6496-19]S5 [6493-30]S8, [6494-10]S4 Salah, Albert A. [6492-42]S9, Schuller, Gerald SC764 Inst Sibiryakov, Alexander Ramachandran, Aishwarya Rizzo, Albert S. [6490A-09]S2 [6499-13]S3 Schultz, Richard R. [6498-07]S2 [6499-03]S1 Ro, Yong-Man [6492-65]SA, Salama, Paul 6508 [6508-52]S10 Siddiqui, Hasib [6498-55]S13 Rambow, Mark [6507-10]S3 [6507-03]S1 ProgComm, [6508-27]S5 Schumann, Heidrun Siewert, Udo [6492-09]S2 Ramchandran, Kannan 6508 Robbins, Kay A. Review Salembier, Philippe 6508 S2 [6505-61]S13, [6507-20]S6, Sigelle, Marc [6498-25]S6, S5 SessChr, 6508 Roberts, Jonathan C. 6495 SessChr, [6508-11]S2 [6507-25]S7 [6500-15]S5 ProgComm, [6508-24]S5, Chr, 6495 S6 SessChr Salis, Ghislain [6503-26]S7 Schuur, Bernhard [6508-37]S8 Sikora, Thomas [6506-08]S3, [6508-107]S13 Robinson, J. Paul Salvi, Joaquim 6503 Schwan, Karsten 6504 [6507-13]S4, [6508-10]S2 Ramponi, Giovanni 6497 [6498-47]SA ProgComm ProgComm, [6504-13]S4, Siler, Ondrej [6498-29]S7 ProgComm, [6497-22]S5 Robinson, Michael G. Samarabandu, Jagath K. 6497 [6504-17]S7, [6504-24]S8 Silva, Frutuoso [6495-11]S4 Rangaswamy, Karthik [6490A-32]S8 ProgComm Schwartz, Edward L. Silva, José Demisio S. [6507-30]S9 Robson, Stuart 6491 Samour, Amjad [6506-08]S3 [6500-17]S6 [6497-27]S7 Rao, A. Ravishankar 6503 ProgComm Sampat, Nitin 6502 Chr, 6502 Schwenke, Derek L. Silva, Samuel [6492-60]SA, ProgComm Roch, Sylvain [6493-02]S1 S3 SessChr [6508-51]S10 Review, [6495-11]S4 Rao, Ashish [6498-37]S11 Rodgers, Peter J. [6495-25]S9 Samset, Eigil [6504-21]S7 Seager, James W. [6491-01]S1 Silvén, Olli J. [6498-05]S2, Rao, Shankar R. [6508-15]S3 Rodricks, Brian G. 6502 Sanahuja, Francesc Sebe, Nicu 6506 Chr 6507 ProgComm, Rasmussen, Rene 6494 S8 ProgComm, 6502 S5 [6504-18]S7 Seetzen, Helge [6492-36]S8 [6507-07]S3 SessChr, 6494 ProgComm, SessChr Sanchez-Reillo, Raul Segura, Josep [6501-20]S3 Silver, Deborah E. Review [6494-18]S7 Roedig, Utz 6504 ProgComm [6491-25]S6 Seidel, Hans-Peter Simoncelli, Eero P. [6492-05]S2 Rauterberg, Matthias 6507 Roemhild, Dieter [6501-33]S6 Sandin, Daniel J. [6490A-03]S1 [6492-40]S8, [6496-21]S5 Simonsen, Erik B. [6491-22]S5 ProgComm Roessler, Peter [6503-19]S6 Sangwan, Raghvinder S. Senthilvelan, M. [6507-08]S3 Singleton, Leo C. [6504-13]S4 Raynor, Jeffrey M. [6501-26]S4 Rogers, Jeffrey L. [6491-18]S4 [6496-09]S2 Seo, Sang-Ho [6501-34]S6, Sitbon, Phillip [6504-26]S8 Razionale, A. V. [6491-27]S7 Rogowitz, Bernice E. 6492 Chr, Sanik, Kevin [6498-30]S7 [6501-35]S6 Sjöström, Mårten Redert, André [6508-14]S3 6492 S1 SessChr, 6492 Sankur, Bülent 6505 Serita, Yasuaki [6501-17]S3 [6508-84]S13 Reed, Matthew [6503-27]S7 S10 SessChr ProgComm, [6505-33]S8, Serranti, Silvia [6503-21]S7 Skipper, Julie A. [6497-02]S1, Reeves, Stanley J. 6498 Rombaut, Joost [6508-62]S13 [6505-54]S12 Sessini, Phillipa [6504-07]S2 [6497-11]S3, [6498-31]S7 ProgComm, [6498-38]S11 Romberg, Justin [6498-17]S4 Santini, Simone 6506 CoChr Setälä, Mikko [6507-35]S9 Sliwa, Tadeusz M. Refai, Hakki H. [6507-22]S7 Ronsin, Joseph [6500-02]S2 Santos, Beatriz S. Seulin, Ralph 6503 [6503-16]S5 Regalia, Phillip A. 6507 Rose, Benjamin M. [6490A-63]SPS ProgComm, [6503-11]S4 Sluss, James J. [6507-22]S7 ProgComm [6490A-71]SPS Sarfraz, Muhammad Review Seuntiens, Pieter J. H. Smeaton, Alan F. 6506 Regazzoni, Carlo S. Rose, Kenneth 6508 Sarhan, Nabil J. [6504-11]S4 [6492-30]S6 ProgComm, 6506 S7 [6508-21]S4 ProgComm Sari-Sarraf, Hamed 6503 Shaked, Doron [6493-37]S10 SessChr, [6506-12]S4 Rehfeld, Nils [6495-12]S5 Rosenbaum, René U. ProgComm, [6503-03]S1 Shang, Junqing [6500-08]S3 Smith, John R. 6506 Rehman, Tauseef U. [6505-61]S13, [6507-20]S6, Sarkar, Anindya [6505-49]S11 Shankar, Ravi [6504-20]S7 ProgComm [6498-32]S8 [6507-25]S7 Särkelä, Heikki [6494-01]S1 Shao, DongXiang [6501-10]S2 Smith, Mark J. T. [6508-02]S1 Reinert-Nash, John R. 6502 Rössler, Peter [6496-17]S4 Sarukkai, Ramesh [6504-16]S6 Sharma, Gaurav 6505 Smith, Michael [6508-46]S9 ProgComm, 6502 S2 Rössler, Roman [6503-12]S4 Sasaki, Takashi [6490B-48]S13 ProgComm, 6508 Smith, Steven L. 6490A SessChr Rothe, Hendrik [6503-07]S2 Sasaki, Tatsuya [6501-22]S3 ProgComm ProgComm, 6490A S7 Reinhart, Craig C. [6502-31]S9 Roudet, Céline [6508-86]S13 Sasaki, Tomoshi [6501-16]S3 Sharp, Gary D. [6490A-32]S8 SessChr Reinheimer, Alice L. 6501 Rourke, Tegan [6490A-30]S8 Sato, Kazuchika [6501-17]S3 Shattuck, David W. Snyder, Franklin D. ProgComm Rouse, David M. [6492-64]SA Sato, Maki [6501-15]S3 [6498-49]S11 [6497-15]S4 Reitmayr, Gerhard [6504-21]S7 Rowlands, John A. Sato, Mitsuru [6501-03]S1 Shaw, Christopher D. 6495 Sochos, Johanna [6502-04]S2 Rejaie, Reza 6504 ProgComm [6501-21]S3 Sato, Tomokazu [6498-35]S10 ProgComm Soler, Pau [6493-28]S8 Remondino, Fabio 6491 Chr, Roy, Sujoy [6508-47]S9, Saucedo, Edgar [6501-32]S6 Shaw, Rufus [6493-31]S8 Son, Chang-Hwan 6491 S4 SessChr, 6491 S5 [6508-93]S13 Sauer, Ken [6498-41]S12 Shawmin, Lei [6508-36]S8 [6493-34]S9 SessChr, [6491-12]S3, Ruckenstein, Gitit [6497-09]S2 Saunders, Ian [6492-13]S3 Shechterman, Mark Song, Daewon [6507-26]S8 [6491-28]S7, [6491-30]S7, Rueger, Stefan [6506-07]S3 Saunier, Gary [6498-50]S1 [6490A-08]S2 Soubret, Antoine [6498-01]S1 [6491-31]S8 Ruiz, Diego [6490A-70]SPS, Sawada, Tadamasa Sheela Rani, B. [6496-08]S2 Souma, Yoshihito [6502-02]S1 Renambot, Luc [6490B-51]S14 [6491-16]S4 [6498-30]S7 Shekhar, Raj [6498-21]S5 Sousa Santos, Beatriz Renshaw, David [6501-26]S4 Rumsey, David [6495-28]S7 Sawant, Amit P. [6495-06]S2 Shen, Chun-Hung [6492-60]SA, [6495-11]S4 Rensing, Christoph 6504 Rusanovskyy, Dmytro Sawchuk, Alexander A. [6508-48]S10 Soutsuka, Kouji ProgComm [6507-09]S3 [6490A-09]S2, Shen, Han-Wei [6495-02]S1 [6490A-56]SPS Repperger, Daniel W. Russo, Michele [6491-06]S2 [6490A-12]S3 Shestak, Sergey A. Spampinato, Giuseppe [6497-02]S1, [6498-31]S7 Ryu, Byong Tae [6493-48]S12 Scala, Paolo [6494-06]S2 [6490A-25]S7, [6502-14]S5 Rest, Armin [6501-02]S1 Ryu, Ho [6494-29]S10 Schafer, Ronald W. [6490A-53]SPS Sparr, Ted M. [6495-14]S6 Restrepo-Martínez, [6498-11]S3 Shetty, Shravya R. Speranza, Filippo Alejandro [6493-52]SA Scheidat, Tobias [6505-38]S9 [6500-30]S11 [6490A-11]S3 Rheingans, Penny L. Review S Schettini, Raimondo Shi, Yun-Qing [6505-27]S6, Sprague, Robert A. Rhodes, Philip J. [6495-14]S6 [6502-07]S3, [6502-12]S4, [6505-58]S12 SympChair Rhody, Harvey [6490A-10]S3 Sabbatini, Maasimo 6506 Chr, 6506 S8 Shigeru, Ando [6497-13]S3 Sproul, Jason [6494-09]S4 Ribarsky, William Review, [6490B-46]S13 SessChr Shiguemori, Elcio H. Sreenan, Cormac J. [6495-21]S8, [6495-29]S3 Saber, Eli [6490A-10]S3 Schill, Kerstin [6492-54]S10 [6497-28]S7 [6504-12]S4 Richardson, Mark A. Sabirin, Muhammad Syah Schmalstieg, Dieter Shih, Loren C. [6494-09]S4, Sreevalsan-Nair, Jaya [6494-23]S8 Houari [6507-15]S4 [6504-21]S7 [6502-10]S4 [6495-05]S2 Ricordel, Vincent [6508-32]S6 Sadasivam, Shankar Schmidt-Sacht, Wulf Shimamoto, Hiroshi Srihari, Sargur N. 6500 Riemens, Bram [6490A-39]S10 [6505-17]S4 [6493-33]S8 [6502-24]S8 ProgComm, [6500-29]S11, Rimey, Raymond D. Safavi-Naini, Rei 6505 Schmiedge, Paul [6495-24]S9 Shin, Dong-Hak [6500-30]S11 [6495-03]S1 ProgComm Schneiderman, Ben Review [6490A-65]SPS, Srinivasan, Harish Rioux, Marc [6491-05]S2, Said, Amir 6508 ProgComm, Schneidewind, Joern Review [6490A-67]SPS [6500-29]S11, [6491-11]S3 6508 S11 SessChr Schonfeld, Dan 6508 Chr, Shin, Hyunho [6490A-27]S7 [6500-30]S11 Rising, Hawley K. 6492 Saini, Gurdial S. [6490A-34]S8 [6508-12]S2, [6508-19]S4, Shin, Jang-Kyoo [6501-34]S6, Stanley, S. [6507-08]S3 ProgComm Saint-Jean, Christophe [6508-63]S13 [6501-35]S6 Starke, Dietmar [6501-33]S6 Rittenschober, Christian [6494-10]S4 Schreck, Tobias Review, Shirai, Toshihito [6501-16]S3 Stasko, John Review [6503-12]S4 Saito, Hideo [6496-24]S1 [6495-04]S1, [6495-18]S8 Steckmann, Sven [6498-22]S5

212 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Participants

Steinbach, Eckehard G. 6508 Suzuki, Motofumi T. Theobalt, Christian Tupin, Florence [6498-25]S6 Venkatachalam, Muthaiah ProgComm [6491-32]S8 [6496-21]S5 Turan, Erman [6506-03]S1 [6504-22]S7 Steinberg, Eran [6502-17]S6 Suzuki, Ryoji [6501-15]S3 Thibault, Jean-Baptiste Turner, Mark I. [6500-04]S2 Venkatasubramanian, Nalini Steinebach, Martin Suzuki, Shirou [6501-06]S2 [6498-41]S12 Tversky, Tal [6492-03]S2 6504 ProgComm [6505-07]S2, [6505-08]S2, Swaminathan, Ashwin Thiemert, Stefan [6505-26]S6 Twidale, Michael B. Review Venkatesh, Svetha 6506 [6505-26]S6, [6505-51]S11 [6505-28]S7, [6505-56]S12 Thieret, Scott [6498-51]S5 Twieg, Donald B. [6498-38]S11 ProgComm Stellbrink, Joe W. [6494-26]S9 Swaminathan, Krishnan Thoma, George R. 6500 Tyler, Christopher W. 6492 Venkatraman, B. [6496-08]S2 Stevens, Richard F. [6499-03]S1 ProgComm, [6500-24]S8, ProgComm, [6492-52]S10 Vera, Esteban [6497-26]S6 [6490A-36]S9 Swan, J. Edward 6495 [6500-31]S11 Verbeek, Fons J. 6506 S6 Stevenson, Robert L. ProgComm Thomas, Benjamin A. SessChr, [6506-16]S6, [6497-04]S1, 6508 [6499-19]S5 U [6506-18]S6 ProgComm, 6508 S7 Thomas, Christopher J. Vese, Luminita A. [6498-28]S6 SessChr, [6508-09]S2, T [6501-27]S4 Uchida, Yoshinori [6501-04]S1 Vetro, Anthony 6508 S10 [6508-33]S7 Thomas, Jean-Baptiste Uehara, Shin-ichi SessChr, 6508 ProgComm, Stiller, Christoph 6503 Taghva, Kazem 6500 [6506-02]S1 [6490A-20]S5 [6508-51]S10 ProgComm ProgComm, [6500-27]S10 Thomas, Naveen M. Uhl, Andreas [6505-23]S5, Vielhauer, Claus SC816 Inst, Stiller, Peter F. 6499 S2 Taguchi, Akira 6497 [6506-09]S3 [6505-66]S15, [6508-35]S7 6505 S9 SessChr, 6505 SessChr, [6499-11]S3, ProgComm Thompson, Paul [6498-49]S11 Um, Gi-Mun [6490A-62]SPS ProgComm, [6505-21]S5, [6508-13]S3 Taira, Kazuki [6490A-42]S12 Tian, Jie [6505-37]S9 Umetani, Keiji [6501-37]S6 [6505-38]S9, [6507-17]S5 Stockhammer, Thomas 6508 Tajbakhsh, Touraj [6502-35]S9 Tian, Jun [6505-52]S11 Unger, Jonas [6501-12]S2 Villan, Renato [6505-67]S15 ProgComm Tak, Sungho [6498-40]S12 Tian, Qi 6506 ProgComm Upadhyay, Akhilesh R. Villasenor, John D. Stolin, Alexander V. Takabe, Yuuji [6499-07]S2 Tian, Yibin [6492-24]S5, [6497-32]S8, [6497-49]SB [6508-46]S9 [6498-02]S1 Takada, Shunji [6502-06]S2 [6502-11]S4 Uragaki, Takeshi [6491-33]S8 Virtanen, Toni [6494-02]S1 Stork, David G. SC814 Inst, Takahara, Junichi Tico, Marius [6498-05]S2, Urban, Bodo [6507-23]S7 Visentin, Gianfranco [6499-02]S1 [6490A-61]SPS [6502-30]S9 Ursu, Danut V. [6502-17]S6 [6490B-46]S13 Strazzullo, Gennaro Takaki, Yasuhiro Timmerer, Christian Utsugi, Kei [6490A-28]S7 Vistuch, Roni [6490A-55]SPS [6491-07]S2 [6490A-29]S7 [6504-18]S7 Uusitalo, Jani J. [6503-20]S6 Vitali, Fabio EI115X Stroinski, Maciej [6504-08]S2 Takala, Jarmo H. 6507 Chr, Tistarelli, Massimo ProgComm Su, Ching-Liang [6499-17]S4, 6507 S7 SessChr, 6507 S8 [6492-42]S9 Vittuari, Luca [6491-28]S7 [6503-17]S5 SessChr, 6507 S4 SessChr, Tobin, Kenneth W. 6503 V Vize, Peter D. [6506-17]S6 [6507-06]S2, [6507-08]S3, Su, Chun-Wei [6494-28]S10 ProgComm Vaahteranoksa, Mikko Vogel, Carl [6505-35]S8 Su, Li [6508-01]S1 [6507-09]S3, [6507-34]S9 Tohme, Michel S. Vogels, Ingrid [6492-30]S6 Takami, Fuminori [6501-28]S4 [6494-02]S1, [6494-30]S1 Su, Qi [6505-37]S9 [6498-39]S12 Vahvilainen, Markku Voigt, Michael [6505-69]S15 Su, Wei [6505-27]S6, Takanashi, Nobuaki Tominaga, Shoji 6493 S2 Voisin, Yvon 6503 ProgComm, [6490A-20]S5 [6497-07]S2 [6505-58]S12 SessChr, 6493 ProgComm, van Buchem, Mark [6503-02]S1, [6503-16]S5 Su, Yeping [6508-39]S8 Takashima, Hajime [6493-38]S10 Volker, Norbert [6499-01]S1, [6501-18]S3 [6506-03]S1 Subramanian, Kalpathi R. 6495 Tominski, Christian Van den Ende, Nele [6499-18]S5 ProgComm Takasu, Yoshifumi [6501-28]S4 [6507-23]S7 Voloshynovskiy, Sviatoslav V. Takatalo, Jari M. [6494-01]S1 [6492-19]S4 Suda, Sachio [6490A-61]SPS Tonry, John L. [6501-41]S1 van der Grond, Jeroen 6505 ProgComm, 6505 S4 Suenaga, Tatsutoshi Takebe, Hiroaki [6500-09]S4 Topkara, Mercan 6505 S8 SessChr, [6505-41]S9, Takeda, Hiroyuki [6498-24]S6 [6506-03]S1 [6501-19]S3 SessChr van der Putten, Peter [6505-67]S15 Sugawa, Shigetoshi Takeda, Takeshi [6501-03]S1 Topkara, Mercan [6505-31]S8 Voltolini, Francesca Takhar, Dharmpal [6498-15]S4 [6506-18]S6 [6501-23]S3, [6501-24]S3 Topkara, Umut [6505-31]S8 van der Veen, Michiel [6491-30]S7 Sugawara, Masayuki Talavage, Thomas 6498 S11 Torres, Raúl E. [6497-42]SB von Spiczak, Jochen SessChr, [6498-37]S11 [6505-22]S5 [6492-53]S10 Tourancheau, Sylvain Van Gool, Luc J. 6491 [6504-21]S7 Sugimoto, Dai [6501-03]S1 Talbar, Sanjay N. [6497-32]S8 [6492-61]SA Vuori, Tero [6494-02]S1 Tam, Wa James [6490A-11]S3 ProgComm, 6506 Sugimoto, Osamu [6494-08]S3 Toutant, Jean-Luc [6499-14]S4 ProgComm Vybornova, Olga [6505-65]S15 Sugita, Yukio [6496-13]S3 Tamaki, Satoshi [6501-19]S3 Tran, Nicholas [6492-51]S10 Tan, Hendry [6507-15]S4 van Meerbergen, Jef Suh, Dongjun [6492-65]SA Travis, Adrian R. L. [6496-23]S3 Sukumar, Vinesh [6492-55]SA, Tang, Jianjun [6497-48]SB [6490A-34]S8 W Tang, Kai Lam [6508-57]S12 van Nes, Floris L. [6493-17]S4 [6494-25]S Trémeau, Alain 6506 van Nieuwenhuyse, Erwin Summers, Kristen M. Tang, Tony K. [6502-36]S6 ProgComm, [6506-02]S1 Wada, Atsushi [6501-18]S3 Tang, Weidong [6498-38]S11 [6495-05]S2 Wagner, Andrew W. [6500-04]S2 Tretter, Daniel R. [6493-21]S6, Vande Moere, Andrew Review Sun, Huifang 6508 ProgComm Tani, Kuniyuki [6501-18]S3 [6506-10]S4, [6508-68]S13 [6508-15]S3 Tani, Yoshiharu [6502-04]S2 Vandewalle, Patrick Walker, Rick [6495-08]S2 Sun, Jun [6500-09]S4 Triantaphillidou, Sophie [6498-06]S2, [6502-09]S3 Sun, Ming-Ting 6508 Taniguchi, Haruyuki [6494-23]S8 Walton, James S. 6491 [6501-15]S3 Vanhoucke, Vincent ProgComm ProgComm Tribulato, Giuseppe [6500-06]S3 Sun, Pei-Li [6493-11]S4, Tanimoto, Masayuki [6506-26]S8 Walvoord, Derek J. [6490A-58]SPS, Vanninen, Matti [6495-06]S2 [6500-07]S3 [6493-15]S4 Trimeche, Mejdi [6497-20]S5 Varela-Benítez, José L. Sun, Qibin [6508-43]S9, [6490A-59]SPS Trocan, Maria S. [6508-31]S6 Walworth, Vivian K. 6490A Tanioka, Kenkichi [6501-06]S2 [6496-20]S5 ProgComm, 6490A S8 [6508-93]S13 Troncoso, Juan Ramón Varshosaz, Masood Sun, Qi-Chao [6508-72]S13 Tannenbaum, Allen R. [6505-10]S3 SessChr [6498-32]S8 [6501-30]S5 Wan, Xiaoxia [6493-51]SA, Sun, Shijun 6508 S8 SessChr, Truchetet, Frédéric Vasic, Bane V. [6508-45]S9 [6508-36]S8 Tao, Dayong [6508-88]S13 [6497-14]S4 [6494-32]S10 Tao, Peining [6494-07]S3 Vasudev, Bhaskaran 6508 Wandell, Brian A. MeetingVIP Sun, Yinlong 6495 ProgComm, Trutschl, Marjan Review ProgComm [6498-47]SA, [6498-48]SA Taruki, Hisayuki [6501-16]S3 Tsai, Chao-Hsu [6490A-31]S8 Wang, Chiao [6490A-12]S3 Taskiran, Cuneyt M. 6506 S4 Vázquez, Carlos A. Wang, Chin-Kiong [6505-16]S4 Sundaram, Hari 6506 Tsai, Yi-Min [6508-50]S10 [6490A-11]S3, ProgComm SessChr, [6506-06]S2, Tsai, Yu-Pao [6490A-60]SPS Wang, Ding [6496-25]S5 [6507-12]S4 [6490A-13]S3 Wang, Jiajun [6508-107]S13 Sung, Dong-Kyu [6501-34]S6 Tsandilas, Theophanis Review Vázquez, Carlos [6490A-14]S3 Suresh, Suneel [6499-06]S2 Tate, Tomoyasu [6501-23]S3 Tseng, Gwo-Feng Wang, Jianning Review Taylor, John M. [6491-05]S2 Vázquez, Carlos A. Wang, Jing [6508-72]S13 Süsstrunk, Sabine E. 6492 [6490A-31]S8 [6508-87]S13 ProgComm, [6492-38]S8, Tekalp, Ahmet M. 6506 Tsoupikova, Daria Wang, Jiong [6498-02]S1 ProgComm Veelaert, Peter 6499 Wang, Qiqi [6498-47]SA, [6498-06]S2, 6502 [6490B-50]S14 ProgComm ProgComm, 6502 S7 Teoh, Soon Tee [6495-09]S4 Tsubaki, Ikuko [6497-45]SB [6498-48]SA Teranishi, Nobukazu 6501 Veeramachaneni, Sriharsha Wang, Shen-Ge [6493-40]S11, SessChr, [6502-09]S3 Tuennermann, Andreas [6500-05]S3 Sutcu, Yagiz [6505-40]S9 ProgComm [6501-11]S2 [6493-50]S12 Tesanovic, Milos [6508-04]S1 Vega, Fernando EI115X Wang, Shengjin [6503-23]S7 Suwa, Misako [6500-09]S4 Tuijn, Chris 6493 ProgComm, ProgComm Suzuki, Fumitsugu Tescher, Andrew G. 6508 6493 S6 SessChr, Wang, Wei [6507-12]S4 ProgComm Vehviläinen, Markku Wang, Wiley [6498-54]S10 [6502-34]S9 [6493-32]S8 [6498-05]S2, [6502-26]S9, Suzuki, Masa-aki [6501-19]S3, Teshima, Tomoaki [6496-24]S1 Tumblin, Jack Review, 6492 S7 Wang, Yao [6508-61]S13 Thang, Truong Cong [6502-30]S9 Wang, Ying [6505-19]S4 [6501-28]S4 SessChr Veldhoven, Pim [6507-24]S7 [6507-03]S1 Tuokko, Reijo O. [6503-20]S6 Wang, Yong [6505-09]S2

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 213 Participants

Wang, Yubing [6504-04]S1 Woods, John W. 6508 Yang, Yongyi 6498 ProgComm, Zhang, Peng [6508-55]S11 Wang, Yuehuan [6496-25]S5 ProgComm, 6508 S6 [6498-03]S1 Zhang, Rong [6508-103]S13 Ward, Chris 6490A SA SessChr Yanikoglu, Berrin A. 6500 Chr, Zhang, Xiaolin [6507-28]S8 SessChr Worring, Marcel 6506 [6500-14]S5 Zhang, Yi-Xin [6493-55]SA Ward, Matthew O. Review ProgComm Yao, Li [6497-48]SB Zhao, Debin [6508-106]S13 Ward, Richard [6491-18]S4 Wright, John [6508-16]S3 Yao, Wei [6508-71]S13 Zhao, Shengdong Review Ware, Colin Review Wu, Angela Y. 6499 Chr Yaron, Avi [6490A-08]S2 Zhao, Zhigang [6508-105]S13 Warren, Penny G. 6501 Wu, Dajun [6508-60]S13, Yaroslavsky, Leonid P. Zheng, Yanyan [6508-106]S13 ProgComm [6508-79]S13 [6490A-55]SPS, Zheng, Yibin 6498 ProgComm, Wartell, Zachary J. Wu, Dapeng [6508-42]S9 [6496-11]S3 [6498-02]S1 [6495-21]S8 Wu, Feng [6508-01]S1, Yavlinsky, Alexei G. Zhou, Minhua [6508-38]S8 Watanabe, Akira [6496-13]S3 [6508-28]S5, [6508-98]S13, [6506-07]S3 Zhou, Steven [6490B-49]S13, Watanabe, Keishiro [6508-106]S13 Ye, Hua [6507-08]S3 [6507-30]S9 [6494-24]S8 Wu, Jianhua [6508-44]S9 Ye, Jong-Chul [6498-40]S12, Zhu, Fengqing [6508-05]S1 Watson, Andrew B. 6492 Wu, Linda M. [6502-33]S9 [6498-46]SA Zhu, Guangxi [6507-29]S8 ProgComm Wu, Min 6505 ProgComm, Yeh, Shih-Ching [6490A-09]S2 Zhu, Tao [6504-25]S8 Watson, Ben [6492-33]S7 6505 S12 SessChr, Yellott, John I. [6492-23]S5 Zhu, Xinglei [6508-47]S9 Wayman, Elizabeth [6505-28]S7, [6505-56]S12 Yellott, John W. [6492-23]S5 Ziegler, Gernot [6496-21]S5 [6500-18]S6 Wu, Mo [6508-75]S13 Yen, Jonathan [6493-21]S6, Zimmermann, Roger 6504 Chr Weaver, Chris [6495-17]S8 Wu, Peng [6506-10]S4 [6506-10]S4, [6508-68]S13 Zink, Michael 6504 ProgComm Weda, Hans [6506-22]S7 Wu, Xiaodong 6499 Yendo, Tomohiro Ziou, Djemel [6499-15]S4 Wedi, Thomas [6508-37]S8 ProgComm [6490A-58]SPS, Zmudzinski, Sascha Wei, Dabi [6497-13]S3 Wu, Yi [6506-27]S8 [6490A-59]SPS [6505-07]S2 Weissman, Michael A. 6490A Wu, Yifeng [6493-25]S7 Yeo, Chuohao [6508-24]S5 Zou, Jie [6500-31]S11 ProgComm, 6490A S2 Wu, Ying 6508 ProgComm, Yezzi, Anthony J. [6498-11]S3, SessChr, [6490A-35]S9 6508 S4 SessChr [6498-13]S3 Wernert, Eric A. [6491-18]S4 Wu, Yongdong [6508-105]S13 Yi, Haoran [6506-27]S8 Westfeld, Andreas Wueller, Dietmar SC753 Inst, Yi, Jonghyon [6493-03]S2, [6505-46]S10 [6494-22]S8, 6502 [6493-04]S2 Wetherill, Leah F. [6491-18]S4 ProgComm, 6502 S8 Yin, Peng [6508-39]S8 Wetzel, Arthur W. [6491-19]S4 SessChr, [6502-21]S7, Yokoya, Naokazu [6491-17]S4, Wharton, Eric J. [6497-16]S4 [6502-22]S7, [6502-29]S9 [6498-35]S10 Whiting, Bruce R. Yoneda, Akira [6491-33]S8 [6498-04]S1 Yoo, Jang Jin [6494-03]S1 Whitton, Mary [6490B-47]S13 X Yoo, Young-Jin [6502-27]S9 Whoang, Inteck [6496-22]S5 Yoon, Bin Yeong [6508-56]S11 Wichmann, Felix A. Xiao, Feng 6502 ProgComm, Yoon, Chunhong [6498-53]S12 [6492-08]S2 [6502-01]S1, [6502-03]S1 Yoshida, Ryuichi [6502-02]S1 Widener, Patrick [6504-24]S8 Xiao, Yingcai Review, 6495 You, Zhixiang [6508-03]S1 Widenhorn, Ralf [6501-02]S1 ProgComm Young, Douglas J. [6501-41]S1 Wiecki, Thomas [6492-08]S2 Xie, Dehong [6493-51]SA, Yu, Hang [6490B-49]S13 Wieser, Volkmar V. W. [6494-32]S10 Yu, Heather H. 6508 [6503-10]S3 Xie, Liangjun [6498-04]S1 ProgComm Wildsoet, Christine F. Xin, Binjian [6503-13]S5 Yu, Jun [6505-11]S3, [6492-24]S5 Xiong, Ruiqin [6508-28]S5 [6505-48]S10 Wilhelms, Jane Review Xu, Beilei [6493-46]S12 Yu, Liu [6498-46]SA Wiliams, Don [6494-19]S7 Xu, Fang [6498-20]S5 Yu, Lu [6508-72]S13, Willemsen, Oscar Xu, Jizheng [6508-28]S5 [6508-73]S13 [6490A-26]S7 Xu, Tao [6508-30]S6 Yu, Mei [6508-76]S13 Willett, Rebecca M. 6498 S4 Xu, Wenli [6508-03]S1, Yu, Xinguo [6508-08]S2 SessChr, [6498-19]S4 [6508-61]S13, Yu, Xuwen [6504-06]S2 Williams, Donald R. SC807 [6508-66]S13 Yu, Zhou [6498-41]S12 Inst, [6494-16]S6, Xu, Youzhi [6508-84]S13 Yuan, JunLi [6508-43]S9, [6494-20]S7 Xuong, Nguyen-Huu [6508-47]S9, [6508-93]S13 Williams, Mark B. [6501-08]S2 Yuasa, Tomoyuki [6502-02]S1 [6498-02]S1 Yurcik, William J. 6495 Wimmer, Peter [6490A-38]S10 ProgComm Windham, William R. Y [6503-04]S1 Yadid-Pecht, Orly 6501 Winkler, Antje [6505-55]S12 ProgComm Z Winkler, Stefan [6490B-49]S13, Yamamoto, Akira [6501-37]S6 [6491-02]S1, [6507-30]S9 Zaffar, Iffat [6496-01]S1 Yamamoto, Keiichi Zakhor, Avideh [6498-16]S4 Winterfors, Emanuel [6496-24]S1 [6498-44]SA Zamfir, Adrian [6502-17]S6 Yamasaki, Masami Zamfir, Marta [6502-17]S6 Wippermann, Frank C. [6490A-28]S7 [6501-07]S2 Zavorin, Ilya [6500-04]S2 Yamashita, Takayuki Zeise, Eric K. 6494 Wittenbrink, Craig M. 6495 [6502-24]S8 ProgComm ProgComm, 6494 S6 Yamashita, Takenori SessChr, [6494-20]S7 Wittenburg, Kent [6506-04]S2 [6501-37]S6 Wittmann, Steffen [6508-37]S8 Zelensky, Alexander A. Yamashita, Tomokazu [6497-43]SB Wolthusen, Stephen D. Review [6501-24]S3 Wong, Nelson [6495-07]S2 Zeng, Wenjun [6505-13]S3 Yamrom, Boris Review Zeng, Zhi [6503-23]S7 Wong, Pak C. 6495 CoChr Yan, Weiqi [6500-23]S8 Wong, Ping Wah 6505 Chr Zetzsche, Christoph Yanaka, Kazuhisa [6492-02]S1, [6492-54]S10 Woodgate, Graham J. [6490A-41]S12 [6490A-19]S5 Zewail, Rami F. [6497-12]S3 Yang, Allen Y. [6508-15]S3 Zhang, Chengyan [6503-24]S7 Woodring, Jonathan Yang, Jing Review [6495-02]S1, [6495-16]S6 Zhang, Ji [6508-18]S3 Yang, Junlan [6508-63]S13 Zhang, Jiacai [6497-48]SB Woods, Andrew J. SC060 Yang, Kai-Chieh [6492-45]S9 Inst, 6490A SA SessChr, Zhang, Ke [6508-73]S13 Yang, Mingqiang [6500-02]S2 Zhang, Lei 6506 ProgComm 6490A S1 SessChr, 6490A Yang, Xin [6505-37]S9 Chr, [6490A-30]S8

214 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] Publication Order Form

■ IS&T/SPIE Member IS&T/SPIE ID #

______For Office Use Only First Name M.I. Last Name Date ______Title Amt. Recd. ______CC Cash Check TC Company Check # ______Address (include Mail Stop) P.O. # ______IDN # ______City State/Province Zip/Postal Code

______ORD # ______Country other than USA

______Phone Fax

______E-Mail Address (IS&T/SPIE does not sell e-mail addresses) Date of Birth (Optional)

■ Check this box if you do not wish to receive information from organizations other than IS&T/SPIE.

Proceedings and Publications Fill in the volume or order number(s) and price(s) of the publications you wish to order below.

QTY. VOL NO. TITLE PRICE (USD)

PUBLICATIONS TOTAL $______USD CA, FL, WA residents add sales tax; Canadian residents must add GST ...... $______USD Shipping/Handling (Books & CD-ROMs) ...... $______USD U.S. 5% of order total [2-3 weeks delivery] Elsewhere 10% of order total [3-5 weeks delivery] Express Shipping: U.S. $15 USD for 1st item; $10 USD each addl item [2-3 days delivery] SUBTOTAL Elsewhere $30 USD for 1st item; $15 USD each addl item [1 week delivery] $______USD

Method of Payment ■ Check enclosed. Payment in U.S. dollars (by draft on a U.S. bank or international money order) is required. Do not send currency. Wire transfers from banks must include a copy of the transfer order.

■ Charge to my: ■ VISA ■ MasterCard ■ Discover ■ American Express ■ Diners Club

Card Number ______

Expiration date ______TOTAL $______USD Signature ______USD ■ Purchase order enclosed (Purchase orders must be preapproved). All orders must be PREPAID in U.S. dollars. Prices subject to change without notice. No returns without written authorization of IS&T/SPIE. ITEMS WILL NOT BE SHIPPED UNLESS PAYMENT IS RECEIVED.

electronicimaging.org • Tel: +1 703 642 9090 • [email protected] 215 Proceedings

Order Proceedings volumes now and receive low prepublication prices

Vol# Title (Editor) Prepublication Price

6490 Stereoscopic Displays and Virtual Reality Systems XIV (A. J. Woods/N. A. Dodgson/ J. O. Merritt/M. T. Bolas/I. E. McDowall) ...... $100 ✔ 6491 Videometrics IX (J. Beraldin/F. Remondino/ M. R. Shortis) ...... $70 Proceedings on CD-ROM 6492 Human Vision and Electronic Imaging XII (B. E. Rogowitz/T. N. Pappas/S. J. Daly) ...... $90 Full-text papers from all 19 Proceedings volumes. PC, Macintosh, and Unix compatible. ✔ 6493 Color Imaging XII: Processing, Hardcopy, and Applications (R. Eschbach/G. G. Marcu) ...... $80 Electronic Imaging 2007 ✔ 6494 Image Quality and System Performance IV (Includes Vols. 6490-6508) (L. C. Cui/Y. Miyake) ...... $60 Order No. CDS248 • Est. pub. March 2007 ✔ 6495 Visualization and Data Analysis 2007 (R. F. Erbacher/ Meeting attendee: $135 J. C. Roberts/M. T. Gröhn/K. Börner) ...... $53 Nonattendee member price: $935 6496 Real-Time Image Processing 2007 (N. Kehtarnavaz/ M. F. Carlsohn) ...... $53 Nonattendee nonmember price: $1230 6497 Image Processing: Algorithms and Systems V (J. T. Astola/K. O. Egiazarian/E. R. Dougherty) ...... $70 6498 Computational Imaging V (C. A. Bouman/ E. L. Miller/I. Pollak) ...... $80 ✔ 6499 Vision Geometry XV (L. Latecki/D. M. Mount/ A. Y. Wu) ...... $53 ✔ 6500 Document Recognition and Retrieval XIV (X. Lin/B. A. Yanikoglu) ...... $60 6501 Sensors, Cameras, and Systems for Scientific/ Industrial Applications VIII (M. Blouke) ...... $60 6502 Digital Photography III (R. A. Martin/J. M. DiCarlo/ N. Sampat) ...... $60 6503 Machine Vision Applications in Industrial Inspection XV (F. Meriaudeau/K. S. Niel) ...... $53 ✔ 6504 Multimedia Computing and Networking 2007 (R. Zimmermann/C. Griwodz) ...... $53 6505 Security, Steganography, and Watermarking of Multimedia Contents IX (E. J. Delp III/P. Wong) . . . $100 ✔ 6506 Multimedia Content Access: Algorithms and Systems (A. Hanjalic/R. Schettini/N. Sebe) ...... $53 6507 Multimedia on Mobile Devices 2007 (R. Creutzburg/ J. Takala/J. Cai) ...... $60 ✔ 6508 Visual Communications and Image Processing 2007 (C. Chen/D. Schonfeld/J. Luo) ...... $125

✔ Indicates volumes that will be available at the meeting. Other Proceedings will be available an average of 6 weeks after the meeting.

216 electronicimaging.org • Tel: +1 703 642 9090 • [email protected] SPIE Membership

Your Resource. Your Society.

Information is increasingly a source of competitive advantage. Through face-to-face interaction, publications and online resources, you gain more from your membership in SPIE.

Join today. spie.org/membership

Tel: +1 360 676 3290 • [email protected] Participate in the preeminent event focused on the

future of imaging.

Get an insider’s look at the technologies and industries

that are having a huge impact on humankind. Meet the

leaders who can help you with your research and your

career. Find the right teaming opportunities to take your

projects to the next level.

Don’t miss being part of the excitement at Electronic

Imaging 2008.

Co-Sponsored by SPIE and IS&T

27-31 January 2008 San Jose Convention Center • San Jose, California USA

electronicimaging.org

Conferences • Courses • Exhibition IS&T/SPIE 19th Annual Symposium

Electronic Imaging Science and T

echnology

28 January–1 February 2007 • February 28 January–1

San Jose Marriott and San Jose Convention Center

, San Jose, California USA