Towards Seamless Interfaces for Human-Robot Interaction: a Vision-Based Approach

Towards Seamless Interfaces for Human-Robot Interaction: a Vision-Based Approach

<p> Computer Science Seminar</p><p> e Towards Seamless Interfaces c for Human-Robot Interaction: A n</p><p> e Vision-Based Approach i by c</p><p>S Dr. Junaed Sattar</p><p> r Date: January 26, 10 a.m.</p><p> e Place: 111 Armes t Abstract u This talk will present a vision-based interface for human-robot interaction and control for autonomous robots in arbitrary environments. The array of sensors for mobile p robots have grown significantly in recent years, and these provide a vast range of diverse modalities for environmental perception in a variety of autonomous agents. </p><p> m Vision has the advantage of being a cost-effective, energy-efficient, unobtrusive sensing modality, and has seen widespread integration in robotic systems. The advent</p><p> o of robust algorithms and a significant increase in computational power are the two most significant reasons for such widespread integration. This research looks at </p><p>C machine vision as an intuitive and uncomplicated method for a human operator to </p><p> communicate in close-range with a mobile robot, and encapsulates them in an </p><p> f interaction framework. The collection of communication paradigms that has been investigated includes, but are not limited to, visual tracking and servoing, programming o of robot behaviors with visual cues, visual feature recognition, mapping and </p><p> identification of individuals through gait characteristics using spatio-temporal visual </p><p> t patterns and quantitative modeling of a human-robot dialog system. The proposed framework enables a human operator to control and program a robot without the need n for any complicated input mechanisms, and also enables the robot to learn about its environment and the operator using a vision-based interface. This work has also relied e on machine learning methods -- supervised learning in particular -- to train the robot vision system to improve robustness. A key aspect of this work is a system for human-</p><p> m robot dialog for safe and efficient task execution under uncertainty. This talk will present extensive validation results performed through a set of human-interface trials, t and through demonstrations in the field on the Aqua amphibious robot platform in the r under water and terrestrial domains. While this framework is not specific to robots operating in the under water domain, vision under water is affected by a number of a issues, such as lighting variations and color degradation, among others. Evaluating the approach in such difficult operating conditions provides a definitive validation of this p approach. e This speaker is a candidate for a tenure-track position in the Department of Computer </p><p>D Science.</p>

View Full Text

Details

  • File Type
    pdf
  • Upload Time
    -
  • Content Languages
    English
  • Upload User
    Anonymous/Not logged-in
  • File Pages
    1 Page
  • File Size
    -

Download

Channel Download Status
Express Download Enable

Copyright

We respect the copyrights and intellectual property rights of all users. All uploaded documents are either original works of the uploader or authorized works of the rightful owners.

  • Not to be reproduced or distributed without explicit permission.
  • Not used for commercial purposes outside of approved use cases.
  • Not used to infringe on the rights of the original creators.
  • If you believe any content infringes your copyright, please contact us immediately.

Support

For help with questions, suggestions, or problems, please contact us