Making Sense of the Physical World with High-resolution Tactile Sensing
This event has passed.
Monday, February 12, 2018 - 12:00pm to 1:00pm
With the rapid progress in robotics, people expect robots to be able to accomplish a wide variety of tasks in the real world, such as working in factories, performing household chores, and caring for elderly. However, it is still very difficult for robots to act in the physical world. A major challenge lies in the lack of adequate tactile sensing. Progress requires advances in the sensing hardware, but also requires advances in the software that can exploit the tactile signals generated when the robot touches an object. The sensor we use is a vision-based tactile sensor called GelSight, which measures the geometry and traction field of the contact surface. For interpreting the high-resolution tactile signal, we utilize both traditional statistical models and deep neural networks. I will describe research on two kinds of tasks: exploration and manipulation. For exploration, I use active touch to estimate the physical properties of the objects. The work has included learning the basic properties (e.g., hardness), of artificial objects, as well as estimating the general properties of natural objects via autonomous tactile exploration. For manipulation, I study the robot's ability to detect slip or incipient slip with tactile sensing during grasping. My research helps robots to better understand and flexibly interact with the physical world.