We study tactile‑driven re‑grasping: given an initial imperfect grasp, a robot should repeatedly adjust its grasp until it reaches a desired contact configuration. We collect teleoperated demonstrations with a front camera and two tactile sensors, encode observations into compact latents, and train an offline RL policy (IQL). A simple tactile success signal (“circle” marker visibility on the tactile pad) lets the policy keep trying until it achieves the intended grasp.
We also found that collecting high quality data was important for Offline RL. We want the policy to learn a behavior of re-grasping. With only successful grasp without re-grasping data can't train the policy to retry. For that when collecting data, we only relyed on the camera and sensor input not watching the scene directly. So when we couldn't get the successful grasp every first try and that led to regrasping motion during data collection. This helped the policy to learn how to try until it properly grasp the right position.