Browsing by Subject "Deep Learning in Grasping and Manipulation"
Now showing 1 - 2 of 2
- Results Per Page
- Sort Options
Item Generalized Environment-Enabled Object Grasping using a Fixture-Aware Double Deep Q-Network(2022-06) Sasagawa, EddieThis thesis expands on the problem of grasping an object that can only be grasped bya single parallel gripper when a fixture (e.g., wall, heavy object) is harnessed. Preceding work that tackle this problem are limited in that the employed networks implicitly learn specific targets and fixtures to leverage. However, the notion of a usable fixture can vary in different environments, at times without any outwardly noticeable differences. In this work, we propose a method to relax this limitation and further handle environments where the fixture location is unknown. The problem is formulated as visual affordance learning in a partially observable setting. We present a self-supervised reinforcement learning algorithm, Fixture-Aware Double Deep Q-Network (FA-DDQN), that processes the scene observation to 1) identify the target object based on a reference image, 2) distinguish possible fixtures based on interaction with the environment, and finally 3) fuse the information to generate a visual affordance map to guide the robot to successful Slide-to-Wall grasps. We demonstrate our proposed solution in simulation and in real robot experiments to show that in addition to achieving higher success than baselines, it also performs zero-shot generalization to novel scenes with unseen object configurations.Item Target-Driven Robotic Manipulation with Visual Attribute Reasoning(2022-07) Yang, YangAs robots move from factories to our daily lives, robotic manipulation for ordinaryusers is attracting more attention from the robotics community. Target-driven manipulation is a necessary function for robots to enter people’s everyday spaces because it enables robots to perform tasks (such as grasping a specific object) driven by user inputs. Three key challenges, however, preclude the development of target-driven robotic manipulation: (1) ambiguity caused by a mismatch between human referring and robot understanding; (2) clutter formed by a target object and its surrounding objects; (3) domain shift referring to the change of data distribution from training and deployment environments. In this thesis, we address these challenges by equipping target-driven robotic manipulation with visual attribute reasoning. People recognize and grasp a target object in daily scenes by remembering the critical properties of the target. Visual attribute reasoning, or the ability to perceive andreason about essential attributes of a target item, enables humans to understand the target object and its surroundings, as well as plan their actions accordingly. In this thesis, we present the categorization of visual attributes of objects and their crucial functions in robotic manipulation. We develop robotic manipulation systems that integrate object attributes in the form of appearances, spatial locations, and local relations. The robotic systems can accomplish target-driven tasks in challenging and unconstrained environments because of the integration of these object attributes. As a result, our research advances target-driven robotic manipulation in terms of clutter handling, model generalization, and human-robot interaction in particular. Our long-term goal is to develop intelligent robots that connect human users withtheir surrounding environments. Future robots are predicted to interact with humans and accomplish complex target-driven tasks that benefit users. This thesis makes a step towards the goal by leveraging visual attribute reasoning. We present robotic grasping systems that can locate an invisible target occluded in clutter, grasp a never-seen object based on appearance attributes, and disambiguate unclear commands guided by object attributes.