Learn More
—This paper addresses the issue of learning and representing object grasp affordances, i.e. object-gripper relative configurations that lead to successful grasps. The purpose of grasp affordances is to organize and store the whole knowledge that an agent has about the grasping of an object, in order to facilitate reasoning on grasping solutions and their(More)
Skeletal trees are commonly used in order to express geometric properties of the shape. Accordingly, tree edit distance is used to compute a dissimilarity between two given shapes. We present a new tree edit based shape matching method which uses a recent coarse skeleton representation. The coarse skeleton representation allows us to represent both shapes(More)
We discuss the need of an elaborated in-between stage bridging early vision and cognitive vision which we call 'Early Cognitive Vi-sion' (ECV). This stage provides semantically rich, disambiguated and largely task independent scene representations which can be used in many contexts. In addition, the ECV stage is important for generalization processes across(More)
—We describe a bootstrapping cognitive robot system that—mainly based on pure exploration—acquires rich object representations and associated object-specific grasp affordances. Such bootstrapping becomes possible by combining innate com-petences and behaviors by which the system gradually enriches its internal representations, and thereby develops an(More)
We develop means of learning and representing object grasp af-fordances probabilistically. By grasp affordance, we refer to an entity that is able to assess whether a given relative object-gripper configuration will yield a stable grasp. These affordances are represented with grasp densities, continuous probability density functions defined on the space of(More)
— We describe an embodied cognitive system based on a three-level architecture that includes a sensorimotor layer, a mid-level layer that stores and reasons about object-action episodes, and a high-level symbolic planner that creates abstract action plans to be realised and possibly further specified by the lower levels. The system works in two modes,(More)
We describe a system for autonomous learning of visual object representations and their grasp affordances on a robot-vision system. It segments objects by grasping and moving 3D scene features, and creates probabilistic visual representations for object detection, recognition and pose estimation, which are then augmented by continuous characterizations of(More)
Keywords: Cognitive vision Contour representation 3D contours Contour relations Perceptual relations 3D reasoning Driver assistance Grasping a b s t r a c t In this work, we make use of 3D contours and relations between them (namely, coplanarity, cocolority, distance and angle) for four different applications in the area of computer vision and vision-based(More)