Organizational Unit:
Institute for Robotics and Intelligent Machines (IRIM)

Research Organization Registry ID
Description
Previous Names
Parent Organization
Parent Organization
Includes Organization(s)
Organizational Unit
ArchiveSpace Name Record

Publication Search Results

Now showing 1 - 10 of 13
  • Item
    Grasp selection strategies for robot manipulation using a superquadric-based object representation
    (Georgia Institute of Technology, 2016-07-29) Huaman, Ana Consuelo
    This thesis presents work on the implementation of a robotic system targeted to perform a set of basic manipulation tasks instructed by a human user. The core motivation on the development of this system was in enabling our robot to achieve these tasks reliably, in a time-efficient manner and under mildly realistic constraints. Robot manipulation as a field has grown exponentially during the last years, presenting us with a vast array of robots exhibiting skills as sophisticated as preparing dinner, making an expresso or operating a drill. These complex tasks are in general achieved by using equally complex frameworks, assuming extensive pre-existing knowledge, such as perfect environment knowledge, sizable amounts of training data or availability of crowdsourcing resources. In this work we postulate that elementary tasks, such as pick-up, pick-and-place and pouring, can be realized with online algorithms and a limited knowledge of the objects to be manipulated. The presented work shows a fully implemented pipeline where each module is designed to meet the core requirements specified above. We present a number of experiments involving a database of 10 household objects used in 3 selected elementary manipulation tasks. Our contributions are distributed in each module of our pipeline: (1) We demonstrate that superquadrics are useful primitive shapes suitable to represent on-the-fly a considerable number of convex household objects; their parametric nature (3 axis and 2 shape parameters) is shown to be helpful to represent simple semantic labels for objects (i.e. for a pouring task) useful for grasp and motion planning. (2) We introduce a hand-and-arm metric that considers both grasp robustness and arm end-comfort to select grasps for simple pick-up tasks. We show with real and simulation results that considering both hand and arm aspects of the manipulation task helps to select grasps that are easier to execute in real environments without sacrificing grasp stability on the process. (3) We present grasp selection and planning strategies that exploit task constraints to select the more appropriate grasp to carry out a manipulation task in an online and efficient manner (in terms of planning and execution time).
  • Item
    Planning in constraint space for multi-body manipulation tasks
    (Georgia Institute of Technology, 2016-04-05) Erdogan, Can
    Robots are inherently limited by physical constraints on their link lengths, motor torques, battery power and structural rigidity. To thrive in circumstances that push these limits, such as in search and rescue scenarios, intelligent agents can use the available objects in their environment as tools. Reasoning about arbitrary objects and how they can be placed together to create useful structures such as ramps, bridges or simple machines is critical to push beyond one's physical limitations. Unfortunately, the solution space is combinatorial in the number of available objects and the configuration space of the chosen objects and the robot that uses the structure is high dimensional. To address these challenges, we propose using constraint satisfaction as a means to test the feasibility of candidate structures and adopt search algorithms in the classical planning literature to find sufficient designs. The key idea is that the interactions between the components of a structure can be encoded as equality and inequality constraints on the configuration spaces of the respective objects. Furthermore, constraints that are induced by a broadly defined action, such as placing an object on another, can be grouped together using logical representations such as Planning Domain Definition Language (PDDL). Then, a classical planning search algorithm can reason about which set of constraints to impose on the available objects, iteratively creating a structure that satisfies the task goals and the robot constraints. To demonstrate the effectiveness of this framework, we present both simulation and real robot results with static structures such as ramps, bridges and stairs, and quasi-static structures such as lever-fulcrum simple machines.
  • Item
    Navigation behavior design and representations for a people aware mobile robot system
    (Georgia Institute of Technology, 2016-01-15) Cosgun, Akansel
    There are millions of robots in operation around the world today, and almost all of them operate on factory floors in isolation from people. However, it is now becoming clear that robots can provide much more value assisting people in daily tasks in human environments. Perhaps the most fundamental capability for a mobile robot is navigating from one location to another. Advances in mapping and motion planning research in the past decades made indoor navigation a commodity for mobile robots. Yet, questions remain on how the robots should move around humans. This thesis advocates the use of semantic maps and spatial rules of engagement to enable non-expert users to effortlessly interact with and control a mobile robot. A core concept explored in this thesis is the Tour Scenario, where the task is to familiarize a mobile robot to a new environment after it is first shipped and unpacked in a home or office setting. During the tour, the robot follows the user and creates a semantic representation of the environment. The user labels objects, landmarks and locations by performing pointing gestures and using the robot's user interface. The spatial semantic information is meaningful to humans, as it allows providing commands to the robot such as ``bring me a cup from the kitchen table". While the robot is navigating towards the goal, it should not treat nearby humans as obstacles and should move in a socially acceptable manner. Three main navigation behaviors are studied in this work. The first behavior is the point-to-point navigation. The navigation planner presented in this thesis borrows ideas from human-human spatial interactions, and takes into account personal spaces as well as reactions of people who are in close proximity to the trajectory of the robot. The second navigation behavior is person following. After the description of a basic following behavior, a user study on person following for telepresence robots is presented. Additionally, situation awareness for person following is demonstrated, where the robot facilitates tasks by predicting the intent of the user and utilizing the semantic map. The third behavior is person guidance. A tour-guide robot is presented with a particular application for visually impaired users.
  • Item
    Semantic mapping for service robots: building and using maps for mobile manipulators in semi-structured environments
    (Georgia Institute of Technology, 2015-04-08) Trevor, Alexander J. B.
    Although much progress has been made in the field of robotic mapping, many challenges remain including: efficient semantic segmentation using RGB-D sensors, map representations that include complex features (structures and objects), and interfaces for interactive annotation of maps. This thesis addresses how prior knowledge of semi-structured human environments can be leveraged to improve segmentation, mapping, and semantic annotation of maps. We present an organized connected component approach for segmenting RGB-D data into planes and clusters. These segments serve as input to our mapping approach that utilizes them as planar landmarks and object landmarks for Simultaneous Localization and Mapping (SLAM), providing necessary information for service robot tasks and improving data association and loop closure. These features are meaningful to humans, enabling annotation of mapped features to establish common ground and simplifying tasking. A modular, open-source software framework, the OmniMapper, is also presented that allows a number of different sensors and features to be combined to generate a combined map representation, and enabling easy addition of new feature types.
  • Item
    Time-optimal sampling-based motion planning for manipulators with acceleration limits
    (Georgia Institute of Technology, 2015-04-07) Kunz, Tobias
    Robot actuators have physical limitations in how fast they can change their velocity. The more accurately planning algorithms consider these limitations, the better the robot is able to perform. Sampling-based algorithms have been successful in geometric domains, which ignore actuator limitations. They are simple, parameter-free, probabilistically complete and fast. Even though some algorithms like RRTs were specifically designed for kinodynamic problems, which take actuator limitations into account, they are less efficient in these domains or are, as we show, not probabilistically complete. A common approach to this problem is to decompose it, first planning a geometric path and then time-parameterizing it such that actuator constraints are satisfied. We improve the reliability of the latter step. However, the decomposition approach can neither deal with non-zero start or goal velocities nor provides an optimal solution. We demonstrate that sampling-based algorithms can be extended to consider actuator limitations in the form of acceleration limits while retaining the same advantageous properties as in geometric domains. We present an asymptotically optimal planner by combining a steering method with the RRT* algorithm. In addition, we present hierarchical rejection sampling to improve the efficiency of informed kinodynamic planning in high-dimensional spaces.
  • Item
    Autonomous environment manipulation to facilitate task completion
    (Georgia Institute of Technology, 2015-03-31) Levihn, Martin
    A robot should be able to autonomously modify and utilize its environment to assist its task completion. While mobile manipulators and humanoid robots have both locomotion and manipulation capabilities, planning systems typically just consider one or the other. In traditional motion planning the planner attempts to find a collision free path from the robot's current configuration to some goal configuration. In general, this process entirely ignores the fact that the robot has manipulation capabilities. This is in contrast to how humans naturally act - utilizing their manipulation capabilities to modify the environment to assist locomotion. If necessary, humans do not hesitate to move objects, such as chairs, out of their way or even place an object, such as a board, on the ground to reach an otherwise unreachable goal. We argue that robots should demonstrate similar behavior. Robots should use their manipulation capabilities to move or even use environment objects. This thesis aims at bringing robots closer to such capabilities. There are two primary challenges in developing practical systems that allow a real robotic system to tightly couple its manipulation and locomotion capabilities: the inevitable inaccuracies in perception as well as actuation that occur on physical systems, and the exponential size of the search space. To address these challenges, this thesis first extends the previously introduced domain of Navigation Among Movable Obstacles (NAMO), which allows a robot to move obstacles out of its way. We extend the NAMO domain to handle the underlying issue of uncertainty. In fact, this thesis introduces the first NAMO framework that allows a real robotic systems to consider sensing and action uncertainties while reasoning about moving objects out of the way. However, the NAMO domain itself has the shortcoming that it only considers a robot's manipulation capabilities in the context of clearing a path. This thesis therefore also generalizes the NAMO domain itself to the Navigation Using Manipulable Obstacles (NUMO) domain. The NUMO domain enables a robot to more generally consider the coupling between manipulation and locomotion capabilities and supports reasoning about using objects in the environment. This thesis shows the relationship between the NAMO and NUMO domain, both in terms of complexity as well as solution approaches, and presents multiple realizations of the NUMO domain. The first NUMO realization enables a robot to use its manipulation capabilities to assist its locomotion by changing the geometry of the environment for scenarios in which obstructions can be overcome through the usage of a single object. The system led a real humanoid robot to autonomously build itself a bridge to cross a gap and a stair step to get on a platform. A second NUMO realization then introduces reasoning about force constraints using knowledge about the mechanical advantages of a lever and battering ram. The discussed system allows a robot to consider increasing its effective force though the use of objects, such as utilizing a rod as a lever. Finally this thesis extends the NUMO framework for geometric constraints to scenarios in which the robot is faced with a substantial lack of initial state information and only has access to onboard sensing. In summary, this thesis enables robots to autonomously modify their environment to achieve task completion in the presence of lack of support for mobility, the need to increase force capabilities and partial knowledge.
  • Item
    A linguistic method for robot verification programming and control
    (Georgia Institute of Technology, 2014-10-29) Dantam, Neil Thomas
    There are many competing techniques for specifying robot policies, each having advantages in different circumstances. To unify these techniques in a single framework, we use formal language as an intermediate representation for robot behavior. This links previously disparate techniques such as temporal logics and learning from demonstration, and it links data driven approaches such as semantic mapping with formal discrete event and hybrid systems models. These formal models enable system verification -- a crucial point for physical robots. We introduce a set of rewrite rules for hybrid systems and apply it automatically build a hybrid model for mobile manipulation from a semantic map. In the manipulation domain, we develop a new workspace interpolation methods which provides direct, non-stop motion through multiple waypoints, and we introduce a filtering technique for online camera registration to avoid static calibration and handle changing camera positions. To handle concurrent communication with embedded robot hardware, we develop a new real-time interprocess communication system which offers lower latency than Linux sockets. Finally, we consider how time constraints affect the execution of systems modeled hierarchically using context-free grammars. Based on these constraints, we modify the LL(1) parser generation algorithm to operate in real-time with bounded memory use.
  • Item
    Visual object perception in unstructured environments
    (Georgia Institute of Technology, 2014-08-25) Choi, Changhyun
    As robotic systems move from well-controlled settings to increasingly unstructured environments, they are required to operate in highly dynamic and cluttered scenarios. Finding an object, estimating its pose, and tracking its pose over time within such scenarios are challenging problems. Although various approaches have been developed to tackle these problems, the scope of objects addressed and the robustness of solutions remain limited. In this thesis, we target a robust object perception using visual sensory information, which spans from the traditional monocular camera to the more recently emerged RGB-D sensor, in unstructured environments. Toward this goal, we address four critical challenges to robust 6-DOF object pose estimation and tracking that current state-of-the-art approaches have, as yet, failed to solve. The first challenge is how to increase the scope of objects by allowing visual perception to handle both textured and textureless objects. A large number of 3D object models are widely available in online object model databases, and these object models provide significant prior information including geometric shapes and photometric appearances. We note that using both geometric and photometric attributes available from these models enables us to handle both textured and textureless objects. This thesis presents our efforts to broaden the spectrum of objects to be handled by combining geometric and photometric features. The second challenge is how to dependably estimate and track the pose of an object despite the clutter in backgrounds. Difficulties in object perception rise with the degree of clutter. Background clutter is likely to lead to false measurements, and false measurements tend to result in inaccurate pose estimates. To tackle significant clutter in backgrounds, we present two multiple pose hypotheses frameworks: a particle filtering framework for tracking and a voting framework for pose estimation. Handling of object discontinuities during tracking, such as severe occlusions, disappearances, and blurring, presents another important challenge. In an ideal scenario, a tracked object is visible throughout the entirety of tracking. However, when an object happens to be occluded by other objects or disappears due to the motions of the object or the camera, difficulties ensue. Because the continuous tracking of an object is critical to robotic manipulation, we propose to devise a method to measure tracking quality and to re-initialize tracking as necessary. The final challenge we address is performing these tasks within real-time constraints. Our particle filtering and voting frameworks, while time-consuming, are composed of repetitive, simple and independent computations. Inspired by that observation, we propose to run massively parallelized frameworks on a GPU for those robotic perception tasks which must operate within strict time constraints.
  • Item
    Knowledge transfer in robot manipulation tasks
    (Georgia Institute of Technology, 2014-04-08) Huckaby, Jacob O.
    Technology today has progressed to the point that the true potential of robotics is beginning to be realized. However, programming robots to be robust across varied environments and objectives, in a way that is accessible and intuitive to most users, is still a difficult task. There remain a number of unmet needs. For example, many existing solutions today are proprietary, which makes widespread adoption of a single solution difficult to achieve. Also, most approaches are highly targeted to a specific implementation. But it is not clear that these approaches will generalize to a wider range of problems and applications. To address these issues, we define the Interaction Space, or the space created by the interaction between robots and humans. This space is used to classify relevant existing work, and to conceptualize these unmet needs. GTax, a knowledge transfer framework, is presented as a solution that is able to span the Interaction Space. The framework is based on SysML, a standard used in many different systems, which provides a formalized representation and verification. Through this work, we demonstrate that by generalizing across the Interaction Space, we can simplify robot programming and enable knowledge transfer between processes, systems and application domains.
  • Item
    Trust and reputation for formation and evolution of multi-robot teams
    (Georgia Institute of Technology, 2013-11-15) Pippin, Charles Everett
    Agents in most types of societies use information about potential partners to determine whether to form mutually beneficial partnerships. We can say that when this information is used to decide to form a partnership that one agent trusts another, and when agents work together for mutual benefit in a partnership, we refer to this as a form of cooperation. Current multi-robot teams typically have the team's goals either explicitly or implicitly encoded into each robot's utility function and are expected to cooperate and perform as designed. However, there are many situations in which robots may not be interested in full cooperation, or may not be capable of performing as expected. In addition, the control strategy for robots may be fixed with no mechanism for modifying the team structure if teammate performance deteriorates. This dissertation investigates the application of trust to multi-robot teams. This research also addresses the problem of how cooperation can be enabled through the use of incentive mechanisms. We posit a framework wherein robot teams may be formed dynamically, using models of trust. These models are used to improve performance on the team, through evolution of the team dynamics. In this context, robots learn online which of their peers are capable and trustworthy to dynamically adjust their teaming strategy. We apply this framework to multi-robot task allocation and patrolling domains and show that performance is improved when this approach is used on teams that may have poorly performing or untrustworthy members. The contributions of this dissertation include algorithms for applying performance characteristics of individual robots to task allocation, methods for monitoring performance of robot team members, and a framework for modeling trust of robot team members. This work also includes experimental results gathered using simulations and on a team of indoor mobile robots to show that the use of a trust model can improve performance on multi-robot teams in the patrolling task.