TechTalks from event: Technical session talks from ICRA 2012

Conference registration code to access these videos can be accessed by visiting this link: PaperPlaza. Step-by-step to access these videos are here: step-by-step process .
Why some of the videos are missing? If you had provided your consent form for your video to be published and still it is missing, please contact

Industrial Robotics

  • Tool Position Estimation of a Flexible Industrial Robot Using Recursive Bayesian Methods Authors: Axelsson, Patrik; Karlsson, Rickard; Norrlöf, Mikael
    A sensor fusion method for state estimation of a flexible industrial robot is presented. By measuring the acceleration at the end-effector, the accuracy of the arm angular position is improved significantly when these measurements are fused with motor angle observation. The problem is formulated in a Bayesian estimation framework and two solutions are proposed; one using the extended Kalman filter (EKF) and one using the particle filter (PF). The technique is verified on experiments on the ABB IRB4600 robot, where the accelerometer method is showing a significant better dynamic performance, even when model errors are present.
  • A Sensor-Based Approach for Error Compensation of Industrial Robotic Workcells Authors: Tao, Pey Yuen; Yang, Guilin; Tomizuka, Masayoshi
    Industrial robotic manipulators have excellent repeatability while accuracy is significantly poorer. Numerous error sources in the robotic workcell contributes to the accuracy problem. Modeling and identification of all the errors to achieve the required levels of accuracy may be difficult. To resolve the accuracy issues, a sensor based indirect error compensation approach is proposed in this paper where the errors are compensated online via measurements of the work object. The sensor captures a point cloud of the work object and with the CAD model of the work object, the actual relative pose of the sensor frame and work object frame can be established via a point cloud registration. Once this relationship has been established, the robot will be able to move the tool accurately relative to the work object frame near the point of compensation. A data pre-processing technique is proposed to reduce computation time and prevent a local minima solution during point cloud registration. A simulation study is presented to illustrate the effectiveness of the proposed solution.
  • Robot End-Effector Sensing with Position Sensitive Detector and Inertial Sensors Authors: Wang, Cong; Chen, Wenjie; Tomizuka, Masayoshi
    For the motion control of industrial robots, the end-effector performance is of the ultimate interest. However, industrial robots are generally only equipped with motor-side encoders. Accurate estimation of the end-effector position and velocity is thus difficult due to complex joint dynamics. To overcome this problem, this paper presents an optical sensor based on position sensitive detector (PSD), referred as PSD camera, for direct end-effector position sensing. PSD features high precision and fast response while being cost-effective, thus is favorable for real-time feedback applications. In addition, to acquire good velocity estimation, a kinematic Kalman filter (KKF) is applied to fuse the measurement from the PSD camera with that from inertial sensors mounted on the end-effector. The performance of the developed PSD camera and the application of the KKF sensor fusion scheme have been validated through experiments on an industrial robot.
  • Experiments towards Automated Sewing with a Multi-Robot System Authors: Schrimpf, Johannes; Wetterwald, Lars Erik
    In this paper a concept for automated multi-robot-aided sewing is presented. The objective of the work is to demonstrate automatic sewing of 3D-shaped covers for recliners, by assembling two different hide parts with different shapes, using two robots to align the parts during sewing. The system consists of an industrial sewing machine and two real-time controlled Universal Robots 6-axis industrial manipulators. A force feedback system combined with optical edge sensors is evaluated for the control of the sewing process. The force sensors are used to synchronize the velocity and feed rate between the robots and the sewing machine. A test cell was built to determine the feasibility of the force feedback control and velocity synchronization. Experiments are presented which investigate the ability of the robot to feed a hide part into the sewing machine using a force sensor and different strategies for velocity synchronization.
  • Automated Throwing and Capturing of Cylinder-Shaped Objects Authors: Frank, Thorsten; Janoske, Uwe; Mittnacht, Anton; Schroedter, Christian
    A new approach for transportation of objects within production systems by automated throwing and capturing is investigated. This paper presents an implementation, consisting of a throwing robot and a capturing robot. The throwing robot uses a linear and the capturing robot a rotary axis. The throwing robot is capable of throwing cylinder-shaped objects onto a target point with high precision. The capturing robot there smoothly grips the cylinders during flight by means of a rotational movement. In order to synchronize the capturing robot and the cylinder’s pose and velocity, its trajectory has to be modeled as well as the motion sequences of both robots. The throwing and capturing tasks are performed by the robots automatically without the use of any external sensor system.

Human Detection and Tracking

  • Iterative Pedestrian Segmentation and Pose Tracking under a Probabilistic Framework Authors: Li, Yanli
    This paper presents a unified probabilistic framework to tackle two closely related visual tasks: pedestrian segmentation and pose tracking along monocular videos. Although the two tasks are complementary in nature, most previous approaches focus on them individually. Here, we resolve the two problems simultaneously by building and inferring a single body model. More specifically, pedestrian segmentation is performed by optimizing body region with constraint of body pose in a Markov Random Field (MRF), and pose parameters are reasoned about through a Bayesian filtering, which takes body silhouette as an observation cue. Since the two processes are inter-related, we resort to an Expectation-Maximization (EM) algorithm to refine them alternatively. Additionally, a template matching scheme is utilized for initialization. Experimental results on challenging videos verify the framework's robustness to non-rigid human segmentation, cluttered backgrounds and moving cameras.
  • A Connectionist-Based Approach for Human Action Identification Authors: Alazrai, Rami; Lee, C. S. George
    This paper presents a hierarchal, two-layer, connectionist-based human-action recognition system (CHARS) as a first step towards developing socially intelligent robots. The first layer is a K-nearest neighbor (K-NN) classifier that categorizes human actions into two classes based on the existence of locomotion, and the second layer consists of two multi-layer recurrent neural networks that distinguish between subclasses within each class. A pyramid of histograms of oriented gradients (PHOG) descriptor is proposed for extracting local and spatial features. The PHOG descriptor reduces the dimensionality of input space drastically, which results in better convergence for the learning and classification processes. Computer simulations were conducted to illustrate the performance of the proposed CHARS and the role of temporal factor in solving this problem. A widely used KTH human-action database and the human-action dataset from our lab were utilized for performance evaluation. The proposed CHARS was found to perform better than other existing human-action recognition methods and achieved a 95.55% recognition rate.
  • Using Dempster’s Rule of Combination to Robustly Estimate Pointed Targets Authors: Pateraki, Maria; Baltzakis, Haris; Trahanias, Panos
    In this paper we address an important issue in human-robot interaction, that of accurately deriving pointing information from a corresponding gesture. Based on the fact that in most applications it is the pointed object rather than the actual pointing direction which is important, we formulate a novel approach which takes into account prior information about the location of possible pointed targets. To decide about the pointed object, the proposed approach uses the Dempster-Shafer theory of evidence to fuse information from two different input streams: head pose, estimated by visually tracking the off-plane rotations of the face, and hand pointing orientation. Detailed experimental results are presented that validate the effectiveness of the method in realistic application setups.
  • Head-To-Shoulder Signature for Person Recognition Authors: Kirchner, Nathan; Alempijevic, Alen; Virgona, Alexander Joseph
    Ensuring that an interaction is initiated with a particular and unsuspecting member of a group is a complex task. As a first step the robot must effectively, expediently and reliably recognise the humans as they carry on with their typical behaviours (in situ). A method for constructing a scale and viewing angle robust feature vector (from analysing a 3D pointcloud) designed to encapsulate the inter-person variations in the size and shape of the people's head to shoulder region (Head-to-shoulder signature - HSS) is presented. Furthermore, a method for utilising said feature vector as the basis of person recognition via a Support-Vector Machine is detailed. An empirical study was performed in which person recognition was attempted on in situ data collected from 25 participants over 5 days in a office environment. The results report a mean accuracy over the 5 days of 78.15% and a peak accuracy 100% for 9 participants. Further, the results show a considerably better-than-random (1/23 = 4.5%) result for when the participants were: in motion and unaware they were being scanned (52.11%), in motion and face directly away from the sensor (36.04%), and post variations in their general appearance. Finally, the results show the HSS has considerable ability to accommodate for a person's head, shoulder and body rotation relative to the sensor - even in cases where the person is faced directly away from the robot.
  • Bigram-Based Natural Language Model and Statistical Motion Symbol Model for Scalable Language of Humanoid Robots Authors: Takano, Wataru; Nakamura, Yoshihiko
    The language is a symbolic system unique to human being. The acquisition of language, which has its meanings in the real world, is important for robots to understand the environment and communicate with us in our daily life. This paper propose a novel approach to establish a fundamental framework for the robots which can understand language through their whole body motions. The proposed framework is composed of three modules : ``motion symbol", ``motion language model", and ``natural language model". In the motion symbol module, motion data is symbolized by Hidden Markov Models (HMMs). Each HMM represents abstract motion patterns. Then the HMMs are defined as motion symbols. The motion language model is stochastically designed for links between motion symbols and words. This model consists of three layers of motion symbols, latent variables and words. The connections between the motion symbol and the latent state, and between the latent state and the words is denoted by two kinds of probabilities respectively. One connection is represented by the probability that the motion symbol generates the latent state, and the other connection is represented by the probability that the latent state generates the words. Therefore, the motion language model can connect the motion symbols to the words through the latent state. The natural language model stochastically represents sequences of words. In this paper, a bigram, which is a special case of N-gram model, is adopted as the natura
  • Cognitive Active Vision for Human Identification Authors: Utsumi, Yuzuko; Sommerlade, Eric; Bellotto, Nicola; Reid, Ian
    We describe an integrated, real-time multi-camera surveillance system that is able to find and track individuals, acquire and archive facial image sequences, and perform face recognition. The system is based around an inference engine that can extract high-level information from an observed scene, and generate appropriate commands for a set of pan-tiltzoom (PTZ) cameras. The incorporation of a reliable facial recognition into the high-level feedback is a main novelty of our work, showing how high-level understanding of a scene can be used to deploy PTZ sensing resources effectively. The system comprises a distributed camera system using SQL tables as virtual communication channels, Situation Graph Trees for knowledge representation, inference and high-level camera control, and a variety of visual processing algorithms including an on-line acquisition of facial images, and on-line recognition of faces by comparing image sets using subspace distance. We provide an extensive evaluation of this method using our system for both acquisition of training data, and later recognition. A set of experiments in a surveillance scenario show the effectiveness of our approach and its potential for real applications of cognitive vision.

Trajectory Planning and Generation

  • Optimal Acceleration-Bounded Trajectory Planning in Dynamic Environments Along a Specified Path Authors: Johnson, Jeffrey; Hauser, Kris
    Vehicles that cross lanes of traffic encounter the problem of navigating around dynamic obstacles under actuation constraints. This paper presents an optimal, exact, polynomial-time planner for optimal bounded-acceleration trajectories along a fixed, given path with dynamic obstacles. The planner constructs reachable sets in the path-velocity-time (PVT) space by propagating reachable velocity sets between obstacle tangent points in the path-time (PT) space. The terminal velocities attainable by endpoint-constrained trajectories in the same homotopic class are proven to span a convex interval, so the planner merges contributions from individual homotopic classes to find the exact range of reachable velocities and times at the goal. A reachability analysis proves that running time is polynomial given reasonable assumptions, and empirical tests demonstrate that it scales well in practice and can handle hundreds of dynamic obstacles in a fraction of a second on a standard PC.
  • Robot Excitation Trajectories for Dynamic Parameter Estimation Using Optimized B-Splines Authors: Rackl, Wolfgang; Lampariello, Roberto; Hirzinger, Gerd
    In this paper we adressed the problem of finding exciting trajectories for the identification of manipulator link inertia parameters. This can be formulated as a constraint nonlinear optimization problem. The new approach in the presented method is the parameterization of the trajectories with optimized B-splines. Experiments are carried out on a 7 joint Light-Weight robot with torque sensoring in each joint. Thus, unmodeled joint friction and noisy motor current measurements must not be taken into account here. The estimated dynamic model is verified on a different validation trajectory. The results show a clearly improvement of the estimated dynamic model compared to a CAD-valued model.
  • On-Line Trajectory Generation: Nonconstant Motion Constraints Authors: Kroeger, Torsten
    A concept of on-line trajectory generation for robot motion control systems enabling instantaneous reactions to unforeseen sensor events was introduced in a former publication. This previously proposed class of algorithms requires constant kinematic motion constraints, and this paper extends the approach by the usage of time-variant motion constraints, such that low-level trajectory parameters can now abruptly be changed, and the system can react instantaneously within the same control cycle (typically one millisecond or less). This feature is important for instantaneous switchings between state spaces and reference frames at sensor-dependent instants of time, and for the usage of the algorithm as a control submodule in a hybrid switched robot motion control system. Real-world experimental results of two sample use-cases highlight the practical relevance of this extension.
  • Setpoint Scheduling for Autonomous Vehicle Controllers Authors: Au, Tsz-Chiu; Quinlan, Michael; Stone, Peter
    This paper considers the problem of controlling an autonomous vehicle to arrive at a specific position on a road at a given time and velocity. This ability is particularly useful for a recently introduced autonomous intersection management protocol, called AIM, which has been shown to lead to lower delays than traffic signals and stop signs. Specifically, we introduce a setpoint scheduling algorithm for generating setpoints for the PID controllers for the brake and throttle actuators of an autonomous vehicle. The algorithm constructs a feasible setpoint schedule such that the vehicle arrives at the position at the correct time and velocity. Our experimental results show that the algorithm outperforms a heuristic-based setpoint scheduler that does not provide any guarantee about the arrival time and velocity.
  • A Real-Time Motion Planner with Trajectory Optimization for Autonomous Vehicles Authors: Xu, Wenda; Wei, Junqing; Dolan, John M.; Zhao, Huijing; Zha, Hongbin
    In this paper, an efficient real-time autonomous driving motion planner with trajectory optimization is proposed. The planner first discretizes the plan space and searches for the best trajectory based on a set of cost functions. Then an iterative optimization is applied to both the path and speed of the resultant trajectory. The post-optimization is of low computational complexity and is able to converge to a higherquality solution within a few iterations. Compared with the planner without optimization, this framework can reduce the planning time by 52% and improve the trajectory quality. The proposed motion planner is implemented and tested both in simulation and on a real autonomous vehicle in three different scenarios. Experiments show that the planner outputs highquality trajectories and performs intelligent driving behaviors.
  • Improved Non-Linear Spline Fitting for Teaching Trajectories to Mobile Robots Authors: Sprunk, Christoph; Lau, Boris; Burgard, Wolfram
    In this paper, we present improved spline fitting techniques with the application of trajectory teaching for mobile robots. Given a recorded reference trajectory, we apply non-linear least-squares optimization to accurately approximate the trajectory using a parametric spline. The fitting process is carried out without fixed correspondences between data points and points along the spline, which improves the fit especially in sharp curves. By using a specific path model, our approach requires substantially fewer free parameters than standard approaches to achieve similar residual errors. Thus, the generated paths are ideal for subsequent optimization to reduce the time of travel or for the combination with autonomous planning to evade obstacles blocking the path. Our experiments on real-world data demonstrate the advantages of our method in comparison with standard approaches.