Accepted by the International Journal of Robotics Research.
Pose and Motion from Contact
Yan-Bin Jia Michael Erdmann
The Rob otics Institute
Scho ol of Computer Science
Carnegie Mellon University
Pittsburgh, PA 15213-3890
June 26, 1998
Abstract
In the absence of vision, grasping an object often relies on tactile feedback from
the ngertips. As the nger pushes the object, the ngertip can feel the contact point
move. If the object is known in advance, from this motion the nger may infer the
location of the contact point on the object and thereby the object pose. This paper
primarily investigates the problem of determining the pose orientation and position
and motion velocity and angular velocity of a planar object with known geometry from
such contact motion generated by pushing.
A dynamic analysis of pushing yields a nonlinear system that relates through contact
the object pose and motion to the nger motion. The contact motion on the ngertip
thus encodes certain information about the object pose. Nonlinear observability theory is
employed to show that such information is sucient for the nger to \observe" not only
the pose but also the motion of the object. Therefore a sensing strategy can berealized
as an observer of the nonlinear dynamical system. Two observers are subsequently
introduced. The rst observer, based on the result of [15], has its \gain" determined
by the solution of a Lyapunov-like equation; it can be activated at any time instant
during a push. The second observer, based on Newton's method, solves for the initial
motionless object pose from three intermediate contact points during a push.
Under the Coulomb friction model, the paper copes with support friction in the plane
and/or contact friction between the nger and the object. Extensive simulations have
been done to demonstrate the feasibility of the two observers. Preliminary experiments
with an Adept robot have also been conducted. Acontact sensor has been implemented
using strain gauges. 1
Figure 1: Di erent motions of contact drawn as dots on an ellipse pushing a quadrilateral
in two di erent initial p oses.
1 Intro duction
Part sensing and grasping are two fundamental op erations in automated assembly. Tra-
ditionally, they are p erformed sequentially in an assembly task. Parts in many assembly
applications are manufactured to high precisions based on their geometric mo dels. This
knowledge of part geometry can sometimes signi cantly facilitate sensing as well as grasp-
ing. It can sometimes also help integrate these two op erations, reducing the assembly time
and cost.
Consider the task of grasping something, say, a p en, on the table while keeping your eyes
closed. Your ngers fumble on the table until one of them touches the p en and inevitably
starts pushing it for a short distance. While feeling the contact move on the ngertip, you
can almost tell which part of the p en is b eing touched. Assume the pushing nger is moving
away from you. If the contact remains almost stable, then the middle of the pen is b eing
touched; if the contact moves counterclo ckwise on the ngertip, then the right end of the
p en is b eing touched; otherwise the left end is b eing touched. Immediately, a picture of the
pen con guration has b een formed in your head so you co ordinate other ngers to quickly
close in for a grip.
The ab ove example tells us that the p ose of a known shap e may be inferred from the
contact motion on a nger pushing the shap e. To b etter illustrate this idea, Figure 1 shows
two motions of a quadrilateral in di erent initial p oses pushed by an ellipse under the same
motion. Although the initial contacts on the ellipse were the same, the nal contacts are
quite far apart. Thinking in reverse leads to the main questions of this pap er:
1. Can we determine the p ose of an ob ject with known geometry and mechanical prop-
erties from the contact motion on a single pushing nger, or simply, from a few inter-
mediate contact p ositions during the pushing? 2 Nonlinear System
State
Object
Contact Dynamics of Kinematics Pushing
Sensor Finger Implementation Input
Output
State Observer
Estimate
Figure 2: Anatomy of pushing.
2. Can we determine any intermediate p ose of the ob ject during the pushing?
3. Furthermore, can we estimate the motion of the ob ject during the pushing?
In this pap er, we will give armative answers to the ab ove questions in the general
case. To accomplish this, we will characterize pushing as a system of nonlinear di erential
equations based on its dynamics. As shown in Figure 2, the state of the system will include
the con gurations p ositions, orientations, and velo cities of the nger and ob ject during
the push at any time instant. The system input will b e the acceleration of the nger. The
system output will b e the contact lo cation on the nger sub ject to the kinematics of contact.
This output will be fed to nonlinear observers, which serve as the sensing algorithms, to
estimate the ob ject's p ose and motion.
Section 2 cop es with the dynamics of pushing and the kinematics of contact, deriving a
system of di erential equations that govern the ob ject and contact motions while resolving
related issues such as supp ort friction in the plane and the initial ob ject motion; Section 3
applies nonlinear control theory to verify the soundness of our sensing approach to be pro-
p osed, establishing the lo cal observabilityof this dynamical pushing system from the nger
contact; Section 4 describ es two nonlinear observers which estimate the ob ject p ose and
motion at any instant and at the start of pushing, resp ectively, and which require di erent
amounts of sensor data; Section 5 extends the results to incorp orate contact friction b etween
the nger and the ob ject; Section 6 presents simulations on b oth observers and the imple-
mentation of a contact sensor, demonstrating that three intermediate contact p oints often
suce to determine the initial p ose for the ngers and ob jects tested; Finally, Section 7
summarizes the pap er and outlines future work. 3
1.1 Related Work
Our work is grounded in rob otics where an abundance of previous work exists. It also draws
up on the part of nonlinear control theory that concerns nonlinear observability and observers.
1.1.1 Rob otics
Dynamics of sliding rigid b o dies was treated by MacMillan [36] for non-uniform pressure
distributions, and by Goyal et al. [19] using geometric metho ds based on the limit surface
description of friction. Howe and Cutkosky [26] exp erimentally showed that the limit sur-
face only approximates the force-motion relationship for sliding b o dies and discussed other
simpli ed practical mo dels for sliding manipulation.
Mason [37] pioneered the study of the mechanics of pushing using quasi-static analysis,
predicting the direction in which an ob ject b eing pushed rotates and plotting out its in-
1
stantaneous rotation center. For unknown centers of friction, Alexander and Maddo cks [2]
reduced the problem of determining the motion of a slider under some applied force to the
case of a bip o d, obtaining analytical solutions for simple sliders. The problem of predicting
the accelerations of multiple 3D ob jects in contact with Coulomb friction has a nonlinear
complementarity formulation due to Pang and Trinkle [40]; the existence of solutions to
mo dels with sliding and rolling contacts has b een established.
Montana [38] derived a set of di erential equations describing the motion of a contact
p ointbetween two rigid b o dies in resp onse to a relative motion of these b o dies, and employed
these equations to sense the lo cal curvature of an unknown ob ject and to follow its surface
while steering the contact p oint to some desired lo cation on the end e ector. The kinematics
of spatial motion with p oint contact was also studied by Cai and Roth [6] who assumed a
tactile sensor capable of measuring the relative motion at the contact p oint. The sp ecial
kinematics of two rigid b o dies rolling on each other was considered by Li and Canny [33]in
view of path planning in the contact con guration space. In our work, contact kinematics
is derived directly from the absolute velo cities of the nger and the ob ject rather than from
their relative velo city at the contact. Also we are concerned with a nger and ob ject only
in the plane not in 3D.
Part of our motivation came from the blind grasping task at the b eginning of the pap er.
The caging work by Rimon and Blake [43] is concerned with constructing the space of all
con gurations of a two- ngered hand controlled by one parameter that con ne a given 2D
ob ject; these con gurations can lead to immobilizing grasps by following continuous paths
in the same space. This work requires an initial image of the ob ject taken by a camera.
Work related to caging includes parts feeder design [42] and xture design [5]. In this pap er,
we are concerned with how to \feel" a known ob ject using only one nger and how to infer
its p ose and motion information rather than how to constrain and grasp the ob ject using
multiple ngers.
A larger part of the motivation of our work was from parts orienting. Orienting mechan-
ical parts was studied early on by Grossman and Blasgen [22]. They used a vibrating box
to constrain a part to a small nite number of p ossible stable p oses and then determined
1
The center of friction is the centroid of the pressure distribution. 4
the particular p ose by a sequence of prob es using a tactile sensor. Inspired by their result,
Erdmann and Mason [12] constructed a planner that employs sensorless tilting op erations
to orient planar ob jects randomly dropp ed into a tray, based on a simple mo del of the
quasi-static mechanics of sliding. Utilizing the theory of limit surfaces [19], Bohringer et
al. [4] develop ed a geometric mo del for the mechanics of an array of micro electromechanical
structures and showed how this structure can be used to uniquely and eciently align a
part up to symmetry. Goldb erg [17] showed that every p olygonal part with unknown initial
orientation can be oriented by a parallel-jaw gripp er up to symmetry in the part's convex
hull. He constructed an algorithm with sub-cubic running time as a pro of of sensorless parts
orienting.
Based on the limit surface mo del and quasi-static analysis, Lynch et al. [35] conducted
active sensing of an ob ject's center of mass during pushing from tactile feedback; and de-
velop ed a control system that translates and orients ob jects. Their work assumes known
contact edge and do es not quantitively estimate the ob ject's motion due to the quasi-static
nature. Also applying quasi-static analysis, Akella and Mason [1] describ ed a complete op en-
lo op planner that can orient and translate p olygonal ob jects in the plane by pushing with a
straight fence.
In our previous work [29], weintro duced the metho ds of cone inscription and p oint sam-
pling that compute the p oses of known shap es from a continuum and a nite number of
p ossibilities, resp ectively, using simple geometric constraints such as coincidence and con-
tainment. Paulos and Canny [41] studied the problem of nding optimal p oint prob es for
re ning the p ose of a p olygonal part with known geometry from an approximate p ose; they
revealed that this problem is dual to the grasping problem of computing optimal nger
placements and gave an ecient near-optimal solution.
Mo del-based recognition and lo calization traditionally solve a constraint satisfaction
problem by searching for a consistent matching b etween sensory data and mo dels. This is
often conducted by pruning the search tree sub ject to pure geometric constraints. Grimson
and Lozano-P erez [21] used tactile measurements of p ositions and surface normals on a 3D
ob ject to identify and lo cate it from a set of known 3D ob jects, based on the geometric
constraints imp osed by these tactile data. Gaston and Lozano-P erez [14] showed how to
identify and lo cate a p olyhedron on a known plane using tactile information that includes
contact p oints and the ranges of surface normals at these p oints. Grimson [20] advo cated
that mo del-based recognition and lo calization should be regarded as a constraint satisfac-
tion problem that searches for a consistent matching between sensory data e.g., 2D and
mo dels e.g., 3D under some geometric constraints.
A ma jor di erence between the sensing approach describ ed in this pap er and the listed
previous work is that our approach combines geometric constraints with dynamics of ma-
nipulation. On the one hand, this would relieve the requirement for sucient geometric
constraints and hence lessen the hardware load. On the other hand and more imp ortantly,
the involvement of dynamics extends sensing to dynamic p ose and motion estimation. Sub-
sequently, the to ols whichwe will apply are from nonlinear control theory rather than from
AI and computational geometry.
One feasible implementation for contact detection is to employ a force or tactile sen-
sor. The pap er by Salisbury [44] prop osed the concept of ngertip force sensing with an 5
approach for determining contact lo cations and orientations from force and moment mea-
surements. Fearing and Binford [13] designed a cylindrical tactile sensor to determine the
principal curvatures of an ob ject through rolling contact. Based on continuum mechanics
and photo elastic stress analysis, Cameron et al. [7] built a tactile sensor using a layer of
photo elastic material along with its mathematical mo del. Allen and Rob erts [3] deployed
rob ot ngers to obtain a number of contact p oints around an ob ject and then t in a least
squares manner the data to a sup erquadric surface representation to reconstruct the ob-
ject's shap e. Howe and Cutkosky [25] intro duced dynamic tactile sensing in which sensors
capture ne surface features during motion, presenting mechanical analysis and exp erimental
p erformance measurements for one typ e of dynamic tactile sensor|the stress rate sensor.
1.1.2 Nonlinear Control
The theoretical foundation of our work comes from the part of control theory concerned with
the observability and observers of nonlinear systems. For a general intro duction to nonlinear
control theory, we refer the reader to Isidori [27] and Nijmeijer and van der Schaft [39].
Hermann and Krener [24] rst studied observability using the notion of observation space.
We will use their observability rank condition to show that the p ose and motion of a p olygonal
ob ject pushed by a disk is lo cally observable from the disk contact. A result due to Crouch[9]
shows that an analytic system is observable if and only if the observation space distinguishes
p oints in the state space.
Luenb erger-like asymptotic observers, rst constructed by Luenb erger [34] for linear sys-
tems, remain likely the most commonly used observer forms for nonlinear systems to day.
Gauthier, Hammouri, and Othman [15] describ ed an observer for ane-control nonlinear
systems whose \gain" is determined via the solution of an appropriate Lyapunov-like equa-
tion. Their observer has a very simple form: it is a copy of the original system, together
with a linear corrective term that dep ends only on the state space dimension. Our rst p ose
observer will b e constructed using the GHO pro cedure.
Ciccarella et al. [8] prop osed a similar observer whose gain vector is controlled by the prop-
erly chosen eigenvalues of a certain matrix obtained from the original system's Brunowsky
canonical form, thus providing more freedom on optimizing the observer b ehavior. Extending
the results of GHO [15], Gauthier and Kupka [16] characterized non-ane control systems
that are observable under any input and constructed a generic exp onential observer for these
systems.
Zimmer [46] presented a state estimator that conducts on-line minimization over some ob-
jective function. His observer, with provable convergence, iteratively uses Newton's metho d
to mo dify its state estimate every xed p erio d of time. Our second p ose observer will also
make use of Newton's metho d but we will estimate the initial motionless p ose of the ob ject,
relieving the task of evaluating the complex second order partial derivatives of the drift eld
as Zimmer had encountered. 6 yF Finger F α x F ' α ( u ) ( s ) Object B vF β β' F y ω B xB
O v dp p
vp
Figure 3: Finger F translating and pushing ob ject B .
1.2 Notation
We abide by the following conventions on notation in this pap er. Every vector x is a column
T
vector written as x ;:::;x for some variables x ;:::;x . The derivative of a vector
1 n 1 n
dx dx
T T
n
1
;:::; . function xt=x t;:::;x t with resp ect to t is denoted by dx=dt =
1 n
dt dt
T
The gradient of a scalar function y x, where x = x ;:::;x , is a row vector @y=@x =
1 n
@y @y
T
;:::; . The partial derivative of a vector eld f x = f x;:::;f x with
1 m
@x @x
n
1
@f
T
i
resp ect to vector x =x ;:::;x is a m n matrix, given by @ f =@ x = .
1 n
ij
@x
j
Toavoid anyambiguity, the notation `_ ' means di erentiation with resp ect to time, while
0
the notation ` ' means di erentiation with resp ect to some curve parameter. For example,
du d
0
_
gives the velo city of a p ointmoving on a curve u. The cross pro duct of = u_ =
du dt
twovectors e.g., v is treated as a scalar wherever ambiguitywould not arise. A scalar
in a cross pro duct e.g., the angular velo city ! in ! acts as a vector of equal magnitude
and orthogonal to the plane.
2 Motion of Contact
Throughout the pap er we consider the two-dimensional problem of a translating nger F
pushing an ob ject B . Coulomb's friction law is assumed and the co ecient of supp ort
friction, that is, friction between B and the plane, is everywhere . For simplicity, let us
assume uniform mass and pressure distributions of B . Let us also assume frictionless contact
between F and B at present and deal with contact friction exclusively in Section 5. Let v
F
be the velo city of F , known to F 's controller, v and ! the velo city and angular velo city of
B , resp ectively, all in the world co ordinate frame Figure 3.
Let F 's b oundary be a smo oth curve and B 's b oundary be a piecewise smo oth closed
curve such that u and s are the two p oints in contact in the lo cal frames of F and
B , resp ectively. Following convention, moving counterclo ckwise along and increases u
and s, resp ectively. Assume that one curve segmentof stays in contact with throughout 7
the pushing.
2.1 Dynamics of Pushing
That F and B maintain contact imp oses a velo city constraint:
0
0
v + u_ = v + ! R + R s;_ 1
F
!
cos sin
where R = is the rotation matrix asso ciated with the orientation of
sin cos
B , which is determined by u, s, and the orientation of F . Newton's and Euler's equations
on rigid b o dy dynamics are stated as:
Z
^ _
F + g v dp = mv ; 2
p
B
Z
^
R F + R p g v dp = I !;_ 3
p
B
where F is the contact force acting on B , g the acceleration of gravity, m the mass, the mass
density, and I the angular inertia ab out the center of mass O all of B . Here v = v + ! R p
p
v
p
2
^
is the velo cityofp 2 B and v = its direction.
p
kv k
p
With no friction at the contact p oint, F acts along the inward normal of B :
0
F R = 0; 4
0
R F > 0: 5
Finally, the normals of F and B at the contact are opp osite to each other; equivalently, we
have
0
0
R = 0;
6
0
0
R < 0:
Given the nger motion v , there are seven equations 1, 2, 3, 4, and 6 with
F
3
seven variables u; s; ! ; v , and F . From these equations, we are now ready to derive the
di erential equations for u; s; ! , and v .
q
R
I m
and = the area and radius of Let a be the acceleration of F , A = dp =
F
B
m
R
0 0
^ ^
R R p v + v dp an integral asso ciated gyration of B , resp ectively, and =
p p
B
with supp ort friction. We have
2
That F is translating implies either v 6=0 or ! 6= 0 after the pushing starts. So v can vanish over at
p
most one p oint p 2 B , which will vanish in the integrals in equations 2 and 3.
3
Note that equations 1 and 2 and variables v and F are each counted twice. 8
Theorem 1 Consider the pushing system described by 1{6. The points of contact evolve
according to
0 00
0 2 0 0