Post Job Free
Sign in

Control Data

Location:
Nashville, TN
Posted:
December 10, 2012

Contact this candidate

Resume:

Learning Sensory Motor Coordination for Grasping

by a Humanoid Robot

Mark Edward Cambron and Richard Alan Peters II

Vanderbilt University School of Engineering

Station B, Box 131

Nashville,TN 37235

Telephone: 615-***-****

Fax: 615-***-****

E-mail: ***@****.**********.***, ****@****.**********.***

Abstract This paper proposes a method for

the learning of Sensory-Motor Coordination (SMC)

through the teleoperation of a humanoid robot de-

signed for human-robot interaction. It is argued that

SMC in a complex environment must be learned rather

than programmed. Schema theory is reviewed as a tool

for the description of animal behavior at the level of

functional modules and higher. SMC is shown to be

necessary for the formation of schema assemblages for

the control of behavior. Examples are given of four

behavior-based robot control architectures that implic-

itly use SMC-driven schema control. It is shown that

while the robots are capable of learning, they all rely,

to a certain extent, on ad hoc choices of SMC by the

designers. A description of the humanoid robot and

the sensors it uses for reaching and grasping is given.

The proposed method of learning via teleoperation is Fig. 1. ISAC, Vanderbilt s humanoid robot.

described. Sensory data acquired through grasping is

presented.

tems that manifest it were known then analogous systems

could be implemented in robots.1 SMC in animals is not

I. Introduction completely understood, but research in it has recently ad-

ESEARCHERS at the Intelligent Robotics Labora- vanced to the point where plausible mechanisms for it have

R tory at Vanderbilt University have been developing a been described. Evidence from studies in neurophysiology

humanoid robot, ISAC, over the past several years (Fig. 1). [5], ontogenesis [6], [7], and cognitive science [8] suggests

The robot was designed expressly for research in human- that to interact e ectively and e ciently with its environ-

robot interaction [1], [2]. ISAC s control architecture is ment, an animal must learn through its own experiences

an agent-based, hybrid deliberative-reactive system. Like the reciprocal causative relationships between sensing and

a behavior-based robot, ISAC s complex behaviors result action that foster its success or survival. (cf. Sec. II.) That

from the interaction of independent computational mod- is, SMC must be learned, or at least re ned, through an

ules that operate asynchronously in parallel [3], [4]. animal s direct experience with acting in the world.

Schema theory [8] can be used to describe the functional

To interact naturally with people in a human-centered

aspects of an animal s behavior without exact speci ca-

environment, a robot must be able to coordinate sensing

tion of the biological systems that support it. Schemas

with action. That is, it must have Sensory-Motor Coor-

exist at a frame of reference higher than that of the in-

dination (SMC). It is possible to program a certain de-

dividual computational elements (neurons in the case of

gree of SMC into a robot prior to its deployment. But it

animals). A schema description of the behavior of an an-

is impossible for a programmer to anticipate every phys-

imal is inherently modular. It provides a framework for

ical contingency that may arise in a robot s interactions

the description of behaviors in terms of the interactions of

with people. This is due to the intrinsic complexity of

modules that control motion, process sensory information,

a human-centered environment. Only animals (including

people) have SMC that permits them to work e ectively 1 Implementation is possible if the functionality of the biological

in a complex natural world. If SMC in animals were well systems can be reproduced in electro-mechanical systems. Schema

understood if the structures and functions of the sys- theory suggests that it can. (cf. Sec. III.)

create and recall memories, etc. In animals, the mod- ated with it occur, the action is performed without resort

ules may more or less directly correspond to speci c net- to modeling or deliberation. Basic behaviors are canoni-

works of neurons. But this separation of function from cal in the sense that all actions exhibited by the robot are

structure a ords the possibility of realizing the behavior generated through the cooperation and competition of ba-

of an animal in a robot by substituting computers and sic behaviors operating concurrently or in sequence. At

electro-mechanical devices for neuron networks and bio- any given point in time, some of the basic behaviors will

mechanical subsystems. Behavior-based robots (BBR) [9], be enabled and others suppressed depending on the task

[10] are particularly amenable to this. BBRs act through and environmental context of the robot. Since a BBR ex-

the combination of basic behaviors, which are motor ac- hibits any and all its behaviors through the combination

tions tightly coupled to sensory stimuli both external to and sequencing of basic behaviors, a BBR is wholly depen-

the robot and internal (i.e. proprioceptic). dent on, and to a large extent de ned by, sensory motor

This paper proposes a method for the learning of coordination.

sensory-motor coordination through the teleoperation of Sensory-motor coordination is fundamental for another

a behavior-based robot. The goal of the work is to enable compelling reason. It forms a foundation for higher level

a robot to learn SMC by nding the correlations between learning and perception. In particular, the categorization

sensory events and motor control events that co-occur dur- of sensory stimuli can be accomplished through SMC [11].

ing task execution. The robot is guided by a human oper- A mobile agent can learn the sensory patterns that corre-

ator through repeated trials of a speci c task while record- spond to an obstacle by associating stimuli with its mo-

ing all its incoming sensory data. The motor and sensory tor responses, as when a characteristic stimulus pattern

data gathered throughout the trials will be analyzed to routinely accompanies the sudden inability to move. Sim-

nd representative couplings between sensory stimuli and ilarly, as Pfeifer has demonstrated, an agent can learn to

motor actions. If successful this will not only permit the distinguish between objects that it can manipulate and

robot to perform the task autonomously, but also (with an those which it cannot [12]. If the internal sensation of a

appropriate control system) enable the robot to adapt to need (a drive or a goal) having been satis ed accompanies

variations in the task or in the environment. a set of actions performed in the presence of speci c stim-

uli, that stimuli can be recognized as being bene cial to

the agent (e.g. an energy source food). Recent experi-

II. Sensory-Motor Coordination

ments by Pfeifer and others have demonstrated that such

Sensory-Motor Coordination underlies the physical be-

SMC events can be used to learn classi cations of objects

havior of an animal in response to its environment. More

and events in the environment more easily and more accu-

than a response, SMC is a feedback loop that changes

rately than can traditional machine sensing strategies such

both the animal and the environment. An animal s mo-

as model-based vision [13], [14].

tions are caused by muscle contractions. These contrac-

tions are elicited by electrochemical signals that are gen-

erated by circuits of motor neurons. When the animal III. Schema Theory

moves, it causes a relative shift in the environment. As Since the behavior of animals is mediated by their ner-

the environment shifts, energy patterns sweep across the vous systems, the understanding of their behavior from

animal s sensory organs. Sensory organs are transducers rst principles requires an understanding of nervous sys-

that, in e ect, transform external, spatio-temporally dy- tems. Neuroscience has provided a structural description

namic energy elds into electrochemical signals carried by that includes neurons (individuals and networks) and lay-

circuits of sensory neurons internal to the animal. These ers, columns, and modules in the brain [15]. But the func-

sensory signals (more or less directly) modulate the signals tion of these structures is not completely understood and it

in the original motor circuits. Learning occurs in the map- is function more than structure that determines behavior.

ping from sensory response signal to motor control signal. Functional analysis is complicated by the fact that many of

Thus, an animal senses the environment and acts. The ac- the neuronal structures participate in di erent functions.

tion changes the environment relative to the animal, which With certain exceptions there are no discernible one-to-

senses those changes and acts accordingly. one mappings of low-level structure to high-level function

SMC is likewise needed by a sensory-guided robot. The [8].

basic behaviors of a BBR are independent units of SMC. Arbib et al. employ schema theory as a framework for

They include what are commonly called re ex actions. the rigorous analysis of [animal] behavior that requires no

When a basic behavior is enabled2 and the stimuli associ- prior commitment to hypotheses on the location of each

schema (unit of functional analysis) but can be linked to a

2 A BBR typically has a suite of basic behaviors, not all of which are

operational at the same time. Depending on the task and environ- structural analysis as and when it becomes appropriate.

mental contexts, various basic behaviors will be enabled or disabled.

[8], (p. 33). Thus schemas are descriptions of functions

If a behavior is enabled made operational it will remain quiescent

that are performed by networks of neurons and the muscles

until its triggering sensory stimuli are present.

and appendages that they control. Schema theory enables object) must be learned. On the other hand, an animal

the top-down analysis of a complex behavior by providing must, to a certain extent, hit the ground running to sur-

a structure for logically dissembling it, that is it facilitates vive. Motion must precede perception so that the animal

the analytical decomposition of a complex behavior into can move at birth and so that the e ects of its motion can

sets of simpler behaviors. On the other hand, schemas also be perceived and learned. Perceptual schemas, must there-

enable the bottom-up analysis of sets co-occurring behav- fore be learned or tuned in concert with motion. Simulta-

iors. The collective behavior of a set of simple schemas can neously, motor schemas must be tuned to enable e cient

be deduced if the framework for their competition, coop- sensing. Thus, sensory-motor coordination requires the

eration, and sequencing is known. This collective behavior coupling of perceptual schemas and motor schemas into

is a higher-level schema called an assemblage. assemblages. Perceptual schemas provide goal and trajec-

tory information to the motor schemas, whereas the latter

Not only are the behaviors of animals describable by

provide a physical framework within which a perceptual

schemas but also are the control systems of behavior-based

schema can extract salient information. Arbib et al. place

robots. BBRs are, given their modular architectures, par-

motor schemas and perceptual schemas at the foundation

ticularly amenable to such description. The theory of

of animal function. Under the in uence of the environ-

behavior-based robotics is grounded on the idea that com-

ment these schemas self-organize to control an animal s

plex behavior in an agent emerges through the competition

behavior.

and cooperation of simple behaviors in the context of an

For the designers of robots the main implication of the

environment, which is precisely the idea of assemblage in

onset of motility prior to sensation in animals is that re-

schema theory.

exes are not primary. (See [8], Sec. 2.1.1, p. 13 f.f.) Put in

To the extent that function can be separated from struc-

another way, basic behaviors are not truly basic. Motion is

ture, a schema representation enables a speci c behav-

primary; it can happen without sensing. Re exes develop

ior to be performed by agents with dissimilar computa-

with the onset of sensing. Then sensory signals modulate

tional hardware. In particular, a behavior observed in

the signaling of motor circuits and re exes emerge.

an animal that can be described accurately by schemas

could be implemented on an appropriately structured

robot. Schemas, therefore, provide for comparative analy- IV. Schemas and SMC in Behavior Based Robots

sis of similar behaviors on dissimilar agents, be they bio- The following four examples of behavior-based robot

chemical or electro-mechanical. control systems depend on SMC and can be described

Arbib et al. group schemas in two categories. Motor through assemblages of schemas. Each of the architectures

schemas are the control systems which can be coordi- has basic behaviors at its foundation. In each case, the

nated to e ect the wide variety of movement. A set of basic behaviors are selected by the designer of the robot.

basic motor schemas is hypothesized to provide simple, Each of the architectures can be designed to learn, and

prototypical patterns of movement. Perceptual schemas as a result exhibit emergent SMC. The learning, however,

are those used for perceptual analysis. They embody the occurs at levels above basic behaviors.

processes whereby the system determines whether a given a) Brooks Subsumption Architecture. : Brooks sub-

domain of interaction is present in the environment. They sumption architecture controls a robot through a collection

not only serve as pattern-recognition routines but can also of augmented nite state machines (AFSM) organized into

provide the appropriate parameters concerning the current layers [9]. A subsumptive robot has no central planner or

relationship of the organism with its environment. [8], (p. controller. Each AFSM can be activated by sensory inputs

42). and produces outputs that drive actuators or are passed

Research in the ontogenesis of animals has demonstrated to the inputs of other modules. Within subsumption, the

that the ability to move exists prior to an animal s abil- AFSMs are motor schemas. The sensory inputs are percep-

ity to sense its environment. Arbib et al. state that this tual schemas. An AFSM with well-de ned sensory input

does not, however, imply that motility is an end in itself. implements a basic behavior. Assemblages are formed dy-

Rather this, motor foundation serves to group the later namically as AFSMs at one level are activated or inhibited

development of sensory maps and sensorimotor represen- by AFSMs at a higher level. Usually the basic behaviors

tations in a self-directed manner. [8], (p. 10). Thus, in in the lowest layer are preprogrammed; the sensory signals

animals the formation of the musculo-skeletal system and that trigger an AFSM are not learned. Learning can take

the neuro-circuits for motor control precedes the develop- place in a subsumption architecture, (e.g., Brooks robot,

ment of perceptual schemas. Such a development sched- Ghengis [16]) but generally this occurs in layers above the

ule makes sense. Perceptual schemas in animals, even if rst.

passed on phylogentically, must be tuned; sensory stim- b) Mataric s action-oriented representations. :

uli is required for a perceptual modality to develop. Other Mataric designed, using subsumption, a mobile robot that

perceptual schemas (e.g., a semantic description of a visual learns to navigate an environment through the use of

action-oriented representations [17]. The robot both gen- tion) and summing their outputs onto the motor con-

erates and operates from an action map of the environ- trollers [12]. The response of each behavior module to

ment. While wandering in the environment and reacting its inputs is preprogrammed. The overall robot system

to sensory input according to its basic behaviors (e.g., wall does learn, however, as it interacts with the environment,

following, object avoidance, etc.) the robot generates the guided by a value system. Values are, essentially, the

map by building up a directed graph. Each node of the preprogrammed re exes and reinforcement schemes, that

graph contains a description of the motor state at the time cause the robot to seek some sensory stimuli and to avoid

of its formation and description of the sensory data that others. Learning occurs through the adaptive modulation

was received as the robot performed the actions described of sensory signals that are fed to the behavior modules.

by the motor state. Adjacent nodes in the graph corre- Pfeifer de nes categorization of an object as the robot s

spond to adjacent areas in the environment. Once the en- appropriate interaction with the object. Through the

vironment has been mapped, the robot can reach a phys- value-based learning scheme the robot learns how to couple

ical location by activating the corresponding node of the sensing with actuation so that appropriate behaviors are

graph. The graph is searched (using spreading activation) learned for di erent stimulus patterns. Thus the objects

back from the goal node to the node that represents the that project the di erent stimulus patterns are classi ed

current position of the robot. The nodes along the short- de facto without forming an abstract model of the object.

est connecting path are enabled. The robot reaches the Pfeifer s robot learns about objects by nding the corre-

goal by moving according to the motor commands of its lations between sensory signals and behaviors that lead to

current node until its sensory input more closely matches favorable results and by decoupling behaviors from stimuli

the data from the next node. Then it executes the motor when that coupling leads to unfavorable results. Thus, an

commands from next node and proceeds successively from appropriate linkage between sensing and action at the task

node to node until the goal is reached. level is learned by trial and error.

Mataric s robot learns while acting by forming a spatio-

temporal sensory-motor description of the environment.

The map indicates the sensory and motor status of the V. Learning Basic Behaviors

robot at a particular point in space at a particular time

Behavior-based robots employ schemas implicitly. Their

relative to the current position. Thus, the robot learns how

complex behaviors emerge through the interaction of a

to sequence and basic behaviors from sensory input. This

canonical set of basic behaviors, each of which is a sensory-

is undoubtedly a form of SMC but it learns the sequencing

driven motor controller. Therefore, in a BBR high-level

of basic behaviors rather than the SMC that de nes the

behavior emerges from assemblages of perceptual schemas

basic behaviors themselves.

linked to motor schemas, just as in animals. In terms of

c) Arkin s Motor Schema. : A robot controlled by

schema theory, the practice of designing BBRs di ers from

Arkin s motor schema3 architecture follows gradients in a

the ontogenesis of animals. The designer of a BBR must

vector eld map of its environment [18]. Computational

decide ad hoc or through trial and error, exactly which

modules such collision detectors and obstacle or object rec-

coupling of sensory data to motor controller constitutes

ognizers are perceptual schemas since they compute the

a useful basic behavior. And the designer must decide

vectors at points in space that serve to impel the robot.

which basic behaviors to include in the canonical set. He

Motor schemas (in Arbib s sense) within Arkin s archi-

or she determines the perceptual to motor schema linkage

tecture are assemblages of motor controllers that respond

at the base level and decides which of these rst-order as-

individually to components of the vector eld map. A mo-

semblages to include on the robot.4 In other words, the

tor controller generally has a xed response to its input

designer programs SMC into the robot at the lowest level.

vector. The response is a function of the magnitude and

BBRs that learn, such as those described in the previ-

direction of the input vector, but that function is generally

ous section, learn at the level above basic behaviors. They

preprogrammed and does not change. Any learning that

learn which behaviors to activate and which to inhibit or

occurs happens in the perceptual schemas that compute

to suppress under various sensory conditions, or they learn

the vector eld.

an appropriate sequence of behaviors in response to sen-

d) Pfeifer s SMC-based categorization. : Pfeifer s

sory input, or they learn a control function that modulates

robots are based on an extended Braitenberg architec-

the sensory signals before they reach the basic behaviors.

ture, another type of BBR [19]. A number of basic be-

While these robots might work well, they are still subject

havior modules (Pfeifer calls these re exes ) operate in

to the errors and oversights of their designers in program-

parallel, receiving sensory inputs (including propriocep-

ming SMC into the functional base-level of the robot.

3 Motor Schema is the name that Arkin has given his control archi-

4 In some BBRs the higher-order assemblages are also completely

tecture. It makes use of both perceptual schemas and motor schemas

in the sense that Arbib describes them. speci ed by the designer. Such robots cannot learn.

S1 : base joint positions and velocities

How, then, does one enable a robot to learn SMC at the

S2 : shoulder joint positions and velocities

level of basic behaviors? There are at least two possibili-

S3 : elbow joint positions and velocities

ties:

S4 : wrist joint positions and velocities

1) Design and implement on the robot the fundamen-

S5 : Finger Positions

tal motor circuits that enable actuation. Have the

S6 : 6-axis FT sensor output

robot move randomly while sensing. Reinforce any

S7 : Proximity sensor 1 output

sensory-motor coupling (a temporal coincidence of

S8 : Proximity sensor 1 output

sensory signals and motor actions) that leads to pur-

S9 : Finger touch sensor 1

poseful motion. An approach such as this is nec-

S10 Finger touch sensor 2

essary for a fully autonomous agent, like an ani-

mal. This approach has been used successfully by

TABLE I

researchers in arti cial life [20]. Learning SMC this

Signals measured during teleoperation.

way with a robot could require much time.

2) Take advantage of the fact that a robot can be tele-

operated. When a person teleoperates a robot, the

person s SMC causes the robot to act purposefully. The other end-e ector, the PneuHand II is similar to

If the robot records all of its sensory signals dur- the rst but has two more actuators (in addition to the

ing repeated teleoperations, through signal analysis pistons), one each on the thumb and fore nger. (Fig. 2).

it should be able to identify the sensory-motor cou- The actuator is a DC micro-motor connected to a lead

plings that accompany purposeful motion. screw by a universal joint. The nut moves the drive link.

These actuators give the thumb and fore nger a contin-

Both of these approaches require signal analysis algorithms

uous range of positions between opened and closed. In a

that will detect signal correlations or coincidences. More-

typical grasp with the PneuHand II, the motors are used

over, the detected sensory-motor couplings must be used

to close the thumb and fore nger on an object. Then the

to construct basic behavior modules. Both of these prob-

pistons are engaged to close the other two ngers and to

lems are open research issues.

add strength to the grasp of the thumb and fore nger.

ISAC s sensor suite includes the stereo camera pair, the

VI. The ISAC Humanoid mics for sonic localization, and the motion detector array.

It also has, between the arms and the end e ectors (at

ISAC is an upper torso humanoid with two 6-degree-

the wrist joints) 6-axis force torque sensors. Two photo-

of-freedom arms actuated by McKibben Arti cial Muscles

electric proximity sensors are installed in the center of each

attached to the arms as antagonistic pairs and with anthro-

end e ector s palm. The sensors generate binary signals.

pomorphic end-e ectors. It has a color, stereo, active vi-

One sensor res when an object is within approximately

sion system with pan, tilt, and verge that implements the 5

100 mm of the palm, the other res at 50 mm. On the

basic oculomotor behaviors of primates: vergence, smooth

end of each digit of the end e ectors is a touch sensor with

pursuit, saccade, vestibular-ocular re ex, and opto-kinetic

binary output which res on contact.

re ex. ISAC employs sonic localization, infrared motion

detection, and speech I/O.

VII. Methods and Procedures

One of the two end e ectors is the PneuHand I. It has a

The objective of this research is to enable a ISAC to

central palm plate to which three ngers and an opposing

learn the sensory-motor control couplings that de ne a

thumb are connected [21]. Each digit consists of two pha-

canonical set of basic behaviors and to learn the sensory

langes and two joints, proximal and distal. Each digit is

signals that precede and follow behavior changes during

individually actuated by a pneumatic piston. The one ac-

task execution. The approach is to have a person teleop-

tuator per digit design was intended to decrease the mass

erate the robot through a task a number of times while

of the hand and to simplify its control. However, as a con-

the robot records the motor control sequence and the sig-

sequence of this, each digit of the PneuHand I has only two

nals from its sensors. For the experiments reported herein,

positions, opened or closed. The primary advantage of the

ISAC s task was to nd, to reach toward, and to grasp a

PneuHand I over other simple grippers is its morphology.

stationary object. This was accomplished through tele-

It has a human form factor that enables it to grasp and

operation, wherein the teleoperator controlled the action

to hold any object (within the weight limits of the arm-

of the robot from a keyboard while observing the robot s

hand system) that has a quasi-cylindrical handle (e.g. a

actions. The signals recorded are listed in Table 1.

hammer, a telephone handset, a drinking glass, etc.) The

The sequence of actions for a single trial was:

PneuHand I can grasp with its thumb and one, two, or

three ngers. But, it is does not have the dexterity to do 1) Perform an active visual analysis of the scene: Grab

things other than grasping and holding. a stereo image pair of the scene. Locate the ob-

ject through the visual attention network. Fixate on

the object. Estimate the object s 3D position in the

workspace. Extract edges from the imagery. Esti-

mate of the medial axis and the limbs of the object.

Combine the position and orientation to estimate the

pose of the object.

2) Reach toward the object: using visual servoing to-

ward the 3d position point, with a wrist position

relative to the rest of the arm chosen visually by the

operator, and with the hand open.

3) Close the hand: at an arm position close to the ob-

ject chosen visually by the operator.

Figures 3 and 4 show the parallel behaviors which com-

prise the control architecture of the teleoperated grasp

agent and the sensory-motor data ow pathways from the Fig. 2. Vanderbilt University s PnueHand2

sensing and motor control agents into the

During this process ISAC records the sensory-motor

data stream. An example of a typical data set is shown

in Figure 5. Upon the completion of the operation, the

From sensors

and effectors

From Other

operator judges its success on a three point scale: 0

Process

unsuccessful, 1 partially or almost successful, and 2

successful.

Camera

VIII. Learning SMC Object Fixation Angles

The motor control sequence within each trial will be

used to determine the motor events the times of tran- Arm

sition between continuous motor operation states. The Position

Visual Servoing

motor events from a trial will be used to partition all the

sensory signals within that trial. Since the same task is re- Gripper

Position

peated by the same operator several times there should be

GraspObject

the same number of motor events in each trial, although

the time between them will vary. After all the trials are

completed, the signals will be time warped to align the Fig. 3. Architecture of the grasp agent.

motor events across trials. Then in a time interval brack-

eting the motor event, the signals from a single sensor will

be correlated across all trials to determine if there is a

From Arm Agent

corresponding sensory event (the signal exhibits a change Arm

Proximity

consistently near the motor event.) Only the signals that Position

From Pneu Agent

Sensors

exhibit a consistent sensory event within an interval of a

motor event will be considered to be salient to that mo-

Force

tor event and analyzed further. (A signal that is constant SMC

From Arm Agent

SMC Data

Torque Sensor Data

or that changes inconsistently near a motor event across

multiple trials of the same task is presumed to be super-

FSR

uous to the SMC of that event.) Through averaging (or From Hand Agent

Hand

some nonlinear combination such as median ltering) a Data

Position

characteristic signal for that sensory event at the given

motor event will be formed. Then the signals from dif- Touch

ferent sensors will be correlated within individual trials Data Flow Data From Hand Agent

to determine which sensors react together near the mo-

Mechanism

tor events. To each motor event, the characteristic signals

Representation

from the salient sensors are coupled to form a sensory-

motor coordination event. An SMC event is, therefore, a

Fig. 4. Data ow.

motor state transition, that is either preceded or followed

by a consistent signals in more than one sensor.

IX. Resultant Data

200

Sensor Data

Figure 6 is a composite graph of the sensor data in the 0

3.5 seconds before visual servoing (VS) algorithm was com- 200

plete. The constant signals are the thumb and index nger 0 5 10 15

1000

Hand Pos

position and the ngertip touch sensors, since the hand is 0

open and not in contact with anything. The proximity sen- 1000

0 5 10 15

sors starting ring during the nal 2 seconds. Moreover, 1000

Finger Pos

dx, dy, and dz show that the hand motion is slowing down 500

as the hand approaches the target location. Therefore the 0

0 5 10 15

200

stability of position and ring of both proximity sensors

Force

0

are a strong indication that the hand has completed the 200

servoing behavior in and is in position to begin grasping. 0 5 10 15

200

Torque

Figure 7 depicts the signals during the grasp procedure. 0

Here, the signals indicate the positions of the closing n- 200

0 5 10 15

gers and show when the touch sensors re. Also, they time (s)

show the hand position stabilizing. The proximity sen- Fig. 5. Typical sensor and motor data.

sors continue to re o and on, essentially at random, due

apparently to noise.

300

X. Conclusions and Future Work

Touch 2

At the time of this writing, the rst experiments in SMC 200

data gathering during teleoperation had been performed. Touch 1

We found that the teleoperation procedure is repeatable in 100 dY

the way needed for the analysis: Having the same number dX

of motor events yet having su cient variability to detect 0

Sensor Data

dZ

true sensory events and to average out the spurious ones. Proximity 1

100

It remains to perform the analysis described in Sec. VIII. Proximity 2

Palm 100mm from Object

200

Acknowledgments Palm 50mm from Object

Thumb Position

This work has been partially funded by the DARPA Index Finger Position

300

Mobile Autonomous Robotics Systems (MARS) Program.

For more information, please visit our web page at 400

0 0.5 1 1.5 2 2.5 3

time (sec)

http://shogun.vuse.vanderbilt.edu

Fig. 6. Composite sensor data prior to end of VS.

References

[1] Kawamura, K., R. A. Peters II, D. M. Wilkes, W. A. Alford, and

T. E. Rogers, Towards Interactive Human-Humanoid Teaming:

Foundations and Achievements, IEEE Intelligent Systems, (to 400

appear), 2000.

Fingers Start to Close

[2] Kawamura, K., R. A. Peters II, S. Bagchi, M. Iskarous, and M. 300

Bishay, Intelligent robotic systems in service of the disabled,

IEEE Transactions on Rehabilitation Engineering, vol. 1, no. 3, Touch 2

pp. 14-21, March 1995 and Erratum, vol. 1, no. 11, November Proximity 2

200

1995. Touch 1

[3] Pack, R. T., IMA: The Intelligent Machine Architecture Ph. D. 100

Sensor Data

Dissertation, Vanderbilt University, May 1999. Proximity 1

[4] Peters, R. A. II, D. M. Wilkes, D. Gaines, and K. Kawamura, dZ

dX

A Software Agent Based Control System for Human-Robot 0

Interaction, Proc. 2nd Int l. Symp. on Humanoid Robotics dY

HURO 99, Tokyo, Japan, October 8-9, 1999. 100 Thumb Position

[5] Hamburger, V, The developmental history of the motor neu-

ron, Neurosciences Research Program Bulletin, no. 15, pp. 1-

37, April, 1977. Index Finger Position

200

[6] Bushnell, E. W. and J. P. Boudreau, Motor development and Fingers Stop Closing

the mind: The potential role of motor abilities as a determi- 300

nant of aspects of perceptual development, Child Development 0 0.5 1 1.5 2 2.5 3

Time (s)

vol. 64, pp. 1005-1021, 1993.

[7] Thelen, E. AND L. B. Smith, A Dynamic Systems Approach

Fig. 7. Composite sensor data prior to end of grasp.)

to the Development of Cognition and Action, MIT Press, Cam-

bridge, MA, 1994.

[8] Arbib M. A., P. Erdi, and J. Szent gothai, Neural Organiza-

a [15] Carter, R. Mapping the Mind University of California Press,

tion: Structure, Function, and Dynamics, MIT Press (Brad- Berkeley, CA, 1998.

ford), Cambridge, MA, 1998. [16] Brooks, R. A., A Robot That Walks; Emergent Behaviors from

a Carefully Evolved Network, MIT AI Lab Memo 1091, MIT

[9] Brooks, R. A., A Robust Layered Control System for a Mo-

Arti cial Intelligence Laboratory, Cambridge, MA, February,

bile Robot, IEEE Journal of Robotics and Automation, vol. 2,

1989.

no. 1, pp. 14-23, March 1986; also MIT AI Memo 864, Septem-

[17] Mataric, M. J., Navigating with a rat brain: a neurobiologi-

ber 1985.

cally inspired model for robot spatial representation,, in From

[10] Arkin, R. C., Behavior-Based Robotics, MIT Press, Cambridge,

Animals to Animats I, J.-A. Meyer and S. Wilson, eds, MIT

MA, 1998.

Press, 1991.

[11] Scheier, C. and R. Pfeifer, Classi cation as sensory-motor coor-

[18] Arkin, R. C., Motor schema-based navigation for a mobile

dination, Proc. 3rd European Conf. on Arti cial Life, pp. 656- robot: An approach to programming by behavior, Proc. IEEE

667, 1995. Conf. on Robotics and Automation, Raleigh, NC, pp. 264-271,

[12] Pfeifer, R. and C. Scheier, Sensory-motor coordination: the 1987.

metaphor and beyond, Robotics and Autonomous Systems, [19] Lambrinos, D. and C. Scheier Extended Braitenberg Archi-

Special Issue on Practice and Future of Autonomous Agents, tectures, AILab Technical Report, No. 95-10, University of

vol. 20, no. 2-4, pp. 157-178, 1997 Zurich, 1995.

[13] Lambrinos, D. and C. Scheier, Building complete autonomous [20] Terzopoulos, D., X. Tu, R. Grzeszczuk, Arti cial shes: Au-

agents: a case study on categorization, Proceedings IROS 96, tonomous locomotion, perception, behavior, and learning in a

IEEE/RSJ Int l Conf. on Intelligent Robots and Systems, simulated physical world, Arti cial Life, vol. 1, no. 4, pp. 327-

November 4-8, 1996. 351, December, 1994.

[21] Christopher, J. L., Jr., A PneuHand for Human-Like Grasping

[14] Pfeifer, R. and C. Scheier Representation in natural and ar-

on a Humanoid Robot, M. S. Thesis, Vanderbilt University, May

ti cial agents: an embodied cognitive science perspective,

1999.

Zeitschrift fr Naturforschung, vol. 53c pp. 480-503, 1998

723.dvi



Contact this candidate