Seashell Effect Pretouch Sensing for Robotic Grasping

13 downloads 0 Views 2MB Size Report
The seashell effect pretouch sensing fingertip on the PR2 robot gripper. Two different designs are presented for different applications: (a) the sensor on the ...
Seashell Effect Pretouch Sensing for Robotic Grasping Liang-Ting Jiang and Joshua R. Smith

Abstract— This paper introduces seashell effect pretouch sensing, and demonstrates application of this new sensing modality to robot grasp control, and also to robot grasp planning. “Pretouch” refers to sensing modalities that are intermediate in range between tactile sensing and vision. The novel pretouch technique presented in this paper is effective on materials that prior pretouch techniques fail on. Seashell effect pretouch is inspired by the phenomenon of “hearing the sea” when a seashell is held to the ear, and in particular relies on the observation that the “sound of the sea” changes as the distance from the seashell to the head varies. To turn the familiar seashell effect into a sensor for robotic manipulation, a cavity and microphone was built into a robot finger. The sensor detects changes in the spectrum of ambient noise that occur when the finger approaches an object. Environmental noise is amplified most (attenuated least) at the cavity’s resonant frequency, which changes as the cavity approaches an object. After introducing the sensing modality and characterizing its performance, the paper describes experiments performed with prototype sensors integrated into a Willow Garage PR2 robot’s gripper. We explore two primary applications: (1) reactive grasp control and (2) pretouch-assisted grasp planning. For reactive grasp control, we demonstrate the advantage of pretouch over pressure sensing for sensing objects that are too compliant for the PR2’s tactile sensors to detect during the grasp execution. We demonstrate the benefit of seashell effect pretouch over prior pretouch methods by demonstrating that the new method succeeds with materials on which prior methods fail. For pretouch-assisted grasp planning, the pointcloud from the PR2’s texture-projection / narrow stereo camera is augmented with additional points collected by recording the position of the robot’s end effector when the pretouch sensor detects the object. This method can compensate for points that are missing from the pointcloud either because of depth camera failure or occlusion. For this application, pretouch is advantageous over contact sensing because contact sensing tends to unintentionally displace the object.

I. INTRODUCTION AND RELATED WORK Long range non-contact sensors such as RGB cameras, depth cameras, and laser scanners are widely used in robotics for object recognition and pose estimation. These long range sensors typically provide the data used for grasp planning prior to manipulation. Contact sensors such as tactile pressure sensors are commonly used during the process of manipulation to provide force feedback information to a grasp controller. “Pretouch” sensors are non-contact sensors with properties intermediate between these two: the range of pretouch sensors is shorter than vision but longer than contact-based tactile sensing. This paper introduces “seashell Liang-Ting Jiang is with the Department of Mechanical Engineering, University of Washington, Seattle, WA. [email protected] Joshua R. Smith is with the Departments of Computer Science and of Electrical Engineering, University of Washington, Seattle, WA.

[email protected]

Fig. 1. The seashell effect pretouch sensing fingertip on the PR2 robot gripper. Two different designs are presented for different applications: (a) the sensor on the finger surface for sensing extremely compliant objects, and (b) the sensor on the fingertip for adding pretouch pointcloud at the unknown portions of the object before grasp planning.

effect pretouch,” a novel form of pretouch that works on a wide variety of materials; the paper demonstrates the application of seashell effect pretouch to both grasp control and grasp planning. A. Robotic Pretouch Sensing The hypothesis that pretouch sensing can be useful for manipulation is being explored actively [1][2][3][4]. It is viewed as potentially beneficial for manipulation because it provides reliable geometric information in the last centimeter before contact. The disadvantage of using tactile sensing for collecting local geometric information (as in [5], [6]) is that contacting the object may displace it—an outcome that is particularly likely when the object’s geometry is initially uncertain. An advantage of pretouch over vision is that pretouch is not subject to the problem of the hand occluding the camera, because the sensor is integrated into the hand; another is that there are no camera-to-hand registration errors since the sensor is in the coordinate frame of the hand. Because of these challenges for vision sensing, grasps planned over point cloud data are typically executed open loop: new point cloud data is usually not collected during execution of a grasp. By contrast, like tactile sensor data, pretouch sensor values can easily be collected during grasp execution and used by a grasp controller. Finally, depth sensors typically provide incomplete point clouds, dropping many points because of failure modes such as: challenging optical properties (transparency, specularity); algorithmic failure to find stereo correspondences; and failure to find projected structure (visible or IR texture or grids). A further distinction between pretouch and depth sensing technologies is that the latter typically fail below some minimum distance. For example, here is a list of depth sensing technologies and their minimum specified range:

Microsoft Kinect, 120 cm [7]; PR2 textured narrow stereo camera, 60 cm [8]; Hokuyo UTM-30LX laser scanner, 10 cm [9]; Parallax PING ultrasound sensor, 2 cm [10]. Compared to tactile sensing, one can think of pretouch as a sensor that detects surface proximity, but has no lower limit on detectable force, and thus is able to sense arbitrarily compliant objects. Electric field pretouch has many desirable properties, but only works with materials that are conductive or have high dielectric constant. Optical pretouch depends on surface albedo, and thus fails in some of the same cases as the long range vision sensors: transparent or specular objects. Thus optical pretouch may fail to complement the 3D visual sensors used to plan grasps: since they rely on similar physical phenomena, they are likely to fail in a correlated fashion. Seashell effect pretouch has the desirable properties of other pretouch mechanisms, but does not depend on electrical or optical material properties. Thus it works on materials that are difficult for electric-field pretouch, optical pretouch, and conventional vision / depth sensing.

improving the grasps that can be planned for un-modeled objects. Petrovskaya [5][6] demonstrated the use of tactile sensors for global object localization, and proposed an efficient Monte Carlo algorithm for realtime applications. Since Petrovskaya relies on tactile sensing, the approach is subject to difficulties such as unintentional displacement of the object by the robot manipulator. Another feature distinguishing our approach is our integration of pretouch sensing with optical depth sensing. This paper demonstrates the use of seashell effect pretouch for both grasp control and grasp planning.

B. Seashell Effect

The radiation impedance of an open-ended pipe has a small but finite value. Its imaginary part acts as an end correction to the geometrical pipe length from which the resonance frequency of the pipe is calculated. Dalmont [16] found an empircal formula for the end correction in this kind of pipe-surface configuration by fitting a function to values produced by a finite element model. The end correction δobj caused by the approaching object is given by:

There is a well-known folk myth that if one holds a seashell to the ear, one can hear the sound of the ocean. The rushing sound that one hears is in fact the noise of the surrounding environment, resonating within the cavity of the shell. The same effect can be produced with any resonant cavity, such as an empty cup or even by simply cupping one’s hand over one’s ear. The resonator is simply amplifying (attenuating) the ambient environmental noise, in a frequency dependent fashion that is determined by the resonant modes of the specific cavity [11]. It is easily verified that the perceived sound depends on the position of the seashell with respect to the head. Inspired by this seashell effect, we propose to measure the distance to nearby objects by detecting changes in the ambient noise spectrum inside an acoustic cavity. A cavity with a microphone is integrated into a robot finger; as the finger approaches an object, the spectrum of the noise collected by the microphone changes.

II. SEASHELL EFFECT PRETOUCH SENSOR The seashell effect pretouch sensor is essentially an openended pipe attached on a microphone. The detailed sensor design and characterization will be described in this section. A. Acoustic Theory

δobj =

3.5(h/a)0.8 (h/a

where a, w are the inner radius and wall thickness of the pipe; h is the distance between the object and pipe opening, and d is the radius of the object. The end correction δpipe caused by the opening geometry of the pipe is given by: δpipe = 0.6133a

C. Robot Grasping Robotic grasping in unstructured human environments is a highly challenging problem for robotics. Besides the inaccuracy of the end-effector motion control, a key difficulty is the lack of reliable perception, caused by sensor calibration errors, or simply the limitation of sensors. One approach to accommodate it is to improve the grasp selection and execution algorithm for unknown objects which is able to deal with uncertainty [12][13][14]. Another approach is to improve the quality of the depth sensor. Depth sensing camera systems, including textured stereo cameras and structured infrared depth cameras, have dramatically improved recently. However, these sensors still frequently fail to provide points, either because of geometric difficulties (such as occlusion), or material difficulties (transparency or specularity). In this paper, we demonstrate the use of pretouch to augment the point clouds provided by depth cameras, with the aim of

a + 3w/a)−0.4 + 30(h/d)2.6

1 + 0.044(ka)2 1 + 0.19(ka)2

for ka < 3.5, where k is wavenumber, which is determined by the frequency of acoustic wave. The fundamental frequency of a one-sides open tube, is approximately given by: f0 =

1 c 4 (L + δobj + δpipe )

where L is the original length of the pipe, and c is the sound of the speed. The formulas above show that the approaching object will increase the end correction, and thus increase the effective acoustical length of the pipe, which results in a decrease in the pipe’s resonance frequency. In order to know the accuracy of the empirical formulas and understand the effect of pipe geometry, the computed resonance frequencies of different geometry are compared with our experimental data and shown in Figure 2 (L is the length of the pipe).

Microphone I (Signal Channel)

Sound Signal in the cavity

Microphone II (Reference Channel)

Ambient Sound Signal

Filtered Resonance Frequency

Fig. 4.

Fig. 2. Comparison of the theoretical and experimental fundamental mode resonance frequencies. The figure also shows the effect of cavity geometry on resonance frequencies. Note that the sensors in this figure have not been optimized for performance: they do not use the reference microphone of our final system. This figure is meant to illustrate the effect of cavity geometry.

11000

Resonance Frequency (Hz)

10500

10000

9500

Pretouch Thresold

9000

8500

Kalman Filter [17]

Mic Preamp

Estimated Resonance Frequency

A/D Converter

Spectral Peak Estimation [16]

Welch Spectrum Estimation

+ -

Subtracted Spectrum

The system architecture of the seashell effect pretouch sensor.

= 1024; overlap ratio = 70%; Hanning data taper). The spectrum of the reference channel is subracted from the spectrum of the sensor signal before finding the peak power, thus the effect of the noise in the environment is rejected. The peak-finding algorithm is finding the frequency having maximum power. Due to the sampled nature of spectra obtained using the Fast Fourier Transform, the peak (location and height) found by finding the maximum-magnitude frequency bin is only accurate to within half a bin. A bin represents a frequency interval of Fs /Ns Hz (in this case is 43Hz). Therefore, an accurate and computationally simple approximation method [18] is used to estimate the actual spectral peak frequency. The frequency spectral peak is then filtered by a Kalman filter [19] with process variance 10−5 and measurement variance 10−4 . The paramters are experimented and determined by the balance of fast response time and the measurement stability. The distance measurement can be carried out by mapping the filtered frequency to the precalibrated frequency-distance function.

8000

C. Integration with the PR2 Robot

Contrast to Noise Ratio (CNR) = 22.04 7500

1

2

3

4

5 6 Distance (mm)

7

8

9

10

Fig. 3. The box and whisker plot of 1000 estimated resonance frequencies at each distance. It represents the sensor characteristics of the sensor with length L=5 mm and radius a=2.5 mm integrated on the robot fingertip. The blue lines at the bottom and top of each box are the 25th and 75th percentile (the lower and upper quartiles, respectively), and the red band near the middle of the box is the 50th percentile (the median). The lower and upper black horizonal lines are the lowest and highest samples still within 1.5 × (upperquartile − lowerquartile) of the lower quartile and the upper quartile, respectively. The green crosses are outlier samples.

B. Sensor System Design To make usable sensors based on the principle described above, our sensor system is comprised of both hardware and software components. Figure 4 shows the schematic of the system. The cavity amplifies the ambient noise in a certain frequency response. The sound in the cavity is collected by an omnnidirectional electret condenser microphone cartridge (WM-61A, Panasonic) with sampling rate (Fs ) of 44,100 Hz. The signal is amplified by 50dB and converted to digital through an audio interface. An additional microphone is used to collect the sound outside the cavity as the reference signal in order to capture only the frequency response of cavity. The power spectrum of the signal is estimated every 0.05 second (N = 2205) using Welch spectrum estimation (Ns

We use Willow Garage PR2 robot as our experiment platform. The current implementation is detached from the exsisting PR2 finger electrical interface. Two different configurations of the seashell effect pretouch sensors are integrated on the PR2 gripper shown in Figure 1. The design in Figure 1(b) is for the pretouch-assisted grasp planning application. The sensor is located on the fingertip for adding pointcloud at the location of fingertip. The design in Figure 1(a) is for the application of grasping. The sensor is located on the surface of the finger, enabling sensing extremely compliant objects during the grasp execution. The cavity used in our system is selected to a 2.5 mm radius / 5 mm length cylindrical pipe attached with the 3mm thick microphone, which has compact size to be embedded on the PR2 gripper fingertip. D. Sensor System characterization The performance of the sensor on the fingertip is evaluated by collecting 1000 sensor readings (filtered spectral peak frequency) at various distance from 1mm to 10mm. A boxand-whisker plot is plotted to present the performance of the sensor (Figure 3). A contrast to noise ratio (CNR) for evaluating the sensor performance is defined as CN R =

mean(f10 ) − mean(f1 ) , i = 1..10 mean(std(fi ))

where i is the distance in millimeter. The box plot shows the filtered resonance frequency starts to decrease from 6 mm. Based on that, we select the threshold to 9500Hz (the lower quartile at 3 mm) as the first distance being able to be measured with confidence. In this way, the upper quartile at 3 mm is smaller than the lower quartile at 6 mm, so the sensor is not getting confused. All the software in this work is implmented in ROS (Robot Operation System). The signal processing and frequency estimation is implemented in Python as a ROS node, which continuously publishes the estimated resonance frequency, total signal energy, and distance in a ROS message at the rate of 20 Hz. E. Material sensitivity Our sensor does not depend on optical or electrical material properties. Instead, it depends on mechanical or acoustic properties. Thus it is likely to complement long range optical depth sensors. For example, it can sense transparent and extremly light-reflective materials which are difficult for optical sensors. The only materials we have found so far on which the sensor fails are open foams (such as very thin bubble wrap), certain rough fabrics, and fur. III. APPLICATION I: REACTIVE GRASPING OF COMPLIANT OBJECTS Tactile sensors have been widely researched for local perception of objects by monitoring the contact force [12][15][5]. One of the cases in which this approach might fail is when the object is too compliant, so the pressure sensor cannot detect it. The first application of our seashell pretouch sensor is exploring the feasibility of sensing these extremely compliant objects during grasp execution. A pregrasp execution experiment was performed, in which the pressure sensor built in to the PR2’s gripper was compared to our seashell effect pretouch sensor for the ability to detect the objects. The two sensors were applied in a pre-grasp task to four different objects with different compliance. The objects used were a cookie box, a disposable cup, a folded paper box, and a folded aluminum foil box. In the force sensing trial, the PR2’s pressure sensor fingertips are installed on both right and left fingers on the PR2 right gripper. The standard PR2 gripper sensor controller is used to detect contact with objects. In order to increase the sensitivity of the pressure sensors, we lowered the contact detection threshold from the default values (0.1 N for the high-pass filtered pressure readings, and 1.2 N for the unfiltered pressure) to the lowest values (0.05 N for the high-passed pressure readings, and 0.4 N for the unfiltered pressure) such that false positive detection is avoided. In the pretouch trial, the seashell effect pretouch sensor is installed on one of the fingers on the PR2 right gripper. The gripper is commanded to close until the fingertip is sufficiently close to the object according to the seashell sensor. A trial is defined as successful if the gripper stops before sqeezing so much that it breaks the object. The pressure

Fig. 5. (a) The pressure sensor doesn’t sense the aluminum foil box and thus squeezes the object. (b) The pretouch sensor on the right finger senses the aluminum foil box and stops before touching the object. (c) The pressure sensor doesn’t sense the paper box and thus squeezes the object. (d) The pretouch sensor on the right finger senses the paper box and stops before touching the object. TABLE I R ESULTS OF THE SENSING FOR COMPLIANT OBJECTS EXPERIMENTS .

Objects

(∨: success; ×: failed) Pressure

cookie box disposable cup folded paper box folded aluminum foil box

∨ ∨ × ×

Seashell pretouch ∨ ∨ ∨ ∨

sensor is able to detect the contact for cookie box and disposable cup, but it is not able to sense the contact with the extremely compliant folded paper and aluminum foil boxes, thus breaks them. On the other hand, the seashell effect pretouch sensor is able to sense all four objects and stop the gripper at an approppriate distance from the object, which shapes the pre-grasp pose (Figure 5 and Table I). An interactive mode in which the gripper dynamically adjusts its opening based on the seashell effect pretouch sensor has also been implemented. This experiment shows the pretouch sensor can be a good complement to tactile sensing when grasping compliant objects. IV. APPLICATION II: PRETOUCH-ASSISTED GRASP PLANNING 3D perception is becoming more and more important in robotics, as well as in other fields. Pointcloud data structures are widely used to represent the sensed spatial and color information. A large community is developing a dedicated open source library, PCL (Point Cloud Library) [17], providing algorithms to process pointclouds efficiently. In robotic grasping, an incomplete pointcloud can result in poor grasp

to probe at the location of interest. Both components are implemented as ROS nodes. The initial pointcloud is generated by the depth data obtained from a single frame of the narrow-field-of-view stereo camera on the PR2. This pointcloud is used to segment the objects from the table, and the pointcloud cluster of the segmented object is then saved on the pointcloud server. If pretouch pointcloud is desired, the pretouch motion planner will use the pointcloud to decide the initial probe location and then probe around the area. The completed workflow of the object grasping with pretouch pointcloud is shown in Figure 6, and each key step will be described in detail below. A. Pretouch Motion Planning

Fig. 6. Pretouch pointcloud grasping workflow. Two new components, pretouch motion planner and pointcloud server, are added to the PR2 grasping framework. The robot can choose to use the pretouch sensor or not to use it. The concatenated pointcloud will be used for grasp planning, followed by the collision map processing and grasp execution.

plans, simply because the planner is operating with incorrect information about the object geometry. Although depth sensing hardware has dramatically improved recently, it is still prone to dropping a many points, and there are still cases in which key object shape information cannot be collected by the depth camera because of geometric constraints; for example, in highly constrained environments, the robot may not be able to move its head enough to collect additional views. We propose to use pretouch sensing in conjunction with the PR2s existing depth sensors, to provide data on portions of the object where information is missing, either because of occlusion or depth sensor failure. Given a partial pointcloud collected by a depth sensor, our pretouch sensor (combined with robot kinematic data) will add additional points to the pointcloud, prior to grasp planning. In this section, we will describe the methdology and how additional pointcloud could be beneficial to robotic grasping. Some practical examples will also be discussed. Currently, our grasping framework is built on top of the PR2 object manipulation framework proposed in [12]. We add two novel components to the framework: (1) a pointcloud server to store both the pointcloud generated from the depth sensor (stereo cameras in this case) and the pretouch pointcloud we are generating later; and (2) a pretouch motion planner to command the robot end-effector

The pretouch motion planner is a component that decides where the robot should move its end-effector to and look for the object surface. The current pretouch motion planning algorithm assumes the object is convex and fits an ellipic cylinder from the segmented object pointcloud, and then the initial probe pose is determined by the farthest point the endeffector can reach in the unknown area of the object allowed by the PR2 arm’s workspace. The algorithm works for most of objects with round and rectangular shapes. The fingertip always approaches in the direction normal to the periphery of the fitted ellipse with a constant velocity. When the pretouch sensor senses the distance closer than the threshold (when the estimated resonance frequency lower than the threshold 9500 Hz), W points are added on the pointcloud server at the location translated by 3 mm in the X-axis from the fingertip frame. The offset 3mm is corresponding to the threshold 9500 Hz determined by the sensor characterization in Figure 3. W is the weight of the point generated by the pretouch sensor. It determines how much the grasp planner will take these new points into account when doing grasp planning. Adding a single point at each location will not alter the grasp plannng results much because the pretouch pointcloud is sparse. The end-effector will continue lowering and probing with a fixed interval in the Z (vertical) axis before the gripper hitting the table. Once the height is too low, the gripper moves back to the initial height and also moves toward to the robot body. The same process continues until the newly detected points are close enough to the existing pointcloud generated from the camera. In our particular implementation, PCL [17] is used for the processing of pointclouds, and a Jacobian inverse controller is used as a low level controller to control the end-effector pose and velocity in the cartesian space. Figure 7 shows an example of the PR2 probing a Coke bottle. The incomplete pointcloud due to the transparency and reflection is filled by the pretouch sensor. B. Grasp Planning and Execution The grasp planner used in this paper relies solely on the observed pointcloud of an object versus fitting object models. The observed pointcloud can be either with the pretouch points added or not. The algorithms presented in

Fig. 7. (a) PR2 robot probing the unseen area of the bottle and adding pretouch pointcloud before grasp planning. (b) The 3-D visualization in rviz. The red points are generated from the narrow stereo camera , and the yellow points are generated by the pretouch sensor during the pretouch motion. (c) The view from the left-narrow stereo camera overlayed with the visualization markers.

detail in [12]. This planner uses a set of heuristics, such as hand alignment with the cluster bounding box, size of the bounding box compared to gripper aperture, and number of observed points inside the gripper, to assess the quality of grasp, and also to populate a list of grasp candidates. When the pointcloud generated by the pretouch sensor is provided, it is concatenated with the pointcloud from the cameras before being fed to the grasp planner. If pretouch pointcloud is not provided, only the pre-stored pointcloud from camera will be used to do the planning. The grasp candidate with the highest quality is used to execute the object grasping. After collision map processing, the robot will try to grasp the object in the planned pose. C. Experiment and Results In order to understand how the additional pointcloud added by the pretouch sensor can affect the planned grasps, object grasping experiments are performed on 4 objects both with and without using the pretouch-assisted grasp planning approach. Figure 8 shows the 3D visualization in rviz, the image seen from the narrow stereo camera on the PR2 sensor head overlayed with the visualization markers, and the planned grasp pose in the real world.

(a) Coke bottle: Without pretouch sensor, the narrow-view stereo camera only captures the pointcloud of the wrapper due to the transparency and the reflection of the light. Also the right and back sides of the red wrapper is missing because of self-occlusion. Since the grasp is planned only using the pointcloud from the vision (red), the robot do not know the complete shape of the bottle, thus the planned grasp is off-centered. This particular grasp failed becuase the robot tries to grasp the bottle from the inclined portion of the bottle. After adding pretouch pointcloud (yellow points), the fitted bounding box now matches the actual shape of the bottle better. The final grasp planned using the concatenated pointcloud is centered and thus successful. This example shows the pretouch pointcloud is useful for grasp planning when the object is self-occluded due to the view of camera. (b) Lego blocks: The robot intends to grasp the yellow block on the table. In this configuration, ideally the a frontsided grasp should be more safe and reliable. However, there is another block in front of the target, which prevents any front grasp. When the robot is forced to do a side or overhead grasp, the lack of information on the backside of the object may result in the wrong decision. such as in this example, the robot decides to grasp from the top and it locates the gripper without knowing the bumps in the back, thus hits the bump during the grasp execution. With the help of the pretouch pointcloud, the robot fingers are located on the side walls in the overhead grasp, which avoids dealing with the bumpy backside. In this example, the depth sensor works properly, but due to the collision constraints, the additional information provided by the pretouch sensor plays an important roll in grasp planning. (c) milk box: Due to the particular inclined shape and the view of the camera, the entire back side of the milk box is occluded. Similar to the case of lego block, there are chances that the planned grasp will collide with the backside of the object without knowing there are things on the back. The added pretouch pointcloud again helps plan a centered grasping pose in this case. (d) Snapple bottle: The transparency and reflection of the plastic material cause the missing information on most of the portions except the wrapper on the bottle. Also the opening of the bottle is outside of the camera’s view, so the initially planned overhead grasp is off-centered. The pretouch sensor adds additional useful information and shift the grasping pose to the center, which makes the grasp sucessful. This example demonstrates the pretouch pointcloud is useful when the object is self-occluded and partially out of the view. We demonstrate general applicability of the approach on some critical cases that the pretouch pointcloud becomes useful to correct the grasp planning by reducing the uncertainty in object’s geometry in the cases of occlusion, depth sensor failure, or when in constrained environment. V. CONCLUSIONS AND FUTURE WORK A. Conclusions This paper demonstrates a novel acoustic pretouch modality for robotic grasping inspired by the well-known seashell

Vision Only

Vision + Pretouch

Vision Only

(a)

(b)

(c)

(d)

Vision + Pretouch

3-D Visualization

Camera View

Real World Grasp

3-D Visualization

Camera View

Real World Grasp

Pointcloud from camera

Pointcloud from pretouch sensor

Fig. 8. Examples of pretouch-assisted grasping. Compare the planned grasp with and without the additional pointcloud generated using the pretouch sensor on 4 objects: (a) Coke bottle, (b) Lego blocks, (c) milk box, (d) Snapple bottle. For each object, the first column shows the results with vision only, and the second column shows results with the additional pointcloud from pretouch sensor. The red pointcloud are from stereo camera, the yellow pointcloud are from pretouch sensor, the green block is the fitted bounding box given the pointcloud, and the gripper model shows the final planned grasp to be excuted. In (a) and (d), part of the incomplete pointcloud in the transparent portions is filled by the pretouch sensor. In (b) and (d), part of the incomplete pointcloud in the backside is filled by pretouch sensor.

effect. As far as we know, this effect has not previously been used to build proximity sensors. Prototypes based on this priciple were built and installed on PR2 gripper. We demonstrate that this technique is useful for sensing extremely compliant objects, which can not be detected by the standard pressure sensors on the PR2 fingertip. We also demonstrate that using pretouch to add points to a pointcloud collected by a depth camera can enable improved grasp plans. Seashell effect pretouch can be used to collect spatial information in regions where optical depth sensing has failed. Unlike existing pretouch modalities, seashell effect pretouch complements optical depth sensing because they rely on fundamentally different physical mechanisms: mechanical/acoustic in the case of seashell effect pretouch, and electromagnetic/optical in the case of depth cameras, electric field pretouch, and optical pretouch. Videos of both pretouch-assisted grasp planning and reactive grasping applications can be seen at: http://sensor.cs.washington.edu/pretouch/icra12/

B. Future Work It is natural to wonder whether seashell effect pretouch could be improved by actively generating sound. It is likely that the answer is yes, but the passive scheme reported here works well enough for the applications we describe, and avoiding active audio generation allows for apparatus that is simpler, smaller, and therefore more easily integrated into robot fingers. Also, our current implementation operates within the band of human hearing, so an active scheme might be annoying to humans. The integration of the seashell effect pretouch fingertip to the existing SPI interface on the PR2 gripper is in progress. Our next generation system will have seashell effect sensors on both fingers without external wires, and have embedded signal processing capability. (In our initial implementation, an external cable to the sensor is used, and only one finger has the pretouch sensor.) Once the pretouch sensors are wellintegrated into the robot, we plan to make them available to other PR2 users. With the sensors in both fingers, we will be able to apply seashell effect pretouch to actually grasp extremely compliant objects (taking the next step beyond the preshaping demonstrated here). This will require moving the entire arm to align the two fingers with the object. We also plan to experiment with multiple sensors on the same finger, to acquire surface orientation information in addition to the distance information we currently have. This will facilitate a new mode of pretouch pointcloud collection, in which the hand moves continuously, servoing to the pretouch data. This is expected to be more efficient than the current discrete pointcloud sampling scheme. Also, a continuously executed pretouch servoing scheme can be readily integrated with grasp control / execution: the grasp controller would try to keep the fingers near the object during grasp execution, making pretouch measurements during grasp execution. This could allow both collection of new

pointcloud information during tentative grasping motions, and could also be used to falsify or invalidate the current pointcloud during the normal reach motions in the case of object motion or depth camera error. For example, if the pretouch sensor detects unexpected obstacles during grasp execution, the grasp could either be aborted or modified. In the first case, the existing pointcloud would be discarded, a new pointcloud would be captured with the depth camera and pretouch, and a new grasp would be planned. In the second case, the points newly detected by pretouch would be added to the existing pointcloud, and realtime grasp re-planning would occur. A further extension of this work would be to also make use of the 2-D image from the PR2’s stereo camera and forearm camera. Even in regions where depth sensing has failed, the 2-D images might provide valuable suggestions for useful trajectories for pretouch exploration. Finally, pretouch-assisted depth perception could also be used to obtain an object’s complete 3-D model, or used for object recognition before grasping. R EFERENCES [1] B. Mayton, L. LeGrand, J.R. Smith, An Electric field pretouch system for grasping and co-manipulation, in ICRA, 2010. [2] R. Wistort, J.R. Smith, Electric field servoing for robotic manipulation, in IROS, 2008. [3] K. Hsiao, P. Nangeroni, M. Huber, A. Saxena, A. Ng, Reactive Grasping Using Optical Proximity Sensors, in ICRA, 2009. [4] R. Platt, A.H. Fagg, R. Grupen, Null space grasp control: Theory and experiments, in IEEE Transactions on Robotics, vol. 26(2), pp. 282295, 2010. [5] A. Petrovskaya, O. Khatib, S. Thrun, A.Y. Ng, Bayesian estimation for autonomous object manipulation based on tactile sensors, in ICRA, 2006. [6] A. Petrovskaya, O. Khatib, Global Localization of Objects via Touch, in IEEE Transactions on Robotics, vol. 27(3), pp. 569-585, 2011. [7] http://www.msxbox-world.com/xbox360/kinect/ faqs/305/kinect-technical-specifications.html [8] This figure is from our own experimental evaluation of the PR2’s narrow stereo camera. [9] http://www.hokuyo-aut.jp/02sensor/07scanner/ utm_30lx.html [10] http://www.parallax.com/tabid/768/ProductID/92/ Default.aspx [11] D.H. Ashmead, R.S. Wall, Auditory perception of walls via spectral variations in the ambient sound field, in J. Vis. Impair. Blindn., vol. 9, pp. 615, 1998. [12] K. Hsiao, S. Chitta, M. Ciocarlie, E.G. Jones, Contact-Reactive Grasping of Objects with Partial Shape Information, in ICRA, 2010. [13] K. Hsiao , M. Ciocarlie, P. Brook, Bayesian Grasp Planning, in ICRA, 2011. [14] P. Brook, M Ciocarlie, K. Hsiao, Collaborative Grasp Planning with Multiple Object Representations, in ICRA, 2011. [15] J.M. Romano, K. Hsiao, G. Niemeyer, S. Chitta, K.J. Kuchenbecker, Human-Inspired Robotic Grasp Control with Tactile Sensing, in IEEE Transaction on Robotics, 2011. [16] J.P. Dalmont, C.J. Nederveen, N. Joly, Radiation impedance of tubes with different flanges: numerical and experimental investigations, in J. Sound and Vibration, vol. 244(3), pp. 505-534, 2001. [17] R.B. Rusu, S. Cousins, 3D is here: Point Cloud Library (PCL), in ICRA, 2011. [18] E. Jacobsen, P. Kootsookos, Fast, Accurate Frequency Estimators, in IEEE Signal Processing Magazine, vol. 24(3), pp. 123-125, 2007. [19] R.E. Kalman, A New Approach to Linear Filtering and Prediction Problems, in Journal of Basic Engineering, vol. 82(D), pp. 35-45, 1960.

Suggest Documents