Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation

Page created by Dennis Herrera
 
CONTINUE READING
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous
                                                                 Robotic Manipulation
                                                                               Branden Romero1 , Filipe Veiga1 , and Edward Adelson1
                                                     1 Massachusetts     Institute of Technology @mit.edu, adelson@csail.mit.edu

                                            Abstract— High resolution tactile sensors are often bulky
                                         and have shape profiles that make them awkward for use
arXiv:2005.09068v1 [cs.RO] 18 May 2020

                                         in manipulation. This becomes important when using such
                                         sensors as fingertips for dexterous multi-fingered hands, where
                                         boxy or planar fingertips limit the available set of smooth
                                         manipulation strategies. High resolution optical based sensors
                                         such as GelSight have until now been constrained to relatively
                                         flat geometries due to constraints on illumination geometry.
                                         Here, we show how to construct a rounded fingertip that utilizes
                                         a form of light piping for directional illumination. Our sensors
                                         can replace the standard rounded fingertips of the Allegro hand.
                                         They can capture high resolution maps of the contact surfaces,
                                         and can be used to support various dexterous manipulation
                                         tasks.

                                                                I. INTRODUCTION
                                            The sense of touch has been shown to greatly contribute
                                         to the dexterity of human manipulation, especially in cases
                                         where high precision is required [1]. The complex ensemble
                                         of mechanoreceptors in the human hand provides extremely
                                         rich tactile sensory signals [2]. These sensory signals encode
                                                                                                                       Fig. 1. An Allegro Hand equipped with four sensors holding a mesh cup.
                                         information such as contact force and contact shape, and can                  Each finger provides a high resolution 3D reconstruction of its respective
                                         be used to detect complex state transitions such as making                    contact areas.
                                         or breaking contact or the occurrence of slippage between
                                                                                                                          In this paper we present a GelSight fingertip sensor, where
                                         the finger and the grasped object [3].
                                                                                                                       we preserve the softness and high signal resolution signals
                                            In recent years, vision based tactile sensors have become
                                                                                                                       associated with the classic GelSight sensor [5], while dra-
                                         very prominent due to their high signal resolutions and
                                                                                                                       matically changing the sensors shape to better suit the needs
                                         the softness of their sensing surfaces [4], [5], [6]. The
                                                                                                                       for dexterous manipulation task. We begin by providing
                                         softness of the sensing surface allows for larger contact
                                                                                                                       a discussion on the advantages of a round finger tip as
                                         regions as it deforms to comply with the object surface. The
                                                                                                                       opposed to a flat one (Sec. II-A), describing the design
                                         resulting contact areas are then characterized in great detail
                                                                                                                       choices made to achieve the target shape (Sec. II-B) and
                                         via the high resolution signals. Together, these properties
                                                                                                                       showcasing the manufacturing procedures (Sec. II-C). We
                                         have enabled the use of these sensors in tackling several
                                                                                                                       follow with a description of the sensor calibration, used for
                                         tasks such as assessing grasp success [7], servoing object
                                                                                                                       recovering 3D surface deformations from the GelSight sensor
                                         surfaces [8], detecting slip and shear force [9], reconstructing
                                                                                                                       images (Sec. II-E) and by a description of the method we
                                         3D surfaces [10] and distinguishing between different cloth
                                                                                                                       use to estimate and track the contact areas (Sec. II-F.3).
                                         materials [11]. The high resolution signals provided by the
                                                                                                                       Finally, we present several experiments where we use the
                                         sensors does come at a cost with sensor design being con-
                                                                                                                       sensor signals and take advantage of the sensor’s shape to
                                         strained in terms shape [5], [6]. Sensors from the TacTip [12]
                                                                                                                       manipulate objects (Sec. III) and discuss the outcome of our
                                         family have been developed in a wide range of geometries.
                                                                                                                       work while also providing some future directions (Sec. IV).
                                         However, the very high resolution sensors based on GelSight
                                         have been constrained to flat or nearly flat designs, due to the
                                                                                                                                 II. DESIGN AND MANUFACTURING
                                         difficulties of providing well controlled directional lighting
                                         with non-planar geometries. Our goal here is to remove                        A. Importance of Shape in Fingertip Sensors
                                         this design constraint and to allow the creation of GelSight                    Consider the task of flipping an object that is resting on
                                         fingertips that are rounded.                                                  a table, as depicted in Fig. 2. Here the index finger of a
                                            *This work was supported by the Toyota Research Institute and the Office   hand rolls an object that is lying on a supporting surface
                                         of Naval Research                                                             towards the thumb. Of the two cases depicted, we can see
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
Fig. 3. Human demonstration illustrating contacts made on different parts
                                                                                 of a cylinder using the finger’s transverse plane. (a) the contact areas made
                                                                                 with a curved finger,(b) the contact areas made with a flat finger. The round
                                                                                 finger keeps a consistent contact area at each contact point unlike the flat
                                                                                 finger.

                                                                                 the limited kinematic structure of each finger. Since curved
                                                                                 fingertips are able to perceive contact patches in a wider
                                                                                 ranged of orientations, explicitly reorienting the fingertip
                                                                                 becomes unnecessary. An example of a grasp configuration
                                                                                 where the differences between the two sensors are visible is
Fig. 2. Human demonstration illustrating rolling a cube along the frontal
plane of the finger. (a) the rolling motion of a curved finger which maintains
                                                                                 also depicted in Fig. 3.
a constant contact area size and has a wider range of motion. (b) shows the
rolling motion with a flat finger which has a shrinking contact area and         B. Design
limited range of motion.
                                                                                    The goal of our design (Fig. 4) is to enable robots
                                                                                 with dexterous manipulators to have rich information about
that when using flat sensors, the contact patch location and                     the contact, in particular dense information about the 3D
size greatly vary throughout the manipulation trajectory, with                   geometry of the contact areas. We propose a novel illumi-
the size of the contact patch being reduced to almost a point                    nation system, that despite more complex geometry, allows
contact when it reaches the edge of the sensor. Having such                      us to perform 3D height map reconstruction despite some
a small contact patch not only reduces the stability of the                      limitations.
object, but also reduces the robots perception of the objects                       To achieve this design we make sacrifices in terms of
state. On the other hand, when a curved sensor is used, while                    illumination quality as seen in the previous sensor [5]
the location of contact still changes in a similar manner, the                   hoping that the reconstructed height map is still suitable for
contact patch size remains relatively consistent throughout                      robotic manipulation. In particular we use an opaque sensing
the object trajectory. This decrease in variation of the contact                 surface that does not have a one-to-one mapping between
patch size makes it much easier to track the object state.                       surface normals and RGB values, and secondly while the
   Another case where fingertip shape clearly impacts the                        previous sensor provides information about the x-gradient
performance of dexterous hand systems is when performing                         and the y-gradient from at least two sources of information
grasp quality assessment. For assessing the quality of a grasp                   via direction lighting, we only provide information about the
it is critical that the contact areas acquired after grasping                    x-gradient with two sources of information and the y-gradient
the object are perceivable by the fingertip sensors. When                        with one source of information.
using flat sensors, in order to maximize the contact patch                          These choices were made in order to provide as uniform
information, the fingers have to be reoriented such that the                     of an illumination pattern as possible along the entire surface
sensing surface is orthogonal to the contact location. As                        of the finger. To do this we relied on a technique called light
previously stated, this can be problematic when considering                      piping. Light piping is inspired by fiber optics in which a
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
Back LED Board

                     35.6 mm

                                        Cover

                               160◦ FOV Camera
                30.5 mm
                          Camera Holder
 28 mm
           Epoxy Resin Shell

    Silicone
                                                               Binder

                                                          Bottom LED Board
                                                                                  Fig. 5. Illustration showing the light piping illumination system along a
                                                                                  single axis. (a) the path in which the light travels without contact. (b) the
                                                                                  path of the light when contact is made.
                                                         Allegro Mounting Plate
                  Semi-Specular Paint
                                                                                  mold. We chose Smooth-On MoldMax XLS II as our mold
                                                                                  material because we will cast epoxy resin into these molds
Fig. 4. CAD model of the sensor illustrating the assembled finger, along
with the exploded view of the sensor showing the internal components.
                                                                                  later, and we found this material robust to multiple castings
                                                                                  compared to other mold materials we used.
light is constrained within a medium via total internal reflec-                      The second mold will be used to cast silicone onto the
tion (TIR). We achieve something similar by using a semi-                         plastic shell, so we create a two piece mold where the base
specular sensing surface and a thin plastic shell. We design                      of the mold is a rigid 3D printed piece that will be rigidly
the sensor in this manner, because semi-specular surfaces                         attached to the shell, and the other piece is a soft silicone
only slightly diffuses light while the lambertian surface used                    mold made from Smooth-On MoldStar 20T.
in the previous sensor completely diffuses the light. This                           2) Casting: We begin the process by casting the plastic
means that rather than the light dissipating as it travels along                  shell. Here we chose to cast the shell with a clear epoxy
the sensor like it would with a lambertian surface, it will                       resin, in particular Smooth-On Epoxacast 690. The shell
keep its directionality and thus more uniformly illuminate                        is only 1mm thick so we chose this material because of
the surfaces beyond the curve of the surface Fig. 5. The thin                     its low-viscosity, clarity, rigidity, and overall ease of use.
plastic shell keeps the light from escaping into the interior                     After casting the shell we let the piece sit for 24 hours
of the sensor via TIR unless contact is made.                                     to completely cure. The next part tries to address some of
                                                                                  the limitations of the previous sensor in terms of durability.
C. Manufacturing                                                                  In particular, in the previous sensor the paint was easy to
   As a consequence of a more complicated geometry we                             remove and the gel easily delaminated from acrylic window.
have a more complicated manufacturing process compared                            These issues stem from the fact that, one it is difficult to
to previous versions of the GelSight sensors. Where the                           get silicone to stick to anything, and secondly that things
previous sensor only required 3D printing, laser cutting,                         were mechanically attached rather than chemically. To start
casting silicone into an open face mold, and pour over                            we begin by priming the surface of our plastic shell with
methods for coating, we rely on manufacturing a series of                         Dow DOWSIL P5200. This promotes silicone’s adhesion to
two-piece molds, casting into said molds, and then applying                       a variety of surfaces, but the caveat is the silicone must
the opaque coating using an airbrush. However, in this                            cure on that surface to form a chemical bond, rather than
process we have created methods that also significantly                           mechanically attaching a cured piece of silicone to a surface
increase the durability and reliability of the sensors compared                   like in the previous Gelsights. The next part is creating
to the previous version.                                                          the opaque coating. After experimenting with a variety of
   1) Mold Making: First, the desired geometry of the plastic                     coatings we decided on creating a custom coating that is
shell, as well as the geometry of the silicone and the shell                      quite durable. The coating is made out of a silicone paint
combined, are 3D printed with the Form Labs Form2 SLA                             base, Smooth-on Psycho Paint, and a non-leafing silver
printer using the clear resin. The cast pieces have to be                         dollar aluminum flake pigment. We spray this coating, so we
optically clear, so we prime the 3D printed pieces with                           dilute it with a silicone solvent, Smooth-On NOVOCS. The
Krylon Crystal Clear and then dip the pieces in a clear UV                        exact ratio used is 1:10:30 pigment, silicone paint base, and
cure resin and let it drip until a thin layer remains. We then                    silicone solvent ratio by mass. We spray the interior of the
cure the resin with a UV lamp. We do this multiple times                          silicone part of our mold with Mann Release Technologies
until the print is smooth and optically clear. The reference                      Ease Release 200, and then spray our opaque coating in with
plastic shell piece is then used to create a two-piece silicone                   an airbrush. We quickly screw our plastic shell onto our mold
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
(a)                                                                         (b)

Fig. 6. (a) the bottom shows the 3D reconstruction from the contact made textbftop from pressing the screw on the sensing surface. (b) Shows the 3D
reconstruction pipeline. Top from left to right shows the raw image, the height map generated from the fast Possion solver, and the image patch that is
used to generate bottom right the point cloud patch. Bottom left image show the complete 3D reconstruction.

base, assemble the mold, and pour our optically clear silicone                  board is screwed into the cover using an M2 screw. The
gel (Silicone Inc. XPS-565 1:15 A:B ratio by mass) into the                     cover is then press fit into the plastic shell. Two M2 screws
mold. We want to cast the silicone before the coating cures                     are inserted into the bottom of the Allegro mounting plate
so that they are chemically bonded. The mold is left out for                    that then pass through the through holes of the bottom LED
6 hours at room temperature and then is placed in the oven                      board, blinder, and plastic shell and then screwed into the
at 95 degrees Celsius.                                                          cover. The two LED boards are soldered together via four
   3) Assembly: The camera holder, cover, blinder, and                          wires. Two power cables are routed to a Raspberry Pi 4 to
mounting plate are 3D printed on the Markforged Onyx One                        power the LEDs with 3.3V, and then the camera is connected
printer with the Onyx filament which is very suitable for                       to a Raspberry Pi via CSI flat connector cable.
creating strong fixtures. The camera, Frank-S15-V1.0 Rasp-
berry Pi camera sensor, is then press fitted into the camera                    D. Software Interface
holder. The camera has a high FOV of 160 degrees which                             The image from the sensor is streamed via HTTP. The
observes a significant area of the sensing surface, while being                 image is streamed 640x480 with a frame rate of 90FPS. The
significantly more compact than previous cameras used in                        latency is low with a measured latency of about 40ms delay.
GelSight. The camera holder is then press fitted into the                       So far no computing happens on the Raspberry Pi. In terms
cover and then screwed in with a M2 screw.The back LED                          of interfacing all four sensors with a host computer, each
                                                                                camera is connected to a Raspberry Pi 4 using the standard
                                                                                CSI flat connector cables, then the four Raspberry Pis are
                                                                                then connected to a gigabit Ethernet switch in which the
                                                                                host computer is also connected to.

                                                                                E. Sensor calibration
                                                                                   The calibration process sets out to solve two things. The
                                                                                first is to map RGB values to gradients; the second is to
                                                                                find the 2D-3D correspondence in the form of pixel to point
                                                                                in point cloud correspondence. In terms of mapping RGB
                                                                                values to gradient we can no longer use a single look-up table
                                                                                like the previous GelSight sensors. Despite trying to achieve
                                                                                uniform lighting throughout the whole sensing surface there
                                                                                are obvious non-uniformities, so we propose constructing a
                                                                                look-up table for a set of regions. Once the look-up tables are
                                                                                constructed and we perform 3D reconstruction, the methods
Fig. 7. The sensor attached to the CNC rig used for calibration. The
sensor pokes the surface of the sensor at a variety of locations to construct
                                                                                know nothing about the curvature of the finger so we propose
the look-up table for 3D reconstruction and map the 3D surface to a 2D          a piece-wise forward projection of our height map onto the
image.                                                                          geometry of the surface.
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
Fig. 8.   Sample begin and end state of each object during the rolling phase of the experiment.

   1) 2D-3D Correspondence: To begin we start by getting                  order to do controlled rolling of an unknown object, we
the 3D geometry of our sensing surface. We break up the                   propose a tracking method along with a reactive controller
surface into a set of quads and get the vertices of those                 to deal with uncertainty in the geometry and dynamics of
quads. We now have to discover where these vertices lie in                the object. We assume that no slip will occur throughout
image space. In order to do this we constructed a CNC rig                 execution of the trajectory and chose an action according to
(Fig.7) in which we attach the finger tip rigidly in a known              the changes in geometry of the sensing surface.
location. We then have a probe with a 4mm diameter sphere                    1) 3D Reconstruction: In order to achieve reconstruction
attached to it. We tell the CNC to poke the sensor at the                 we need to have fast enough feedback about the geometry
calculated vertices. On each poke we take a picture of the                of the sensing surface. Using the tables constructed from
sensing surface from the sensors camera and perform Hough                 the calibration procedure we can now perform real time 3D
Circle Transform [13] to find the centroid of the sphere in               reconstruction as shown in Fig. 6. At each time step we create
image space. That point is then added to a table with its                 a difference image between the current sensor reading and an
corresponding location in the surfaces 3D space. Once all                 image of the sensor without contact to filter out everything
vertices have been probed we construct the reference point                in the image except the contact area. We convert the RGB
cloud. For each quad, we take the corresponding image patch               values in the difference image to gradients using our look-up
and calculate the perspective transform, since the image is               tables. This is passed to our fast Possion solver and we get a
taken from a perspective view. We then warp the image                     height map. Once we receive the height map we extract each
patch and get its resolution. Back in the surfaces space we               image patch corresponding to each quad, warp the image
create a linearly spaced grid in the quad with the same                   to get rid of the perspective view and add the calculated
resolution as the image patch and project it onto the surface             height to the corresponding point in the point cloud. Our 3D
geometry. Now when we receive a height map image during                   reconstruction runs at 40hz when the image is down-sampled
reconstruction we just take each image patch, warp it, and                to 320x240.
then change the depth of the corresponding points based-off                  2) Determining Contact Area: As mentioned in Sec. II-
the depth at each pixel.                                                  B, one limitation of the sensor is that we only provide
   2) RGB to Gradient Correspondence: We construct a                      information about the gradient in the x-direction with two
look-up table mapping each RGB value to a gradient for                    sources of information and the gradient in the y-direction
each image patch determined in the previous section. To                   with one source of information. This results in inaccuracies
construct the gradient, we begin by poking the finger tip                 about the heightmap along the y-axis. So, basic thresholding
in each quad several times at varying locations. For each                 on the depth of the heightmap does not result in an accurate
poke we calculate the centroid and radius of the poke in                  contact patch. To address this issue, the contact area is
image space, again using Hough Circle Transform. Since                    determined by selecting the points with a subset of points
the geometry of the probe is known we can map each pixel                  with the largest displacement.
intensity to a gradient. For each quad we take the average                   3) Tracking Contact Area: To track the contact area we
pixel intensity of all the pokes in that region and map that              use Iterative Closest Point (ICP) [14]. On each time-step
to the gradient. For RGB values not in the look-up table we               we calculate the current contact area and calculate its convex
assign it a gradient by linearly interpolating the gradients              hull. We get the points from the previous contact area that
mapped to the nearest RGB values.                                         lie within the convex hull of the current contact area and
                                                                          perform ICP to get the change in the contact area. We only
F. Controlled Rolling                                                     perform ICP with the points within the convex hull since on
  To validate the sensors and the sensor geometry we will                 each time step while rolling we lose contact with areas of
perform controlled rolling of a set of unknown objects. In                the previous contact area.
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
Fig. 9. Top from left to right Sequence of the experiment being performed on the plastic sphere during the rolling stage. Bottom from left to right
The corresponding point clouds showing the evolution of the point cloud as the object rolls.

                                                                            varying smoothness, hardness, and geometries. We show a
                                                                            sample of each object being rolled in Fig. 8, Fig. 9, and
                                                                            Fig. 10. Looking at some of the objects, in particular the
                                                                            roll of tape and yo-yo, the object was successfully rolled
                                                                            to the desired position but exhibits unwanted rotational slip.
                                                                            The object that is the source of the unsuccessful trial was the
                                                                            golf ball. During the execution of rolling the object rolled out
                                                                            of the fingers. This might be a result of the dimples reducing
Fig. 10. Illustration showing the experimental set-up. From left to right   the total contact area made with the sensing surface. Fig. 9
The placement of the object, when the object makes contact with both
fingers, the end state of the object when the object reaches the desired    serves as visual verification of the 3D reconstruction.
position.

   4) Controller: We use a hybrid velocity/force controller                                       IV. CONCLUSION
[15]. This allows the finger to perform a compliant motion
where the finger moves up in task space while maintaining a                    Most applications of robotic manipulation rely on the
consistent force normal to the the contact patch, resulting in              use of suction or parallel grips due to the simplicity of
a rolling motion. The maximum displacement of the contact                   their controls. This however comes at the cost of requiring
surface is used as a proxy for force in the controller.                     high precision vision-based perception systems in order to
                                                                            complete a task. Further more, the use of such grippers limit
                          III. RESULTS
                                                                            applications to pick-and-place. Tasks that require the use of
A. Experimental Setup                                                       an object, such as cutting with a knife, require not only to
   To validate our sensors capabilities we perform controlled               pick up an object but to also reorient it. For aforementioned
rolling on a set of unknown objects as shown in Fig. 8, Fig. 9,             grippers this requires special rigs, two-arm manipulators,
and Fig. 10. The experiment is broken up into three stages.                 or interacting with the environment. These issues constrain
In the first stage we manually place an unknown object in                   robotic manipulation to structured environments. Dexterous
between the index finger and the thumb of the Allegro Hand.                 manipulation is seen as a way forward, but most systems
While the thumb of the Allegro Hand stays stationary the                    controlling dexterous hands rely heavily on vision, which
index finger moves towards the thumb until contact is made.                 compound the difficulty of the problem. While there are a
After contact is made, the index finger will continue to move               variety dexterous manipulators equipped with tactile sensors,
towards the thumb until the desired maximum displacement                    they are either too rigid or do not provide information.
of the sensing surface is achieved. We then use the controller                 In this paper, we have presented a high-resolution, compli-
described Sec. II-F to roll the object until contact is made                ant, and round tactile sensor for dexterous manipulators. This
in a desired region of the finger as shown in Fig. 10. A                    required designing a new illumination system, that despite its
trial is considered a success if the finger is able to roll the             limitations, is still suitable for manipulation. We justify this
object until contact is made within the target contact region.              design by exploring the importance of the geometry of the
If the object falls out of the grasp at any stage in the trial,             sensor for dexterous manipulation. We also show that it is
overshoots the desired contact region, or does not reach the                capable of being deployed into applications requiring real-
contact area within 3 seconds after the rolling stage of the                time feedback by performing a set of in-hand manipulation
trial begins the trial is considered a failure. We perform 10               in the form of controlled rolling on a diverse set of objects.
trials for each object.                                                        Future work involves reducing the size of the sensor,
   1) Experimental Results: As illustrated in table (add table              making improvements in the illumination and 3D reconstruc-
here) our method and sensor were able to successfully                       tion, adding additional sensing modalities, and exploring the
perform 99 out of 100 controlled rolls into the desired contact             use of these sensors in more in-hand manipulation task and
region despite being presented a diverse set of objects with                grasping.
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
R EFERENCES
 [1] R. Johansson. The effects of anesthesia on motor skills -
     youtube. [Online]. Available: https://www.youtube.com/watch?v=
     0LfJ3M3Kn80
 [2] R. S. Johansson and A. Vallbo, “Tactile sensibility in the human hand:
     relative and absolute densities of four types of mechanoreceptive units
     in glabrous skin.” The Journal of physiology, vol. 286, no. 1, pp. 283–
     300, 1979.
 [3] R. S. Johansson and J. R. Flanagan, “Coding and use of tactile signals
     from the fingertips in object manipulation tasks,” Nature Reviews
     Neuroscience, vol. 10, no. 5, p. 345, 2009.
 [4] N. F. Lepora and B. Ward-Cherrier, “Superresolution with an optical
     tactile sensor,” in 2015 IEEE/RSJ International Conference on Intel-
     ligent Robots and Systems (IROS). IEEE, 2015, pp. 2686–2691.
 [5] W. Yuan, S. Dong, and E. Adelson, “Gelsight: High-resolution robot
     tactile sensors for estimating geometry and force,” Sensors, vol. 17,
     no. 12, p. 2762, 2017.
 [6] A. Yamaguchi and C. G. Atkeson, “Combining finger vision and
     optical tactile sensing: Reducing and handling errors while cutting
     vegetables,” in 2016 IEEE-RAS 16th International Conference on
     Humanoid Robots (Humanoids). IEEE, 2016, pp. 1045–1051.
 [7] R. Calandra, A. Owens, D. Jayaraman, J. Lin, W. Yuan, J. Malik,
     E. H. Adelson, and S. Levine, “More than a feeling: Learning to grasp
     and regrasp using vision and touch,” IEEE Robotics and Automation
     Letters, vol. 3, no. 4, pp. 3300–3307, 2018.
 [8] N. F. Lepora, K. Aquilina, and L. Cramphorn, “Exploratory tactile
     servoing with active touch,” IEEE Robotics and Automation Letters,
     vol. 2, no. 2, pp. 1156–1163, 2017.
 [9] W. Yuan, R. Li, M. A. Srinivasan, and E. H. Adelson, “Measurement
     of shear and slip with a gelsight tactile sensor,” in 2015 IEEE
     International Conference on Robotics and Automation (ICRA). IEEE,
     2015, pp. 304–311.
[10] S. Wang, J. Wu, X. Sun, W. Yuan, W. T. Freeman, J. B. Tenenbaum,
     and E. H. Adelson, “3d shape perception from monocular vision,
     touch, and shape priors,” in 2018 IEEE/RSJ International Conference
     on Intelligent Robots and Systems (IROS). IEEE, 2018, pp. 1606–
     1613.
[11] W. Yuan, Y. Mo, S. Wang, and E. H. Adelson, “Active clothing material
     perception using tactile sensing and deep learning,” in 2018 IEEE
     International Conference on Robotics and Automation (ICRA). IEEE,
     2018, pp. 1–8.
[12] N. Pestell, L. Cramphorn, F. Papadopoulos, and N. F. Lepora, “A
     sense of touch for the shadow modular grasper,” IEEE Robotics and
     Automation Letters, vol. 4, no. 2, pp. 2220–2226, April 2019.
[13] D. H. Ballard, “Generalizing the hough transform to detect arbitrary
     shapes,” Pattern recognition, vol. 13, no. 2, pp. 111–122, 1981.
[14] Y. Chen and G. Medioni, “Object modeling by registration of multiple
     range images,” in Proceedings. 1991 IEEE International Conference
     on Robotics and Automation, April 1991, pp. 2724–2729 vol.3.
[15] M. T. Mason, “Compliance and force control for computer controlled
     manipulators,” IEEE Transactions on Systems, Man, and Cybernetics,
     vol. 11, no. 6, pp. 418–432, June 1981.
Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation Soft, Round, High Resolution Tactile Fingertip Sensors for Dexterous Robotic Manipulation
You can also read