PreprintPDF Available

Deep Functional Predictive Control for Strawberry Cluster Manipulation using Tactile Prediction

Authors:
Preprints and early-stage research may not have been peer reviewed yet.

Abstract

This paper introduces a novel approach to address the problem of Physical Robot Interaction (PRI) during robot pushing tasks. The approach uses a data-driven forward model based on tactile predictions to inform the controller about potential future movements of the object being pushed, such as a strawberry stem, using a robot tactile finger. The model is integrated into a Deep Functional Predictive Control (d-FPC) system to control the displacement of the stem on the tactile finger during pushes. Pushing an object with a robot finger along a desired trajectory in 3D is a highly nonlinear and complex physical robot interaction, especially when the object is not stably grasped. The proposed approach controls the stem movements on the tactile finger in a prediction horizon. The effectiveness of the proposed FPC is demonstrated in a series of tests involving a real robot pushing a strawberry in a cluster. The results indicate that the d-FPC controller can successfully control PRI in robotic manipulation tasks beyond the handling of strawberries. The proposed approach offers a promising direction for addressing the challenging PRI problem in robotic manipulation tasks. Future work will explore the generalisation of the approach to other objects and tasks.
Deep Functional Predictive Control (deep-FPC):
Robot Pushing 3-D Cluster using Tactile Prediction
Kiyanoush Nazari1, Gabriele Gandolfi2, Zeynab Talebpour, Vishnu Rajendran3,
Willow Mandil1, Paolo Rocco2and Amir Ghalamzan-E.3
Abstract This paper introduces a novel approach to address
the problem of Physical Robot Interaction (PRI) during robot
pushing tasks. The approach uses a data-driven forward model
based on tactile predictions to inform the controller about
potential future movements of the object being pushed, such as
a strawberry stem, using a robot tactile finger. The model is
integrated into a Deep Functional Predictive Control (d-FPC)
system to control the displacement of the stem on the tactile
finger during pushes. Pushing an object with a robot finger
along a desired trajectory in 3D is a highly nonlinear and
complex physical robot interaction, especially when the object
is not stably grasped. The proposed approach controls the stem
movements on the tactile finger in a prediction horizon. The
effectiveness of the proposed FPC is demonstrated in a series
of tests involving a real robot pushing a strawberry in a cluster.
The results indicate that the d-FPC controller can successfully
control PRI in robotic manipulation tasks beyond the handling
of strawberries. The proposed approach offers a promising
direction for addressing the challenging PRI problem in robotic
manipulation tasks.
I. INTRODUCTION
In the field of Physical Robot Interaction (PRI), successful
manipulation tasks rely on accurate interaction models that
utilise rich sensory information and intelligent control strate-
gies. While advanced learning from demonstration methods,
such as deep Movement Primitives (MPs) [1] and Deep
Probabilistic MPs [2], plan the robot’s movements given
an image of the robot’s workspace, they are not effective
for PRI tasks. Tactile feedback is a particularly effective
sensing modality for such cases especially when vision-based
control, such as visual servoing [3], is not feasible due to
occlusion [4]. For example, pushing a ripe strawberry that
is occluded by plant stems, leaves, or unripe fruits in a
cluster [5] can require tactile feedback for effective control.
Effective object manipulation under uncertainty [6], pre-
grasp manipulation to align the object with the fingers [7],
and agile soccer ball pushing by a mobile robot [8] are a few
examples of pushing. Analytical models for pushing require
complete knowledge of the environment, including physical
and geometric properties such as object pose, shape [9],
friction parameters, and mass. Developing analytical models
for unstructured environments characterised by high degrees
of freedom, non-linearity, and stochasticity, such as the case
of pushing a flexible stem to reach a strawberry, can be a
challenging task [10].
1School of Computer Science, University of Lincoln, UK. 2DEIB,
Politecnico di Milano, Italy. 3Lincoln Institute for Agri-food Technology,
University of Lincoln, UK. This work was partially supported by EPSRC
(AgriFoRwArdS) Grant reference: EP/S023917/
Fig. 1: Strawberry pushing setup: a Franka Emika robotic
arm is pushing a cluster of strawberries from right to left
where the nearest strawberry stem comes in contact with its
tactile finger. (Top Right) the robot at the beginning of the
pushing action, (Top Left) the robot, and the cluster at the end
of pushing. Sample tactile finger images are shown from right
to left for the initial and final point in the pushing trajectory,
respectively. Our tactile finger design features a deformable
half-conic membrane with an integrated miniature camera
and LED light.
Most existing pushing methods are designed for 2-D
scenarios in which an object is moving on a flat surface, but
in the case of strawberry picking, a 3-D pushing scenario is
more relevant [11]. Pushing a strawberry in a 3-dimensional
space is more challenging than pushing an object on a
table (i.e. a 2D problem). While interactive movement prim-
itives [12] can be used to plan pushing actions, an accurate
interaction model is crucial for effectively controlling the
planned motion of the strawberry during pushing in this
scenario.
In this paper, we presented a novel deep functional pre-
dictive control pipeline for the manipulation of strawberries
grown on a table. Our pipeline consists of three key modules:
a deep action-conditioned Tactile Forward Model (TFM), a
deep Contact Localisation Model (CLM), and an online deep
Functional Predictive Control (d-FPC) to generate control
actions. We collected a dataset of plastic strawberries being
pushed in our lab setting to train TFM, which is the state-
of-the-art tactile prediction model. We also trained CLM
to calibrate our tactile sensor using a dataset of strawberry
pushing. Finally, d-FPC uses real-time predictions from TFM
and CLM to generate robot actions based on future error
signal estimations to control the stem pose on the sensor
surface. We compared our proposed functional predictive
controller’s performance with a PD control-based system that
only uses CLM and demonstrated that the predictive system
outperforms this baseline model. This study addresses the
challenge of pushing flexible objects in 3D, and to the best
of our knowledge, this is the first study to do so. Our results
demonstrate the effectiveness of our proposed approach and
pave the way for future research in the manipulation of
flexible objects using deep functional predictive control.
II. REL ATE D WORKS
Cluster manipulation in fruit harvesting is a challenging
task from both motion planning and motion control per-
spectives [14], [12]. One of the challenges is avoiding the
slip of a grasped object, which can be addressed through
closed-loop robot trajectory adaptation [15]. Deformable
object manipulation, such as cloth, has been modelled using
simplified mass-spring models or 3D mesh generation [16],
while heuristic feature spaces have been used for flexible
cable manipulation with dual robot arms [17]. However,
analytical modelling methods are limited to specific object
sets and are not scalable to larger object and action sets.
In contrast, our proposed approach uses a time-series model
for action-conditioned tactile prediction for pushing control
which can be applied in unstructured settings without the
knowledge about the model of the individual objects.
Tactile feedback is mostly used for grasp control in robotic
object clutter manipulation [4] and detecting a grip on
fruits [18]. However, the use of tactile sensors has been
limited to grip control and has not been applied for any
cluster manipulation. In our work, we exploit tactile feedback
for trajectory-level control for pushing a flexible plant stem.
Tactile prediction models are used for controlling ma-
nipulation tasks, from the simple task of rolling a marble
on a table [19] to the complex task of slip control [15].
The core of such controllers is a forward model that can
generate predicted tactile readings (i.e. tactile images). For
instance, action-conditioned tactile predictive models are
utilised with a taxel-based tactile sensor in pick and place
tasks [13], demonstrating the approach performs well only
for flat surface objects.
Our approach uses a time-series model for tactile predic-
tion based on [13]. We form a deep Predictive Functional
Control (d-FPC) [20] which enables the robot to control
the strawberry pushing actions. Deep models have been
extensively used for learning lower dimensional state spaces
for Model Predictive Control (MPC). These methods have
also been used for learning visual dynamic models for
control [21]. In a simplified task of rolling a mini sphere,
the tactile prediction was used in an MPC controller [19].
We form a Proportional-Derivative (PD) control over the
error in the prediction horizon to control the contact state
of a flexible object on a robot hand. Unlike previous work
that used trajectory adaptation to minimise the likelihood of
predicted binary slip signal in a prediction horizon [15], our
model learns the complex contact behaviour and generates
actions to control the movements of the stem on the tactile
finger to keep it stable.
III. METHODOLOGY
a) Camera-based tactile sensor: We use a customised
camera-based tactile sensor for pushing strawberries similar
to Tactip [22]. This sensor has a camera and an LED light
looking at a deformable membrane with embedded white
markers (Fig. 1). The applied pressure on the sensor yields
a deformation that is captured by the camera. The sharp tip
and low membrane stiffness make this tactile sensor suitable
for the task of strawberry cluster manipulation.
b) Contact Localisation Model (CLM): The motions of
the marker array printed on the sensor are indicative of the
magnitude and location of the applied force. For the current
problem setting, we are more interested in force localisation
for doing stem contact state control. To find the mapping
from raw tactile images to contact location in 1-dimensional
space, we use a Convolutional Neural Network with the
architecture shown in Fig.2 (red box). CLM consists of two
convolutional and three dense layers. The output of CLM
is the distance of the contact force from the sensor camera
lens along the sensor conic axis. The data set for training
CLM consists of applying forces to the fixed sensor by a
rod (mimicking strawberry stem) attached to the robot end-
effector (EE) with a 5mm distance step. At each step, the
robot applies forces on the membrane toward the sensor base
by a 1mm penetration step. Overall, 150 pushing samples in
10 locations are collected to train CLM.
c) Tactile Forward Model (TFM): Tactile prediction
aims to estimate future tactile images based on a set of
previous tactile images x0, ..., xc1obtained from physical
interactions, where cis the length of the context window.
Specifically, the objective is to sample from the conditional
distribution p(xc:T|x0:c1), where xidenotes the ith tactile
image in the sequence and Tis the sum of the context
window length and the prediction horizon length.
Since the robot’s actions alter the environment during
physical interaction, we incorporate action conditioning
to predict tactile sensation more accurately. The action-
conditioned tactile prediction problem is formulated as pre-
dicting the future tactile images xc:Tgiven a sequence of
previous robot actions a0:c1, previous tactile images x0:c1,
and a sequence of future/planned robot actions/trajectory
ac:T. Here, a robot action, aR6, refers to the end-
effector task space position and orientation (Euler angles)
with respect to the robot base, while a tactile image is
represented by xR64×64×3, which captures the surface
deformation caused by the applied force. The conditional
distribution will be:
p(xc:T|x0:c1,a0:T)(1)
Factorising this we can define the model as
ΠT
t=cpθ(xt|x0:t1,a0:t). Learning now involves training
the parameters of the factors θ.
Contact
Localisation
Model
Tactile Forward Model
Contact
Localisation
Model
d-FPC
Conv2D
ReLu
MaxPool
Flatten
GeMM
ReLu
x 2 x 2
CLM
x 2x 2
Skip Connection
ConvLSTM
Conv2D
ReLu
MaxPool
Conv2D
ReLu
UpSample
ConvLSTM
Tanh
TFM
Fig. 2: The block diagram of the proposed data-driven functional predictive control for pushing strawberries. The model
consists of (1) tactile forward model (TFM) which is based on [13], contact localisation model (CLM), and the functional
predictive controller (d-FPC) that generates future actions resulting in the minimum stem displacement on the tactile finger.
The model architecture is depicted in Fig.2 (blue box).
We extract scene features from the input tactile image by
convolutional filters in the first two layers of the network
as the encoder. Each convolution operation is followed by
the Relu activation function and 2D maxpooling operations.
Robot action sequences are concatenated with latent tactile
features after the convolutional layers. These latent space
features with downsampled width and height and a larger
number of channels are fed to the Conv-LSTM chain.
These layers process the spatiotemporal dependencies among
the latent features. After this point, we need to upscale
the features to reach the tactile image size. As such, two
convolutional layers, each one followed by Relu activation
and 2D upsampling, are applied to ConvLSTM outputs. To
apply the pixel motion changes to the input, we use the
skip connection for the input tactile image and apply tanh
activation to construct the next tactile images in the sequence.
d) deep-Functional Predictive Control (d-FPC): We
denote the predicted stem location (from CLM) on the sensor
at time tby st. The goal of our d-FPC is to control the
stem displacement on the tactile finger. Hence, this allows
the robot to keep the contact fixed with the strawberry
stem during pushing actions and avoid the contact location
approaching the tip or the base of the sensor. These are sensor
surface boundary zones and approaching them increases the
probability of losing contact with the stem. We use the stem-
finger contact point at time tas the reference for our d-FPC
controller. We define an error signal as the distance of the
contact point from the reference point:
ei,t = ˆsist, i =c, ..., T (2)
where ˆsiis the predicted stem location for a sequence of
planned robot movements. We formulate our d-FPC over the
error signal as follows:
at,res =X
i=c:T
(kpi×ei,t +kdi×˙ei,t)(3)
where at,res is the residual action value to be added to the
reference trajectory at,ref to generate the control action At.
Atis a rotational velocity around the contact line axis. Fig.2
(green box) shows the schematic of the d-FPC. The generated
control output is a rotational velocity proportional to the
distance of the stem from the reference line. The derivative
term avoids overshooting and having large instant rotations.
IV. EXP ERI MEN TAL SETUP AND DATA SE T
a) Tactile sensor and manipulation task: Various types
of tactile sensors are discussed in the literature, including
in [4]. In this work, we use a custom-made camera-based
tactile sensor that has a half-conic geometry and a tapered
tip (shown in Fig.1) designed to allow for easier penetration
among stems and fruits, providing valuable tactile feedback.
The deformable membrane of the sensor is 3D-printed and
dot features are printed with a linear pattern on its conic
inner surface. Changes in the marker pattern resulting from
contact forces provide information about contact force value,
geometry, and location. The camera, which is located on
the sensor base, and the LED, used for illuminating the
markers, are powered by an onboard Raspberry Pi, and
tactile images are transmitted at a frequency of 60 Hz. The
sensor is mounted on a Franka Emika gripper, providing an
effective and versatile tool for physical interaction in a range
of applications.
We have collected the data from a series of strawberry-
pushing tasks in 3-D. The pushing dataset includes data
for single strawberry pushing and pushing a cluster of
strawberries. To simulate the table-top strawberry growing
scenario, we attached each plastic strawberry to a thin wire
that makes a nonlinear elastic behaviour similar to those
usually observed in tabletop-grown strawberries. To simulate
realistic tactile feedback, we added knots on the stalk of
each strawberry (Fig.1) and injected silicone to increase their
weight (each strawberry weighs c. 20 g to 30 g).
We generate the pushing trajectories for the training data
collection phase by two methods: first by Pilz industrial
motion planner by specifying initial and target robot poses,
and second by defining a minimum time reference trajectory
using the robot’s Cartesian velocity controller. We use the
second method to be able to regenerate comparably similar
trajectories in test time, as opposed to the first case where
trajectories are generated by the motion planning library.
Trajectories include linear and circular motion patterns to
perform the pushing tasks. Arc trajectories were used to
collect more tactile-conditioned robot movements, where the
finger followed the motion of the pushed stem/strawberry.
These pushes started at a position p0and orientation q0,
followed an arc trajectory, and ended at a final position
pfwith a value of zcoordinate larger than initial position.
The final orientation qfis selected to maintain contact with
the elements pushed. The pushing actions were performed
from right to left and vice versa, and they involved single or
multiple stems (Fig. 1), generating greater deformations on
the membrane.
We collected a total of 430 mixed linear/circular motion
tasks containing (i) tactile images from the finger at 60 Hz
and (ii) robot state data sampled at 1000 Hz, representing
the position and orientation of the end effector in the
planned trajectory. These readings were synchronised using
the ROS ApproximateTime policy and fed into the tactile
forward model both in training and test times.
Considering the robot’s motion, slip occurred mainly on
the width and length of the finger but could also happen in
other directions depending on the motion of the stems during
the pushing actions.
V. RESULTS AND DISCUSSION
We test the performance of our proposed control pipeline
in real-time on pushing tasks of strawberry stems and
compare the performance with a baseline controller and an
open-loop system. The tactile sensor is mounted on Franka
Emika robot connected to a PC with Intel® Core™ i7-8700K
CPU @ 3.70GHz × 12 and 64GB RAM running Ubuntu
20.04 and ROS Noetic. Torch library is used for offline
training and online testing of the neural network models. Test
manipulation tasks consist of performing pushing trajectories
with linear and circular motion patterns using the robot’s
Cartesian velocity controller.
Performance metrics include: (I) Stem max displacement
and (II) the number of stem slip instances on the sensor
surface. If we denote stem location at time tby siwhere
i(0,1, ..., T )for a pushing trial, metric (I) is defined as the
absolute value of the difference of maximum and minimum
stem location in a trial |max(si)min(si)|:i= 1, ..., T .
Metric (II) is defined as the number of time steps where the
differential values ˙siwere larger than threshold γ. While
metric (I) shows full stem displacement, metric (II) shows
the stem’s sudden large motion instances or slippage on the
sensor surface. We also present the area under the curve of
stem displacement and generated action. We repeat each test
case 5 times and present the mean and standard deviation
of the metric values. Overall we conducted 100 test-pushing
trials.
To evaluate the effectiveness of d-FPC for pushing control,
we compare the control performance with a PD control-based
tactile servoing system as the baseline model. Both models’
results are presented against the open-loop system with a
pre-specified reference trajectory.
In this paper, we utilise a minimum-time reference trajec-
tory (such as bang-bang) for the open-loop system, although
any desired reference trajectory can be used. To make valid
comparisons among trials, we consider three initial contact
zones for the stem including Zone-1 where the contact point
is between the middle and tip of the sensor, Zone-2 has the
contact point between the middle and base of the sensor, and
Zone-3 where the contact point is close to sensor centre line.
Since the tactile sensor has varying deformation limits across
its conic axis we compare the trials with corresponding initial
contact zones together.
We conduct a comparison test with a one-degree-of-
freedom (DOF) horizontal pushing along the Y-axis of the
robot’s base frame. Both PD and d-FPC controllers generate
control actions for the robot hand’s rotation around the
contact line to prevent stem slip on the sensor surface.
The results are presented in Table I, where test cases are
conducted separately for each initial contact zone. Both
PD and d-FPC controllers decrease the stem’s maximum
displacement. We observe that d-FPC outperforms the PD
controller for Zone-1 and Zone-3, but PD shows better
performance for Zone-2 very close to the sensor base. This
is because the sensor has its largest deformation limit in the
Base zone, resulting in relatively large initial deformation
after making contact, making it difficult for TFM to predict
future stem states. The prediction of the error signal helps
d-FPC to have more reaction time than PD.
We find that d-FPC is the most effective controller to
reduce the number of stem slip instances, with the smallest
area under the curve of displacement compared to the PD
controller. We also present the computation time to show
the relative computation complexity of each system. Since
d-FPC has two stacked deep models, the computation time
is larger than the PD controller.
To compare the performance of different controllers in a
qualitative manner, we present the stem location obtained in
two trials (shown in Fig. 3a): Trial-1, where the stem-finger
initial contact point is in Zone-1, is shown with solid lines,
and Trial-2, with the contact point in Zone-2, is shown with
a dashed line. Our results show that d-FPC outperforms PD
controller and open loop in maintaining the stem contact,
resulting in the smallest displacement of the stem. Further-
more, Fig. 3b shows the control actions generated by each
Model Contact
zone
Stem max
disp.
Stem slip
instances
Disp.
integral
Action
integral
Comp.
time (ms)
Open-loop
1 0.80 ±0.2 31.23 ±4.3 0.83 ±0.1 - -
2 1.35 ±0.2 50.19 ±5.7 0.91 ±0.1 - -
3 0.91 ±0.1 39.83 ±3.2 0.86 ±0.2 - -
PD
1 0.65 ±0.1 27.2 ±6.5 0.75 ±0.1 2.93 ±0.7 18.73 ±2
20.36 ±0.0 10.2 ±2.4 0.48 ±0.0 5.12 ±3.8 20.30 ±1
3 0.63 ±0.1 24.2 ±1.6 0.47 ±0.1 9.73 ±5.4 19.73 ±1
d-FPC
10.20 ±0.0 5.0 ±1.2 0.12 ±0.0 3.74 ±0.8 60.49 ±6
2 0.43 ±0.0 7.2 ±0.7 0.18 ±0.0 4.27 ±1.2 55.02 ±2
30.25 ±0.1 6.0 ±0.6 0.09 ±0.0 4.57 ±2.4 58.54 ±3
TABLE I: Control performance for the PD and d-FPC in pushing a single strawberry along a linear trajectory.
Model Robot
trajectory
Stem max
disp.
Stem slip
instances
Disp.
integral
Action
integral
Open-loop Linear 1.21 ±0.18 44.38 ±10.3 0.88 ±0.4 -
Circular 1.35 ±0.46 48.18 ±5.2 1.02 ±0.5 -
PD Linear 0.58 ±0.21 25.53 ±4.2 0.63 ±0.1 5.39 ±6.2
Circular 1.20 ±0.01 17.6 ±2.0 0.44 ±0.0 9.89 ±0.8
d-FPC Linear 0.29 ±0.04 8.11 ±1.4 0.13 ±0.0 4.49 ±2.5
Circular 0.54 ±0.05 5.0 ±1.5 0.22 ±0.0 6.66 ±0.8
TABLE II: Comparison of the controllers in linear and circular pushing trajectories.
Model Robot
trajectory
Stem max
disp.
Stem slip
instances
Disp.
integral
Open-loop Linear 1.43 ±0.30 49.33 ±15.64 1.39 ±0.33
Circular 1.29 ±0.67 47.98 ±6.33 1.19 ±0.23
PD Linear 0.79 ±0.21 29.4 ±6.52 0.66 ±0.21
Circular 1.14 ±0.23 20.5 ±2.69 0.56 ±0.84
d-FPC Linear 0.31 ±0.08 17.1 ±2.39 0.25 ±0.03
Circular 0.61 ±0.11 9.5 ±4.58 0.27 ±0.18
TABLE III: Controller and open loop performances for Pushing a cluster of strawberries.
controller. We observe that d-FPC generates actions of larger
magnitude in Trial-1 because the likelihood of losing the
stem in Zone-1 (namely closer to the tip) is larger than in
Zone-2. In Trial-2, the magnitude of d-FPC and PD controller
actions is similar since the contact between the stem and
sensor membrane is tighter due to a larger deformation of
the sensor closer to the sensor base.
We test the performance of the systems in a three DOF
task with a bang-bang reference for translation along Y,
Z, and rotation Wxof Cartesian velocity space. This is a
more challenging task because the robot wrist will rotate
45 degrees along the pushing trajectory which causes larger
deformation of the stem and more slip instances. Based on
Table II d-FPC is the most effective controller in decreasing
the stem displacement and slip instances. PD has a smaller
improvement in max displacement for the circular motion
than the linear motion compared to the open-loop system.
This indicates that not having enough reaction time in this
task can lead to failure in achieving the control objective.
We test the generalisation performance of the pushing
controller when pushing a stem in a cluster of strawberries.
In this task additional to the target stem, other stems, leaves,
or strawberries come into contact with the sensor which
makes both tactile prediction and control more challenging.
Table III shows the results for pushing a stem in a cluster.
Although the control performance of PD and d-FPC degrades
compared to pushing an isolated stem, both systems improve
the performance metrics relative to the open-loop system.
Fig.3c shows cluster pushing results for sample trials
of linear and circular pushing trajectories. For the linear
push, PD has slight improvement compared to the open-
loop system but d-FPC reduces stem displacement more
effectively. For the circular push, while the open-loop system
loses contact with the stem because of large stem slippage
in the last part of the trial, both PD and d-FPC reduce the
stem displacement to avoid large slips. d-FPC keeps the
displacement more bounded relative to the PD controller
does.
(a) stem pose (b) control action (c) strawberry cluster pushing
Fig. 3: Comparison of control performance between d-FPC, open loop, and PD controller in maintaining the location of the
stem constant on the finger surface (Trial-1 (T1) solid and Trial-2 (T2) dashed lines) (a) At time 0.85 s, the stem makes
contact with the tactile finger and the controllers activate. d-FPC can maintain the stem contact point during the pushing
action, while the open loop result shows the stem moving out of the tactile finger surface. (b) The magnitude of the control
input shows d-FPC provides larger wrist rotation to avoid stem contact displacement. (c) Strawberry cluster pushing results.
VI. CONCLUSION
We presented a novel deep Predictive Functional Control
(d-PFC) framework to control the contact location of a straw-
berry stem on our tactile finger. Our proposed method lever-
ages a time-series model for generating action-conditioned
tactile predictions and a convolutional neural network (CNN)
model converting the tactile images to contact location. We
demonstrated the effectiveness of our approach through a
series of experiments with a Franka Emika robot and a
customised tactile finger, showing that our model can learn
complex contact behaviours and generate actions to control
the movements of flexible objects to keep them stable,
e.g. pushing a cluster of strawberries. Overall, our work
highlights the potential of deep learning-based approaches
in addressing the challenges of tactile sensing-based manip-
ulation tasks and lays the foundation for future research in
this field.
REFERENCES
[1] O. Sanni, G. Bonvicini, M. A. Khan, P. C. L´
opez-Custodio, K. Nazari
et al., “Deep movement primitives: toward breast cancer examination
robot,” in Proceedings of the AAAI Conference on Artificial Intelli-
gence, vol. 36, no. 11, 2022, pp. 12 126–12 134.
[2] A. Tafuro, B. Debnath, A. M. Zanchettin et al., “dpmp-deep prob-
abilistic motion planning: A use case in strawberry picking robot,”
arXiv preprint arXiv:2208.09074, 2022.
[3] S. S. Mehta, W. MacKunis, and T. F. Burks, “Robust visual servo
control in the presence of fruit motion for robotic citrus harvesting,”
Computers and Electronics in Agriculture, vol. 123, pp. 362–375,
2016.
[4] A. Yamaguchi and C. G. Atkeson, “Recent progress in tactile sensing
and sensors for robotic manipulation: can we turn tactile sensing into
vision?” Advanced Robotics, vol. 33, no. 14, pp. 661–673, 2019.
[5] Y. Xiong, C. Peng, L. Grimstad, P. J. From, and V. Isler, “Development
and field evaluation of a strawberry harvesting robot with a cable-
driven gripper, Computers and electronics in agriculture, vol. 157,
pp. 392–402, 2019.
[6] M. R. Dogar and S. S. Srinivasa, “Push-grasping with dexterous hands:
Mechanics and a method,” in 2010 IEEE/RSJ International Conference
on Intelligent Robots and Systems. IEEE, 2010, pp. 2123–2130.
[7] J. E. King, M. Klingensmith, C. M. Dellin, M. R. Dogar, P. Velagapudi,
N. S. Pollard, and S. S. Srinivasa, “Pregrasp manipulation as trajectory
optimization.” in Robotics: Science and Systems. Berlin, 2013.
[8] R. Emery and T. Balch, “Behavior-based control of a non-holonomic
robot in pushing tasks,” in Proceedings 2001 ICRA. IEEE International
Conference on Robotics and Automation (Cat. No. 01CH37164),
vol. 3. IEEE, 2001, pp. 2381–2388.
[9] G. Amirkhani, A. Goodridge, M. Esfandiari, H. Phalen, J. H. Ma,
I. Iordachita, and M. Armand, “Design and fabrication of a fiber
bragg grating shape sensor for shape reconstruction of a continuum
manipulator, IEEE Sensors Journal, 2023.
[10] M. Q. Mohammed, L. C. Kwek, S. C. Chua, A. Al-Dhaqm, S. Na-
havandi, T. A. E. Eisa, M. F. Miskon, M. N. Al-Mhiqani, A. Ali,
M. Abaker et al., “Review of learning-based robotic manipulation in
cluttered environments, Sensors, vol. 22, no. 20, p. 7938, 2022.
[11] J. St¨
uber, C. Zito, and R. Stolkin, “Let’s push things forward: A survey
on robot pushing,” Frontiers in Robotics and AI, p. 8, 2020.
[12] S. Mghames, M. Hanheide, and A. Ghalamzan, “Interactive movement
primitives: Planning to push occluding pieces for fruit picking, in
2020 IEEE/RSJ International Conference on Intelligent Robots and
Systems (IROS). IEEE, 2020, pp. 2616–2623.
[13] W. Mandil, K. Nazari et al., “Action conditioned tactile prediction: a
case study on slip prediction,” Robotics: Science and Systems, 2022.
[14] H. Zhou, X. Wang, W. Au, H. Kang, and C. Chen, “Intelligent robots
for fruit harvesting: Recent developments and future challenges,
Precision Agriculture, vol. 23, no. 5, pp. 1856–1907, 2022.
[15] K. Nazari, W. Mandil, and A. M. G. Esfahani, “Proactive slip control
by learned slip model and trajectory adaptation,” in Conference on
Robot Learning. PMLR, 2023, pp. 751–761.
[16] V. E. Arriola-Rios, P. Guler, F. Ficuciello, D. Kragic, B. Siciliano, and
J. L. Wyatt, “Modeling of deformable objects for robotic manipulation:
A tutorial and review,” Frontiers in Robotics and AI, vol. 7, p. 82,
2020.
[17] J. Zhu, B. Navarro, P. Fraisse, A. Crosnier, and A. Cherubini, “Dual-
arm robotic manipulation of flexible cables,” in 2018 IEEE/RSJ
International Conference on Intelligent Robots and Systems (IROS).
IEEE, 2018, pp. 479–484.
[18] H. Zhou, H. Kang, X. Wang, W. Au, M. Y. Wang, and C. Chen,
“Branch interference sensing and handling by tactile enabled robotic
apple harvesting,” Agronomy, vol. 13, no. 2, p. 503, 2023.
[19] S. Tian, F. Ebert, D. Jayaraman, M. Mudigonda, C. Finn, R. Calandra,
and S. Levine, “Manipulation by feel: Touch-based control with deep
predictive models, in 2019 International Conference on Robotics and
Automation (ICRA). IEEE, 2019, pp. 818–824.
[20] J. A. Rossiter and M. S. Aftab, “Recent developments in tuning
methods for predictive functional control, Processes, vol. 10, no. 7,
p. 1398, 2022.
[21] A. Nagabandi, K. Konolige, S. Levine, and V. Kumar, “Deep dynamics
models for learning dexterous manipulation,” in Conference on Robot
Learning. PMLR, 2020, pp. 1101–1112.
[22] J. Lloyd and N. F. Lepora, “Goal-driven robotic pushing using tactile
and proprioceptive feedback, IEEE Transactions on Robotics, vol. 38,
no. 2, pp. 1201–1212, 2021.
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
In the dynamic and unstructured environment where horticultural crops grow, obstacles and interference frequently occur but are rarely addressed, which poses significant challenges for robotic harvesting. This work proposed a tactile-enabled robotic grasping method that combines deep learning, tactile sensing, and soft robots. By integrating fin-ray fingers with embedded tactile sensing arrays and customized perception algorithms, the robot gains the ability to sense and handle branch interference during the harvesting process and thus reduce potential mechanical fruit damage. Through experimental validations, an overall 83.3–87.0% grasping status detection success rate, and a promising interference handling method have been demonstrated. The proposed grasping method can also be extended to broader robotic grasping applications wherever undesirable foreign object intrusion needs to be addressed.
Article
Full-text available
Robotic manipulation refers to how robots intelligently interact with the objects in their surroundings, such as grasping and carrying an object from one place to another. Dexterous manipulating skills enable robots to assist humans in accomplishing various tasks that might be too dangerous or difficult to do. This requires robots to intelligently plan and control the actions of their hands and arms. Object manipulation is a vital skill in several robotic tasks. However, it poses a challenge to robotics. The motivation behind this review paper is to review and analyze the most relevant studies on learning-based object manipulation in clutter. Unlike other reviews, this review paper provides valuable insights into the manipulation of objects using deep reinforcement learning (deep RL) in dense clutter. Various studies are examined by surveying existing literature and investigating various aspects, namely, the intended applications, the techniques applied, the challenges faced by researchers, and the recommendations adopted to overcome these obstacles. In this review, we divide deep RL-based robotic manipulation tasks in cluttered environments into three categories, namely, object removal, assembly and rearrangement, and object retrieval and singulation tasks. We then discuss the challenges and potential prospects of object manipulation in clutter. The findings of this review are intended to assist in establishing important guidelines and directions for academics and researchers in the future.
Article
Full-text available
Predictive functional control (PFC) is a popular alternative to PID because it exploits model information better and enables systematic constraint handling while also being cheap and computationally efficient. A recent overview paper reviewed some recent proposals for improving the tuning efficacy. This paper extends and develops upon that review paper by introducing some exciting new proposals for how to making tuning more intuitive and, thus, easier for unskilled operators. Moreover, there are early indications that these proposals are easily modified for use in nonlinear cases while maintaining a very low cost and a simple and fast online computation.
Conference Paper
Full-text available
Tactile predictive models can be useful across several robotic manipulation tasks, e.g. robotic pushing, robotic grasping, slip avoidance, and in-hand manipulation. However, available tactile prediction models are mostly studied for image-based tactile sensors and there is no comparison study indicating the best performing models. In this paper, we presented two novel data-driven action-conditioned models for predicting tactile signals during real-world physical robot interaction tasks (1) action condition tactile prediction and (2) action conditioned tactile-video prediction models. We use a magnetic-based tactile sensor that is challenging to analyse and test state-of-the-art predictive models and the only existing bespoke tactile prediction model. We compare the performance of these models with those of our proposed models. We perform the comparison study using our novel tactile enabled dataset containing 51,000 tactile frames of a real-world robotic manipulation task with 11 flat-surfaced household objects. Our experimental results demonstrate the superiority of our proposed tactile prediction models in terms of qualitative, quantitative and slip prediction scores.
Article
Full-text available
Intelligent robots for fruit harvesting have been actively developed over the past decades to bridge the increasing gap between feeding a rapidly growing population and limited labour resources. Despite significant advancements in this field, widespread use of harvesting robots in orchards is yet to be seen. To identify the challenges and formulate future research and development directions, this work reviews the state-of-the-art of intelligent fruit harvesting robots by comparing their system architectures, visual perception approaches, fruit detachment methods and system performances. The potential reasons behind the inadequate performance of existing harvesting robots are analysed and a novel map of challenges and potential research directions is created, considering both environmental factors and user requirements.
Conference Paper
Full-text available
During postharvest storage, fruit firmness usually decreases due to respiration and bruise, the former of which indicates the fruit ripeness while the latter negatively influence consumers’ taste preference. This paper presents a portable and low-cost device using vision-based tactile information to evaluate fruit firmness in a non-destructive manner. The device consists of a camera, LED lights, and a soft sensing layer with small bumps to capture detailed tactile information of the fruit. Two working modes are designed and a CNN-LSTM architecture is developed to relate the tactile information to fruit overall firmness or detect local firmness distortion. According to the experimental results, an R^2 up to 92.9% was achieved for the evaluation of the overall firmness of Cuixiang kiwifruit, and accuracy of 98.0% was obtained for the detection of local firmness distortion of Fuji apples. These results demonstrate the efficacy of the proposed solution to evaluate fruit firmness featuring high precision, and its non-destructive and potable nature is also anticipated to be favorable by the fresh fruit market.
Article
Continuum dexterous manipulators (CDMs) are suitable for performing tasks in a constrained environment due to their high dexterity and maneuverability. Despite the inherent advantages of CDMs in minimally invasive surgery, real-time control of CDMs’ shape during non-constant curvature bending is still challenging. This study presents a novel approach for the design and fabrication of a large deflection fiber Bragg grating (FBG) shape sensor embedded within the lumens inside the walls of a CDM with a large instrument channel. The shape sensor consisted of two fibers, each with three FBG nodes. A shape-sensing model was introduced to reconstruct the centerline of the CDM based on FBG wavelengths. Different experiments, including shape sensor tests and CDM shape reconstruction tests, were conducted to assess the overall accuracy of the shape sensing. The FBG sensor evaluation results revealed the linear curvature-wavelength relationship with the large curvature detection of 0.045 mm and a high wavelength shift of up to 5.50 nm at a 90° bending angle in both bending directions. The CDM’s shape reconstruction experiments in a free environment demonstrated the shape tracking accuracy of 0.216±0.126 mm for positive/negative deflections. Also, the CDM shape reconstruction error for three cases of bending with obstacles were observed to be 0.436±0.370 mm for the proximal case, 0.485±0.418 mm for the middle case, and 0.312±0.261 mm for the distal case. This study indicates the adequate performance of the FBG sensor and the effectiveness of the model for tracking the shape of the large-deflection CDM with nonconstant-curvature bending for minimally-invasive orthopaedic applications.
Conference Paper
Studies have shown that picking techniques play an important role in determining fruit quality at harvest (e.g. bruising, stem retention, etc). When picking fruit such as apples and pears, professional pickers use active perception, incorporating both visual and tactile input about fruit orientation, stem location, and the fruit’s immediate surroundings. This combination of tactile, visual, and force feedback is what enables human workers to execute dynamic movements that quickly and efficiently remove fruit from the tree without damage. However, much of the prior work on robotic fruit picking has formulated the harvesting problem as a position-control problem, using visual feedback for closed-loop end-effector placement while disregarding feedback on physical contact. As a first step towards more intelligent fruit picking — combining proprioception, localized sensing, and observed forces — we have developed a custom end-effector with multiple in-hand sensors, including tactile sensors on the fingertips. This paper presents the mechatronic design of the device as well as results from multiple outdoor picking trials with a Honeycrisp apple tree. Preliminary results show that, with multi-modal sensing, fruit slip, fruit separation from the tree, and fruit release from the hand can be detected.
Article
In robots, nonprehensile manipulation operations such as pushing are a useful way of moving large, heavy, or unwieldy objects, moving multiple objects at once, or reducing uncertainty in the location or pose of objects. In this study, we propose a reactive and adaptive method for robotic pushing that uses rich feedback from a high-resolution optical tactile sensor to control push movements instead of relying on analytical or data-driven models of push interactions. Specifically, we use goal-driven tactile exploration to actively search for stable pushing configurations that cause the object to maintain its pose relative to the pusher while incrementally moving the pusher and object toward the target. We evaluate our method by pushing objects across planar and curved surfaces. For planar surfaces, we show that the method is accurate and robust to variations in initial contact position/angle, object shape, and start position; for curved surfaces, the performance is degraded slightly. An immediate consequence of our work is that it shows that explicit models of push interactions might be sufficient but are not necessary for this type of task. It also raises the interesting question of which aspects of the system should be modeled to achieve the best performance and generalization across a wide range of scenarios. Finally, it highlights the importance of testing on nonplanar surfaces and in other more complex environments when developing new methods for robotic pushing.