Content uploaded by Amir Ghalamzan
Author content
All content in this area was uploaded by Amir Ghalamzan on Sep 07, 2023
Content may be subject to copyright.
Available via license: CC BY-SA 4.0
Content may be subject to copyright.
Content uploaded by Kiyanoush Nazari
Author content
All content in this area was uploaded by Kiyanoush Nazari on Aug 06, 2023
Content may be subject to copyright.
Deep Functional Predictive Control (deep-FPC):
Robot Pushing 3-D Cluster using Tactile Prediction
Kiyanoush Nazari1, Gabriele Gandolfi2, Zeynab Talebpour, Vishnu Rajendran3,
Willow Mandil1, Paolo Rocco2and Amir Ghalamzan-E.3
Abstract— This paper introduces a novel approach to address
the problem of Physical Robot Interaction (PRI) during robot
pushing tasks. The approach uses a data-driven forward model
based on tactile predictions to inform the controller about
potential future movements of the object being pushed, such as
a strawberry stem, using a robot tactile finger. The model is
integrated into a Deep Functional Predictive Control (d-FPC)
system to control the displacement of the stem on the tactile
finger during pushes. Pushing an object with a robot finger
along a desired trajectory in 3D is a highly nonlinear and
complex physical robot interaction, especially when the object
is not stably grasped. The proposed approach controls the stem
movements on the tactile finger in a prediction horizon. The
effectiveness of the proposed FPC is demonstrated in a series
of tests involving a real robot pushing a strawberry in a cluster.
The results indicate that the d-FPC controller can successfully
control PRI in robotic manipulation tasks beyond the handling
of strawberries. The proposed approach offers a promising
direction for addressing the challenging PRI problem in robotic
manipulation tasks.
I. INTRODUCTION
In the field of Physical Robot Interaction (PRI), successful
manipulation tasks rely on accurate interaction models that
utilise rich sensory information and intelligent control strate-
gies. While advanced learning from demonstration methods,
such as deep Movement Primitives (MPs) [1] and Deep
Probabilistic MPs [2], plan the robot’s movements given
an image of the robot’s workspace, they are not effective
for PRI tasks. Tactile feedback is a particularly effective
sensing modality for such cases especially when vision-based
control, such as visual servoing [3], is not feasible due to
occlusion [4]. For example, pushing a ripe strawberry that
is occluded by plant stems, leaves, or unripe fruits in a
cluster [5] can require tactile feedback for effective control.
Effective object manipulation under uncertainty [6], pre-
grasp manipulation to align the object with the fingers [7],
and agile soccer ball pushing by a mobile robot [8] are a few
examples of pushing. Analytical models for pushing require
complete knowledge of the environment, including physical
and geometric properties such as object pose, shape [9],
friction parameters, and mass. Developing analytical models
for unstructured environments characterised by high degrees
of freedom, non-linearity, and stochasticity, such as the case
of pushing a flexible stem to reach a strawberry, can be a
challenging task [10].
1School of Computer Science, University of Lincoln, UK. 2DEIB,
Politecnico di Milano, Italy. 3Lincoln Institute for Agri-food Technology,
University of Lincoln, UK. This work was partially supported by EPSRC
(AgriFoRwArdS) Grant reference: EP/S023917/
Fig. 1: Strawberry pushing setup: a Franka Emika robotic
arm is pushing a cluster of strawberries from right to left
where the nearest strawberry stem comes in contact with its
tactile finger. (Top Right) the robot at the beginning of the
pushing action, (Top Left) the robot, and the cluster at the end
of pushing. Sample tactile finger images are shown from right
to left for the initial and final point in the pushing trajectory,
respectively. Our tactile finger design features a deformable
half-conic membrane with an integrated miniature camera
and LED light.
Most existing pushing methods are designed for 2-D
scenarios in which an object is moving on a flat surface, but
in the case of strawberry picking, a 3-D pushing scenario is
more relevant [11]. Pushing a strawberry in a 3-dimensional
space is more challenging than pushing an object on a
table (i.e. a 2D problem). While interactive movement prim-
itives [12] can be used to plan pushing actions, an accurate
interaction model is crucial for effectively controlling the
planned motion of the strawberry during pushing in this
scenario.
In this paper, we presented a novel deep functional pre-
dictive control pipeline for the manipulation of strawberries
grown on a table. Our pipeline consists of three key modules:
a deep action-conditioned Tactile Forward Model (TFM), a
deep Contact Localisation Model (CLM), and an online deep
Functional Predictive Control (d-FPC) to generate control
actions. We collected a dataset of plastic strawberries being
pushed in our lab setting to train TFM, which is the state-
of-the-art tactile prediction model. We also trained CLM
to calibrate our tactile sensor using a dataset of strawberry
pushing. Finally, d-FPC uses real-time predictions from TFM
and CLM to generate robot actions based on future error
signal estimations to control the stem pose on the sensor
surface. We compared our proposed functional predictive
controller’s performance with a PD control-based system that
only uses CLM and demonstrated that the predictive system
outperforms this baseline model. This study addresses the
challenge of pushing flexible objects in 3D, and to the best
of our knowledge, this is the first study to do so. Our results
demonstrate the effectiveness of our proposed approach and
pave the way for future research in the manipulation of
flexible objects using deep functional predictive control.
II. REL ATE D WORKS
Cluster manipulation in fruit harvesting is a challenging
task from both motion planning and motion control per-
spectives [14], [12]. One of the challenges is avoiding the
slip of a grasped object, which can be addressed through
closed-loop robot trajectory adaptation [15]. Deformable
object manipulation, such as cloth, has been modelled using
simplified mass-spring models or 3D mesh generation [16],
while heuristic feature spaces have been used for flexible
cable manipulation with dual robot arms [17]. However,
analytical modelling methods are limited to specific object
sets and are not scalable to larger object and action sets.
In contrast, our proposed approach uses a time-series model
for action-conditioned tactile prediction for pushing control
which can be applied in unstructured settings without the
knowledge about the model of the individual objects.
Tactile feedback is mostly used for grasp control in robotic
object clutter manipulation [4] and detecting a grip on
fruits [18]. However, the use of tactile sensors has been
limited to grip control and has not been applied for any
cluster manipulation. In our work, we exploit tactile feedback
for trajectory-level control for pushing a flexible plant stem.
Tactile prediction models are used for controlling ma-
nipulation tasks, from the simple task of rolling a marble
on a table [19] to the complex task of slip control [15].
The core of such controllers is a forward model that can
generate predicted tactile readings (i.e. tactile images). For
instance, action-conditioned tactile predictive models are
utilised with a taxel-based tactile sensor in pick and place
tasks [13], demonstrating the approach performs well only
for flat surface objects.
Our approach uses a time-series model for tactile predic-
tion based on [13]. We form a deep Predictive Functional
Control (d-FPC) [20] which enables the robot to control
the strawberry pushing actions. Deep models have been
extensively used for learning lower dimensional state spaces
for Model Predictive Control (MPC). These methods have
also been used for learning visual dynamic models for
control [21]. In a simplified task of rolling a mini sphere,
the tactile prediction was used in an MPC controller [19].
We form a Proportional-Derivative (PD) control over the
error in the prediction horizon to control the contact state
of a flexible object on a robot hand. Unlike previous work
that used trajectory adaptation to minimise the likelihood of
predicted binary slip signal in a prediction horizon [15], our
model learns the complex contact behaviour and generates
actions to control the movements of the stem on the tactile
finger to keep it stable.
III. METHODOLOGY
a) Camera-based tactile sensor: We use a customised
camera-based tactile sensor for pushing strawberries similar
to Tactip [22]. This sensor has a camera and an LED light
looking at a deformable membrane with embedded white
markers (Fig. 1). The applied pressure on the sensor yields
a deformation that is captured by the camera. The sharp tip
and low membrane stiffness make this tactile sensor suitable
for the task of strawberry cluster manipulation.
b) Contact Localisation Model (CLM): The motions of
the marker array printed on the sensor are indicative of the
magnitude and location of the applied force. For the current
problem setting, we are more interested in force localisation
for doing stem contact state control. To find the mapping
from raw tactile images to contact location in 1-dimensional
space, we use a Convolutional Neural Network with the
architecture shown in Fig.2 (red box). CLM consists of two
convolutional and three dense layers. The output of CLM
is the distance of the contact force from the sensor camera
lens along the sensor conic axis. The data set for training
CLM consists of applying forces to the fixed sensor by a
rod (mimicking strawberry stem) attached to the robot end-
effector (EE) with a 5mm distance step. At each step, the
robot applies forces on the membrane toward the sensor base
by a 1mm penetration step. Overall, 150 pushing samples in
10 locations are collected to train CLM.
c) Tactile Forward Model (TFM): Tactile prediction
aims to estimate future tactile images based on a set of
previous tactile images x0, ..., xc−1obtained from physical
interactions, where cis the length of the context window.
Specifically, the objective is to sample from the conditional
distribution p(xc:T|x0:c−1), where xidenotes the ith tactile
image in the sequence and Tis the sum of the context
window length and the prediction horizon length.
Since the robot’s actions alter the environment during
physical interaction, we incorporate action conditioning
to predict tactile sensation more accurately. The action-
conditioned tactile prediction problem is formulated as pre-
dicting the future tactile images xc:Tgiven a sequence of
previous robot actions a0:c−1, previous tactile images x0:c−1,
and a sequence of future/planned robot actions/trajectory
ac:T. Here, a robot action, a∈R6, refers to the end-
effector task space position and orientation (Euler angles)
with respect to the robot base, while a tactile image is
represented by x∈R64×64×3, which captures the surface
deformation caused by the applied force. The conditional
distribution will be:
p(xc:T|x0:c−1,a0:T)(1)
Factorising this we can define the model as
ΠT
t=cpθ(xt|x0:t−1,a0:t). Learning now involves training
the parameters of the factors θ.
Contact
Localisation
Model
Tactile Forward Model
Contact
Localisation
Model
d-FPC
Conv2D
ReLu
MaxPool
Flatten
GeMM
ReLu
x 2 x 2
CLM
x 2x 2
Skip Connection
ConvLSTM
Conv2D
ReLu
MaxPool
Conv2D
ReLu
UpSample
ConvLSTM
Tanh
TFM
Fig. 2: The block diagram of the proposed data-driven functional predictive control for pushing strawberries. The model
consists of (1) tactile forward model (TFM) which is based on [13], contact localisation model (CLM), and the functional
predictive controller (d-FPC) that generates future actions resulting in the minimum stem displacement on the tactile finger.
The model architecture is depicted in Fig.2 (blue box).
We extract scene features from the input tactile image by
convolutional filters in the first two layers of the network
as the encoder. Each convolution operation is followed by
the Relu activation function and 2D maxpooling operations.
Robot action sequences are concatenated with latent tactile
features after the convolutional layers. These latent space
features with downsampled width and height and a larger
number of channels are fed to the Conv-LSTM chain.
These layers process the spatiotemporal dependencies among
the latent features. After this point, we need to upscale
the features to reach the tactile image size. As such, two
convolutional layers, each one followed by Relu activation
and 2D upsampling, are applied to ConvLSTM outputs. To
apply the pixel motion changes to the input, we use the
skip connection for the input tactile image and apply tanh
activation to construct the next tactile images in the sequence.
d) deep-Functional Predictive Control (d-FPC): We
denote the predicted stem location (from CLM) on the sensor
at time tby st. The goal of our d-FPC is to control the
stem displacement on the tactile finger. Hence, this allows
the robot to keep the contact fixed with the strawberry
stem during pushing actions and avoid the contact location
approaching the tip or the base of the sensor. These are sensor
surface boundary zones and approaching them increases the
probability of losing contact with the stem. We use the stem-
finger contact point at time tas the reference for our d-FPC
controller. We define an error signal as the distance of the
contact point from the reference point:
ei,t = ˆsi−st, i =c, ..., T (2)
where ˆsiis the predicted stem location for a sequence of
planned robot movements. We formulate our d-FPC over the
error signal as follows:
at,res =−X
i=c:T
(kpi×ei,t +kdi×˙ei,t)(3)
where at,res is the residual action value to be added to the
reference trajectory at,ref to generate the control action At.
Atis a rotational velocity around the contact line axis. Fig.2
(green box) shows the schematic of the d-FPC. The generated
control output is a rotational velocity proportional to the
distance of the stem from the reference line. The derivative
term avoids overshooting and having large instant rotations.
IV. EXP ERI MEN TAL SETUP AND DATA SE T
a) Tactile sensor and manipulation task: Various types
of tactile sensors are discussed in the literature, including
in [4]. In this work, we use a custom-made camera-based
tactile sensor that has a half-conic geometry and a tapered
tip (shown in Fig.1) designed to allow for easier penetration
among stems and fruits, providing valuable tactile feedback.
The deformable membrane of the sensor is 3D-printed and
dot features are printed with a linear pattern on its conic
inner surface. Changes in the marker pattern resulting from
contact forces provide information about contact force value,
geometry, and location. The camera, which is located on
the sensor base, and the LED, used for illuminating the
markers, are powered by an onboard Raspberry Pi, and
tactile images are transmitted at a frequency of 60 Hz. The
sensor is mounted on a Franka Emika gripper, providing an
effective and versatile tool for physical interaction in a range
of applications.
We have collected the data from a series of strawberry-
pushing tasks in 3-D. The pushing dataset includes data
for single strawberry pushing and pushing a cluster of
strawberries. To simulate the table-top strawberry growing
scenario, we attached each plastic strawberry to a thin wire
that makes a nonlinear elastic behaviour similar to those
usually observed in tabletop-grown strawberries. To simulate
realistic tactile feedback, we added knots on the stalk of
each strawberry (Fig.1) and injected silicone to increase their
weight (each strawberry weighs c. 20 g to 30 g).
We generate the pushing trajectories for the training data
collection phase by two methods: first by Pilz industrial
motion planner by specifying initial and target robot poses,
and second by defining a minimum time reference trajectory
using the robot’s Cartesian velocity controller. We use the
second method to be able to regenerate comparably similar
trajectories in test time, as opposed to the first case where
trajectories are generated by the motion planning library.
Trajectories include linear and circular motion patterns to
perform the pushing tasks. Arc trajectories were used to
collect more tactile-conditioned robot movements, where the
finger followed the motion of the pushed stem/strawberry.
These pushes started at a position p0and orientation q0,
followed an arc trajectory, and ended at a final position
pfwith a value of zcoordinate larger than initial position.
The final orientation qfis selected to maintain contact with
the elements pushed. The pushing actions were performed
from right to left and vice versa, and they involved single or
multiple stems (Fig. 1), generating greater deformations on
the membrane.
We collected a total of 430 mixed linear/circular motion
tasks containing (i) tactile images from the finger at 60 Hz
and (ii) robot state data sampled at 1000 Hz, representing
the position and orientation of the end effector in the
planned trajectory. These readings were synchronised using
the ROS ApproximateTime policy and fed into the tactile
forward model both in training and test times.
Considering the robot’s motion, slip occurred mainly on
the width and length of the finger but could also happen in
other directions depending on the motion of the stems during
the pushing actions.
V. RESULTS AND DISCUSSION
We test the performance of our proposed control pipeline
in real-time on pushing tasks of strawberry stems and
compare the performance with a baseline controller and an
open-loop system. The tactile sensor is mounted on Franka
Emika robot connected to a PC with Intel® Core™ i7-8700K
CPU @ 3.70GHz × 12 and 64GB RAM running Ubuntu
20.04 and ROS Noetic. Torch library is used for offline
training and online testing of the neural network models. Test
manipulation tasks consist of performing pushing trajectories
with linear and circular motion patterns using the robot’s
Cartesian velocity controller.
Performance metrics include: (I) Stem max displacement
and (II) the number of stem slip instances on the sensor
surface. If we denote stem location at time tby siwhere
i∈(0,1, ..., T )for a pushing trial, metric (I) is defined as the
absolute value of the difference of maximum and minimum
stem location in a trial |max(si)−min(si)|:i= 1, ..., T .
Metric (II) is defined as the number of time steps where the
differential values ˙siwere larger than threshold γ. While
metric (I) shows full stem displacement, metric (II) shows
the stem’s sudden large motion instances or slippage on the
sensor surface. We also present the area under the curve of
stem displacement and generated action. We repeat each test
case 5 times and present the mean and standard deviation
of the metric values. Overall we conducted 100 test-pushing
trials.
To evaluate the effectiveness of d-FPC for pushing control,
we compare the control performance with a PD control-based
tactile servoing system as the baseline model. Both models’
results are presented against the open-loop system with a
pre-specified reference trajectory.
In this paper, we utilise a minimum-time reference trajec-
tory (such as bang-bang) for the open-loop system, although
any desired reference trajectory can be used. To make valid
comparisons among trials, we consider three initial contact
zones for the stem including Zone-1 where the contact point
is between the middle and tip of the sensor, Zone-2 has the
contact point between the middle and base of the sensor, and
Zone-3 where the contact point is close to sensor centre line.
Since the tactile sensor has varying deformation limits across
its conic axis we compare the trials with corresponding initial
contact zones together.
We conduct a comparison test with a one-degree-of-
freedom (DOF) horizontal pushing along the Y-axis of the
robot’s base frame. Both PD and d-FPC controllers generate
control actions for the robot hand’s rotation around the
contact line to prevent stem slip on the sensor surface.
The results are presented in Table I, where test cases are
conducted separately for each initial contact zone. Both
PD and d-FPC controllers decrease the stem’s maximum
displacement. We observe that d-FPC outperforms the PD
controller for Zone-1 and Zone-3, but PD shows better
performance for Zone-2 very close to the sensor base. This
is because the sensor has its largest deformation limit in the
Base zone, resulting in relatively large initial deformation
after making contact, making it difficult for TFM to predict
future stem states. The prediction of the error signal helps
d-FPC to have more reaction time than PD.
We find that d-FPC is the most effective controller to
reduce the number of stem slip instances, with the smallest
area under the curve of displacement compared to the PD
controller. We also present the computation time to show
the relative computation complexity of each system. Since
d-FPC has two stacked deep models, the computation time
is larger than the PD controller.
To compare the performance of different controllers in a
qualitative manner, we present the stem location obtained in
two trials (shown in Fig. 3a): Trial-1, where the stem-finger
initial contact point is in Zone-1, is shown with solid lines,
and Trial-2, with the contact point in Zone-2, is shown with
a dashed line. Our results show that d-FPC outperforms PD
controller and open loop in maintaining the stem contact,
resulting in the smallest displacement of the stem. Further-
more, Fig. 3b shows the control actions generated by each
Model Contact
zone
Stem max
disp.
Stem slip
instances
Disp.
integral
Action
integral
Comp.
time (ms)
Open-loop
1 0.80 ±0.2 31.23 ±4.3 0.83 ±0.1 - -
2 1.35 ±0.2 50.19 ±5.7 0.91 ±0.1 - -
3 0.91 ±0.1 39.83 ±3.2 0.86 ±0.2 - -
PD
1 0.65 ±0.1 27.2 ±6.5 0.75 ±0.1 2.93 ±0.7 18.73 ±2
20.36 ±0.0 10.2 ±2.4 0.48 ±0.0 5.12 ±3.8 20.30 ±1
3 0.63 ±0.1 24.2 ±1.6 0.47 ±0.1 9.73 ±5.4 19.73 ±1
d-FPC
10.20 ±0.0 5.0 ±1.2 0.12 ±0.0 3.74 ±0.8 60.49 ±6
2 0.43 ±0.0 7.2 ±0.7 0.18 ±0.0 4.27 ±1.2 55.02 ±2
30.25 ±0.1 6.0 ±0.6 0.09 ±0.0 4.57 ±2.4 58.54 ±3
TABLE I: Control performance for the PD and d-FPC in pushing a single strawberry along a linear trajectory.
Model Robot
trajectory
Stem max
disp.
Stem slip
instances
Disp.
integral
Action
integral
Open-loop Linear 1.21 ±0.18 44.38 ±10.3 0.88 ±0.4 -
Circular 1.35 ±0.46 48.18 ±5.2 1.02 ±0.5 -
PD Linear 0.58 ±0.21 25.53 ±4.2 0.63 ±0.1 5.39 ±6.2
Circular 1.20 ±0.01 17.6 ±2.0 0.44 ±0.0 9.89 ±0.8
d-FPC Linear 0.29 ±0.04 8.11 ±1.4 0.13 ±0.0 4.49 ±2.5
Circular 0.54 ±0.05 5.0 ±1.5 0.22 ±0.0 6.66 ±0.8
TABLE II: Comparison of the controllers in linear and circular pushing trajectories.
Model Robot
trajectory
Stem max
disp.
Stem slip
instances
Disp.
integral
Open-loop Linear 1.43 ±0.30 49.33 ±15.64 1.39 ±0.33
Circular 1.29 ±0.67 47.98 ±6.33 1.19 ±0.23
PD Linear 0.79 ±0.21 29.4 ±6.52 0.66 ±0.21
Circular 1.14 ±0.23 20.5 ±2.69 0.56 ±0.84
d-FPC Linear 0.31 ±0.08 17.1 ±2.39 0.25 ±0.03
Circular 0.61 ±0.11 9.5 ±4.58 0.27 ±0.18
TABLE III: Controller and open loop performances for Pushing a cluster of strawberries.
controller. We observe that d-FPC generates actions of larger
magnitude in Trial-1 because the likelihood of losing the
stem in Zone-1 (namely closer to the tip) is larger than in
Zone-2. In Trial-2, the magnitude of d-FPC and PD controller
actions is similar since the contact between the stem and
sensor membrane is tighter due to a larger deformation of
the sensor closer to the sensor base.
We test the performance of the systems in a three DOF
task with a bang-bang reference for translation along Y,
Z, and rotation Wxof Cartesian velocity space. This is a
more challenging task because the robot wrist will rotate
45 degrees along the pushing trajectory which causes larger
deformation of the stem and more slip instances. Based on
Table II d-FPC is the most effective controller in decreasing
the stem displacement and slip instances. PD has a smaller
improvement in max displacement for the circular motion
than the linear motion compared to the open-loop system.
This indicates that not having enough reaction time in this
task can lead to failure in achieving the control objective.
We test the generalisation performance of the pushing
controller when pushing a stem in a cluster of strawberries.
In this task additional to the target stem, other stems, leaves,
or strawberries come into contact with the sensor which
makes both tactile prediction and control more challenging.
Table III shows the results for pushing a stem in a cluster.
Although the control performance of PD and d-FPC degrades
compared to pushing an isolated stem, both systems improve
the performance metrics relative to the open-loop system.
Fig.3c shows cluster pushing results for sample trials
of linear and circular pushing trajectories. For the linear
push, PD has slight improvement compared to the open-
loop system but d-FPC reduces stem displacement more
effectively. For the circular push, while the open-loop system
loses contact with the stem because of large stem slippage
in the last part of the trial, both PD and d-FPC reduce the
stem displacement to avoid large slips. d-FPC keeps the
displacement more bounded relative to the PD controller
does.
(a) stem pose (b) control action (c) strawberry cluster pushing
Fig. 3: Comparison of control performance between d-FPC, open loop, and PD controller in maintaining the location of the
stem constant on the finger surface (Trial-1 (T1) solid and Trial-2 (T2) dashed lines) (a) At time 0.85 s, the stem makes
contact with the tactile finger and the controllers activate. d-FPC can maintain the stem contact point during the pushing
action, while the open loop result shows the stem moving out of the tactile finger surface. (b) The magnitude of the control
input shows d-FPC provides larger wrist rotation to avoid stem contact displacement. (c) Strawberry cluster pushing results.
VI. CONCLUSION
We presented a novel deep Predictive Functional Control
(d-PFC) framework to control the contact location of a straw-
berry stem on our tactile finger. Our proposed method lever-
ages a time-series model for generating action-conditioned
tactile predictions and a convolutional neural network (CNN)
model converting the tactile images to contact location. We
demonstrated the effectiveness of our approach through a
series of experiments with a Franka Emika robot and a
customised tactile finger, showing that our model can learn
complex contact behaviours and generate actions to control
the movements of flexible objects to keep them stable,
e.g. pushing a cluster of strawberries. Overall, our work
highlights the potential of deep learning-based approaches
in addressing the challenges of tactile sensing-based manip-
ulation tasks and lays the foundation for future research in
this field.
REFERENCES
[1] O. Sanni, G. Bonvicini, M. A. Khan, P. C. L´
opez-Custodio, K. Nazari
et al., “Deep movement primitives: toward breast cancer examination
robot,” in Proceedings of the AAAI Conference on Artificial Intelli-
gence, vol. 36, no. 11, 2022, pp. 12 126–12 134.
[2] A. Tafuro, B. Debnath, A. M. Zanchettin et al., “dpmp-deep prob-
abilistic motion planning: A use case in strawberry picking robot,”
arXiv preprint arXiv:2208.09074, 2022.
[3] S. S. Mehta, W. MacKunis, and T. F. Burks, “Robust visual servo
control in the presence of fruit motion for robotic citrus harvesting,”
Computers and Electronics in Agriculture, vol. 123, pp. 362–375,
2016.
[4] A. Yamaguchi and C. G. Atkeson, “Recent progress in tactile sensing
and sensors for robotic manipulation: can we turn tactile sensing into
vision?” Advanced Robotics, vol. 33, no. 14, pp. 661–673, 2019.
[5] Y. Xiong, C. Peng, L. Grimstad, P. J. From, and V. Isler, “Development
and field evaluation of a strawberry harvesting robot with a cable-
driven gripper,” Computers and electronics in agriculture, vol. 157,
pp. 392–402, 2019.
[6] M. R. Dogar and S. S. Srinivasa, “Push-grasping with dexterous hands:
Mechanics and a method,” in 2010 IEEE/RSJ International Conference
on Intelligent Robots and Systems. IEEE, 2010, pp. 2123–2130.
[7] J. E. King, M. Klingensmith, C. M. Dellin, M. R. Dogar, P. Velagapudi,
N. S. Pollard, and S. S. Srinivasa, “Pregrasp manipulation as trajectory
optimization.” in Robotics: Science and Systems. Berlin, 2013.
[8] R. Emery and T. Balch, “Behavior-based control of a non-holonomic
robot in pushing tasks,” in Proceedings 2001 ICRA. IEEE International
Conference on Robotics and Automation (Cat. No. 01CH37164),
vol. 3. IEEE, 2001, pp. 2381–2388.
[9] G. Amirkhani, A. Goodridge, M. Esfandiari, H. Phalen, J. H. Ma,
I. Iordachita, and M. Armand, “Design and fabrication of a fiber
bragg grating shape sensor for shape reconstruction of a continuum
manipulator,” IEEE Sensors Journal, 2023.
[10] M. Q. Mohammed, L. C. Kwek, S. C. Chua, A. Al-Dhaqm, S. Na-
havandi, T. A. E. Eisa, M. F. Miskon, M. N. Al-Mhiqani, A. Ali,
M. Abaker et al., “Review of learning-based robotic manipulation in
cluttered environments,” Sensors, vol. 22, no. 20, p. 7938, 2022.
[11] J. St¨
uber, C. Zito, and R. Stolkin, “Let’s push things forward: A survey
on robot pushing,” Frontiers in Robotics and AI, p. 8, 2020.
[12] S. Mghames, M. Hanheide, and A. Ghalamzan, “Interactive movement
primitives: Planning to push occluding pieces for fruit picking,” in
2020 IEEE/RSJ International Conference on Intelligent Robots and
Systems (IROS). IEEE, 2020, pp. 2616–2623.
[13] W. Mandil, K. Nazari et al., “Action conditioned tactile prediction: a
case study on slip prediction,” Robotics: Science and Systems, 2022.
[14] H. Zhou, X. Wang, W. Au, H. Kang, and C. Chen, “Intelligent robots
for fruit harvesting: Recent developments and future challenges,”
Precision Agriculture, vol. 23, no. 5, pp. 1856–1907, 2022.
[15] K. Nazari, W. Mandil, and A. M. G. Esfahani, “Proactive slip control
by learned slip model and trajectory adaptation,” in Conference on
Robot Learning. PMLR, 2023, pp. 751–761.
[16] V. E. Arriola-Rios, P. Guler, F. Ficuciello, D. Kragic, B. Siciliano, and
J. L. Wyatt, “Modeling of deformable objects for robotic manipulation:
A tutorial and review,” Frontiers in Robotics and AI, vol. 7, p. 82,
2020.
[17] J. Zhu, B. Navarro, P. Fraisse, A. Crosnier, and A. Cherubini, “Dual-
arm robotic manipulation of flexible cables,” in 2018 IEEE/RSJ
International Conference on Intelligent Robots and Systems (IROS).
IEEE, 2018, pp. 479–484.
[18] H. Zhou, H. Kang, X. Wang, W. Au, M. Y. Wang, and C. Chen,
“Branch interference sensing and handling by tactile enabled robotic
apple harvesting,” Agronomy, vol. 13, no. 2, p. 503, 2023.
[19] S. Tian, F. Ebert, D. Jayaraman, M. Mudigonda, C. Finn, R. Calandra,
and S. Levine, “Manipulation by feel: Touch-based control with deep
predictive models,” in 2019 International Conference on Robotics and
Automation (ICRA). IEEE, 2019, pp. 818–824.
[20] J. A. Rossiter and M. S. Aftab, “Recent developments in tuning
methods for predictive functional control,” Processes, vol. 10, no. 7,
p. 1398, 2022.
[21] A. Nagabandi, K. Konolige, S. Levine, and V. Kumar, “Deep dynamics
models for learning dexterous manipulation,” in Conference on Robot
Learning. PMLR, 2020, pp. 1101–1112.
[22] J. Lloyd and N. F. Lepora, “Goal-driven robotic pushing using tactile
and proprioceptive feedback,” IEEE Transactions on Robotics, vol. 38,
no. 2, pp. 1201–1212, 2021.