Mir Feroskhan’s research while affiliated with Nanyang Normal University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (56)


Dynamics-Driven Visual Servoing of Over-Actuated Quadrotors
  • Conference Paper

December 2024

·

3 Reads

Archit Krishna Kamath

·

·

Mir Feroskhan

Learning Multi-Pursuit Evasion for Safe Targeted Navigation of Drones

December 2024

·

26 Reads

·

6 Citations

IEEE Transactions on Artificial Intelligence

Safe navigation of drones in the presence of adversarial physical attacks from multiple pursuers is a challenging task. This paper proposes a novel approach, asynchronous multistage deep reinforcement learning (AMS-DRL), to train adversarial neural networks that can learn from the actions of multiple evolved pursuers and adapt quickly to their behavior, enabling the drone to avoid attacks and reach its target. Specifically, AMS-DRL evolves adversarial agents in a pursuit-evasion game where the pursuers and the evader are asynchronously trained in a bipartite graph way during multiple stages. Our approach guarantees convergence by ensuring Nash equilibrium among agents from the game-theory analysis. We evaluate our method in extensive simulations and show that it outperforms baselines with higher navigation success rates. We also analyze how parameters such as the relative maximum speed affect navigation performance. Furthermore, we have conducted physical experiments and validated the effectiveness of the trained policies in real-time flights. A success rate heatmap is introduced to elucidate how spatial geometry influences navigation outcomes. Project website: https://github.com/NTU-ICG/AMS-DRL-for-Pursuit-Evasion .



Ensuring Safety in Target Pursuit Control: A CBF-Safe Reinforcement Learning Approach
  • Preprint
  • File available

November 2024

·

18 Reads

This paper addresses the pursuit control problem for multi-agent systems, aiming to ensure collision-free tracking under input saturation and external disturbances. We propose a novel Control Barrier Function (CBF)-Safe Reinforcement Learning (CSRL) algorithm, which integrates model-free reinforcement learning with a safety filter to guarantee system safety. The framework introduces an input-constrained CBF that dynamically adjusts control bounds, enabling robust target tracking even during evasive maneuvers. A safety filter is designed to transform unsafe RL actions into safe control signals by solving a Quadratic Program (QP), ensuring the safety for sensing, collision avoidance, and input constraints of pursuers. Theoretical analysis proves the feasibility of the CBF-QP using the Karush-Kuhn-Tucker (KKT) conditions. Simulation results validate the effectiveness of the CSRL algorithm, demonstrating its ability to handle complex pursuit scenarios while maintaining safety and improving control performance.

Download


The perching maneuver by Harris’ hawks and avian-inspired drone
a Schematic overview of a typical perching maneuver consisting of a dive phase (light blue line) and of an agile climb phase (blue line) (see also Supplementary Movie 1, adapted from ref. ²⁶); the illustrated bird shows the optimal control strategies displayed by the Harris' hawk (light brown) and by the avian-inspired drone (black/blue). b Comparison of Harris' hawks of the study²⁶ and the avian-inspired LisEagle drone¹⁷. We indicate the wing and tail sweep joints in red and list the mass, wing area, and wing span of the two.
Avian-inspired drone
a The avian-inspired drone, LisEagle, used in this work. b Illustration of its eight degrees of freedom. This study focuses on longitudinal motion, which is affected by symmetric sweeping of left and right wing, tail incidence, tail sweep, and thrust. The control of these degrees of freedom is given to the trajectory optimization method. Instead, lateral displacements induced by mechanical asymmetries of the drone feathers and actuator responses are stabilized by reactive Proportional-Derivative (PD) controllers where tail yaw correct yaw motion and asymmetric left and right wing twist correct roll motion.
Perching optimization in simulation
a Visualization of the simulation experiment constraints. The drone is initialized in a straight flight condition at a velocity (brown arrow) of 10 ms⁻¹ and varying distance from the target point (brown cross). The drone is required to reach the target point and is allowed to use up to 1 m of space downward, while its motion upwards is not restricted. The transition state is defined at the lowest point of the trajectory, where all the energy is kinetic. b Perching flight trajectories that result from the optimization algorithm with the objective of minimizing distance flown at high angle of attack²⁶. Starting points are in the range of 9–15 m, spaced in intervals of 1 m and are shown by an arrow. The climb phase is indicated by the shaded area, on the right of which, we show kinetic, potential, and dissipated energy at the target point. c Effect of a limited range of motion in wing and tail sweep on flight path and key characteristic metrics. The central panel shows the resulting flight paths, the left panel (E kinetic) displays the relative kinetic energy on impact, and the right panel (distance required) indicates the horizontal distance required for the perching maneuver (both surfaces linearly interpolated between experiments, indicated by black dots). d Drone trajectory during the climb phase in the morphological space defined by tail incidence, wing sweep, and tail sweep. The prominent 3-dimensional line shows the actuation sequence, with color change indicating time flow. The lighter lines depict the 2-dimensional projections on each parameter plane. The time scale at the bottom highlights the three distinct configurations of the drone during the climb phase. Source data are provided as a Source Data file.
Validation on drone
a Experimental setup consisting of a launcher accelerating the drone to the transition state to fly along the planned trajectory while being tracked by a motion-tracking system and eventually landing in a protective net for recovery. b The planned optimal trajectory, the trajectories from the 12 LisEagle flights, and the 537 Harris' Hawks flights²⁶. We align the optimal trajectory by aligning the point of impact of the Harris' Hawk with the corresponding point of the optimized drone trajectory. Beyond the impact point, each of the trajectories is shown in a faded color. c Energy ratios at impact of the simulated drone, the real drone, and Harris' hawks. Source data are provided as a Source Data file.
Qualitative comparison of the perching maneuver of the drone, simulation, and bird
a Perched flight trajectory of the drone. b Simulation of the optimal trajectory highlighting the three configurations of the climb phase (also shown with the same color scheme in Fig. 3d). On the top, the sequence is shown observed from a virtual camera at an angle matching the camera’s angle used to record the drone’s flight trajectory and at the bottom, from a virtual camera at an angle matching (c) the Harris' hawk flight recording²⁶. Each overlayed image in the sequence was captured 0.15 s apart.
Agile perching maneuvers in birds and morphing-wing drones

September 2024

·

374 Reads

·

3 Citations

Avian perching maneuvers are one of the most frequent and agile flight scenarios, where highly optimized flight trajectories, produced by rapid wing and tail morphing that generate high angular rates and accelerations, reduce kinetic energy at impact. While the behavioral, anatomical, and aerodynamic factors involved in these maneuvers are well described, the underlying control strategies are poorly understood. Here, we use optimal control methods on an avian-inspired drone with morphing wing and tail to test a recent hypothesis derived from perching maneuver experiments of Harris’ hawks that birds minimize the distance flown at high angles of attack to dissipate kinetic energy before impact. The resulting drone flight trajectories, morphing sequence, and kinetic energy distribution resemble those measured in birds. Furthermore, experimental manipulation of the wings that would be difficult or unethical with animals reveals the morphing factors that are critical for optimal perching maneuver performance of birds and morphing-wing drones.


NPE-DRL: Enhancing Perception Constrained Obstacle Avoidance with Non-Expert Policy Guided Reinforcement Learning

September 2024

·

94 Reads

·

2 Citations

IEEE Transactions on Artificial Intelligence

Obstacle avoidance under constrained visual perception presents a significant challenge, requiring rapid detection and decision-making within partially observable environments, particularly for unmanned aerial vehicles (UAVs) maneuvering agilely in three-dimensional space. Compared to traditional methods, obstacle avoidance algorithms based on deep reinforcement learning (DRL) offer a better comprehension of the uncertain operational environment in an end-to-end manner, reducing computational complexity and enhancing flexibility and scalability. However, the inherent trial-and-error learning mechanism of DRL necessitates numerous iterations for policy convergence, leading to sample inefficiency issues. Meanwhile, existing sample-efficient obstacle avoidance approaches that leverage imitation learning often heavily rely on offline expert demonstrations, which are not always feasible in hazardous environments. To address these challenges, we propose a novel obstacle avoidance approach based on Non-Expert Policy Enhanced DRL (NPE-DRL). This approach integrates a fundamental DRL framework with prior knowledge derived from a non-expert policy-guided imitation learning. During the training phase, the agent starts by online imitating the actions generated by the non-expert policy during interactions and progressively shifts toward autonomously exploring the environment to generate the optimal policy. Both simulation and physical experiments validate that our approach improves sample efficiency and achieves a better exploration-exploitation balance in both virtual and real-world flights. Additionally, our NPE-DRL-based obstacle avoidance approach shows better adaptability in complex environments characterized by larger scales and denser obstacle configurations, demonstrating a significant improvement in UAVs’ obstacle avoidance capability. Code available at https://github.com/zzzzzyh111/NonExpert-Guided-Visual-UAV-Navigation-Gazebo .</p


Learning Resilient Formation Control of Drones with Graph Attention Network

September 2024

·

39 Reads

The rapid advancement of drone technology has significantly impacted various sectors, including search and rescue, environmental surveillance, and industrial inspection. Multidrone systems offer notable advantages such as enhanced efficiency, scalability, and redundancy over single-drone operations. Despite these benefits, ensuring resilient formation control in dynamic and adversarial environments, such as under communication loss or cyberattacks, remains a significant challenge. Classical approaches to resilient formation control, while effective in certain scenarios, often struggle with complex modeling and the curse of dimensionality, particularly as the number of agents increases. This paper proposes a novel, learning-based formation control for enhancing the adaptability and resilience of multidrone formations using graph attention networks (GATs). By leveraging GAT's dynamic capabilities to extract internode relationships based on the attention mechanism, this GAT-based formation controller significantly improves the robustness of drone formations against various threats, such as Denial of Service (DoS) attacks. Our approach not only improves formation performance in normal conditions but also ensures the resilience of multidrone systems in variable and adversarial environments. Extensive simulation results demonstrate the superior performance of our method over baseline formation controllers. Furthermore, the physical experiments validate the effectiveness of the trained control policy in real-world flights.




Citations (32)


... Image correction is useful for target detection in distorted images [8,23,27,28], as it can enhance both detection success rate and accuracy. Recent research [27] integrates image correction into the camera calibration process in an iterative manner. ...

Reference:

Calibration Method for Ultra‐Wide FOV Fisheye Cameras Based on Improved Camera Model and SE(3) Image Pre‐Correction
Omnidrone-Det: Omnidirectional 3D Drone Detection in Flight
  • Citing Conference Paper
  • August 2024

... As an extreme example, perching on a vertical wall can enable the UAV to achieve a zero-distance landing roll. Biomimetic flying robots can achieve landings similar to the agility of birds, but their complex mechanical structures are more susceptible to fatigue damage [2]. Therefore, expanding the feasible flight trajectory at high AOA based on the structural performance of UAV landing gear is crucial for achieving successful perching. ...

Agile perching maneuvers in birds and morphing-wing drones

... Heuristic-based approaches leverage empirical rules to simplify search processes and reduce computational complexity [148], [149]. Besides, learning-based methods, such as imitation learning (IL) [150]- [152], reinforcement learning (RL) [153]- [155], and both of them [156]- [158], are also applied to path planning. ...

NPE-DRL: Enhancing Perception Constrained Obstacle Avoidance with Non-Expert Policy Guided Reinforcement Learning

IEEE Transactions on Artificial Intelligence

... A common alternative to the former techniques consists of fusing laser sensors and cameras. Recent works [37,38] have studied the fusion of point cloud and image features using advanced architectures, like transformers, achieving state-of-the-art performance in ...

AYDIV: Adaptable Yielding 3D Object Detection via Integrated Contextual Vision Transformer
  • Citing Conference Paper
  • May 2024

... The emergence of drone technology has significantly impacted various sectors, including search and rescue, environmental surveillance, and industrial inspection. Multidrone systems, in particular, offer distinct advantages such as enhanced efficiency, scalability, and redundancy compared to singledrone operations [1]- [3]. To fully exploit these advantages, drones typically need to collaborate with each other and maintain desired formations [4], which necessitates reliable inter-agent communication and egocentric observations during flight. ...

Toward Collaborative Multitarget Search and Navigation with Attention‐Enhanced Local Observation

... A typical Z-number is in the form of Z = (A, B), where A and B represent fuzzy restriction and reliability, respectively [19,20]. Given the advantages of Z-numbers in representing uncertain information, scholars have focused on this field [21][22][23]. Aliev [24] introduced generalized decision theories into Z-numbers. Later, an approximate reasoning method was proposed based on the rule "IF-THEN" and Z-numbers [25]. ...

Multitarget Assignment Under Uncertain Information Through Decision Support Systems
  • Citing Article
  • August 2024

IEEE Transactions on Industrial Informatics

... Terminal constraint guidance (TCG), which can ensure that the controlled object reaches the target position with the desired angle or velocity constraints, can be regarded as a powerful solution. TCG can be achieved by various approaches, which include proportional navigation (PN) guidance strategy [14][15][16], optimal guidance law [17][18][19], and sliding mode control scheme [20,22,24,25]. For TCG design, the convergence rate is a significant research topic. ...

Three-Dimensional Terminal Angle Constraint Guidance Law with Class K ∞ Function-Based Adaptive Sliding Mode Control
  • Citing Article
  • February 2024

Aerospace Science and Technology

... The results highlight the importance of considering individual aircraft configurations in conflict avoidance to prevent collisions. Jiaping et al. (2024) proposed an Asynchronous Multi-Stage Deep Reinforcement Learning (AMS-DRL) approach for drone navigation under adversarial attacks from multiple pursuers [21]. By evolving adversarial agents and ensuring Nash equilibrium, the method enables drones to evade attacks and reach targets, outperforming baselines in simulations and real-time tests. ...

Learning Multi-Pursuit Evasion for Safe Targeted Navigation of Drones
  • Citing Article
  • December 2024

IEEE Transactions on Artificial Intelligence

... Among its variants for this study, we chose YOLOv5s as our baseline model due to its compact size, competent performance, and real-time inference capabilities, making it suitable for potential future deployment on edge devices like Jetson hardware. Recent research [34] has demonstrated YOLOv5s' effectiveness in real-time multi-object detection and tracking, particularly within edge-accelerated systems, making it suitable for applications such as multi-drone operations. ...

Real-Time Multi-Drone Detection and Tracking for Pursuit-Evasion With Parameter Search
  • Citing Article
  • January 2024

IEEE Transactions on Intelligent Vehicles

... From a control perspective, in the design of the control structure for coaxial multirotors, aerodynamic effects due to rotor proximity are often neglected [17][18][19][20] or treated simplistically. One common choice in this regard is the inclusion of a constant penalization gain in the control allocation design to consider the thrust loss due to rotor proximity [21][22][23]. However, this conventional approach fails to capture the complex and nonlinear interactions between lower and upper rotors by just considering a linear model. ...

Design, Modeling, and Control of a Coaxial Drone
  • Citing Article
  • January 2024

IEEE Transactions on Robotics