In this study, we introduce an innovative Robot State Estimation (RSE) methodology incorporating a learning-based contact estimation framework for legged robots, which obviates the need for external physical contact sensors. This approach integrates multimodal proprioceptive sensory data, employing a Physics-Informed Neural Network (PINN) in conjunction with an Unscented Kalman Filter (UKF) to enhance the state estimation process. The primary objective of this RSE technique is to calibrate the Inertial Measurement Unit (IMU) effectively and furnish a detailed representation of the robot’s dynamic state. Our methodology exploits the PINN to mitigate IMU drift issues by imposing constraints on the loss function via Ordinary Differential Equations (ODEs). The advantages of utilizing a contact estimator based on proprioceptive sensory data are multifold. Unlike vision-based state estimators, our proprioceptive approach is immune to visual impairments such as obscured or ambiguous environments. Moreover, it circumvents the necessity for dedicated contact sensors—components not universally present on robotic platforms and challenging to integrate without substantial hardware modifications. The contact estimator within our network is trained to discern contact events across various terrains, thereby facilitating resilient proprioceptive odometry. This enables the contact-aided invariant Kalman Filter to produce precise odometric trajectories. Subsequently, the UKF algorithm estimates the robot’s three-dimensional attitude, velocity, and position. Experimental validation of our proposed PINN-based method illustrates its capacity to assimilate data from multiple sensors, effectively reducing the influence of sensor biases by enforcing ODE constraints, all while preserving intrinsic sensor characteristics. When juxtaposed with the employment of the UKF algorithm in isolation, our integrated RSE model demonstrates superior performance in state estimation. This enhanced capability automatically reduces sensor drift impacts during operational deployment, rendering our proposed solution applicable to real-world scenarios.
This paper proposes an approach to design a quadruped robot with a dynamic spine using Reservoir Computing. Four bending sensors are integrated into the continuous dynamics spine to adjust spine force for balancing and enable closed-loop control of the Center of Mass (CoMs) mapping onto the dynamic surface. Passively modeling the system enabled a quadrupedal robot with a flexible spine to navigate rough terrain with improved efficiency, agility, and minimized impact on explosive power. Additionally, paper introduce to develop a novel, intelligent control mechanism for the quadrupedal robot to operate effectively on dynamic, flexible surfaces. Our approach focuses on a hierarchical distributed control mechanism developed for leg to cooperatively change centralized frames for a better functional reflection.
Quadruped locomotion and gait patterns as trotting, galloping, trot running has been investigated and applied to a variety of existing quadruped robots such as Big Dog from Boston Dynamics, A1 Robot dog from Unitree. Most of these studies are based either on biology inspired gaits or the best possible locomotion that can be performed by the individual robot with its pre-set mechanics and its availability of the degree of freedoms. While these are already available as their basic modes, a wide number of researchers are investigating locomotion via deep neural nets. These are making headlines in the research community for efficiency of use, and yet the explainability is lacking in most cases. Just like a Large Language Model might give spurious results here and there for basic common sense questions, these deep neural nets also make errors with unknown interpretability to the inputs. Regarding training, they require careful tuning of hyperparameters and training with a number of parameters unknown to user predictions. For example, on the field we might have a terrain which is flat for a certain length, in addition to a rocky climb, followed by a slippery slope. The combinations are as many as possible and the existing state of the art is heavily depending on human intervention and training predictions to handle the change of modes of the gait patterns that can fit into the terrain underneath. In this paper, we develop a novel embodied explainable machine learning algorithm which can help minimize the training as well as human intervention when autonomous operations are required. Specifically, we utilize the Markov Decision Process (MDP) along with rules set forth by DARPA in the Explainable AI (XAI) research. The XAI research enables us to generate textual explanations of the behavior by utilizing the MDP and reinforcement learning to generate mission oriented and situation aware cost functions along with the ones which are already pre-programmed. We validate our hypothesis in real hardware across different conditions.
Space debris in the 1-10 cm diameter regime presents a particular hazard to future operations in low Earth orbit (LEO). These objects are too small to be reliably detected by terrestrial radar or telescopes, but too large for their impacts to be mitigated with Whipple shielding. This paper addresses this hazard by presenting a methodology for space-based detection and active debris removal (ADR) via novel intelligent collaborative sensing and efficient decentralized data fusion. In simulation, we demonstrate that a small heterogeneous constellation of co-orbiting satellites using attentional neural networks can autonomously share information and maneuver in a cooperative game to de-orbit space debris (e.g. with a mechanical arm or pulsed laser) while minimizing total energy consumption and collision risk in the constellation. Particular focus is given to the robustness and effectiveness of the developed collaborative sensing system with respect to sensor uncertainties and transient changes in observability between nodes.
Rise of unmanned vehicles and autonomous robots has been accompanied by study of path planning, navigation, and decision-making algorithms. Current state-of-the-art employs deep neural nets to extract the required features. This technology, though successful in most cases, fails where the training has not been done for unseen conditions. For such unlabeled training data, transfer learning approaches have been proposed. A major drawback of using transfer learning approaches is that the actions and/or state spaces are reactive only to present circumstances. A truly intelligent autonomous operation has to consider a subordinate-to-a-human approach for its mission risks that vary with topography, path planning as well as mission goals. To address these complex combinatorial problems, DARPA has initiative a novel Explainable AI (XAI) technology in the past few years. In XAI, machine learning is paired with human intervention to make decisions by generating textual explanations of all the available relevant information / decision. In this paper, we propose to use available information along with human intelligence in a feedback loop for helping the unlabeled data to be trained and generate cost functions which were previously not programmed. We study this context/situation awareness to generate list of decision available from explanations on combinatorial tasks. Moreover, we employ this approach to a quadruped robot to learn its environment and the AI model starts in its infancy to mimic human cognitive architecture. We show that the learning process can be improved in a way that suits a particular mission in mind.
This paper presents a methodology to study the need and implementation of a GPS-denied navigation system that gives position, velocity and time (PVT) graph. We discuss one such technology that uses an inertial measurement unit (IMU) comprising of accelerometer, gyroscope, magnetometer, and altimeter. We investigate the input and output relationship between a GPS available navigation system (Google Maps) and a GPS-denied navigation system (IMU based AI board). We delineate how to make such a system in a block-wise fashion so that future researchers can get a head start to the area of research that is termed pedestrian dead reckoning (PDR). We show our implementation here which is currently better than the bulk of the research that was found during our time of literature survey. This implementation takes ideas from liquid (NN) machine learning, hybrid convolutional neural network deep learning based online PDR navigation, and generative adversarial network (GAN) based motion transfer learning.
This paper studies the problem of real-time routing in a multi-autonomous robot enhanced network at the uncertain and vulnerable tactical edge. In practical harsh environment such as a battlefield, the uncertainty of social mobility and complexity of vulnerable environment due to unpredictable physical and cyber-attacks from the enemy would seriously affect the effectiveness and practicality of these emerging network protocols. This paper presents a GT-SaRE-MANET (Game-Theoretic Situation-aware Robot Enhanced Mobile Ad-hoc Network) routing protocol that adopts the online reinforcement learning technique to supervise the mobility of multi-robots as well as handle the uncertainty and potential physical and cyber-attack at the tactical edge. The proposed design can better support the virtual, augmented, and mixed reality technology in the future battlefield.
In this paper, finite horizon intelligent decision-making problem has been investigated for self-organized autonomous systems especially under unstructured environment. According to the latest studies, the uncertainty of environment will seriously affect the effectiveness of decision making especially for autonomous systems. To handle these issues, transfer learning, and deep reinforcement learning has been presented recently. However, those existing Learning algorithms commonly needs a large set of state-space which cause the algorithm to be time-consuming and not suitable for real-time application. Therefore, in this paper, a library of polices trained using Deep Q-Learning under similar environments is built and implemented.
In this paper, multi-player sequential game with an unknown non-stationary irrational player is investigated for cooperative autonomous robots decision-making applications. In practice, the irrationality of agents can seriously degrade the effectiveness of decision making especially for distributed cooperative tasks with applications to multi-robot systems. Specifically, The irrationality can be caused by the cooperation agent's mechanical failure or sensor flaw. To handle this issue, a novel dynamic evaluation system, which includes two important parameters, i.e. cooperation index and competitive flag, is designed to efficiently quantify the player's level of cooperation or competition firstly. Then, the continuous deep Q network space is proposed to predict the action value with respect to a continuous cooperation index.
In this paper, a novel decentralized intelligent adaptive optimal strategy has been developed to solve the pursuit-evasion game for massive Multi-Agent Systems (MAS) under uncertain environment. Existing strategies for pursuit-evasion games are neither efficient nor practical for large population multi-agent system due to the notorious ``Curse of dimensionality" and communication limit while the agent population is large. To overcome these challenges, the emerging mean field game theory is adopted and further integrated with reinforcement learning to develop a novel decentralized intelligent adaptive strategy with a new type of adaptive dynamic programing architecture named the Actor-Critic-Mass (ACM). Through online approximating the solution of the coupled mean field equations, the developed strategy can obtain the optimal pursuit-evasion policy even for massive MAS under uncertain environment.
It is very challenging to develop a mission critical control for networked heterogeneous UAS that is intelligent and resilient even when implemented into a complex environment with practical constraints such as limited communication, uncertainty system dynamics etc. Meanwhile, lacking applicable decentralized control seriously limits the usage of networked UAS into critical military and civilian missions. Recently, many learning-based decentralized control algorithms have been developed. However, there are two significant limitations, i.e. 1) slow learning convergence speed which cannot catch the environmental changing rate and 2) The gap between mission planning and real-time control. Our proposed algorithm will overcome these challenges and reap the advantages from networked UAS.
Deeply integrating online fast reinforcement learning with real-time networked control, a novel mission critical decentralized resilient and intelligent control algorithm will be developed for network heterogeneous unmanned autonomous systems (UAS) in presence of limited communication, system uncertainties and harsh environment. Different from traditional decentralized control and learning algorithms, proposed design is a real-time applicable optimal and resilient solution that has particularly considered real-time mission completeness, the convergence speed of learning algorithm and the impacts from limited communication, system uncertainties and harsh environment.
In this paper, intelligent path planning and coordination control of Networked Unmanned Autonomous Systems (NUAS) in dynamic environments is presented. A biologically-inspired approach based on a computational model of emotional learning in mammalian limbic system is employed. The methodology, known as Brain Emotional Learning (BEL), is implemented in this application for the first time. The multi-objective properties and learning
capabilities added by BEL to the path planning and coordination co-design of Networked Unmanned Autonomous Systems (NUAS) are very useful, especially while dealing with challenges caused by dynamic environments
with moving obstacles. Furthermore, the proposed method is very promising for implementation in real-time applications due to its low computational complexity. Numerical results of the BEL-based path planner and intelligent controller for NUAS demonstrate the effectiveness of the proposed approach.
The main contribution of this paper is to utilize the computational model of emotional learning in mammal’s
brain, i.e., BEL, for developing a novel path planning and intelligent control method for practical real-time NUAS. To
the best of the authors knowledge, this is the first time that BEL is implemented for accomplishing intelligent path planning and coordination control of NUAS. The learning capabilities added by the proposed approach to the path
planning and coordination of MAS enhances the overall path planning strategy, which is very useful especially while dealing with challenges caused by dynamic and uncertain environments with unpredictable and unknown moving obstacles.
Due to the highly uncertain and dynamic nature of military conflict and planetary exploration missions, enabling aerial and terrestrial unmanned autonomous systems (UAS) to gracefully adapt to mission and environmental changes is a very challenging task. In particular, the United States Army, Air Force, Navy, and NASA have recently shown interest in the task of load transportation by means of UAS, which rely heavily on the knowledge of both the UAS model and the load dynamics to function. Most of the currently available autopilot systems for UAS were built without suspended load transportation capabilities and are thus not appropriate, for example, to assist soldiers or planetary explorers in the tasks of carrying and deploying supplies, transporting injured people, or warfare. This research provides knowledge to the problem of autonomous suspended load transportation, attending national agencies expectations that UAS will perform in a reliable manner even in the challenging situation when loads of uncertain characteristics are transported and deployed, which heavily modify the UAS dynamic during the execution of the task. This work presents a novel model-free adaptive wavenet PID (AWPID)-based controller for enabling aerial UAS to transport cable suspended loads of unknown characteristics. In order to accomplish this goal, a control design is presented which enables the UAS to perform a trajectory tracking task, based solely on the knowledge of the UAS position. The methodology proposes a novel structure, which identifies inverse error dynamics using a radial basis neural network with daughter Mexican hat wavelets activation function. A real-time load transportation mission consisting of a multi-rotorcraft UAS carrying a cable suspended load of unknown characteristics validates the effectiveness of the trajectory tracking control strategy, showing smooth control signals even when the mathematical model of the aerial UAS and load dynamics are not known.
In this paper, an advanced wireless mobile collaborative sensing network will be developed. Through properly combining wireless sensor network, emerging mobile robots and multi-antenna sensing/communication techniques, we could demonstrate superiority of developed sensing network. To be concrete, heterogeneous mobile robots including unmanned aerial vehicle (UAV) and unmanned ground vehicle (UGV) are equipped with multi-model sensors and wireless transceiver antennas. Through real-time collaborative formation control, multiple mobile robots can team the best formation that can provide most accurate sensing results. Also, formatting multiple mobile robots can also construct a multiple-input multiple-output (MIMO) communication system that can provide a reliable and high performance communication network.
Military applications require networked multi-UAV system to perform practically, optimally and reliably under changing mission requirements. Lacking the effective control and communication algorithms is impeding the development of multi-UAV systems significantly. In this paper, distributed optimal flocking control and network co-design problem has been investigated for networked multi-UAV system in presence of uncertain harsh environment and unknown dynamics. First, the mathematical interaction between network imperfections and practical wireless network channel performance has been investigate. Then, a novel co-model has been developed for networked multi-UAV combining effects from physical system and network channel model effectively. Then, adopting neuro dynamics programming (NDP) technique and actor-critic-identifier (ACI) design architecture, a novel online finite horizon optimal flocking control and network co-design has been proposed. The developed algorithm cannot merely obtain the distributed optimal co-design within finite time, but also relax the stringent requirement about physical UAV system and network dynamics. In addition, developed novel co-design can satisfy the practical constraints, e.g. transmit power constraint etc. The Lyapunov stability analysis is used to validate the effectiveness of developed scheme. With the proper NN weight update law, proposed co-design can ensure all closed-loop signals and NN weights are uniformly ultimately bounded (UUB). Furthermore, simulation results have been provided to demonstrate the effectiveness of the developed scheme.
A 24 GHz medium-range human detecting sensor, using the Doppler Radar Physiological Sensing (DRPS) technique, which can also detect unmanned aerial vehicles (UAVs or drones), is currently under development for potential rescue and anti-drone applications. DRPS systems are specifically designed to remotely monitor small movements of non-metallic human tissues such as cardiopulmonary activity and respiration. Once optimized, the unique capabilities of DRPS could be used to detect UAVs. Initial measurements have shown that DRPS technology is able to detect moving and stationary humans, as well as largely non-metallic multi-rotor drone helicopters. Further data processing will incorporate pattern recognition to detect multiple signatures (motor vibration and hovering patterns) of UAVs.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.