Model-Driven Requirements for Humans-on-the-Loop Multi-UAV Missions

  title={Model-Driven Requirements for Humans-on-the-Loop Multi-UAV Missions},
  author={Ankit Agrawal and Jan-Philipp Stegh{\"o}fer and Jane Cleland-Huang},
  journal={2020 IEEE Tenth International Model-Driven Requirements Engineering (MoDRE)},
The use of semi-autonomous Unmanned Aerial Vehicles (UAVs or drones) to support emergency response scenarios, such as fire surveillance and search-and-rescue, has the potential for huge societal benefits. Onboard sensors and artificial intelligence (AI) allow these UAVs to operate autonomously in the environment. However, human intelligence and domain expertise are crucial in planning and guiding UAVs to accomplish the mission. Therefore, humans and multiple UAVs need to collaborate as a team… 

Figures and Tables from this paper

Explaining Autonomous Decisions in Swarms of Human-on-the-Loop Small Unmanned Aerial Systems
This paper examines the User Interface (UI) design trade-offs associated with providing timely and detailed explanations of autonomous behavior for swarms of small Unmanned Aerial Systems (sUAS) or drones and analyses the impact of UI design choices on awareness of the situation.
Adaptive Autonomy in Human-on-the-Loop Vision-Based Robotics Systems
A solution based upon adaptive autonomy levels is proposed, whereby the system detects loss of reliability of these models and responds by temporarily lowering its own autonomy levels and increasing engagement of the human in the decision-making process.
RescueAR: Augmented Reality Supported Collaboration for UAV Driven Emergency Response Systems
RescueAR utilizes location-based augmented reality technology to facilitate the exchange of information among rescue team members during a UAV-driven emergency response to identify injured victim locations using AR glasses or mobile devices.
Extending MAPE-K to support Human-Machine Teaming
This paper identifies critical human-machine teaming factors and describes the infrastructure needed across the various phases of the MAPE-K loop in order to effectively support HMT, the Human Machine Teaming paradigm, and presents guidelines for integrating HMT into MAPe-K.
Towards a Model-Integrated Runtime Monitoring Infrastructure for Cyber-Physical Systems
A novel framework for model-integrated runtime monitoring is proposed that combines model-driven techniques and runtime monitoring to automatically generate large parts of the monitoring framework and to reduce the maintenance effort necessary when part of the monitored system change.
Hazard analysis for human-on-the-loop interactions in sUAS systems
The study showed that the hazard trees provided effective coverage for a wide variety of sUAS application scenarios and were useful for stimulating safety thinking and helping users to identify and potentially mitigate human-interaction hazards.


Monitoring CPS at Runtime - A Case Study in the UAV Domain
A case study on monitoring in UAVs, CPS-specific monitoring challenges, and a prototype implemented by extending \reminds, a framework for software monitoring so far mainly used in the domain of metallurgical plants are reported.
Deep Reinforcement Learning Approach with Multiple Experience Pools for UAV’s Autonomous Motion Planning in Complex Unknown Environments
A multiple experience pools (MEPs) framework leveraging human expert experiences for DRL to speed up the learning process and indicate that UAVs trained using MEP–DDPG can stably complete a variety of tasks in complex, unknown environments.
A swarm of wheeled and aerial robots for environmental monitoring
This research proposes indeed a monitoring system composed by wheeled rover, obviously extendable to more vehicles, that maps the state and the evolution of the pollution by means of a set of sensors purposely developed.
A Survey on Aerial Swarm Robotics
The main sections of this paper focus on major results covering trajectory generation, task allocation, adversarial control, distributed sensing, monitoring, and mapping, and dynamic modeling and conditions for stability and controllability that are essential in order to achieve cooperative flight and distributed sensing.
Trusted Autonomy Between Humans and Robots: Toward Human-on-the-Loop in Robotics and Autonomous Systems
  • S. Nahavandi
  • Political Science
    IEEE Systems, Man, and Cybernetics Magazine
  • 2017
Autonomy refers to the capability of a machine to perform a task, or part of it, with no-or substantially reduced-human intervention.
The State-of-the-Art of Human–Drone Interaction: A Survey
This paper is the first survey on the emerging field of human-drone interaction focusing on multi-rotor systems, providing an overview of existing literature and the current state of the art in the field.
Human-Drone Interactions with Semi-Autonomous Cohorts of Collaborating Drones
A third scenario in which multiple humans and drones collaborate in an emergency response scenario is explored, and different types of interactions are discussed, and examples from current DroneResponse project are drawn.
Dronology: An Incubator for Cyber-Physical Systems Research
A CPS research incubator is introduced – based upon a system, and its associated project environment, for managing and coordinating the flight of small Unmanned Aerial Systems (sUAS).
SAHRTA: A Supervisory-Based Adaptive Human-Robot Teaming Architecture
A Supervisory-Based Adaptive Human-Robot Teaming Architecture (SAHRTA) that selects an appropriate level of autonomy or system interaction based on a complete realtime multi-dimensional workload estimate and predicted future task performance is introduced.
Situation awareness-based agent transparency and human-autonomy teaming effectiveness
It is shown that the SAT model continues to be an effective tool for facilitating shared understanding and proper calibration of trust in human–agent teams and an expansion of the model is necessary to support teamwork paradigms, which require bidirectional transparency.