<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Human-Drone Interactions with Semi-Autonomous Cohorts of Collaborating Drones</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Jane</forename><surname>Cleland-Huang</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Notre Dame Notre Dame</orgName>
								<address>
									<postCode>46556</postCode>
									<region>IN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Ankit</forename><surname>Agrawal</surname></persName>
							<email>aagrawa2@nd.edu</email>
							<affiliation key="aff1">
								<orgName type="institution">University of Notre Dame Notre Dame</orgName>
								<address>
									<postCode>46556</postCode>
									<region>IN</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Human-Drone Interactions with Semi-Autonomous Cohorts of Collaborating Drones</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">56B7FBBD05ACFA4B1857739C143CC84F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T01:09+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>Human-drone collaboration, emergency response</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Research in human-drone interactions has primarily focused on cases in which a person interacts with a single drone as an active controller, recipient of information, or a social companion; or cases in which an individual, or a team of operators interacts with a swarm of drones as they perform some coordinated flight patterns. In this position paper we explore a third scenario in which multiple humans and drones collaborate in an emergency response scenario. We discuss different types of interactions, and draw examples from current DroneResponse project.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Introduction</head><p>Small Unmanned Aerial Systems, which we refer to here as drones, can be effectively deployed to support emergency responders for diverse scenarios such as search-andrescue, accident surveillance, and flood inspections. Currently, emergency responders tend to operate drones manually or using off-the-shelf applications that allow them to preprogram sets of waypoints. However, equipping drones with onboard intelligence allows them to perform tasks autonomously and to contribute more fully to the emergency response.</p><p>In our DroneResponse project we are designing and developing a system to deploy and coordinate the efforts of multiple semi-autonomous drones for use in emergency situations <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b3">3]</ref>. Our vision is for humans and drones to work closely together as part of a complex mission -for example to monitor air quality following a chemical explosion, to perform search and rescue, to deliver medical supplies, or to support firefighters during structural fires. As depicted in Figure <ref type="figure" target="#fig_0">1</ref>, there are several facets to human-drone interaction in such scenarios. Humans need to communicate mission goals and directives to groups of drones as well as to individuals, while drones need to keep humans informed of their current state and progress, and at times, need to seek permission or guidance to perform specific tasks. In addition, both humans and drones need to communicate between themselves (i.e., drone-to-drone and human-tohuman) to coordinate their activities. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>An Interaction Example</head><p>We provide examples of such interactions in the sequence diagram depicted in Figure <ref type="figure" target="#fig_1">2</ref>. The Incident Commander first defines a search area and sends a request to the hive controller to start the search (E1). This is an example of human-to-drone interactions (H2D). The hive-controller then creates a search plan and assigns search routes to drones (E2). The coordination between drones represents droneto-drone (D2D) interaction. In the modeled sequence of actions, Drone n detects a potential drowning victim. It then notifies the human incident commander and starts streaming annotated video to the ground (E3), thereby illustrating drone-to-human communication (D2H). The part of the sequence diagram highlighted in yellow provides an example of a more complex bi-directional human-drone conversation in which the drone uses its sensing (image detection) abilities to detect a victim. It then autonomously switches to track-victim mode, raises a victim-found alert, and streams annotated video to the incident commander. Finally, the incident commander uses the information relayed by the drone, to confirm the victim sighting and to push information from the drone to a physical rescue team (E4). This final step is an example of human-to-human (H2H) interaction, triggered by the initial D2H exchange. This sequence of events illustrates the complex socio-technical aspects of emergent multi-user, multi-drone interaction spaces.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Drone-to-Human Communication (D2H)</head><p>There are numerous challenges that must be addressed in order to achieve efficient human-drone collaboration. In our concurrently published work <ref type="bibr" target="#b0">[1]</ref>, we have focused on designing a user interface that enables situational-awareness (SA) <ref type="bibr" target="#b5">[5]</ref> for human first-responders. SA involves perception (i.e., recognizing and monitoring), comprehension (i.e., interpreting and synthesizing information), and projection (i.e., understanding the situation, projecting future outcomes) so that a user can make effective and actionable decisions. The key to designing D2H communication is identifying information that is needed by different user roles within specific contexts. As an example, a drone may be ascribed the ability to autonomously decide its speed and altitude during a search. If visibility is good, the drone might fly higher and faster in order to cover the search area more quickly, while still returning accurate results. On the other hand, if visibility is lower, the drone might need to fly lower and slower, and adapt its flight plan to compensate for a reduced field of view. In this scenario, the operator needs visual cues and awareness of why a drone behaves as it does. As an outcome of a four month co-design process with our local fire department, we identified two design strategies to address this specific scenario. First, we designed our DroneResponse GUI to depict any environmental factors that were likely to impact drone behavior -for example, low visibility, high-winds, or prohibited airspace. Secondly, we made the drones explain themselves upon demand by describing their current strategies and permissions. In the case of searching for a victim in inclement weather, the drone might explain "flying lower than normal at 10 meters due to low visibility" or "searching river banks at a greater distance than normal due to high wind gusts and moving branches." We report outcomes from our co-design experience, especially with respect to D2H interactions and achieving situational awareness in our related paper <ref type="bibr" target="#b0">[1]</ref>.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Human to Drone Mission Directives</head><p>Achieving effective H2D communication is challenging in systems with multiple drones and complex missions. Several research groups have explored ways to specify drone missions using formal commands, often embedded in domain specific languages <ref type="bibr" target="#b7">[7]</ref>. However, it is infeasible for emergency responders to write such mission specifica-tions under stressful time-constraints of a life-and-death response. A user interface is therefore needed that enables quick mission planning and configuration and which supports high-level directives addressed to the cohort of drones, as well as specific directives addressed to individual drones.</p><p>Researchers have previously explored diverse solutions for issuing commands to drones, such as the use of gestures and voice commands <ref type="bibr" target="#b6">[6,</ref><ref type="bibr" target="#b10">10]</ref> or airplane-like cockpits for controlling large military-style drones <ref type="bibr" target="#b8">[8]</ref>. We have prototyped the use of gestures and voice commands; however, they have several shortcomings that inhibit their use in emergency response scenarios. Voice commands, while appealing, are impractical due to the noise inherent to a rescue scene. This includes sirens, constant radio-chatter, and now the additional noise of drone motors. Gestures are similarly impractical. They have been shown to work effectively in controlled near-distance environments, which is far from the case for an emergency response scenario <ref type="bibr" target="#b2">[2]</ref>. Furthermore, they introduce significant room for error, which is unacceptable in an emergency response environment, where mistakes could cost the lives of both the victims and the rescuers. Domain experts, collaborating in the co-design of DroneResponse soundly ruled out either of these approaches <ref type="bibr" target="#b0">[1]</ref>. We therefore have opted to create a GUI-based solution for HD2 commands for emergency response missions.   In our GUI, which is currently under development, users can initially select a mission type from a high-level list of missions as depicted in Figure <ref type="figure" target="#fig_7">9</ref>. They then perform a series of configurations such as marking a search area. Each predefined mission type will have a corresponding underlying mission plan with configuration points. This plan is sufficient for allowing the mission to proceed through a series of predefined stages and tasks (e.g., search, track, return home). However, users will also need to configure or tweak the mission dynamically as it evolves, by providing additional directives.</p><p>Each of these directives must specify who, what, where, and how a task is to be accomplished. 'Who' refers to whether the command is addressed to the entire cohort or an individual drone. In the case of the cohort, then the hive coordination is empowered to autonomously figure out which drones are best fit to respond. 'Who' could also be specified with constraints such as 3 drones or drones with thermal cameras onboard. Finally 'who' could be addressed to a specific drone if it were the case, that the Incident Commander wished to assign a task to a specific drone. This is more risky, as the selected drone might be unfit for service (e.g., due to low battery or current critical service). 'What' refers to the specific task to be completed -for example re- connaissance, delivery, or serving as a communication relay if drones are communicating using onboard communication channels such as ad-hoc wifi. 'Where' refers to a region or point of interest defined by GPS coordinates. For example, in the case of reconnaissance, the user might need to direct the drone to a certain part of a wooded river bank where somebody has sighted a piece of clothing; while in the case of establishing a communication relay, the user could either specify coordinates or allow it to dynamically position itself so as to optimize communication between all drones. Finally, 'how' enables specific directives for how the task is to be completed. In some cases, the drones could be given significant autonomy to complete well-defined tasks, while in other cases more specific guidelines might be required. We are currently working closely with several emergency response organizations to better discover their needs and to formally model diverse mission plans. The GUI therefore provides a human-facing interface to an underlying mission plan specified using a more formal approach such as belief-desire-intent <ref type="bibr" target="#b9">[9]</ref>. Drones are able to interpret the more formal specification. In our initial prototype we are experimenting with a flow-chart of buttons to enable humans to configure mission directives in a known space of options. These are depicted in Figures <ref type="figure" target="#fig_6">6-8</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Incorrect position</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>GUI versus Physical Devices</head><p>The discussion in this position paper has focused almost entirely on human-drone interfaces based on the use of graphical interfaces; however, drones can also be controlled using physical hand-held devices. In a multi-drone scenario, humans might need to switch between graphical and physical interfaces for several reasons including taking manual control of a malfunctioning drone or temporarily using manual controls for a specific task that is currently beyond the capabilities of a drone to perform autonomously. Our prior work has shown that misalignment of GUI's and physical controllers can easily lead to accidents <ref type="bibr" target="#b3">[3]</ref> (see Figure <ref type="figure" target="#fig_8">10</ref>. For example, when control is passed from a computer to a handheld device, the physical switches on the hand-held controller must be set to stable 'flightmode' positions, otherwise accidents, including crashlandings, could occur. Furthermore, when humans take-over control of remote drones, it can be exceedingly difficult to figure out which direction the drone is facing. Commands are interpreted relative to the drones position, which means that issuing a 'forward' command would cause the drone to fly forward, but without clear orientation from the remote-pilot's perception, that could actually be in any direction. A simple design solution might be to provide a feature to autonomously reorient the drone with respect to the remote pilot so that physical and GUI controls become aligned relative to the drone's and pilot's positions. Given this reorientation, a forward command would then consistently send the drone away from the pilot, and a moveRight command would make it move right. For deployment in emergency situations, more though should be invested in the use of both graphical and physical interfaces, the interactions between them, and transitions of control across devices and between different operators.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusion</head><p>This position paper has presented an informal framework for considering human-drone interactions along the dimensions of H2D, D2H, D2D, and H2H communication in multiuser, multi-drone environments where drones are permitted to operate with some degree of autonomy. We have described some of the challenges we are facing in the design of DroneResponse and some initial ideas for addressing those challenges. Our prior work <ref type="bibr" target="#b0">[1]</ref> has focused primarily on the D2H challenge of supporting situational awareness, while our ongoing work focuses on providing a meaningful interface for more complex bidirectional human and drone interactions.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Humans and drones interact in many different ways.</figDesc><graphic coords="2,38.00,233.18,129.60,116.53" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: A Sequence Diagram showing the interactions between humans and drones for the scenario in which a drone searches for, and detects, a potential drowning victim.</figDesc><graphic coords="3,35.46,251.39,129.60,103.03" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :Figure 4 :</head><label>34</label><figDesc>Figure 3: Firefighters in South Bend manually operate drones</figDesc><graphic coords="3,38.00,112.65,129.60,97.20" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Defibrillator delivery by drone requires remote UI</figDesc><graphic coords="3,35.46,406.92,129.60,81.68" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Mission commands can be addressed to specific groups of drones.</figDesc><graphic coords="4,35.46,202.57,129.60,59.22" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: High-level commands, such as search, deliver, or relay, are domain specific. These ones target river search and rescue.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: The user can mark an area on the map to define either a region or a point of interest to be targeted by the H2D directive.</figDesc><graphic coords="4,35.46,323.10,129.60,97.14" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 9 :</head><label>9</label><figDesc>Figure 9: A user initially selects a predefined mission type which include mission objectives and task specifications.</figDesc><graphic coords="4,459.00,112.46,243.00,195.13" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 10 :</head><label>10</label><figDesc>Figure 10: The throttle on the physical user interface was incorrectly positioned during a transition from software-controlled flight to manually operated flight, causing the UAV to crash.</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>The work described in this position paper has primarily been funded by the National Science Foundation under grants CNS-1737496, CNS-1931962, CCF-1647342, and CCF-1741781. We also thank the Firefighters of South Bend for closely collaborating on the DroneResponse project.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">A</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Abraham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Burger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Christine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fraser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hoeksema</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hwang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Travnik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Scheirer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cleland-Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vierhauser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The Next Generation of Human-Drone Partnerships: Co-Designing an Emergency Response System</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><surname>Cox</surname></persName>
		</author>
		<idno type="DOI">10.1145/3313831.3376825</idno>
		<ptr target="http://dx.doi.org/10.1145/3313831.3376825" />
	</analytic>
	<monogr>
		<title level="m">Conference on Human Factors in Computing Systems (Chi)</title>
				<meeting><address><addrLine>Hawai</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Drone Me: An Exploration into Natural Human-Drone Interaction</title>
		<author>
			<persName><forename type="first">Jessica</forename><forename type="middle">R</forename><surname>Cauchard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jane</forename><forename type="middle">L E</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Kevin</forename><forename type="middle">Y</forename><surname>Zhai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">James</forename><forename type="middle">A</forename><surname>Landay</surname></persName>
		</author>
		<idno type="DOI">10.1145/2750858.2805823</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1145/2750858.2805823" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp &apos;15)</title>
				<meeting>the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp &apos;15)<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="361" to="365" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Discovering, Analyzing, and Managing Safety Stories in Agile Projects</title>
		<author>
			<persName><forename type="first">Jane</forename><surname>Cleland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">-</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Vierhauser</surname></persName>
		</author>
		<idno type="DOI">10.1109/RE.2018.00034</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1109/RE.2018.00034" />
	</analytic>
	<monogr>
		<title level="m">26th IEEE International Requirements Engineering Conference, RE 2018</title>
				<meeting><address><addrLine>Banff, AB, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-08-20">2018. August 20-24, 2018</date>
			<biblScope unit="page" from="262" to="273" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Dronology: an incubator for cyber-physical systems research</title>
		<author>
			<persName><forename type="first">Jane</forename><surname>Cleland-Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Vierhauser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sean</forename><surname>Bayley</surname></persName>
		</author>
		<idno type="DOI">10.1145/3183399.3183408</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1145/3183399.3183408" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 40th International Conference on Software Engineering: New Ideas and Emerging Results, ICSE (NIER) 2018</title>
				<meeting>the 40th International Conference on Software Engineering: New Ideas and Emerging Results, ICSE (NIER) 2018<address><addrLine>Gothenburg, Sweden</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018-05-27">2018. May 27 -June 03, 2018</date>
			<biblScope unit="page" from="109" to="112" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Designing for Situation Awareness: An Approach to User-Centered Design</title>
		<author>
			<persName><forename type="first">Mica</forename><forename type="middle">R</forename><surname>Endsley</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2011">2011</date>
			<publisher>CRC Press, Inc</publisher>
			<pubPlace>Boca Raton, FL, USA</pubPlace>
		</imprint>
	</monogr>
	<note>Second Edition. 2nd ed.</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Human-drone interaction: Let&apos;s get ready for flying user interfaces</title>
		<author>
			<persName><forename type="first">Markus</forename><surname>Funk</surname></persName>
		</author>
		<idno type="DOI">10.1145/3194317</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1145/3194317" />
	</analytic>
	<monogr>
		<title level="j">Interactions</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="78" to="81" />
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">High-Level Mission Specification for Multiple Robots</title>
		<author>
			<persName><forename type="first">Sergio</forename><surname>García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Patrizio</forename><surname>Pelliccione</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Claudio</forename><surname>Menghi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thorsten</forename><surname>Berger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tomas</forename><surname>Bures</surname></persName>
		</author>
		<idno type="DOI">10.1145/3357766.3359535</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1145/3357766.3359535" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th ACM SIGPLAN International Conference on Software Language Engineering (SLE 2019)</title>
				<meeting>the 12th ACM SIGPLAN International Conference on Software Language Engineering (SLE 2019)<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="127" to="140" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Human Factors Guidelines for Unmanned Aircraft Systems</title>
		<author>
			<persName><forename type="first">Alan</forename><surname>Hobbs</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Lyall</surname></persName>
		</author>
		<idno type="DOI">10.1177/1064804616640632</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1177/1064804616640632" />
	</analytic>
	<monogr>
		<title level="j">Ergonomics in Design: The Quarterly of Human Factors Applications</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">04</biblScope>
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">BDI Agents: From Theory to Practice</title>
		<author>
			<persName><forename type="first">Anand</forename><forename type="middle">S</forename><surname>Rao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><forename type="middle">P</forename><surname>Georgeff</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the First International Conference on Multiagent Systems</title>
				<meeting>the First International Conference on Multiagent Systems<address><addrLine>San Francisco, California, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1995-06-12">1995. June 12-14, 1995</date>
			<biblScope unit="page" from="312" to="319" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The State-of-the-Art of Human-Drone Interaction: A Survey</title>
		<author>
			<persName><forename type="first">D</forename><surname>Tezza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Andujar</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2019.2953900</idno>
		<idno>DOI:</idno>
		<ptr target="http://dx.doi.org/10.1109/ACCESS.2019.2953900" />
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="167438" to="167454" />
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
