<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Introducing Multiagent Systems to AV Visual Perception Sub-tasks: A proof-of-concept implementation for bounding-box improvement</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alaa</forename><surname>Daoud</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">INSA Rouen Normandie</orgName>
								<orgName type="institution" key="instit2">Univ Rouen Normandie</orgName>
								<orgName type="institution" key="instit3">Univ Le Havre Normandie</orgName>
								<orgName type="institution" key="instit4">Normandie Univ</orgName>
								<address>
									<addrLine>LITIS UR 4108</addrLine>
									<postCode>F-76000</postCode>
									<settlement>Rouen</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Corentin</forename><surname>Bunel</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">INSA Rouen Normandie</orgName>
								<orgName type="institution" key="instit2">Univ Rouen Normandie</orgName>
								<orgName type="institution" key="instit3">Univ Le Havre Normandie</orgName>
								<orgName type="institution" key="instit4">Normandie Univ</orgName>
								<address>
									<addrLine>LITIS UR 4108</addrLine>
									<postCode>F-76000</postCode>
									<settlement>Rouen</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maxime</forename><surname>Guériau</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">INSA Rouen Normandie</orgName>
								<orgName type="institution" key="instit2">Univ Rouen Normandie</orgName>
								<orgName type="institution" key="instit3">Univ Le Havre Normandie</orgName>
								<orgName type="institution" key="instit4">Normandie Univ</orgName>
								<address>
									<addrLine>LITIS UR 4108</addrLine>
									<postCode>F-76000</postCode>
									<settlement>Rouen</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Introducing Multiagent Systems to AV Visual Perception Sub-tasks: A proof-of-concept implementation for bounding-box improvement</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">27A196BD2326C1D729C924DA5526DEE7</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:28+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Autonomous driving</term>
					<term>perception systems</term>
					<term>bounding-box refinement</term>
					<term>Multiagent Systems</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Object detection is a pivotal task in computer vision, with applications spanning from autonomous driving to surveillance. Traditionally, methods like Non-Maximum Suppression (NMS) and its variants have been used to refine object detection outputs. Fusing predictions from different object detection models using confidence scores to average overlapping bounding boxes from multiple detection models has demonstrated superior performance over conventional methods. In this work, we employ multiple agents, each responsible for handling individual bounding boxes, to generate an improved fused prediction. This agent-based adaptation aims to leverage decentralized processing to potentially increase the system's efficiency and adaptability across various object detection scenarios, particularly in autonomous vehicle (AV) perception systems. We develop two distinct behaviors for the bounding box agents: one replicating the state-of-the-art Weighted Boxes Fusion (WBF) method in a decentralized manner, and the other introducing competitive behavior where agents interact based on Intersection over Union (IoU) and confidence values. We evaluate the performance of our approach using the COCO dataset, demonstrating the flexibility and potential of integrating MAS into object detection workflows including those for AV perception systems.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Autonomous vehicles and intelligent transport systems depend on advanced computer vision technologies, with object detection being a critical task. This enables vehicles to recognize and respond to surrounding objects effectively, with region proposal identifying potential object locations early in the detection process, crucial for timely responses in autonomous driving <ref type="bibr" target="#b0">[1]</ref>. Traditional techniques like Non-Maximum Suppression (NMS) often struggle to balance precision and recall, especially in dynamic environments. Solovyev et al. <ref type="bibr" target="#b1">[2]</ref> introduced Weighted Boxes Fusion (WBF), using confidence scores to average overlapping bounding boxes from multiple detection models, demonstrating superior performance over conventional methods.</p><p>The integration of Multiagent Systems (MAS) into object detection workflows offers new perspectives to address traditional challenges <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4]</ref>. MAS provide dynamic and adaptable decision-making capabilities, enhancing autonomous vehicles' ability to handle complex, unpredictable road conditions. MAS support distributed and adaptive processing <ref type="bibr" target="#b4">[5]</ref>, complementing modern GPU-based computer vision. By distributing tasks across agents, MAS enhances system flexibility and resilience, especially in dynamic environments like autonomous driving or video surveillance <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8]</ref>. Each agent manages a subset of tasks, improving resilience to errors <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10]</ref>.</p><p>MAS can adjust strategies based on scenarios <ref type="bibr" target="#b10">[11,</ref><ref type="bibr" target="#b11">12]</ref>, adapting parameters for bounding box fusion based on context, scene complexity, or environmental changes <ref type="bibr" target="#b12">[13]</ref>. Agents operate independently on different hardware, optimizing processing power and allowing system scalability <ref type="bibr" target="#b13">[14]</ref>. Local decisions are combined through a global process, enhancing accuracy <ref type="bibr" target="#b14">[15]</ref>. MAS can continually learn from 13th International Workshop on Agents in Traffic and Transportation (ATT 2024) held in conjunction with <ref type="bibr">ECAI 2024</ref> alaa.daoud@insa-rouen.fr (A. Daoud); corentin.bunel@insa-rouen.fr (C. Bunel); maxime.gueriau@insa-rouen.fr (M. Guériau) 0000-0002-3640-327X (A. Daoud); 0000-0001-6637-9795 (C. Bunel); 0000-0002-8742-6623 (M. Guériau) their environment and from the interactions between agents <ref type="bibr" target="#b15">[16]</ref>. This potential for adaptive learning motivates the agentification approach, as it opens the possibility for future enhancements. By achieving an agentified method, we can later integrate learning capabilities to further improve adaptability and performance in evolving object detection scenarios. Agent-based approaches are well-suited for integrating diverse models and data sources <ref type="bibr" target="#b16">[17]</ref>, which is essential for the ensemble approaches used in WBF where predictions from different models are combined.</p><p>Agentifying output refinement methods such as NMS or WBF involves assigning individual agents to handle specific bounding boxes, enabling dynamic adjustment based on individual box characteristics. This approach addresses real-time processing requirements and improves scalability and fault tolerance by decentralizing the decision-making process <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b18">19,</ref><ref type="bibr" target="#b19">20]</ref>. In this work, we aim to design and implement a proof-of-concept system integrating MASs into the process of improving bounding boxes in object detection. We will develop two behaviors for the bounding box agents: one replicating the state-ofthe-art Weighted Boxes Fusion (WBF) method in a decentralized manner, and the other introducing competitive behavior where agents interact based on Intersection over Union (IoU) and confidence values. Finally, we will deploy the system and assess its performance using the COCO dataset, testing various levels of competition and cooperation between agents. The remainder of this paper is structured as follows: Section 2 presents the related work in object detection, multiagent systems, and their integration. Section 3 details the system architecture and design principles of the AWBF method. Section 4 describes the implementation of the proof-of-concept system and the development of agent behaviors. Section 5 discusses the experimental setup and evaluation using the COCO dataset. Section 6 presents the results and analysis of the experimental evaluation. Section 7 concludes the paper with a summary of findings and future work directions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Object detection is a fundamental task in computer vision, critical for intelligent transportation systems (ITS) applications such as autonomous driving, traffic monitoring, and surveillance. The integration of MASs into object detection workflows offers significant potential to enhance system efficiency, robustness, and adaptability. This section reviews recent advancements in object detection techniques relevant to the ITS, with a focus on bounding box fusion and the role of MASs. <ref type="bibr">Wang et al. (2019)</ref> presented the Multi-Stage Complementary Fusion (MCF3D) network, an end-toend architecture for 3D object detection that integrates LiDAR and RGB data. This network employs attention mechanisms and prior knowledge to achieve state-of-the-art results, enhancing the detection accuracy necessary for autonomous driving applications <ref type="bibr" target="#b20">[21]</ref>. <ref type="bibr" target="#b21">Qian et al. (2020)</ref> proposed an improved object detection method for remote sensing images, incorporating a novel bounding box regression loss and a multi-level features fusion module. This method enhances the precision of object localization, which is crucial for applications such as traffic monitoring and vehicle detection <ref type="bibr" target="#b21">[22]</ref>. <ref type="bibr" target="#b1">Solovyev et al. (2021)</ref> introduced the Weighted Boxes Fusion (WBF) method, which averages overlapping bounding boxes from multiple detection models using confidence scores. This approach demonstrated superior performance over traditional techniques, highlighting the effectiveness of fusion methods in improving object detection accuracy <ref type="bibr" target="#b1">[2]</ref>. This method is particularly relevant for ITS applications where robust and accurate object detection is paramount for safety and efficiency.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Bounding Box Improvement Techniques</head><p>Zhang and Wu (2022) proposed a multi-view feature adaptive fusion framework that enhances 3D object detection by optimizing depth feature fusion and loss function design. This approach improves the regression accuracy of bounding boxes, which is essential for ITS applications where precise object localization is critical <ref type="bibr" target="#b22">[23]</ref>. <ref type="bibr" target="#b23">Liu et al. (2023)</ref> developed the Fusion network by Box Matching (FBMNet) for multi-modal 3D detection. This method aligns features at the bounding box level, providing stability in challenging scenarios such as asynchronous sensors and misaligned sensor placements, common issues in ITS applications <ref type="bibr" target="#b23">[24]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Multiagent Systems in Object Detection</head><p>Introducing MAS to object detection and computer vision systems is not a new idea. For example <ref type="bibr" target="#b24">Choksuriwong et al. (2005)</ref> developed a MAS for image understanding that localizes and recognizes objects using a distributed system implemented on a cluster computer. This approach leverages invariant features and supervised classification to improve object recognition accuracy, which is vital for traffic monitoring systems <ref type="bibr" target="#b24">[25]</ref>. However, the application of MAS in these areas has decreased recently with the advancements in machine learning techniques and their improved performance in handling object detection tasks. Despite this shift, some researchers have continued to explore the potential of MAS in object detection through various approaches.</p><p>Jiang et al. ( <ref type="formula">2019</ref>) proposed a multi-agent deep reinforcement learning (MADRL) approach for multi-object tracking, using YOLO V3 for object detection and Independent Q-Learners (IQL) for policy learning. This method achieves better performance in precision, accuracy, and robustness compared to other state-of-the-art methods, which is particularly beneficial for real-time traffic monitoring and surveillance <ref type="bibr" target="#b25">[26]</ref>.</p><p>Fekir and Benamrane (2015) introduced a MAS for boundary detection and object tracking using active contours and multi-resolution treatment. This system improves object boundary detection and tracking through cooperative agent strategies, enhancing the accuracy and efficiency of ITS applications such as vehicle and pedestrian tracking <ref type="bibr" target="#b26">[27]</ref>.</p><p>Vincent et al. ( <ref type="formula">2022</ref>) described a MAS using stereovision for perception, enabling agents to collaborate and enhance scene understanding through graph matching algorithms. This approach addresses challenges in correspondence identification and non-covisibility, critical for ITS applications such as multi-vehicle coordination and traffic management <ref type="bibr" target="#b27">[28]</ref>.</p><p>Mahmoudi et al. ( <ref type="formula">2013</ref>) utilized a MAS for object recognition in complex urban areas, leveraging WorldView-2 satellite imagery and digital surface models. This system improves object recognition accuracy through knowledge-based reasoning and cooperative agent capabilities, essential for urban traffic monitoring and smart city applications <ref type="bibr" target="#b28">[29]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Positioning Our Proposal</head><p>In light of the existing work, our proposal aims to integrate the strengths of both bounding box fusion techniques and MASs to develop a more robust and efficient object detection framework tailored for ITS applications. Our approach leverages the distributed processing capabilities of MASs to enhance the accuracy and scalability of bounding box fusion methods. By incorporating advanced fusion techniques and adaptive agent strategies, our system aims to address the limitations of existing methods, such as handling dynamic environments and improving detection precision. Our contributions include:</p><p>1. A multi-agent based framework for bounding box improvement that dynamically assigns agents to handle specific bounding boxes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Integration of advanced fusion techniques, such as Weighted Box Fusion (WBF) and Non-</head><p>Maximum Suppression (NMS), to enhance detection accuracy in various ITS scenarios. 3. Implementation of adaptive agent strategies / behaviors that allow the switch between cooperation and competition dynamically, ensuring robust performance in real-world ITS applications.</p><p>To the best of our knowledge, we are among the first to propose integrating MAS into specific computer vision sub-tasks such as bounding box filtering and fusion. This approach aims to exploit the advantages of MAS to enhance the accuracy, efficiency, and adaptability of object detection systems in ITS applications.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">System Architecture for AWBF</head><p>The agentified Weighted Boxes Fusion (WBF) system integrates multiple agents, each handling individual bounding boxes from various detection models. This Multiagent System (MAS) enhances the efficiency and accuracy of bounding box fusion through distributed processing and specialized agent roles. A central blackboard mechanism facilitates information sharing and coordination.</p><p>MAS offers decentralized decision-making and dynamic adaptability, enhancing resilience and flexibility in handling varied scenarios <ref type="bibr" target="#b29">[30]</ref>. The blackboard acts as a global communication hub, simplifying data interactions and providing a robust framework for synchronized information exchange among agents <ref type="bibr" target="#b30">[31]</ref>. Specific agent roles, from bounding box processing to model-specific adaptations, optimize performance and accuracy by leveraging domain-specific knowledge and algorithms <ref type="bibr" target="#b31">[32]</ref>. Feedback mechanisms enable dynamic adaptation, allowing agents to adjust strategies based on performance and data input changes, maintaining high accuracy in dynamic environments <ref type="bibr" target="#b32">[33]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Overview of Agent Roles</head><p>The system includes various agents with specific responsibilities:</p><p>• Bounding-Box Agents: Handle individual bounding boxes, analyze, and propose fusions with overlapping boxes. • Model-specific Agents: Manage bounding boxes from specific detection models. Can be seen as interfaces between the MAS and CV models. Each agent extracts bounding box proposals from its respective model to ensure compatibility and apply model-specific behaviors and adjustments. • Coordinator Agents: Oversee the fusion process, resolve conflicts between bounding-box agents, and make final decisions on merged bounding boxes. • Data Processing Agents: Optionally handle image preprocessing and result postprocessing.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Blackboard Information Sharing System</head><p>The blackboard serves as a shared information space for communication and data exchange:</p><p>• Data Repository: Central storage for bounding box data, including coordinates, confidence scores, and model origins. • Communication Medium: Allows agents to read and write data, maintaining system modularity and scalability. • Coordination Facilitator: Coordinates actions among agents, especially in resolving fusion conflicts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Processing Workflow</head><p>The workflow involves: </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Implementation and Development of Agent Behaviors</head><p>Our implementation is developed in Python, utilizing the existing WBF codebase to maintain consistency in data processing. By forking the original WBF repository, we leverage developed libraries, utilities, and functions, ensuring the use of the exact same logic in data processing. This allowed us to focus on integrating MAS features without reinventing the core bounding box fusion logic. We built an ad-hoc MAS framework tailored to our requirements. The agents interact via a shared blackboard for communication, and the system supports both centralized and decentralized processing. Following the system architecture described in the previous section, one can implement diverse behaviors and a variety of solution method logics by only changing the decision logic of the bounding box agent and adjusting the coordination mechanism. Model-specific Agents interact with existing object detection models (e.g., YOLO, Faster R-CNN) to receive and process bounding boxes. Model-specific agents convert detection outputs into a standard format used by the system.</p><p>Main implementation challenges included managing computation time, communication overhead, and integrating the MAS with existing computer vision models. Future improvements will focus on developing variety of agent behaviors with optimized parameters for computational and accuracy performance, enhancing the system's scalability, robustness, and adaptability, exploring further integration with advanced machine learning models and real-world deployment scenarios.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Agent Behaviors</head><p>We developed two distinct agent behaviors to demonstrate the versatility and potential of MAS in object detection. The first behavior replicates the Weighted Boxes Fusion (WBF) in a decentralized manner, while the second introduces a competitive interaction among agents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.1.">Behavior 1: Decentralized Weighted Boxes Fusion (WBF)</head><p>This behavior replicates the state-of-the-art WBF method in a decentralized manner. Each agent processes bounding boxes independently and posts results to a shared blackboard (see Algorithm 1), improving system resilience. The agent determines overlapping boxes as candidates for fusion by calculating the Intersection over Union (IoU). Boxes are considered for fusion if their IoU exceeds a certain threshold. The IoU calculation is given by:</p><formula xml:id="formula_0">𝐼𝑜𝑈 (𝐵𝑜𝑥 1 , 𝐵𝑜𝑥 2 ) = 𝑎𝑟𝑒𝑎(𝐵𝑜𝑥 1 ∩ 𝐵𝑜𝑥 2 ) 𝑎𝑟𝑒𝑎(𝐵𝑜𝑥 1 ∪ 𝐵𝑜𝑥 2 )</formula><p>Algorithm 1 Decentralized WBF Algorithm (AWBF) -BoundingBox Agent behavior Fuse 𝐴 𝑖 and 𝐵 𝑗 using weighted average 14:</p><p>end if 15: end for 16: 𝐴 𝑖 posts result to the blackboard 17: Output: Final bounding boxes</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.2.">Behavior 2: Competitive Interaction</head><p>In this behavior, agents compete based on a new metric that we introduce as Intersection over Box area(IoB). The IoBs for two boxes, 𝐴 and 𝐵, are calculated separately as:</p><formula xml:id="formula_1">𝐼𝑜𝐵 𝐴|𝐵 = 𝑎𝑟𝑒𝑎(𝐴 ∩ 𝐵) 𝑎𝑟𝑒𝑎(𝐴) , 𝐼𝑜𝐵 𝐵|𝐴 = 𝑎𝑟𝑒𝑎(𝐴 ∩ 𝐵) 𝑎𝑟𝑒𝑎(𝐵)</formula><p>Attacking or cooperating with other agents depending on calculated strengths. The strength of an attack of 𝐴 on 𝐵 and defense of 𝐵 against 𝐴 are defined by:</p><formula xml:id="formula_2">𝑆 attack (𝐴, 𝐵) = confidence 𝐴 × 𝐼𝑜𝐵 𝐵|𝐴 , 𝑆 defense (𝐵, 𝐴) = confidence 𝐵 × 𝐼𝑜𝐵 𝐴|𝐵</formula><p>The decision rule is based on the difference between attack and defense strengths and a decision threshold </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Illustrative Example: Bounding Box Fusion for Bicycle Detection</head><p>To illustrate the AWBF and competitive behavior in action, we consider the detection of bicycles in image "138639" from COCO dataset using two ad-hoc models. (see Figure <ref type="figure" target="#fig_3">2</ref>)  The bounding boxes from the two models are as follows:</p><p>• For competitive behavior, the attack and defence strengths rely on the calculation of the Intersection over Box values:</p><formula xml:id="formula_3">IoB 𝐵1|𝐵2 = area(𝐵1 ∩ 𝐵2) area(𝐵1) ≈ 0.89 , IoB 𝐵2|𝐵1 = area(𝐵1 ∩ 𝐵2) area(𝐵2) ≈ 0.91 Attack 𝐵1|𝐵2 = 0.9 • 0.91 ≈ 0.819 , Defense 𝐵2|𝐵1 = 0.5 • 0.89 ≈ 0.445 Result = Attack 𝐵1|𝐵2 − Defense 𝐵2|𝐵1 = 0.374</formula><p>Given (𝑇 = 0.3), the agent 𝐵1 wins and 𝐵2 is removed as (Result &gt; 𝑇 ). Increasing the 𝑇 value to 4, the conflict result will fall into the cooperation range and thus we return back to AWBF.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experimental Evaluation</head><p>To evaluate our methods, we conducted extensive tests using the COCO dataset. Our primary objective was to demonstrate the proof of concept without optimizing parameters or model weights beyond the default settings provided by the WBF code. Therefore, our results focus on comparing performance metrics rather than optimizing for maximum accuracy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Evaluation Metrics</head><p>The evaluation metrics used were those recommended and specified by COCO dataset. Namely, Average Precision and Average Recall:</p><p>-Average Precision (AP) Reveals the model's ability to make accurate positive predictions. It is calculated at different Intersection over Union (IoU) thresholds.</p><p>• AP@[IoU=0.50:0.95]: This is the average AP over ten IoU thresholds (0.50 to 0.95 with a step size of 0.05). • AP@0.50: This is the AP at an IoU threshold of 0.5.</p><p>• AP@0.75: This is the AP at an IoU threshold of 0.75.</p><p>• AP[small]: AP for small objects (area &lt; 32 pixels).</p><p>• AP[medium]: AP for medium sized objects (32 ≤ area ≤ 96 pixels).</p><p>• AP[large]: AP for large objects ( area ≥ 96 pixels).</p><p>-Average Recall (AR) Measuring the sensitivity by focusing on the model's ability to correctly identify positive samples from the entire pool of positive instances.</p><p>• AR@[IoU=0.50:0.95]: This is the average recall over ten IoU thresholds (0.50 to 0.95 with a step size of 0.05). • AR@0.50: This is the average recall at an IoU threshold of 0.5.</p><p>• AR@0.75: This is the average recall at an IoU threshold of 0.75.</p><p>• AR[small]: AR for small objects.</p><p>• AR[medium]: AR for medium objects.</p><p>• AR[large]: AR for large objects.</p><p>The results from test runs over the entire dataset are shown in the Table <ref type="table" target="#tab_0">1</ref>. Notably, the results demonstrate that AWBF outperforms individual models whose outputs were used in the fusion process. Although our results did not surpass those of the centralized WBF, they were mostly comparable. Specifically, our approach performed better than WBF on AP-small and AR@10 at an IoU of 0.5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Avg. Precision</head><p>Avg. Recall @[0.5 , 0.95] @[0.5] @[0.75] Small Medium Large @[0. Wen running experiments on subsets of the COCO dataset with different sizes, we observed that the centralized WBF method performs better with larger datasets but shows reduced efficiency on smaller datasets (see Figure <ref type="figure" target="#fig_5">3</ref>). This can be explained by several factors:</p><p>• Law of Large Numbers: As the dataset size increases, the averaging process tends to smooth out random errors and fluctuations, leading to improved performance for the centralized WBF method. • Error Compensation: With more data points, errors in individual detections can compensate for each other, leading to more accurate fusion results. • Increased Data Redundancy: Larger datasets contain more redundant information, reinforcing correct detections and diluting the impact of incorrect ones. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">AWBF Performance on Different Dataset Sizes:</head><p>The AWBF method exhibited more robust and stable performance across varying dataset sizes as Figure <ref type="figure" target="#fig_5">3</ref> shows, which can be attributed to the distributed processing and the redundancy: Each agent processes bounding boxes independently and in parallel, reducing the impact of individual errors and improving overall robustness. Also, Each agent's localized decision-making can lead to better performance, especially in smaller datasets where individual detections have a higher impact.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Competitive Behavior Experiments:</head><p>We also evaluated the competitive behavior using the default parameters. While the initial results did not match the quality of WBF, they demonstrated the potential for diverse agent behaviors. By adjusting the value of T, which controls the level of cooperativeness (1 -competitiveness). We conducted multiple tests on a subset of 500 COCO images, varying the competitiveness level. We observed an interesting trend, as the results showed that increasing competitiveness improved precision (see Figure <ref type="figure" target="#fig_6">4</ref>). AP increased with higher competitiveness, likely because competition removed lower-scoring boxes, reducing false positives and improving precision. Recall remained stable as even with fewer boxes, sufficient accurate boxes were retained.</p><p>To summarize, theses evaluations demonstrated the flexibility and potential of integrating MAS into object detection workflows. While the competitive agent behavior requires further optimization, the initial results validate our approach and open avenues for more sophisticated multi-agent behaviors in future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this work, we presented a proof-of-concept implementation integrating MAS into object detection workflows, specifically focusing on improving bounding box predictions, an essential component of autonomous vehicle perception systems. By leveraging the decentralized processing capabilities of MAS, we demonstrated two distinct agent behaviors: Decentralized (Agentified) Weighted Boxes Fusion and Competitive Interaction. Our experimental evaluation using the COCO dataset showed that while the decentralized WBF approach performed comparably to the centralized WBF, the competitive behavior illustrated the potential for further optimization and innovation in agent-based object detection systems. The results indicate that MAS can offer robust and adaptable solutions for object detection tasks, particularly in dynamic and complex environments like AV perception and intelligent transportation systems. Future work will focus on refining agent behaviors, enhancing system scalability, and integrating more advanced machine learning models to further improve performance and adaptability for AV applications.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: AWBF Agents, colors correspond to the models who generate initial bounding boxes</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>(𝑇 ): Result(𝐴, 𝐵) = 𝑆 attack (𝐴, 𝐵) − 𝑆 defense (𝐵, 𝐴) , 𝐵) &gt; 𝑇 : 𝐴 wins and 𝐵 is removed Result(𝐴, 𝐵) &lt; −𝑇 : 𝐵 wins and 𝐴 is removed otherwise : 𝐴 and 𝐵 fuse using WBF The least case represent the area where agents can cooperate as their strengths are close. Threshold 𝑇 can determine the level of cooperativeness, and thus the value (1 − 𝑇 ) refers to the competitiveness level. (𝑇 = 1) indicates full cooperativeness settings, reverting to AWBF. Contrarily, (𝑇 = 0) indicates full competitiveness unless attack and defense strengths are equal.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Visualization of Bounding Box Proposals and Fusion Results on COCO#138639 (images are cropped to emphasize the area of interest)</figDesc><graphic coords="7,325.23,232.29,135.39,88.40" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Model 1 :</head><label>1</label><figDesc>{'box': [0.192, 0.752, 0.312, 0.873], 'score': 0.9, 'label': 2} • Model 2: {'box': [0.203, 0.756, 0.314, 0.875], 'score': 0.5, 'label': 2} The Intersection over Union for the bounding boxes is: IoU = area of overlap area of union ≈ 0.85 If we apply the WBF method: WBF box = 0.9 • [0.192, 0.752, 0.312, 0.873] + 0.5 • [0.203, 0.756, 0.314, 0.875] 0.9 + 0.5 WBF box ≈ [0.196, 0.754, 0.313, 0.874] , WBF score = 0</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: COCO evaluation metrics evolution as function of the size of test set -Evaluating the AWBF method performance on subsets of COCO dataset</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Precision and recall values evolution with the increase of competition level (decrease in cooperation threshold T) in the agent behavior</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>1 .</head><label>1</label><figDesc>Data Input and Distribution: Model-specific Agents extract bounding boxes from different models and transfer them to Data processing agents who prepare and distribute bounding boxes to bounding box agents. 2. Bounding Box Analysis and Posting: Bounding box agents analyze and post findings to the blackboard, proposing fusions. 3. Review and Fusion: Coordinator agents review and finalize fusion decisions, consulting modelspecific agents as needed. 4. Final Processing and Output: Data processing agents optimize the fused bounding boxes for downstream applications. 5. Feedback and Adaptation: The system adapts to changes by updating agent strategies or parameters based on performance metrics.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>Bounding boxes 𝐵, confidence scores 𝑆, labels 𝐿 2: 𝐴 𝑖 reads overlapping boxes from the blackboard 3: for each overlapping box 𝐵 𝑖 do Calculate attack strength 𝑆 attack = confidence 𝐴 𝑖 × 𝐼𝑜𝐵 𝐵 𝑗 6: Calculate defense strength 𝑆 defense = confidence 𝐵 𝑗 × 𝐼𝑜𝐵 𝐴 𝑖</figDesc><table><row><cell cols="2">2: Read 𝐵, 𝑆, 𝐿 from the blackboard</cell></row><row><cell cols="2">3: Determine overlapping boxes as candidates for fusion</cell></row><row><cell cols="2">4: Filter candidates using IoU metric</cell></row><row><cell cols="2">5: Apply WBF on the final set of candidates</cell></row><row><cell cols="2">6: Post fused boxes to the blackboard</cell></row><row><cell cols="2">7: Output: Fused bounding boxes</cell></row><row><cell cols="2">Algorithm 2 Competitive Interaction Algorithm -BoundingBox Agent 𝐴 𝑖 behavior</cell></row><row><cell>4:</cell><cell>Calculate 𝐼𝑜𝑈 and 𝐼𝑜𝐵 between 𝐴 𝑖 and 𝐵 𝑗</cell></row><row><cell>5:</cell><cell></cell></row></table><note>1: Input: Bounding boxes 𝐵, confidence scores 𝑆, labels 𝐿 1: Input: 7: Calculate result 𝑅 = 𝑆 attack − 𝑆 defense 8: if 𝑅 &gt; 𝑇 then 9: 𝐴 𝑖 wins and 𝐵 𝑗 is removed 10: else if 𝑅 &lt; −𝑇 then 11: 𝐵 𝑗 wins and 𝐴 𝑖 is removed 12: else 13:</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>WBF Performance on Different Dataset Sizes:</head><label></label><figDesc></figDesc><table><row><cell>5.1.</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="6">5 , 0.95] @[0.5] @[0.75] Small Medium Large</cell></row><row><cell>EffNetB0</cell><cell>0.336</cell><cell>0.515</cell><cell>0.354</cell><cell>0.125</cell><cell>0.388</cell><cell>0.528</cell><cell>0.288</cell><cell>0.44</cell><cell>0.467</cell><cell>0.193</cell><cell>0.55</cell><cell>0.688</cell></row><row><cell>EffNetB0-m</cell><cell>0.335</cell><cell>0.516</cell><cell>0.351</cell><cell>0.129</cell><cell>0.389</cell><cell>0.524</cell><cell>0.288</cell><cell>0.441</cell><cell>0.467</cell><cell>0.198</cell><cell>0.55</cell><cell>0.687</cell></row><row><cell>EffNetB1</cell><cell>0.392</cell><cell>0.581</cell><cell>0.418</cell><cell>0.186</cell><cell>0.447</cell><cell>0.571</cell><cell>0.322</cell><cell>0.501</cell><cell>0.532</cell><cell>0.294</cell><cell>0.599</cell><cell>0.735</cell></row><row><cell>EffNetB1-m</cell><cell>0.392</cell><cell>0.581</cell><cell>0.417</cell><cell>0.184</cell><cell>0.447</cell><cell>0.571</cell><cell>0.323</cell><cell>0.502</cell><cell>0.531</cell><cell>0.279</cell><cell>0.602</cell><cell>0.735</cell></row><row><cell>EffNetB2</cell><cell>0.425</cell><cell>0.617</cell><cell>0.453</cell><cell>0.238</cell><cell>0.479</cell><cell>0.591</cell><cell>0.34</cell><cell>0.537</cell><cell>0.569</cell><cell>0.347</cell><cell>0.632</cell><cell>0.75</cell></row><row><cell>EffNetB2-m</cell><cell>0.426</cell><cell>0.617</cell><cell>0.454</cell><cell>0.24</cell><cell>0.481</cell><cell>0.593</cell><cell>0.341</cell><cell>0.537</cell><cell>0.569</cell><cell>0.358</cell><cell>0.634</cell><cell>0.748</cell></row><row><cell>EffNetB3</cell><cell>0.459</cell><cell>0.65</cell><cell>0.491</cell><cell>0.28</cell><cell>0.503</cell><cell>0.616</cell><cell>0.359</cell><cell>0.569</cell><cell>0.604</cell><cell>0.404</cell><cell>0.654</cell><cell>0.77</cell></row><row><cell>EffNetB3-m</cell><cell>0.455</cell><cell>0.646</cell><cell>0.487</cell><cell>0.282</cell><cell>0.494</cell><cell>0.618</cell><cell>0.357</cell><cell>0.566</cell><cell>0.6</cell><cell>0.412</cell><cell>0.65</cell><cell>0.766</cell></row><row><cell>EffNetB4</cell><cell>0.49</cell><cell>0.685</cell><cell>0.529</cell><cell>0.334</cell><cell>0.538</cell><cell>0.64</cell><cell>0.375</cell><cell>0.598</cell><cell>0.634</cell><cell>0.464</cell><cell>0.682</cell><cell>0.782</cell></row><row><cell>EffNetB4-m</cell><cell>0.488</cell><cell>0.684</cell><cell>0.524</cell><cell>0.33</cell><cell>0.533</cell><cell>0.642</cell><cell>0.373</cell><cell>0.596</cell><cell>0.633</cell><cell>0.468</cell><cell>0.68</cell><cell>0.783</cell></row><row><cell>EffNetB5</cell><cell>0.505</cell><cell>0.7</cell><cell>0.544</cell><cell>0.343</cell><cell>0.549</cell><cell>0.646</cell><cell>0.383</cell><cell>0.619</cell><cell>0.656</cell><cell>0.5</cell><cell>0.698</cell><cell>0.791</cell></row><row><cell>EffNetB5-m</cell><cell>0.502</cell><cell>0.696</cell><cell>0.539</cell><cell>0.335</cell><cell>0.546</cell><cell>0.645</cell><cell>0.379</cell><cell>0.614</cell><cell>0.651</cell><cell>0.484</cell><cell>0.692</cell><cell>0.789</cell></row><row><cell>EffNetB6</cell><cell>0.513</cell><cell>0.705</cell><cell>0.555</cell><cell>0.352</cell><cell>0.556</cell><cell>0.652</cell><cell>0.387</cell><cell>0.626</cell><cell>0.664</cell><cell>0.505</cell><cell>0.703</cell><cell>0.795</cell></row><row><cell>EffNetB6-m</cell><cell>0.511</cell><cell>0.701</cell><cell>0.551</cell><cell>0.341</cell><cell>0.555</cell><cell>0.654</cell><cell>0.384</cell><cell>0.623</cell><cell>0.66</cell><cell>0.489</cell><cell>0.704</cell><cell>0.805</cell></row><row><cell>EffNetB7</cell><cell>0.521</cell><cell>0.71</cell><cell>0.562</cell><cell>0.37</cell><cell>0.562</cell><cell>0.66</cell><cell>0.39</cell><cell>0.633</cell><cell>0.671</cell><cell>0.517</cell><cell>0.711</cell><cell>0.801</cell></row><row><cell>EffNetB7-m</cell><cell>0.519</cell><cell>0.71</cell><cell>0.558</cell><cell>0.364</cell><cell>0.562</cell><cell>0.659</cell><cell>0.388</cell><cell>0.63</cell><cell>0.668</cell><cell>0.509</cell><cell>0.71</cell><cell>0.803</cell></row><row><cell>DetRS</cell><cell>0.515</cell><cell>0.71</cell><cell>0.654</cell><cell>0.318</cell><cell>0.565</cell><cell>0.676</cell><cell>0.384</cell><cell>0.628</cell><cell>0.671</cell><cell>0.479</cell><cell>0.723</cell><cell>0.828</cell></row><row><cell>DetRS-m</cell><cell>0.515</cell><cell>0.707</cell><cell>0.564</cell><cell>0.316</cell><cell>0.563</cell><cell>0.677</cell><cell>0.384</cell><cell>0.629</cell><cell>0.673</cell><cell>0.486</cell><cell>0.721</cell><cell>0.834</cell></row><row><cell>resnet50</cell><cell>0.496</cell><cell>0.697</cell><cell>0.538</cell><cell>0.299</cell><cell>0.543</cell><cell>0.656</cell><cell>0.378</cell><cell>0.607</cell><cell>0.64</cell><cell>0.457</cell><cell>0.686</cell><cell>0.8</cell></row><row><cell>resnet50-m</cell><cell>0.496</cell><cell>0.694</cell><cell>0.535</cell><cell>0.296</cell><cell>0.545</cell><cell>0.657</cell><cell>0.379</cell><cell>0.61</cell><cell>0.642</cell><cell>0.464</cell><cell>0.689</cell><cell>0.799</cell></row><row><cell>yolo</cell><cell>0.5</cell><cell>0.678</cell><cell>0.546</cell><cell>0.336</cell><cell>0.544</cell><cell>0.644</cell><cell>0.381</cell><cell>0.628</cell><cell>0.688</cell><cell>0.533</cell><cell>0.734</cell><cell>0.826</cell></row><row><cell>WBF</cell><cell>0.673</cell><cell>0.894</cell><cell>0.709</cell><cell>0.605</cell><cell>0.731</cell><cell>0.846</cell><cell>0.471</cell><cell>0.627</cell><cell>0.846</cell><cell>0.8</cell><cell>0.85</cell><cell>0.867</cell></row><row><cell>AWBF</cell><cell>0.61</cell><cell>0.66</cell><cell>0.625</cell><cell>0.61</cell><cell>0.766</cell><cell>0.675</cell><cell>0.395</cell><cell>0.676</cell><cell>0.745</cell><cell>0.664</cell><cell>0.706</cell><cell>0.819</cell></row><row><cell>Table 1</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">Benchmarking on COCO dataset</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is funded by the French National Research Agency as part of the MultiTrans project under reference ANR-21-CE23-0032.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Faster r-cnn: Towards real-time object detection with region proposal networks</title>
		<author>
			<persName><forename type="first">S</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Girshick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<idno type="DOI">10.1109/TPAMI.2016.2577031</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Pattern Analysis and Machine Intelligence</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<biblScope unit="page" from="1137" to="1149" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Weighted boxes fusion: Ensembling boxes from different object detection models</title>
		<author>
			<persName><forename type="first">R</forename><surname>Solovyev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gabruseva</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.imavis.2021.104117</idno>
		<ptr target="https://doi.org/10.1016/j.imavis.2021.104117" />
	</analytic>
	<monogr>
		<title level="j">Image and Vision Computing</title>
		<imprint>
			<biblScope unit="volume">107</biblScope>
			<biblScope unit="page">104117</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Alphastar: Mastering the real-time strategy game starcraft ii</title>
		<author>
			<persName><forename type="first">O</forename><surname>Vinyals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Babuschkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mathieu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jaderberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">M</forename><surname>Czarnecki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dudzik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Georgiev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Powell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">DeepMind blog</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">20</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Real-time object detection and tracking for unmanned aerial vehicles based on convolutional neural networks</title>
		<author>
			<persName><forename type="first">S.-Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-Y</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-C</forename><surname>Yu</surname></persName>
		</author>
		<idno type="DOI">10.3390/electronics12244928</idno>
	</analytic>
	<monogr>
		<title level="j">Electronics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page">4928</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A multiagent perspective of parallel and distributed machine learning</title>
		<author>
			<persName><forename type="first">G</forename><surname>Weiß</surname></persName>
		</author>
		<idno type="DOI">10.1145/280765.280806</idno>
		<idno>doi:10.1145/280765.280806</idno>
		<ptr target="https://doi.org/10.1145/280765.280806" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Second International Conference on Autonomous Agents, AGENTS &apos;98</title>
				<meeting>the Second International Conference on Autonomous Agents, AGENTS &apos;98<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="226" to="230" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">A Real-Time Multi-Camera Depth Estimation ASIC with Custom On-Chip Embedded DRAM</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">E D</forename><surname>Narinx</surname></persName>
		</author>
		<idno type="DOI">10.5075/epfl-thesis-7163</idno>
		<ptr target="https://doi.org/10.5075/epfl-thesis-7163" />
		<imprint>
			<date type="published" when="2019">2019</date>
			<pubPlace>Lausanne</pubPlace>
		</imprint>
		<respStmt>
			<orgName>École Polytechnique Fédérale de Lausanne</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Vehicle video surveillance system based on image fusion and parallel computing</title>
		<author>
			<persName><forename type="first">S</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gong</surname></persName>
		</author>
		<idno type="DOI">10.1002/cta.2907</idno>
	</analytic>
	<monogr>
		<title level="j">International Journal of Circuit Theory and Applications</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="page" from="1532" to="1547" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Special issue on parallel computing for real-time image processing</title>
		<author>
			<persName><forename type="first">M</forename><surname>Akil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Perroton</surname></persName>
		</author>
		<idno type="DOI">10.1007/s11554-011-0192-y</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Real-Time Image Processing</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="1" to="2" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Distributed multiagent control approach for multitarget tracking</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1155/2015/903682</idno>
	</analytic>
	<monogr>
		<title level="j">Mathematical Problems in Engineering</title>
		<imprint>
			<biblScope unit="volume">2015</biblScope>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Networked distributed fusion estimation under uncertain outputs with random transmission delays, packet losses and multi-packet processing</title>
		<author>
			<persName><forename type="first">R</forename><surname>Caballero-Águila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hermoso-Carazo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Linares-Pérez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Signal Processing</title>
		<imprint>
			<biblScope unit="volume">156</biblScope>
			<biblScope unit="page" from="71" to="83" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Dynamic task allocation method for heterogenous multiagent system in uncertain scenarios of agricultural field operation</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.1088/1742-6596/2356/1/012049</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of Physics: Conference Series</title>
		<imprint>
			<biblScope unit="volume">2356</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Neural networks-based distributed adaptive control of nonlinear multiagent systems</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Shi</surname></persName>
		</author>
		<idno type="DOI">10.1109/TNNLS.2019.2915376</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Networks and Learning Systems</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="page" from="1010" to="1021" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Multi-view fusion-based 3d object detection for robot indoor scene perception</title>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Soon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">K</forename><surname>Quah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Tandianus</surname></persName>
		</author>
		<idno type="DOI">10.3390/s19194092</idno>
	</analytic>
	<monogr>
		<title level="j">Sensors</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Multiagent-based optimal microgrid control using fully distributed diffusion strategy</title>
		<author>
			<persName><forename type="first">R</forename><surname>Azevedo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cintuglu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mohammed</surname></persName>
		</author>
		<idno type="DOI">10.1109/TSG.2016.2587741</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Smart Grid</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="1997" to="2008" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Behavior prediction for unmanned driving based on dual fusions of feature and decision</title>
		<author>
			<persName><forename type="first">S</forename><surname>Zhong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yin</surname></persName>
		</author>
		<idno type="DOI">10.1109/TITS.2020.3037926</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Intelligent Transportation Systems</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="3687" to="3696" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Improving energy-efficiency of scientific computing clusters</title>
		<author>
			<persName><forename type="first">N</forename><surname>Kaabouch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-C</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Niemi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kommeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-P</forename><surname>Hameri</surname></persName>
		</author>
		<idno type="DOI">10.4018/978-1-4666-1842-8.ch001</idno>
	</analytic>
	<monogr>
		<title level="m">Energy-Aware Systems and Networking for Sustainable Initiatives</title>
				<imprint>
			<publisher>IGI Global</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="1" to="19" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Multi-scale analysis strategies in prnu-based tampering localization</title>
		<author>
			<persName><forename type="first">P</forename><surname>Korus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huang</surname></persName>
		</author>
		<idno type="DOI">10.1109/TIFS.2016.2636089</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Information Forensics and Security</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="809" to="824" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Multiagent systems: A survey from a machine learning perspective</title>
		<author>
			<persName><forename type="first">P</forename><surname>Stone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Veloso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Autonomous Robots</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="345" to="383" />
			<date type="published" when="2000">2000</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">A survey on fault-tolerant consensus control of multi-agent systems: trends, methodologies and prospects</title>
		<author>
			<persName><forename type="first">C</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lyu</surname></persName>
		</author>
		<idno type="DOI">10.1080/00207721.2022.2056772</idno>
		<idno>arXiv:</idno>
		<ptr target="https://doi.org/10.1080/00207721.2022.2056772" />
	</analytic>
	<monogr>
		<title level="j">International Journal of Systems Science</title>
		<imprint>
			<biblScope unit="volume">53</biblScope>
			<biblScope unit="page" from="2800" to="2813" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Scalable distributed decision-making and coordination in large and complex systems: Methods, techniques, and models</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lujak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Giordani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Omicini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ossowski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Complexity</title>
		<imprint>
			<biblScope unit="page" from="1" to="3" />
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Mcf3d: Multi-stage complementary fusion for multi-sensor 3d object detection</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wei</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2019.2927012</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="90801" to="90814" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Object detection in remote sensing images based on improved bounding box regression and multi-level features fusion</title>
		<author>
			<persName><forename type="first">X</forename><surname>Qian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.3390/rs12010143</idno>
		<ptr target="https://www.mdpi.com/2072-4292/12/1/143.doi:10.3390/rs12010143" />
	</analytic>
	<monogr>
		<title level="j">Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">3d object detection based on multi-view adaptive fusion</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wu</surname></persName>
		</author>
		<idno type="DOI">10.1109/IPEC54454.2022.9777488</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC)</title>
				<imprint>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="page" from="743" to="748" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Bai</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2305.07713</idno>
		<title level="m">Multi-modal 3d object detection by box matching</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Multi-agents system for image understanding</title>
		<author>
			<persName><forename type="first">A</forename><surname>Choksuriwong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Rosenberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Smari</surname></persName>
		</author>
		<idno type="DOI">10.1109/KIMAS.2005.1427070</idno>
	</analytic>
	<monogr>
		<title level="m">International Conference on Integration of Knowledge Intensive Multi-Agent Systems</title>
				<imprint>
			<date type="published" when="2005">2005. 2005</date>
			<biblScope unit="page" from="149" to="154" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Multi-agent deep reinforcement learning for multi-object tracker</title>
		<author>
			<persName><forename type="first">M</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Hai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Deng</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2019.2901300</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="32400" to="32407" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Multi agent system for boundary detection and object tracking in image sequence based on active contours</title>
		<author>
			<persName><forename type="first">A</forename><surname>Fekir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Benamrane</surname></persName>
		</author>
		<idno type="DOI">10.3233/MGS-150230</idno>
		<ptr target="https://doi.org/10.3233/MGS-150230.doi:10.3233/MGS-150230" />
	</analytic>
	<monogr>
		<title level="j">Multiagent Grid Syst</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="81" to="93" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Multi-agent system perception with stereovision</title>
		<author>
			<persName><forename type="first">G</forename><surname>Vincent</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Patten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Ohmes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Couch</surname></persName>
		</author>
		<idno type="DOI">10.1145/3545947.3573289</idno>
		<idno>doi:10.1145/3545947.3573289</idno>
		<ptr target="https://doi.org/10.1145/3545947.3573289" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 54th ACM Technical Symposium on Computer Science Education V. 2, SIGCSE 2023</title>
				<meeting>the 54th ACM Technical Symposium on Computer Science Education V. 2, SIGCSE 2023<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page">1235</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Object oriented image analysis based on multi-agent recognition system</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Tabib</forename><surname>Mahmoudi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Samadzadegan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Reinartz</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.cageo.2012.12.007</idno>
		<ptr target="https://doi.org/10.1016/j.cageo.2012.12.007" />
	</analytic>
	<monogr>
		<title level="j">Computers &amp; Geosciences</title>
		<imprint>
			<biblScope unit="volume">54</biblScope>
			<biblScope unit="page" from="219" to="230" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Coulouris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dollimore</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kindberg</surname></persName>
		</author>
		<title level="m">Distributed Systems: Concepts and Design</title>
				<imprint>
			<publisher>Addison-Wesley</publisher>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
	<note>5 ed</note>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">The blackboard model of problem solving and the evolution of blackboard architectures</title>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">P</forename><surname>Nii</surname></persName>
		</author>
		<idno type="DOI">10.1609/aimag.v7i2.537</idno>
		<ptr target="https://ojs.aaai.org/aimagazine/index.php/aimagazine/article/view/537.doi:10.1609/aimag.v7i2.537" />
	</analytic>
	<monogr>
		<title level="j">AI Magazine</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page">38</biblScope>
			<date type="published" when="1986">1986</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<monogr>
		<title level="m" type="main">An Introduction to MultiAgent Systems</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wooldridge</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>John Wiley &amp; Sons</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Russell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Norvig</surname></persName>
		</author>
		<title level="m">Artificial Intelligence: A Modern Approach</title>
				<imprint>
			<publisher>Prentice Hall</publisher>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
	<note>3 ed</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
