<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Domain-centric ADAS datasets</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Václav</forename><surname>Diviš</surname></persName>
							<email>divisvaclav@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">University of West Bohemia</orgName>
								<address>
									<addrLine>Sedláčkova 214</addrLine>
									<postCode>3 301 00</postCode>
									<settlement>Pilsen</settlement>
									<country key="CZ">Czech Republic</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tobias</forename><surname>Schuster</surname></persName>
							<email>tobias.schuster@siemens.com</email>
							<affiliation key="aff1">
								<orgName type="department">Siemens Technology, Software Systems &amp; Processes</orgName>
								<orgName type="laboratory">Research in Verification Test</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marek</forename><surname>Hrúz</surname></persName>
							<affiliation key="aff2">
								<orgName type="department" key="dep1">Faculty of Applied Sciences</orgName>
								<orgName type="department" key="dep2">Department of Cybernetics and New Technologies for the Information Society</orgName>
								<orgName type="institution">University of West Bohemia</orgName>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<address>
									<addrLine>Feb 13-14</addrLine>
									<postCode>2023</postCode>
									<settlement>Washington</settlement>
									<region>D.C</region>
									<country>US</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Domain-centric ADAS datasets</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">90CC43A6FFD817835AC96C4B47549DB9</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-04-29T06:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Advanced Driver-Assistance Systems</term>
					<term>Trajectory Planning</term>
					<term>Domain-centric Datasets</term>
					<term>Object Detection</term>
					<term>mean Average Precision</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Since the rise of Deep Learning methods in the automotive field, multiple initiatives have been collecting datasets in order to train neural networks on different levels of autonomous driving. This requires collecting relevant data and precisely annotating objects, which should represent uniformly distributed features for each specific use case. In this paper, we analyze several large-scale autonomous driving datasets with 2D and 3D annotations in regard to their statistics of appearance and their suitability for training robust object detection neural networks. We discovered that despite spending huge effort on driving hundreds of hours in different regions of the world, merely any focus is spent on analyzing the quality of the collected data, from an operational domain perspective. The analysis of safety-relevant aspects of autonomous driving functions, in particular trajectory planning with relation to time-to-collision feature, showed that most datasets lack annotated objects at further distances and that the distributions of bounding boxes and object positions are unbalanced. We therefore propose a set of rules which help find objects or scenes with inconsistent annotation styles. Lastly, we questioned the relevance of mean Average Precision (mAP) without relation to the object size or distance.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction and Motivation</head><p>The shift from classical programming paradigms to Machine Learning-driven approaches (ML) is significant. Humans are nowadays often relying on decisions and recommendations made by Artificial Neural Networks (ANNs), however most of the applications are harmless. This is not the case of Advanced Driver-Assistance Systems (ADAS), where wrong decisions can lead to severe injuries <ref type="bibr" target="#b0">[1]</ref>. The development of an ML-driven application within this field follows strict processes <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>, from acquiring data necessary for training to deploying and evaluating the models. These processes comply with functional safety standards <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref> but do not propose specific measures, nor concrete thresholds which the system should pass before being publicly released. Moreover, it is generally not feasible to collect the full variation of information in a stochastic environment such as public roads. This is why it is important to look deeper into the possibilities of diagnosing and analyzing unbalanced or missing information within large-scale datasets related to the operation domain, i.e. a data-centric approach <ref type="bibr" target="#b5">[6]</ref>.</p><p>A data-centric approach lays the focus on ensuring that the data clearly conveys what the AI must learn. In the case of higher levels of ADAS <ref type="bibr" target="#b6">[7]</ref>, the vehicle can control the movement in longitudinal and lateral directions, while relying predominantly on cameras, radars, and nowadays more frequently on LIDAR systems. One example of an ADAS function is the Adaptive Cruise Control (ACC). The task of ACC is to drive within the lane at a certain speed and keep a safe distance from any potential obstacle. The Time-To-Collision (TTC) for each object is calculated for the purpose of keeping a safe distance. If the TTC value decreases below a defined threshold, a braking or evasive maneuver will be initiated. Based on linear kinematic equations, where distance is equal to speed over time, the greater the speed difference between the subject-automated vehicle (ego vehicle) and the object in front of it, the shorter the TTC will be, hence the ACC needs to plan in advance. As a consequence, on highways, the ACC must incorporate objects at a further distance (on camera images objects will appear smaller) into the trajectory planning <ref type="bibr" target="#b7">[8]</ref>. On the contrary, in cities, where the maximum speed is limited to 50km/h, the average area of an encountered object will be larger. Motivated by these physical dependencies, we took a deeper look at the annotated objects and analyzed their statistical appearance. The examination was done in regard to the functionality of ACC, in state-of-the-art (SOTA) large-scale automotive datasets. Our contributions are as follows:</p><p>• We define a minimum safe distance 𝑠𝑚𝑖𝑛 = 𝑓 (𝑣, 𝑎𝑚𝑎𝑥) for a variety of scenarios (ego speed 𝑣 and weather dependent deceleration 𝑎𝑚𝑎𝑥). We calculate the safe distance based on German legislation, but the process can easily be adapted to any other legislation.</p><p>• We analyze the distribution of objects' bounding boxes' (BB) relative sizes, distances to ego vehicle, and positions in the datasets.</p><p>• We define a set of standardizable sanity checkers which help verify the quality of the collected data and mark ambiguously labeled data.</p><p>• We highlight the concrete missing information which is not part of the datasets and diagnose the cause.</p><p>• We propose an automotive mean Average Precision (amAP) metric, which is now related to the distance to the object, or relative BB size.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Datasets and Related Work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Automotive Datasets</head><p>As mentioned in Section 1, the prerequisite for the correct functionality of the ACC is information about object classes, sizes (used in case of overtaking or evasive maneuver), and distances of the objects to the ego vehicle. Several common automotive datasets are therefore not suitable for such a task, despite some of them providing LIDAR data, namely BDD100K <ref type="bibr" target="#b8">[9]</ref> (no LIDAR data), CityScapes <ref type="bibr" target="#b9">[10]</ref> (no bounding box annotations), Perl <ref type="bibr" target="#b10">[11]</ref> (no 3D annotations) or Apollo Scape <ref type="bibr" target="#b11">[12]</ref> (no images, only LIDAR). The other group of datasets contains images and LIDAR point clouds including the 2D as well as the 3D annotations. It is for this reason that we decided to take the following large-scale automotive datasets into account: KITTI <ref type="bibr" target="#b12">[13]</ref>, Audi Autonomous Driving Dataset (A2D2) <ref type="bibr" target="#b13">[14]</ref>, Lyft Level 5 dataset <ref type="bibr" target="#b14">[15]</ref>, nuScenes dataset <ref type="bibr" target="#b15">[16]</ref>, Waymo Open Dataset <ref type="bibr" target="#b16">[17]</ref> and the ONCE dataset <ref type="bibr" target="#b17">[18]</ref>. Each dataset contains a various number of labeled objects like small vehicles (vehicle, ego vehicle, SUV, motorcycle, etc.), large vehicles (truck, bus, tram), pedestrians, and cyclists. Especially KITTI and nuScenes show high-class imbalance for some classes due to fine-grained classes. Furthermore, the selected datasets contain labeled camera images with 2D and 3D bounding boxes and the corresponding LIDAR point cloud information which provides distance information for each object. However, the size of the datasets, in terms of the number of labeled frames and captured ambient conditions, varies. For instance, A2D2 provides a dataset of 2D labeled images, but only a small part contains 3D bounding boxes. KITTI, nuScenes, LyftLevel5 and Waymo reflect only urban areas whereas A2D2 and ONCE also contain highways, country roads, tunnels etc. On top of that, the sensor setups vary as well, e.g. different camera resolutions. For the KITTI dataset, the authors used a LIDAR sensor and two stereo cameras (left and right), whereas for Waymo five LIDARs (restricted to 75m) and five cameras were used. nuScenes used six cameras and one LIDAR sensor as well as five radars, ONCE uses one LIDAR and seven cameras, and A2D2 five LIDARs and six cameras. However, for the ACC, only the front cameras are taken into account, resulting in a smaller amount of usable images.</p><p>The KITTI dataset is the smallest in terms of scenes and the least diverse, containing only sunny and cloudy daytime scenes. For a short period of time, the Waymo and nuScenes datasets provided the largest variety and amount of data and annotations; they are among the most widely-used autonomous driving datasets. Although the ONCE dataset recently set a new benchmark for the amount of driving hours and frames, Waymo contains the highest amount of 3D bounding boxes because ONCE focuses on self-supervised learning without labels. Table <ref type="table" target="#tab_0">1</ref> gives an overview of important general information per dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Dataset Analysis</head><p>The authors of the above-mentioned datasets compared their works based on the common aspects of the datasets as shown in Table <ref type="table" target="#tab_0">1</ref>. General properties like the number of driving hours are often used to compare datasets and to state an improvement. Moreover, the number of scenes, images, or annotations is often used to determine the quality of the datasets.</p><p>For instance, the authors of the ONCE <ref type="bibr" target="#b17">[18]</ref> and A2D2 <ref type="bibr" target="#b13">[14]</ref> dataset focus on the number of annotations, the amount of driving hours, the adverse weather conditions, the time (day/night) and different locations (urban, highway, country roads) as well as countries/cities where the data was captured. However, specific requirements of driver assistance systems were not considered while creating or evaluating any of those datasets. The intent was rather to generate general datasets for a wide range of supervised and unsupervised learning tasks as well as driving functions.</p><p>The authors of A2D2, as well as nuScenes, focused on statistics relevant to the ACC and other assistance systems. Their work provides information about the distribution of the object distances for different classes as well as the absolute number of objects within the dataset. Additionally, the authors of nuScenes analyzed the distributions of the velocities of common objects like vehicles and bikes as well as bounding box dimensions.</p><p>The KITTI benchmark <ref type="bibr" target="#b12">[13]</ref> is based on the performance analysis of neural networks on the size of BBs in pixels as a proxy for the distance of the ego vehicle to the object. Their work in general follows the COCO evaluation methodology <ref type="bibr" target="#b18">[19]</ref>, but no physical distance information is used. The authors of nuScenes and Waymo did set a baseline for various detection tasks, yet without considering the distances to the different objects explicitly. The analysis closest to ours is done by the authors of the ONCE dataset. They analyzed the collected data regarding distance-wise mean Average Precision performance for 3D object detection using only point clouds. However, their distance thresholds were selected rather intuitively, whereas we specifically derive the distance from the domain safety requirements. Additionally, we analyze the spatial distribution of objects within the images as well as the bounding box/object size compared to the image size.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Background</head><p>Since the majority of related works only analyze datasets from a general ML perspective, omitting the point of data-centric paradigm, we decided to verify the SOTA automotive datasets in regard to a trajectory planning task. One part of our motivation is that forecasting the trajectory planning is conditioned by ego's vehicle velocity, which in case of higher value takes the further-distance objects into account. In order to be able to evaluate the sufficiency and quality of annotated objects within the datasets, we've chosen the Time-To-Collision as an instance to calculate the minimal safe distance from the ego vehicle.</p><p>For the sake of simplicity, we do not consider any obstacle heading from the opposite direction (on the collision course), since we are working with static images and thus don't have the information about the relative motion of the objects. As described in <ref type="bibr" target="#b19">[20]</ref>: "The TTC value at instant 𝑡 is defined as the time for two objects to collide if they continue at their present velocity and on the same path". Let's define the first object to be an obstacle (anything else than the ego vehicle) and the second object to be the ego vehicle. We consider the velocity of an obstacle to be equal to 0km/h (representing the stand-still object and therefore the worst-case scenario), and the ego vehicle's deceleration to be 7 𝑚/𝑠 2 (can variate within a range from 7 𝑚/𝑠 2 till 10 𝑚/𝑠 2 on dry roads <ref type="bibr" target="#b20">[21]</ref>, <ref type="bibr" target="#b21">[22]</ref>). Vehicle deceleration can be seen as a function of adhesion between the tires and the road, which depends on the material used in the tires, material of the road, temperature, weather conditions, mounted braking system, and the mass of the vehicle.</p><p>Based on the definition of TTC, let us consider three driving scenarios:</p><p>• highway (recommended speed 130𝑘𝑚/ℎ ≈ 36𝑚/𝑠)</p><p>• country road (maximum speed 100𝑘𝑚/ℎ ≈ 28𝑚/𝑠)</p><p>• city (maximum speed 50𝑘𝑚/ℎ ≈ 14𝑚/𝑠)</p><p>We now compute the minimal safe distance which needs to be ensured in order to brake in time (without initiating any evasive maneuver), for the following case: ego vehicle is driving on the highway, the possible deceleration is equal to 7𝑚/𝑠 2 and reaction delay is 0.0𝑠. Based on the kinematic equations of a linearly decelerating object, the distance which the object will travel is a function of time 𝑠 = 𝑠0 + 𝑣0𝑡 − 1 2 𝑎𝑡 2 , where time 𝑡 is a function of deceleration 𝑡 = 𝑣 0 −𝑣 𝑎 . With a linear deceleration of 7𝑚/𝑠 2 , the vehicle, moving within the legal limits, will reach its standstill state in time 𝑡 = 𝑣 0 −𝑣 𝑎 = 36−0 7 = 5.14𝑠. Within this time frame, the ego vehicle will travel a distance of 𝑠 = 0 + 36 • 5.14 − 1 2 • 7 • 5.14 2 = 185.04 − 92.46 = 92.58𝑚. For completeness, the braking distance under the same weather conditions on country roads is 55.11𝑚 and in the city 14𝑚 as can be seen in Table <ref type="table" target="#tab_1">2</ref>. As mentioned earlier, this process can be generalized and repeated for It is noticeable that the highway's maximum foresight boundary will be in reality limited by the physical properties of the camera or the maximum speed difference between the ego vehicle and the object. But most importantly, objects within those safe ranges must be part of a dataset (training and testing), otherwise, the system will deal with an epistemic uncertainty <ref type="bibr" target="#b22">[23]</ref>. In order to be able to investigate the statistics of objects' appearances, we need a dataset that provides information about the object's distance. As mentioned in Section 2, some datasets such as Perl and BKK100, etc. are not suitable for this task. Consequently, we have chosen the following large-scale automotive datasets: KITTI, Waymo open Dataset, A2D2 from Audi, nuScenes, LyftLevel5 and ONCE, which contain 3D annotations and distances to the objects.</p><p>In regard to the functionality of trajectory planning, we focused on the following in-dataset object characteristics:</p><p>• distribution of a BB's relative size: in order to verify that a variety of objects' sizes is captured within the dataset,</p><p>• distribution of the distance between obstacles and the ego vehicle (with relation to minimum safe distances): in order to verify that objects in further distances are incorporated within the dataset,</p><p>• relation between BB's relative size and object distance from ego vehicle: to discover abnormality within the dependency,</p><p>• heatmap of an object's appearance density: to visualize the potential asymmetrical appearance of an object with relation to the ego vehicle,</p><p>• an optical flow between consecutive images; in order to identify a series of static images, which can lead to class imbalanced dataset.</p><p>As originally presented in <ref type="bibr" target="#b18">[19]</ref> and further explored in <ref type="bibr" target="#b17">[18]</ref>, it seems to be reasonable to observe the mean Average Precision with relation to a specific object's size. Since the original authors clustered the object groups rather generally (small, middle, big), we propose to have clearly specified operational domain dependencies and incorporate the minimum safe distances as thresholds. Furthermore, the relation of AP to the objects' relative BB sizes or distance highlights detailed discrepancies between the model's performance trained on the same dataset. We therefore incorporate the minimal safe distance 𝑑𝑠 of each scenario as a threshold for the creation of test subsets 𝒜 ⊂ ℬ ⊂ 𝒞 from original test dataset 𝒟. For instance 𝒜 contains objects only in distance &gt; 𝑑𝑠(ℎ𝑖𝑔ℎ𝑤𝑎𝑦). For each subset the average precision can be calculated, representing a concrete value for a specific operational domain e.g. driving in the city. Equation 1 represents a different perspective on the average precision metric, which we call automotive mean AP.</p><formula xml:id="formula_0">𝑎𝑢𝑡𝑜𝑚𝑜𝑡𝑖𝑣𝑒 𝑚𝐴𝑃 = 1 𝑛𝑐 𝑐 ∑︁ 𝑖=1 𝐴𝑃𝑛 𝑖 ,<label>(1)</label></formula><p>where 𝑛𝑐 is a number of domain specific test-subsets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Analysis of datasets</head><p>In this chapter we analyze all datasets from Table <ref type="table" target="#tab_0">1</ref> for the characteristics mentioned in Section 3. In order to evaluate the model's generalization ability, the collected data has to be divided into two parts, namely training and validation. The training part is used for extracting the relevant features and finding a reasonable combination in order to build a high-level feature representation, whereas the validation part is used to evaluate the loss after each training cycle (epoch). The training loop usually ends when the loss, on a validation set, stagnates for several epochs <ref type="bibr" target="#b23">[24]</ref>. Logic demands that both the training and validation parts should have uniformly distributed objects densities and their properties (e.g.size, distance). We therefore decided to evaluate the correlation between the theoretical uniform distribution and observed one by calculating the Wasserstein distance <ref type="bibr" target="#b24">[25]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Analysis of outliers</head><p>There is no reason to assume that all data are flawlessly annotated since the task is usually done by several people, which increases the uncertainty of an inconsistent annotation style. We therefore designed plausibility check functions which allow us to indicate the potentially wrong annotated data and analyze them later on.</p><p>• 𝑓1 : 𝑟𝑒𝑡𝑢𝑟𝑛 (0.0 ≥ 𝑟𝑒𝑙.𝑏𝑏.𝑠𝑖𝑧𝑒 &gt; 1.0)</p><formula xml:id="formula_1">• 𝑓2 : 𝑟𝑒𝑡𝑢𝑟𝑛 (𝑒 − 𝑑𝑖𝑠𝑡 𝑑 +𝑙𝑜𝑛𝑔.𝑐𝑎𝑚.𝑜𝑓 𝑓 ) • 𝑓3 : 𝑟𝑒𝑡𝑢𝑟𝑛 (𝑖𝑛𝑡𝑒𝑟𝑠𝑒𝑐(𝑏𝑏𝑖, 𝑏𝑏𝑗) &gt; 𝑡ℎ𝑟 𝑜𝑣𝑒𝑟𝑙𝑎𝑝 ) • 𝑓4 : 𝑟𝑒𝑡𝑢𝑟𝑛 (𝑜𝑝𝑡.𝑓 𝑙𝑜𝑤(𝑖𝑚𝑔𝑖−1, 𝑖𝑚𝑔𝑖) &gt; 𝑡ℎ𝑟𝑠𝑡𝑎𝑡𝑖𝑐)</formula><p>The function 𝑓1 returns a bounding box whose relative size is out of the range (0.0, 1.0]. The outliers of otherwise exponentially decayed objects' size with respect to the distance to the ego are marked by function 𝑓2. This function covers the quadratic dependency of the BB area and the mapping of the captured object within the real world to the pixel coordinates. The parameter 𝑙𝑜𝑛𝑔.𝑐𝑎𝑚.𝑜𝑓 𝑓 and the denominator 𝑑 were found with the least squares optimization from the analyzed data and therefore are unique for each dataset and each class. The function 𝑓3 highlights objects whose bounding boxes significantly overlap. The threshold of the relative intersection area can be defined by 𝑡ℎ𝑟 𝑜𝑣𝑒𝑟𝑙𝑎𝑝 .</p><p>Examples of the outcomes for functions 𝑓1 − 𝑓3 are given in Figure <ref type="figure" target="#fig_0">1</ref>. We further found that most of the datasets contain sequences of "stop and go" in a traffic jam, or idling at crossroads. These situations result in the recording of many similar images without objects or surrounding variation. Therefore we added which calculates a dense optical flow <ref type="bibr" target="#b25">[26]</ref> from previous and actual images in the sequence and returns a positive flag in case the magnitude sinks bellow empirically defined threshold as seen in Figure <ref type="figure" target="#fig_2">2</ref>. The higher the value of the magnitude, the more objects were moving from frame to frame. With this method, we could even identify if the data were recorded repeatedly in the same place <ref type="bibr" target="#b26">[27]</ref> (when recorded in one session), but we only used it to discover static scenarios. We further encourage to analyse such a subset of filtered images and adapt the parameters of the outliers function 𝑓 1 − 𝑓 4 to make them suitable to the target domain.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Overall results</head><p>Exemplary results of the analysis of the objects' distances distribution as well as the relation between the relative BB size and object distance can be seen in Figures <ref type="figure" target="#fig_4">3 and 4</ref>. Moreover, we show an example of object appearance variation (heatmap) of class Vehicle in the A2D2's dataset in Figure <ref type="figure" target="#fig_5">5</ref>. The information from subset (0.0 to 14.0) meters will be used in scenarios where a vehicle is driving less than 50km/h (idling on a crossroads, for instance). However, the A2D2 dataset doesn't contain any objects with the minimal safe distance necessary for highway driving (92.58m and more).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Concluding Remarks</head><p>To summarize, most of the SOTA autonomous driving datasets are generated with a focus on a large amount of scenes/driving hours/frames while considering different weather, location, and daylight conditions. However, as we discovered, none of the datasets contains a sufficient amount of information for safe autonomous driving on the highway (described in Section 3). Every analyzed dataset was lacking high-distance annotated objects, as can be seen in Table <ref type="table" target="#tab_2">3</ref>. Such a gap can be explained by the physical limitations of the camera (too low resolution),   the annotation style (objects under a certain pixel area were excluded from the annotation process), and the ambient conditions in which the dataset was recorded. In addition, a system trained on such a dataset would have to deal with epistemic uncertainty and look for additional sources of information (namely LIDAR).</p><p>Furthermore, all datasets contain predominantly smallsized objects (the highest MEAN value of the relative BB size of the class Person was 0.091 in the case of the KITTI dataset). For comparison, the same can be stated for the well-known COCO dataset <ref type="bibr" target="#b18">[19]</ref>, where the class Person has a MEAN relative BB size equal to 0.089. By generating heatmaps, we discovered that 99.8% of the objects appear only in the two lower quadrants of the image. Such information can lead to a significant downsizing of the field of view and the thereof acceleration of the detectors' inference time. The majority of overlapping BBs, with potentially wrong annotation styles, were extracted from a sequence of streams on crossroads. Such a static data sequence (9.55% in case of nuScenes dataset) contains a lot of similar features (the majority of surrounding objects are not moving) and could be removed from the dataset.</p><p>Finally, we defined and evaluated a reasonable set of rules, described in Section 4.1, which automatically proves the quality of the collected data from a domainrelated perspective. We encourage the community to use our "domain-centric" approach in order to create a dataset under concrete functional constraints and train detectors on it. Our code and additional results are published on GitLab and can be publicly accessed. 1  This work deepened our vision of a domain-centric ML approach in the automotive industry. To conclude, we outline some research directions which we are currently investigating: (a) Analysis of automotive mAP based on 1 https://gitlab.com/arrk-fi/ObjectDetectionCriticality/-/tree/ dependency_branch. </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Examples of outliers (potentially wrongly annotated data) within the analysed datasets. Images (a, b) capture overlapping BBs, whereas images (c, d) contain objects where the relative BB size doesn't correlate with the object distance.We further encourage to analyse such a subset of filtered images and adapt the parameters of the outliers function 𝑓 1 − 𝑓 4 to make them suitable to the target domain.</figDesc><graphic coords="5,89.94,476.22,99.99,62.91" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>optical flow Magnitude of dense optical flow over dataset: nuScenes actual value median value, win_size = 4 static scene threshold</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: As visible on the figure, 2.3% of the nuScenes dataset are "stop and go" scenarios, where the magnitude of the dense optical flow drops below a static scene threshold. The threshold was recalculated for each dataset based on the image resolution.</figDesc><graphic coords="5,89.94,396.52,99.99,62.91" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3:The information from subset (0.0 to 14.0) meters will be used in scenarios where a vehicle is driving less than 50km/h (idling on a crossroads, for instance). However, the A2D2 dataset doesn't contain any objects with the minimal safe distance necessary for highway driving (92.58m and more).</figDesc><graphic coords="5,192.00,476.22,99.99,62.91" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The black curve indicates the outliers' decision boundary of otherwise exponentially-decayed objects size with respect to the distance to the ego. Outliers can indicate rotated, or wrongly annotated (with unnecessary big margin) objects.</figDesc><graphic coords="6,122.26,462.79,127.44,80.16" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: The number in the top left corner of each quadrant indicates the number of objects' appearances in the respective area. Contrary to the expected symmetrical heatmap, the vehicle appears with different quantities and sizes on both sides of the ego. The distribution is obviously unbalanced in the vertical and horizontal directions containing the majority of the objects in the lower half of the image. Such a statistical information can be used in post-processing by plausibility check of objects appearance.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Left: We picked a Small Vehicle class where a various BB relative size density can be seen. It is visible that the majority of large objects has very small relative BB size. Right: We show the distance to object distributions of class pedestrian for each dataset, where 𝑛𝑠𝑎𝑚𝑝 is number of samples and 𝑤 𝑑𝑖𝑠𝑡 is Wasserstein Distance between Uniform distribution 𝒰 and the dataset distance to object distributions. The lower the value, the closer the two distributions are.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Comparison of analyzed datasets. Cells with "-" indicate not mentioned in the original paper.</figDesc><table><row><cell>Dataset</cell><cell cols="2">Scenes # images</cell><cell># 3D Bounding Boxes</cell><cell>Ambient conditions</cell><cell># classes</cell><cell>Year of release</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>urban, highways,</cell><cell></cell><cell></cell></row><row><cell>ONCE [18]</cell><cell>1M</cell><cell>7M</cell><cell>417k</cell><cell>country roads;</cell><cell>5</cell><cell>2021</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>day/night; various weather</cell><cell></cell><cell></cell></row><row><cell>nuScenes [16]</cell><cell>1k</cell><cell>1.4M</cell><cell>1.4M</cell><cell></cell><cell>23</cell><cell>2020</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>urban, highways,</cell><cell></cell><cell></cell></row><row><cell>A2D2 [14]</cell><cell>-</cell><cell>-</cell><cell>-</cell><cell>country roads;</cell><cell>14</cell><cell>2020</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>day; various weather</cell><cell></cell><cell></cell></row><row><cell>LyftLevel5 [15]</cell><cell>366</cell><cell>323k</cell><cell>1.3M</cell><cell>urban; day; various weather</cell><cell>9</cell><cell>2019</cell></row><row><cell>Waymo [17]</cell><cell>1150</cell><cell>1M</cell><cell>12M</cell><cell>urban; day/night; various weather</cell><cell>4</cell><cell>2019</cell></row><row><cell>KITTI [13]</cell><cell>22</cell><cell>15k</cell><cell>200k</cell><cell>urban; day; sunny, cloudy</cell><cell>8</cell><cell>2012</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>All scenarios with calculated minimum safe distances.</figDesc><table><row><cell></cell><cell>Max. allow</cell><cell>Time to</cell><cell>Min. safe</cell></row><row><cell>Scenario</cell><cell>speed</cell><cell>stand still</cell><cell>distance 𝑑𝑠</cell></row><row><cell></cell><cell>[km/h]</cell><cell>[s]</cell><cell>[m]</cell></row><row><cell>highway</cell><cell>130</cell><cell>5.14</cell><cell>92.58</cell></row><row><cell>country road</cell><cell>100</cell><cell>4.00</cell><cell>55.11</cell></row><row><cell>city</cell><cell>50</cell><cell>2.00</cell><cell>14</cell></row><row><cell cols="4">any ambient conditions, type of vehicle, and speed limi-</cell></row><row><cell>tations.</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Comparison of the analysed datasets highlighting the best results in regard to the training of an object detector used for trajectory planning.</figDesc><table><row><cell cols="3">Research Questions</cell><cell></cell><cell cols="2">Class</cell><cell>ONCE</cell><cell>nuScenes</cell><cell>A2D2</cell><cell>LyftLevel5</cell><cell>Waymo</cell><cell>KITTI</cell></row><row><cell cols="4">Which dataset contains the most distant objects (driving ≥ 130𝑘𝑚/ℎ)?</cell><cell cols="2">Ped. Veh.</cell><cell>0 0</cell><cell>6 2464</cell><cell>0 4</cell><cell>55 1574</cell><cell>0 0</cell><cell>0 0</cell></row><row><cell cols="4">Which datasets has the most uniform distribution according to the Wasserstein distance?</cell><cell cols="2">Ped. Veh.</cell><cell>0.86 0.74</cell><cell>0.82 0.57</cell><cell>0.80 0.81</cell><cell>0.60 0.50</cell><cell>0.65 0.53</cell><cell>0.89 0.70</cell></row><row><cell cols="4">Which datasets has the least</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">outliers based on our sanity</cell><cell></cell><cell>All</cell><cell>212</cell><cell>4742</cell><cell>563</cell><cell>146</cell><cell>6935</cell><cell>288</cell></row><row><cell cols="3">check filters (f1 -f3)?</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell cols="3">Which dataset contains the</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="4">least static images according</cell><cell></cell><cell>All</cell><cell>358</cell><cell>786</cell><cell>157</cell><cell>59</cell><cell>895</cell><cell>77</cell></row><row><cell cols="3">to optical flow f4?</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell cols="6">Relative BB size as a function of distance to object for class: Truck</cell><cell></cell></row><row><cell></cell><cell>14.0</cell><cell></cell><cell>55.11</cell><cell></cell><cell></cell><cell>92.58</cell><cell></cell></row><row><cell></cell><cell>1.0</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>Relative BB size [-]</cell><cell>0.4 0.6 0.8</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">Minimum safe distance in [m] e x 6.0 + 0.075 for 130km/h for 100km/h for 50km/h</cell></row><row><cell></cell><cell>0.2</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>0.0</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>0</cell><cell>20</cell><cell cols="2">40 Distance to object [m] 60</cell><cell>80</cell><cell>100</cell><cell>120</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work is partly funded by ARRK Engineering GmbH. The work has also been supported by the grant of the University of West Bohemia, project No. SGS-2022-017 and by the Technology Agency of the Czech Republic, project No. CK03000179.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">E</forename><surname>Schwalb</surname></persName>
		</author>
		<title level="m">Analysis of safety of the intended use</title>
				<imprint>
			<publisher>sotif</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Managing the development of large software systems: concepts and techniques</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">W</forename><surname>Royce</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th international conference on Software Engineering</title>
				<meeting>the 9th international conference on Software Engineering</meeting>
		<imprint>
			<date type="published" when="1987">1987</date>
			<biblScope unit="page" from="328" to="338" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The waterfall model in large-scale development</title>
		<author>
			<persName><forename type="first">K</forename><surname>Petersen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wohlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Baca</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Product-Focused Software Process Improvement</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="386" to="400" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<idno>ISO 26262</idno>
		<title level="m">Road vehicles -Functional safety (ISO 26262)</title>
				<imprint>
			<publisher>International Organization for Standardization</publisher>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<idno>ISO/PAS 21448</idno>
		<title level="m">Road vehicles -Safety of the intended functionality, Standard, International Organization for Standardization</title>
				<imprint>
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Ng</surname></persName>
		</author>
		<ptr target="https://https-deeplearning-ai.github.io/data-centric-comp/" />
		<title level="m">Deep Learning AI data-centric ai competition</title>
				<imprint>
			<date type="published" when="2021">2021. 2022-08-16</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m">Taxonomy and definitions for terms related to on-road motor vehicle automated driving systems</title>
				<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">3016</biblScope>
		</imprint>
		<respStmt>
			<orgName>SAE International</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Path planning and trajectory planning algorithms: A general overview, Motion and operation planning of robotic systems</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gasparetto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Boscariol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lanzutti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Vidoni</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="3" to="27" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<author>
			<persName><forename type="first">F</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Xian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Madhavan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Darrell</surname></persName>
		</author>
		<idno>CoRR abs/1805.04687</idno>
		<ptr target="http://arxiv.org/abs/1805.04687.arXiv:1805.04687" />
		<title level="m">BDD100K: A diverse driving video database with scalable annotation tooling</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">The cityscapes dataset for semantic urban scene understanding</title>
		<author>
			<persName><forename type="first">M</forename><surname>Cordts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Omran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ramos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rehfeld</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Enzweiler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Benenson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Franke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Roth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Schiele</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<meeting>of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="3213" to="3223" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Ford campus vision and lidar data set</title>
		<author>
			<persName><forename type="first">G</forename><surname>Pandey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Mcbride</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Eustice</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The International Journal of Robotics Research</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="1543" to="1552" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Trafficpredict: Trajectory prediction for heterogeneous traffic-agents</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Manocha</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI Conference on Artificial Intelligence</title>
				<meeting>the AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="6120" to="6127" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Vision meets robotics: The kitti dataset</title>
		<author>
			<persName><forename type="first">A</forename><surname>Geiger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lenz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Urtasun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The International Journal of Robotics Research</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="1231" to="1237" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">A2d2: Audi autonomous driving dataset</title>
		<author>
			<persName><forename type="first">J</forename><surname>Geyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kassahun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mahmudi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ricou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Durgesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hauswald</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">H</forename><surname>Pham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mühlegg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dorn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Fernandez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jänicke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mirashi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Savani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sturm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Vorobiov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Oelker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Garreis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Schuberth</surname></persName>
		</author>
		<idno type="DOI">10.48550/ARXIV.2004.06320</idno>
		<ptr target="https://arxiv.org/abs/2004.06320.doi:10.48550/ARXIV.2004.06320" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Kesten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Usman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Houston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Pandya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Nadhamuni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ferreira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Low</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ondruska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Omari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kulkarni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kazakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Platinsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Shet</surname></persName>
		</author>
		<ptr target="https://level-5.global/level5/data/" />
		<title level="m">Level 5 perception dataset 2020</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">nuscenes: A multimodal dataset for autonomous driving</title>
		<author>
			<persName><forename type="first">H</forename><surname>Caesar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Bankiti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">H</forename><surname>Lang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Vora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">E</forename><surname>Liong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Krishnan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Baldan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Beijbom</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CVPR</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="11621" to="11631" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Scalability in perception for autonomous driving: Waymo open dataset</title>
		<author>
			<persName><forename type="first">P</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kretzschmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Dotiwalla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chouard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Patnaik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tsui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Caine</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE/CVF conference on computer vision and pattern recognition</title>
				<meeting>the IEEE/CVF conference on computer vision and pattern recognition</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="2446" to="2454" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Niu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<title level="m">One million scenes for autonomous driving: Once dataset</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Microsoft coco: Common objects in context</title>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Maire</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Belongie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hays</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Perona</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ramanan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dollár</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Zitnick</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ECCV</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="740" to="755" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Van Der</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Horst</surname></persName>
		</author>
		<author>
			<persName><surname>Hogema</surname></persName>
		</author>
		<title level="m">Time-to-collision and collision avoidance systems</title>
				<imprint>
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">I</forename></persName>
		</author>
		<ptr target="https://unfallanalyse.hamburg/index.php/ifu-lexikon/bremsen/bremstabelle-a/" />
		<title level="m">Bremstabelle hamburg -institut für unfallanalysen -bremstabelle</title>
				<imprint>
			<date type="published" when="2022-09-14">2022. 2022-09-14</date>
		</imprint>
	</monogr>
	<note>für Unfallanalysen</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Experimental research of effectiveness of brakes in passenger cars under selected conditions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Erd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jaśkiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Koralewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Rutkowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Stokłosa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2018 Xi International Science-Technical Conference Automotive Safety, IEEE</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1" to="5" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">What uncertainties do we need in bayesian deep learning for computer vision?</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kendall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gal</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NIPS</title>
		<imprint>
			<biblScope unit="page" from="5574" to="5584" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Early stopping-but when?</title>
		<author>
			<persName><forename type="first">L</forename><surname>Prechelt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Neural Networks: Tricks of the trade</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="1998">1998</date>
			<biblScope unit="page" from="55" to="69" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Optimal transport, old and new. notes for the 2005 saint-flour summer school</title>
		<author>
			<persName><forename type="first">C</forename><surname>Villani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Fundamental Principles of Mathematical Sciences</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<date type="published" when="2008">2008</date>
		</imprint>
	</monogr>
	<note>Springer</note>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Two-frame motion estimation based on polynomial expansion</title>
		<author>
			<persName><forename type="first">G</forename><surname>Farnebäck</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Scandinavian conference on Image analysis</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="363" to="370" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Real-time simultaneous localisation and mapping with a single camera</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Davison</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computer Vision, IEEE International Conference on</title>
				<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1403" to="1403" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
