<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Providing brands visibility data in live sports videos using deep learning algorithms *</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Julius</forename><surname>Gudauskas</surname></persName>
							<email>julius.gudauskas@ktu.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Applied Informatics</orgName>
								<orgName type="institution">Kaunas University of Technology</orgName>
								<address>
									<addrLine>Studentu 50</addrLine>
									<settlement>Kaunas</settlement>
									<country key="LT">Lithuania</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">Information Society</orgName>
								<orgName type="institution">University Studies</orgName>
								<address>
									<addrLine>2024, May 17</addrLine>
									<settlement>Kaunas</settlement>
									<country key="LT">Lithuania</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Providing brands visibility data in live sports videos using deep learning algorithms *</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">7427189287C5C417EA73C2AC4058E201</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:27+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Brands visibility</term>
					<term>logo detection</term>
					<term>YOLO</term>
					<term>Faster R-CNN</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the dynamic landscape of marketing and advertising, assessing brand visibility in live sports events plays a pivotal role in understanding brand exposure and impact. Traditional methods of manual annotation and analysis are time-consuming and subjective, necessitating automated solutions for efficient and objective evaluation. In this study proposed a novel approach leveraging deep learning algorithms to evaluate brand visibility in live sports videos. This research employs state-of-the-art object detection models, such as YOLO (You Only Look Once) and Faster R-CNN, to detect and localize brand logos within video frames. By training these models on annotated open-source logo datasets, we can extract valuable insights about the brands. The experimental results demonstrate the effectiveness of the proposed methodology in detecting logos and providing a valuable data about the positions for brand owners.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the world of sports live video streaming, a considerable number of brands are trying to get noticed by the audience using various popular visibility materials: posters, stickers, billboards, etc. The research evaluating the impact and effectiveness of advertisements in sports arenas emphasizes that people notice at least some of the advertisements they are exposed to and usually remembers a part of them that were the most noticeable <ref type="bibr" target="#b0">[1]</ref>. Typically, clients engage in negotiations with advertising executives to determine the conditions that will govern brand placement in the arena -specifying factors like coverage, frequency of display on advertisement billboards, and overall visibility strategies. A study conducted by Eventmarketer <ref type="bibr" target="#b1">[2]</ref> found that 72% of the audience are captivated by the brand when they see it during the events like music festivals or sports competitions where they are provided with good emotions and excitement. These occasions, characterized by heightened emotions and excitement, offer a unique opportunity for brands to establish a connection with a vast and diverse audience, potentially converting them into new users. However, to reach a wider audience, the brand must be placed in a visible location. Studies has shown that locations, such as boundary line hoarding are considered as the perfect place to display brand logos without irritating the viewers and getting maximum visibility <ref type="bibr" target="#b2">[3]</ref>  <ref type="bibr" target="#b3">[4]</ref>. As sponsorship agreements comes with a significant cost, brand owners are interested to know if their investment is paying off. But measuring the effectiveness of different brand placement can prove challenging, time consuming and requiring manual work, leaving brand owners uncertain about the true impact of their investment in sponsorship deals.</p><p>A comprehensive understanding of the effectiveness of brand advertising through the integration of deep learning techniques in visual material analysis remains an evolving area, prompting the need for further research to refine methodologies and uncover insights that can inform strategic marketing decisions. To calculate the brand visibility metrics, the logos detection algorithm, that has the ability to detect all the logos, needs to be created. Current logo detection methods often focus on a limited set of logo categories, necessitating extensive training data that includes annotations for object bounding boxes <ref type="bibr" target="#b4">[5]</ref>. But the main challenge remains the steep growth of the existing brand's amount and the brand image changes in the current ones. In this research, we addressed an open logo detection challenge and provided a unified brand's logo detection and recognition approach, using up to date machine learning algorithms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related works</head><p>Simple methods, that were developed as the early approaches for specific logo detections, relies on manually engineered visual features and traditional classification models <ref type="bibr">[6] [7]</ref>. But such methods have a huge flaw: the region selection search algorithm based on sliding windows struggles to provide high accuracy in the high time complexity, manually created features lack robustness for logo diversity. Recent advancements in deep learning have revolutionized the field of visual material analysis, providing novel avenues for detecting and recognizing various objects, including logos. Many researchers have explored the application of deep neural networks in various image recognition tasks, demonstrating their capability to extract complex features and patterns. In the context of logo detection, the solution is usually determined by the size of the logos that has to be detected. Compared to a bigger size logos, smaller ones are more difficult to detect for several reasons: small logos with the low resolution contain little visual information, small logos covers a small area, it's surrounding box is more challenging to locate and there are usually less small logo samples <ref type="bibr" target="#b7">[8]</ref>. In the small vehicles logo detection problem proceeding <ref type="bibr" target="#b8">[9]</ref>, researchers solved the issue using a YOLO <ref type="bibr" target="#b9">[10]</ref> algorithm. In contrast to traditional methods relying on manual feature extraction, this system offers the benefits of selflearning features and direct image input. It can efficiently achieve both vehicle logo positioning and recognition functions. The researchers also introduce the Fast RCNN approach, that employs deep neural networks, utilizing convolutional layers to progressively extract abstract feature representations learned from previous convolutions <ref type="bibr" target="#b10">[11]</ref>  <ref type="bibr" target="#b11">[12]</ref>.</p><p>Natural visual scenes usually exhibit complexity and diversity -logos face various challenges, including object interference, shape distortion, different lighting, and limited perspective effects, which increases the difficulty of logo detection. In a recent development, researchers introduced a transfer learning approach, leveraging Densely Connected Convolutional Networks (DenseNet) for logo recognition <ref type="bibr" target="#b12">[13]</ref>. They tested their method on the FlickerLogos-32 dataset and reached the accuracy higher than 92%. The visibility can also be impacted by the bad weather conditions. In the challenge, where the logos has to be detected in bad weather, authors presented an object proposal generation system AttentionMask <ref type="bibr" target="#b13">[14]</ref>. The experimental findings indicate that the suggested approach demonstrates strong capabilities in identifying logos within intricate real-world settings. Nonetheless, data gathered from real-world scenarios may not match the quality of artificially augmented data, leading to a decline in the model's performance when detecting images in such real-world conditions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>This work proposes a solution: a model that can identifying and predicting the bounding box around any logo within an image, irrespective of brand. Given the constant influx of new businesses, maintaining an up-to-date model with sufficient data for each brand proves challenging, potentially leading to difficulties in detecting logos of newly established brands during inference. To address this, the aim is to develop a model focused solely on detecting logos in images, regardless of brand affiliation. This approach eliminates the need to specifically train the model on individual brands, ensuring accurate detection of logos irrespective of their origin. The solution is focused on creating the model with highest accuracy, so the implementation is done using two different object detection algorithm pipelines: one -stage, that utilizes a single pass of the input image and enables the processing of the entire image in one go and two-stage models, that uses two passes of the image to make a prediction, where first pass is used to generate a set of proposals or potential object locations, and the second pass is used to refine these proposals and make final predictions <ref type="bibr" target="#b14">[15]</ref>. In the category of one -stage models, YOLOv7 <ref type="bibr" target="#b15">[16]</ref> visualized in Figure <ref type="figure" target="#fig_0">1</ref>, were selected, because of its advantage in detecting smaller objects compared to a single -shot detection approach, and for two -stage -selected Faster R-CNN, that is visualized in Figure <ref type="figure" target="#fig_1">2</ref>.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Data</head><p>In conducting research on existing open-source logo datasets in Table <ref type="table" target="#tab_0">1</ref>, the attention was dedicated to comprehensively evaluating the diversity of logos within these repositories. Given the diverse nature of logos, ranging from graphic designs to text-based representations, the consideration was given to ensure that the selected datasets contain a wide spectrum of logo types.  QMUL-OpenLogo <ref type="bibr" target="#b17">[18]</ref>. This dataset contains more than 27000 images with 352 different logos.It is a benchmark dataset for logo detection, formed by combining seven existing datasets and establishing an open protocol for evaluating detection performance. This dataset demonstrates a significant imbalance in distribution and notable variations in scale, crucial aspects for evaluating the effectiveness of detection algorithms.</p><p>LogoDet-3k <ref type="bibr" target="#b17">[18]</ref>: This dataset contains more than 3000 unique classes, 158652 images with labeled logo symbols (dataset example in Figure <ref type="figure">3</ref>.). This dataset divides logos into 9 different sub -categories: food, clothes, necessities, electronics, transportation, leisure time equipment, sports, and medicine and other (described in Table <ref type="table" target="#tab_2">3</ref>.). The main advantage of this data set is the large number of variations of the same brands -positions, lighting conditions, angles. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Data preprocessing</head><p>In object detection models, the training data encompasses crucial elements, such as the images itself, the coordinates of object bounding boxes, and their respective labels. Brand logo datasets commonly feature annotations tailored to individual brands -each brand is labeled by the name of it. However, this approach presents a notable challenge: models must be trained separately to detect each brand and will require additional fine -tuning to recognize newly introduced brands. Moreover, some brands will contain more training images than the others, so it also introduces a class disbalance problem that will potentially impact the model's performance.</p><p>In response to these challenges, introduced a data preprocessing step, that involves categorizing brands into two distinct groups: logos with textual elements and graphicbased logos. The result is achieved in this workflow visualized in Figure <ref type="figure" target="#fig_2">4</ref>.:</p><p> Each image is cropped by the bounding box coordinates.  The image is processed using pytesseract -one of the most popular Python libraries for optical character recognition.  If the optical characters were detected -assign the logo with the "TextLogo" label.</p><p>If not -assign the logo with the "GraphicsLogo" label. The newly created dataset (in Table <ref type="table" target="#tab_3">4</ref>.) contains a noticeable class imbalance This might have a huge impact in the model's ability to accurately assign the correct labels, leading to poor overall predictions. To address this issue, more graphic logos from other datasets were added, leaving the final dataset with a similar amount of each category. By ensuring a balanced distribution of samples across both categories, the model is equipped to learn effectively from a diverse range of examples, enhancing its capacity to make accurate predictions across all classes The final dataset was formed using the same amount graphics logos and text logos. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>This research presents a comparative analysis between two object detection frameworks: YOLO (You Only Look Once) and Faster R-CNN. Specifically focusing on their efficiency to detect logo's bounding boxes, the aim is to provide an accuracy comparison between onestage and two-stage detection methodologies. The comparison analysis is essential for understanding the trade-offs between accuracy and efficiency. While one-stage detectors are generally faster, they might sacrifice some accuracy compared to two-stage detectors. By quantitatively comparing the accuracy of bounding box detection between the two methods, determination can be done whether the sacrifice in accuracy is acceptable given the efficiency gains.</p><p>The experiments were done using specifically crafted dataset (3.2) with the model's parameters that provided the best accuracy results:</p><p>• Faster R-CNN with ResNet-50 V1 FNP backbone, leveraging pre-trained weights, a learning rate (lr) set at 10-4, a momentum (m) of 0.9, and a weight decay (wd) of 10-6 and a batch size of 64.</p><p>• YOLOv7 with a learning rate (lr) set at 10-4, L2 regularization at 10-4 and batch size of 64.</p><p>The evaluation of deep learning models was made using an mAP metric. Mean Average Precision (mAP) is a widely used metric for evaluating the performance of machine learning models, particularly in object detection and recognition tasks. It provides the assessment of a model's ability to accurately identify objects within an image dataset. The mAP metric calculates the average precision for each class of objects across all images, then averages these values to produce a single score. This score reflects both the precision (the ratio of true positive predictions to all positive predictions) and recall (the ratio of true positive predictions to all actual positives) of the model. A higher mAP indicates better performance, with a score of 100% representing perfect detection accuracy. By utilizing mAP, we can quantify and compare the effectiveness of different models, aiding in the advancement of computer vision technologies.</p><p>Both model's data were augmented using random sized box crop with probability of 0.5, preserving the integrity of logos bounding boxes during random cropping and resizing. Horizontal Flip and Vertical Flip operations were applied independently, each with a 0.3 probability, introducing variations in viewpoint by flipping images horizontally and vertically. Data augmentation helped to increase YOLO model performance by 3% mAP and Faster R-CNN by 2% mAP. Both model accuracies in the last 70 epochs presented in the chart Figure <ref type="figure" target="#fig_3">5</ref>. Even though, the Faster R-CNN achieved better accuracy results, the speed cost is significant compared both methods. During the performance analysis noticed, that YOLOv7 performs 13% faster. The Faster R-CNN model is more favorable when more time is available, but the YOLO algorithm is more favorable for the real-time task. During the comparison analysis between different Faster R-CNN FNP architectures noticed that ResNet-50 v1 demonstrates noticeable accuracy improvement over MobileNet-V3 and VGG16. The evaluation based on mAP metric shows that ResNet-50v1 provides up to 6% better accuracy results, compared to MobileNet and up to 5% compared to VGG16. Different FNP accuracies in the last 70 epochs presented in chart Figure <ref type="figure" target="#fig_4">6</ref>.  The real-world testing were done using 100 custom annotated frames from sport video footage with 664 logos in it. The model detected 523 logos bounding boxes -78% of all logos that were in testing data. The results presented in the confusion matrix Figure <ref type="figure" target="#fig_5">7</ref>. In this study explored a deep learning algorithms to evaluate brand visibility in live sports videos, presenting a novel approach to address the challenges of manual annotation and subjective analysis. Through the utilization of advanced object detection models like YOLO and Faster R-CNN, the final model demonstrated the capability of automated methods to accurately detect and localize brand logos within the dynamic context of sport videos. Even though the final model demonstrated the accuracy of detecting eight out of ten logos in real -world video, this finding emphasizes the significant potential of automated solution in overcoming the limitations associated with manual annotation, offering more objective and more efficient evaluation of brand positioning. Looking forward, the continued development and refinement of deep learning methodologies, coupled with advancements in real-time monitoring capabilities, hold promise for further enhancing the accuracy and effectiveness of brand visibility evaluation in live sports videos. Additionally, the availability of larger and more diverse annotated datasets will be necessary in improving model performance and generalization.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: YOLOv7 workflow [16].</figDesc><graphic coords="3,103.80,354.60,387.70,202.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Faster R-CNN workflow.</figDesc><graphic coords="3,100.65,591.75,394.10,153.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Data pre-processing pipeline Rather than developing a new dataset, this approach leverages existing ones. Such approach not only saves time and resources, but also ensures that the model benefits from a diverse range of logo samples. Additionally, providing a minimum of two classes for training is essential for any object detection model to effectively learn and generalize. The final dataset for model training is stored in COCO (Common Objects in Context) format.The newly created dataset (in Table4.) contains a noticeable class imbalance This might have a huge impact in the model's ability to accurately assign the correct labels, leading to poor overall predictions. To address this issue, more graphic logos from other datasets were added, leaving the final dataset with a similar amount of each category. By ensuring a balanced distribution of samples across both categories, the model is equipped to learn effectively from a diverse range of examples, enhancing its capacity to make accurate predictions across all classes The final dataset was formed using the same amount graphics logos and text logos.</figDesc><graphic coords="6,88.55,78.75,419.20,79.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Faster R-CNN and YOLOv7 comparison.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Faster R-CNN different FNP comparison.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Testing results confusion matrix.</figDesc><graphic coords="8,157.10,368.95,271.50,144.50" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Open-source logo datasets. This dataset includes 32 different logo classes from various domains. Since it contains about 2240 images with marked boundaries coordinates, it is well suited for building a model for brand detection and recognition (Table2.)</figDesc><table><row><cell>Dataset name</cell><cell>Logos</cell><cell>Images</cell><cell>Year</cell></row><row><cell>BelaLogos [15] FlickrLogos-32 [16]</cell><cell>37 27</cell><cell>10,000 2240</cell><cell>2009 2011</cell></row><row><cell>SportLogo [17]</cell><cell>31</cell><cell>2836</cell><cell>2020</cell></row><row><cell>QMUL-OpenLogo [18]</cell><cell>352</cell><cell>27,083</cell><cell>2018</cell></row><row><cell>FoodLogoDet-1500 [19]</cell><cell>1500</cell><cell>99,768</cell><cell>2021</cell></row><row><cell>LogoDet-3K [18]</cell><cell>3000</cell><cell>158,652</cell><cell>2020</cell></row><row><cell>FlickrLogos-32 [15]:</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc></figDesc><table><row><cell>FlickrLogos-32</cell><cell></cell><cell></cell><cell></cell></row><row><cell>analysis. Datase t part</cell><cell>Description</cell><cell>Images</cell><cell>Total images count</cell></row><row><cell>P1</cell><cell>Selected pictures with clean backgrounds. One brand logo in each image.</cell><cell>10 for each logo</cell><cell>320</cell></row><row><cell>P2 P3</cell><cell>Different images showing at least one brand logo in different conditions. Pictures without logos. Different images showing at least one brand logo in different conditions.</cell><cell>30 for each logo 3000 without any logo 30 for each logo 3000 without any</cell><cell>3960 3960</cell></row><row><cell></cell><cell>Pictures without logos.</cell><cell>logo</cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>LogoDet-3k analysis.</figDesc><table><row><cell></cell><cell>Figure 3: LogoDet-3k visualisation.</cell><cell></cell></row><row><cell>Category</cell><cell>Images count</cell><cell>Total brands count</cell></row><row><cell>Food brands Clothes brands</cell><cell>53,350 31,266</cell><cell>64,276 37,601</cell></row><row><cell>Necessities brands</cell><cell>24,822</cell><cell>30,643</cell></row><row><cell>Electronics brands</cell><cell>9,675</cell><cell>12,139</cell></row><row><cell>Transportation brands</cell><cell>10,445</cell><cell>12,791</cell></row><row><cell>Leisure time equipment brands</cell><cell>5,685</cell><cell>6,573</cell></row><row><cell>Sports brands</cell><cell>3,953</cell><cell>5,041</cell></row><row><cell>Medicine brands</cell><cell>3,945</cell><cell>5,185</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Logo categories results after preprocessing.</figDesc><table><row><cell>Dataset</cell><cell>Text logo</cell><cell>Graphics logo</cell></row><row><cell>LogoDet-3K FlickrLogos-32 and others Total</cell><cell>168,416 113 281,416</cell><cell>25,845 3,154 29,008</cell></row></table></figure>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The impact and effectiveness of advertisements in a sports arena</title>
		<author>
			<persName><forename type="first">L</forename><surname>Turley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">Richard</forename><surname>Shannon</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of services marketing</title>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Shaz</surname></persName>
		</author>
		<title level="m">Experiential Marketing. A practical guide to interactive brand experiences</title>
				<meeting><address><addrLine>London and Philadelphia</addrLine></address></meeting>
		<imprint>
			<publisher>Kogan Page</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Impact of Elements of Ad&apos;s on Sports Fan Attitude during a Live Sporting Event</title>
		<author>
			<persName><forename type="first">G</forename><surname>Raghav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Aradhana</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scholarly Journaly</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="page" from="867" to="876" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">An analytical study of consumer awareness</title>
		<author>
			<persName><forename type="first">R</forename><surname>Arora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chawla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sachdeva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Advanced Research in Management and Socal Sciences</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">8</biblScope>
			<biblScope unit="page" from="137" to="153" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Mutual enhancement for detection of multiple logos in sports videos</title>
		<author>
			<persName><forename type="first">L</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xiaoqing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chengcui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yongtao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Computer Vision</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Elliptical ASIFT agglomeration in class prototype for logo detection</title>
		<author>
			<persName><forename type="first">R</forename><surname>Boia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Florea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Florea</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the British Machine Vision Conference</title>
				<meeting>the British Machine Vision Conference</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Tree-basedShapeDescriptorforscalablelogodetection</title>
		<author>
			<persName><forename type="first">C</forename><surname>Wan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Visual Communications and Image Processing</title>
				<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Deep Learning for Logo Detection: A Survey</title>
		<author>
			<persName><forename type="first">H</forename><surname>Sujuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jiacheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Weiqing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">Y Z</forename><surname>Qiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yuanjie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shuqiang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Multimedia Comput. Commun</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">72</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A real-time vehicle logo detection method based on improved YOLOv2</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Kangning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Shaoqi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Guangqiang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Wireless Algorithms, Systems, and Applications</title>
				<meeting>the International Conference on Wireless Algorithms, Systems, and Applications</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">You only look once: Unified, real-time object detection</title>
		<author>
			<persName><forename type="first">J</forename><surname>Redmon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kumar Divvala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">B</forename><surname>Girshick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Farhadi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Improving Small Object Proposals for Company Logo Detection</title>
		<author>
			<persName><forename type="first">C</forename><surname>Egget</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zecha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brehm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lienhart</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACM on International Conference on Multimedia Retrieval</title>
				<meeting><address><addrLine>New York</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">A closer look: Small object detection in faster R-CNN</title>
		<author>
			<persName><forename type="first">C</forename><surname>Egget</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brehm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Winschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zecha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lienhart</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Multimedia and Expo</title>
				<meeting><address><addrLine>Hong Kong</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Logo Recognition with the Use of Deep Convolutional Neural Networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alsheikhy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sald</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
		<respStmt>
			<orgName>Engineering, Technology and Applied Science Research</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Which airline is this? Airline logo detection in real-world weather conditions</title>
		<author>
			<persName><forename type="first">C</forename><surname>Wilms</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Heid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sadeghi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ribbrock</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Frintrop</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 25th International Conference on Pattern Recognition</title>
				<meeting>the 25th International Conference on Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A Review of Object Detection Models based on Convolutional Neural Network</title>
		<author>
			<persName><forename type="first">F</forename><surname>Sultana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Abu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Paramartha</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ntelligent computing: image processing based applications</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1" to="16" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">PP-YOLO: An effective and efficient implementation of object detector</title>
		<author>
			<persName><forename type="first">X</forename><surname>Long</surname></persName>
		</author>
		<author>
			<persName><forename type="first">.</forename></persName>
		</author>
		<idno type="arXiv">arXiv:2007.12099</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">rXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Logo retrieval with a contrario visual query expansion</title>
		<author>
			<persName><forename type="first">A</forename><surname>Joly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Buisson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">roceedings of the 17th ACM International Conference on Multimedia</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Scalable logo recognition in real -world images</title>
		<author>
			<persName><forename type="first">S</forename><surname>Romberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Garcia Pueyo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lienhart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">Z</forename><surname>Roelof</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st ACM International Conference on Multimedia Retrieval</title>
				<meeting>the 1st ACM International Conference on Multimedia Retrieval</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">A new sport teams logo dataset for detection tasks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kuznetsov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Savchenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Computer Vision and Graphics</title>
				<meeting>the International Conference on Computer Vision and Graphics</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Open logo detection challenge</title>
		<author>
			<persName><forename type="first">H</forename><surname>Su</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the British Machine Vision Conference</title>
				<meeting>the British Machine Vision Conference</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">FoodLogoDet-1500: A dataset for large-scale food logo detection via multi-scale feature decoupling network</title>
		<author>
			<persName><forename type="first">H</forename><surname>Qiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Weiqing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Jing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sujuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shuqiang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th ACM International Conference on Multimedia</title>
				<meeting>the 29th ACM International Conference on Multimedia</meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">LogoDet-3K: Alarge-scale image dataset for logo detection</title>
		<author>
			<persName><forename type="first">W</forename><surname>Jing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Weiqing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sujuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Shengnan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yuanjie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Shuqiang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on multimedia</title>
		<imprint/>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
