<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">HCMUS at MediaEval2021: PointRend with Attention Fusion Refinement for Polyps Segmentation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">E-Ro</forename><surname>Nguyen</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">University of Science</orgName>
								<orgName type="institution" key="instit2">VNU-HCM</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Vietnam National University</orgName>
								<address>
									<addrLine>Ho Chi</addrLine>
									<settlement>Minh city</settlement>
									<country key="VN">Vietnam</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hai-Dang</forename><surname>Nguyen</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">University of Science</orgName>
								<orgName type="institution" key="instit2">VNU-HCM</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Vietnam National University</orgName>
								<address>
									<addrLine>Ho Chi</addrLine>
									<settlement>Minh city</settlement>
									<country key="VN">Vietnam</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Minh-Triet</forename><surname>Tran</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution" key="instit1">University of Science</orgName>
								<orgName type="institution" key="instit2">VNU-HCM</orgName>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="department">John von Neumann Institute</orgName>
								<orgName type="institution">VNU-HCM</orgName>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">Vietnam National University</orgName>
								<address>
									<addrLine>Ho Chi</addrLine>
									<settlement>Minh city</settlement>
									<country key="VN">Vietnam</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">HCMUS at MediaEval2021: PointRend with Attention Fusion Refinement for Polyps Segmentation</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">A89D11590FE987BDE55B13E8D53D6199</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T19:59+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The Medico task in MediaEval 2021 explores the challenge of building accurate and high-performance algorithms to detect all types of polyps in endoscopic images. This paper introduces our approach for the automatic segmentation of polyp images. We employ a ResNeXt as an encoder backbone with a UNet decoder. Further, the addition of PointRend and Attention Fusion Refinement on the network improves our segmentation performance. The experimental results show the efficiency of the proposed method, which achieves a Jaccard index of 0.7572, an accuracy of 0.9634, and a dice score of 0.8326.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>Medico: Transparency in Medical Image Segmentation 2021 <ref type="bibr" target="#b5">[6]</ref> task aims to develop automatic segmentation systems for segmenting polyps in images taken from endoscopies that are transparent and explainable, and reduce the chance that diagnosticians overlook a polyp during a colonoscopy. A modified version of the segmentation part of HyperKvasir <ref type="bibr" target="#b1">[2]</ref> is given with more than 1000 training polyp images with their corresponding masks labeled by medical experts and 200 testing polyp images to challenge the participants for the robust, transparent, and efficient algorithms for polyp segmentation.</p><p>In recent years, the task of automatic polyp segmentation using deep learning-based <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4]</ref> methods has gained a lot of achievements. Especially, the appearance of attention strategies <ref type="bibr" target="#b2">[3]</ref> effectively improves polyp detection and segmentation performance. However, it still has some challenges, including the varieties of polyp's appearance (size, texture, and color). The boundary between a polyp and its neighbor regions is usually blurred and hard to be segmented.</p><p>In this paper, we propose an accurate and real-time framework PointRend with Attention Fusion Refinement (PRAFNet) for the polyp segmentation. Fig. <ref type="figure">1</ref> shows the overview of our proposed framework. PRAFNet utilizes the Attention Fusion Refinement to decode an effective high-level semantic segmentation, and the PointRend <ref type="bibr" target="#b7">[8]</ref> module to generate high-quality polyp segmentation from the colonoscopy images. The following section will introduce our approach and elaborate details about our network.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">APPROACH 2.1 Attention Fusion Refinement</head><p>Current popular medical image segmentation networks usually rely on a U-Net architecture (e.g., U-Net <ref type="bibr" target="#b8">[9]</ref>, U-Net++ <ref type="bibr" target="#b12">[13]</ref>, ResUNet <ref type="bibr" target="#b6">[7]</ref>, etc). These models are essentially encoder-decoder frameworks, which aggregate all multi-level features extracted with a simple decoder, which does not effectively leverage these features. Woo et al. introduce a Convolutional Block Attention Module (CBAM) <ref type="bibr" target="#b10">[11]</ref>, which applies attention-based feature refinement with two distinctive modules, channel and spatial, to learn what and where to emphasize or suppress and refines intermediate features effectively.</p><p>We propose an Attention Fusion Refinement(AFR) module to better aggregate high-level features and focus on important regions, combining high-level features with upsampled features by CBAM as a core module. More specifically, for an input image, five levels of features {𝑓 𝑖 , 𝑖 = 1, .., 5} can be extracted from a ResNeXt <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b11">12]</ref> backbone network. We introduce a new decoder component, AFR, to aggregate the high-level features with upsampled features. As shown in Fig. <ref type="figure">1</ref>, An AFR module inputs a high-level feature 𝑓 𝑖 with the previous upsampled feature 𝑑 𝑖+1 and we obtain the upsampled feature 𝑑 𝑖 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">PointRend</head><p>The U-Net <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b12">13]</ref> model gives decent accuracy. However, it still has some drawbacks like predicting classes with very near distinguishable features, not being able to predict precise boundaries, etc. We have used the PointRend <ref type="bibr" target="#b7">[8]</ref> module to address these drawbacks.</p><p>PointRend constructs point-wise features at selected points by concatenating two features, fine-grained to render fine segmentation details and coarse prediction features to gain more contextual and semantic information. We use the features 𝑓 2 as our fine-grained features and select top 𝐾 = 3136 uncertain points in each subdivision step. In general, the uncertain points are located near the boundary of classes, so it can help refine the polyp's boundary effectively. As shown in Fig. <ref type="figure">1</ref>, we use two subdivision steps of PointRend to obtain the final segmentation, which is the same size as the input image. We plot the uncertain points used in PointRend as blue dots in the coarse predictions 𝑚 2 , 𝑚 1 .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Training strategy</head><p>We apply the Bootstrapped Cross Entropy loss to prevent the models from overfitting on simple pixels and force them to focus on more challenging cases. With the Bootstrapped Cross Entropy, we calculate the loss for the top 𝐾 percent pixels with the largest losses at each step in the training process. We would also add a "warm-up" MediaEval'21, December 13-15 2021, Online E-Ro et al. period to the loss with 𝐾 = 100 such that the network can learn to adapt to the easy regions first. Then transit to the harder areas by gradually decaying K to 15 in a polynomial manner.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Attention Fusion Refinement</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">RESULTS AND ANALYSIS</head><p>We performed experiments on six different settings for two tasks: Method 1 uses the UNet with ResNeXt50 <ref type="bibr" target="#b11">[12]</ref> backbone as a baseline model. Method 2 extends Method 1 with the PointRend. Method 3 extends Method 2 with the Attention Fusion Refinement. Method 4 uses ResNeXt101 as a backbone with the same settings as Method 3. Method 5 uses EfficientNetB6 <ref type="bibr" target="#b9">[10]</ref> as as backbone with the same setting as Method 3. Method 6 ensembles the results of Method 3, Method 4, and Method 5 together.</p><p>For task 1, we submit five runs from Method 2 to Method 6. For task 2, we submit two runs. In the first run, we use Method 4. And the second run is Method 1 for the lightweight architecture.</p><p>Table <ref type="table" target="#tab_0">1</ref> and 2 shows our results on task 1 and task 2, respectively. Method 2 is slightly better than method 1 in all metrics, which shows that PointRend helps improve the results. In method 3, we use AFR, and the results also improve compared to method 2. With a stronger backbone (ResNeXt101 instead of ResNeXt50) in method 4, the results are improved with a Jaccard index of 0.7441. Method 5 with an EfficientNetB6 backbone is better than method 4 in several metrics except for precision. In method 6, we ensemble our methods 3, 4, 5 to achieve our best result in this task with the Jaccard index of 0.7572.</p><p>In task 2, although our method 1 is 1.5 faster than method 2, method 2 has higher accuracy with a real-time efficiency (48 FPS)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">CONCLUSION</head><p>This paper presents a fast and accurate method for automatic polyps segmentation. The proposed methods use an encoder-decoder architecture. ResNeXt is used as an encoder backbone with the UNet decoder. Further, PointRend and Attention Fusion Refinement are applied to improve the segmentation result. PointRend helps refine the uncertainty points, especially with the boundary regions. The Attention Fusion Refinement enhances the fusion between highlevel features and upsampled features in the decoder. In the future, we plan to apply better architecture such as ResUnet++ or PraNet for our work and further improve the results.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Medico polyp segmentation task 1's result. Acc denotes the accuracy, R and P denote the recall and precision, respectively.</figDesc><table><row><cell>Concat</cell><cell>Res Block</cell><cell>CB AM</cell><cell>Res Block</cell></row></table><note>𝑝𝑟𝑒𝑑𝑖𝑐𝑡𝑖𝑜𝑛Figure1: Overview of our proposed method PRAFNet, which consists of three attention fusion refinement (AFR) modules with two adaptive subdividion steps of PointRend module. Please refer to section 2 for more details.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Medico polyp segmentation task 2's result.</figDesc><table /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ACKNOWLEDGMENTS</head><p>This work was funded by Gia Lam Urban Development and Investment Company Limited, Vingroup and supported by Vingroup Innovation Foundation (VINIF) under project code VINIF.2019.DA19. Medico: Transparency in Medical Image Segmentation MediaEval'21, December 13-15 2021, Online</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Polyp Segmentation in Colonoscopy Images Using Fully Convolutional Network</title>
		<author>
			<persName><forename type="first">Mojtaba</forename><surname>Akbari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Majid</forename><surname>Mohrekesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Ebrahim Nasr Esfahani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nader</forename><surname>Reza Soroushmehr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shadrokh</forename><surname>Karimi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kayvan</forename><surname>Samavi</surname></persName>
		</author>
		<author>
			<persName><surname>Najarian</surname></persName>
		</author>
		<idno type="DOI">10.1109/EMBC.2018.8512197</idno>
		<ptr target="https://doi.org/10.1109/EMBC.2018.8512197" />
	</analytic>
	<monogr>
		<title level="m">Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Conference</title>
				<imprint>
			<date type="published" when="2018">2018. 2018</date>
			<biblScope unit="page" from="69" to="72" />
		</imprint>
	</monogr>
	<note>Conference proceedings</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy</title>
		<author>
			<persName><forename type="first">Hanna</forename><surname>Borgli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vajira</forename><surname>Thambawita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pia</forename><forename type="middle">H</forename><surname>Smedsrud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Steven</forename><surname>Hicks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Debesh</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kristin</forename><forename type="middle">Ranheim</forename><surname>Sigrun L Eskeland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Konstantin</forename><surname>Randel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Pogorelov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Duc</forename><surname>Lux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dag</forename><surname>Tien Dang Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Carsten</forename><surname>Johansen</surname></persName>
		</author>
		<author>
			<persName><surname>Griwodz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Håkon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Enrique</forename><surname>Stensland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><forename type="middle">T</forename><surname>Garcia-Ceja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hugo</forename><forename type="middle">L</forename><surname>Schmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><forename type="middle">A</forename><surname>Hammer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pål</forename><surname>Riegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Halvorsen</surname></persName>
		</author>
		<author>
			<persName><surname>De Lange</surname></persName>
		</author>
		<idno type="DOI">10.1038/s41597-020-00622-y</idno>
		<ptr target="https://doi.org/10.1038/s41597-020-00622-y" />
	</analytic>
	<monogr>
		<title level="j">Scientific Data</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">283</biblScope>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">PraNet: Parallel Reverse Attention Network for Polyp Segmentation</title>
		<author>
			<persName><forename type="first">Deng-Ping</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ge-Peng</forename><surname>Ji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tao</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Geng</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Huazhu</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianbing</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ling</forename><surname>Shao</surname></persName>
		</author>
		<idno>arXiv:eess.IV/2006.11392</idno>
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Selective Feature Aggregation Network with Area-Boundary Constraints for Polyp Segmentation</title>
		<author>
			<persName><forename type="first">Yuqi</forename><surname>Fang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Cheng</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yixuan</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kai-Yu</forename><surname>Tong</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-32239-7_34</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-32239-7_34" />
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="302" to="310" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Deep Residual Learning for Image Recognition</title>
		<author>
			<persName><forename type="first">Kaiming</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiangyu</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shaoqing</forename><surname>Ren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jian</forename><surname>Sun</surname></persName>
		</author>
		<idno>arXiv:cs.CV/1512.03385</idno>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Medico Multimedia Task at MediaEval 2021: Transparency in Medical Image Segmentation</title>
		<author>
			<persName><forename type="first">Steven</forename><surname>Hicks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Debesh</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vajira</forename><surname>Thambawita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hugo</forename><surname>Hammer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>De Lange</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sravanthi</forename><surname>Parasa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><surname>Riegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pål</forename><surname>Halvorsen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of MediaEval 2021 CEUR Workshop</title>
				<meeting>MediaEval 2021 CEUR Workshop</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Re-sUNet++: An Advanced Architecture for Medical Image Segmentation</title>
		<author>
			<persName><forename type="first">Debesh</forename><surname>Jha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pia</forename><forename type="middle">H</forename><surname>Smedsrud</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Michael</forename><forename type="middle">A</forename><surname>Riegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dag</forename><surname>Johansen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><forename type="middle">De</forename><surname>Lange</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pål</forename><surname>Halvorsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Håvard</forename><forename type="middle">D</forename><surname>Johansen</surname></persName>
		</author>
		<idno type="DOI">10.1109/ISM46123.2019.00049</idno>
		<ptr target="https://doi.org/10.1109/ISM46123.2019.00049" />
	</analytic>
	<monogr>
		<title level="m">IEEE International Symposium on Multimedia (ISM)</title>
				<imprint>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="225" to="2255" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">PointRend: Image Segmentation as Rendering</title>
		<author>
			<persName><forename type="first">Alexander</forename><surname>Kirillov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yuxin</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kaiming</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ross</forename><surname>Girshick</surname></persName>
		</author>
		<idno>arXiv:cs.CV/1912.08193</idno>
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">U-Net: Convolutional Networks for Biomedical Image Segmentation</title>
		<author>
			<persName><forename type="first">Olaf</forename><surname>Ronneberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Philipp</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><surname>Brox</surname></persName>
		</author>
		<idno>arXiv:cs.CV/1505.04597</idno>
		<imprint>
			<date type="published" when="2015">2015. 2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks</title>
		<author>
			<persName><forename type="first">Mingxing</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Quoc</surname></persName>
		</author>
		<author>
			<persName><surname>Le</surname></persName>
		</author>
		<idno>arXiv:cs.LG/1905.11946</idno>
		<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">CBAM: Convolutional Block Attention Module</title>
		<author>
			<persName><forename type="first">Sanghyun</forename><surname>Woo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jongchan</forename><surname>Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Joon-Young</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">In</forename><surname>So Kweon</surname></persName>
		</author>
		<idno>arXiv:cs.CV/1807.06521</idno>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Aggregated Residual Transformations for Deep Neural Networks</title>
		<author>
			<persName><forename type="first">Saining</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ross</forename><surname>Girshick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Piotr</forename><surname>Dollár</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhuowen</forename><surname>Tu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kaiming</forename><surname>He</surname></persName>
		</author>
		<idno>arXiv:cs.CV/1611.05431</idno>
		<imprint>
			<date type="published" when="2017">2017. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">UNet++: A Nested U-Net Architecture for Medical Image Segmentation</title>
		<author>
			<persName><forename type="first">Zongwei</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Md</forename><surname>Mahfuzur Rahman Siddiquee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nima</forename><surname>Tajbakhsh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jianming</forename><surname>Liang</surname></persName>
		</author>
		<idno>arXiv:cs.CV/1807.10165</idno>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
