<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A Possible Optimisation Procedure for US and MRI Tongue Contours</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Réka</forename><surname>Trencsényi</surname></persName>
							<email>trencsenyi.reka@science.unideb.hu</email>
							<affiliation key="aff0">
								<orgName type="department">Department of Electrical and Electronic Engineering</orgName>
								<orgName type="institution">University of Debrecen</orgName>
								<address>
									<settlement>Debrecen</settlement>
									<country key="HU">Hungary</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">László</forename><surname>Czap</surname></persName>
							<email>czap@uni-miskolc.hu</email>
							<affiliation key="aff1">
								<orgName type="department">Institute of Automation and Infocommunication</orgName>
								<orgName type="institution">University of Miskolc</orgName>
								<address>
									<settlement>Miskolc</settlement>
									<country key="HU">Hungary</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A Possible Optimisation Procedure for US and MRI Tongue Contours</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">D7140FDFDBAB0BF8A143DF765484ECD4</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T13:38+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Data visualisation</term>
					<term>computational linguistics</term>
					<term>speech research</term>
					<term>dynamic US and MRI records</term>
					<term>automatic tongue contour tracking AMS Subject Classification: 68N19</term>
					<term>68P05</term>
					<term>68T50</term>
					<term>68U10</term>
					<term>68W99</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The topic of the article is speech research. The main instruments of the study are US and MRI records of human beings, which were made during speech. In the dynamic records, primarily, the motion of the tongue is analysed and followed by automatic tongue contour tracking algorithms. The tongue contours are used to elaborate geometric transformations between US and MRI frames, which are the starting points of the optimisation of the matching of US and MRI tongue contours belonging to the same speech sound. As a result, the radial US geometry and the rectangular MRI geometry are embedded into each other in a biunique way.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>One of the fundamental tools of the study of speech production is the analysis of dynamic records of human speakers, made by ultrasound (US) <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b8">9]</ref> and magnetic resonance imaging (MRI) <ref type="bibr" target="#b6">[7]</ref> techniques. Investigating and processing these twodimensional records created in the so-called sagittal plane resulting in a side view of the human body, relevant qualitative and quantitative information can be gained about the main features of articulation. Qualitative statements mainly refer to the relative position of the tongue and palate in the case of different speech sounds and sound transitions, while quantitative descriptions focus on the recognition and connection of the geometric parameters which have high importance in the understanding of the relationships between the acoustic and articulatory characteristics of speech. Quantitative analyses can be performed in several ways with a wide variety <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref>. The starting points of the investigations of our present study are tongue contours fitted to the frames of US <ref type="bibr" target="#b9">[10]</ref> and MRI <ref type="bibr" target="#b7">[8]</ref> records by automatic algorithms <ref type="bibr" target="#b10">[11]</ref>. The used US and MRI sources differ from each other in many details, such as the gender and nationality of the speakers, the geometry, resolution, and scale of the images, and the visually evaluable anatomic segments of the vocal tract. The aim of our research work is to match the US and MRI sources by elaborating, applying, and optimising the proper geometric transformations between the US and MRI tongue contours in a biunique way.</p><p>In the literature, several publications can be found that deal with the fusion of information arising from sources produced by different imaging techniques. The demand for automatic tongue contour tracking algorithms emerged even in the previous decade <ref type="bibr" target="#b1">[2]</ref>, confirming the necessity of fully automated procedures like our algorithm that does not require any manual actions, as it is based on dynamic programming <ref type="bibr" target="#b10">[11]</ref>. Another benefit of our present results that we have been working with dynamic US and MRI records instead of static frames belonging to sustained sounds exclusively <ref type="bibr" target="#b0">[1]</ref>. The US videos were made by the Micro system of the MTA-ELTE Lendület Lingual Articulation Research Group of the Hungarian Academy of Sciences, and the MRI videos made by fast MRI were downloaded from the website of the University of Southern California. Also, such studies appeared that aim to perform transformations between coordinate systems connected to US and MRI frames relying on the optimisation of distances measured between special points of the human head. <ref type="bibr" target="#b0">[1]</ref>. In comparison with <ref type="bibr" target="#b0">[1]</ref>, it must be emphasised that our transformations relate directly to the tongue contours, and the transformation is carried out in one step without any intermediate coordinate system, so starting from the US contour, one gets to the MRI contour immediately. Furthermore, the optimisation procedure minimises the global distance between the linked US and MRI tongue contours in the case of more than one sounds simultaneously.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Transformation and Optimisation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">The Geometrical Considerations and Mathematical Formulas of the Transformations for Tongue Contours</head><p>When writing the exact mathematical form of the transformation, we relied on the special geometry of the available US records. Namely, the imaging US head scans such a radial region of the oral cavity which is seen at an angle of 90 ∘ measured from a fixed centre 𝐶. Consequently, it is obvious to treat the US images and the points of the belonging tongue contours in such a polar coordinate system of 𝐶 origin where the position of each pixel is given by radius 𝑟 measured from point 𝐶 and the signed angle 𝜙 measured from the central vertical axis of the image unambiguously. The aim of the transformation is to embed the radial geometry of the US frames to the rectangular geometry of the MRI records described by the two-dimensional Cartesian coordinates so that the US and MRI tongue contours assigned to the same sound should overlap with each other as much as possible.</p><p>The transformation of the US tongue contours can include three basic operations: the scaling of the radial range, the scaling of the angular range, and the translation of the angular range. The three operations can be realised mathematically by the formulas</p><formula xml:id="formula_0">𝑟 ′ = 𝑟 • 𝑅, 𝜙 ′ = 𝐹 𝐼 • 𝜙, 𝜙 ′ 0 = 𝜙 0 + 𝐹 𝐼 0 ,<label>(2.1)</label></formula><p>where the scale factors 𝑅 and 𝐹 𝐼 allow the normalisation of the radial and angular ranges, and the term 𝐹 𝐼 0 performs the translation of the initial angle of the angular range. The mathematical operations of (2.1) can be interpreted in the physical plane of the images in the following way: by scaling the radial range, the magnification of the tongue contours can be modified. The scaling of the angular range creates the possibility to change the width of the angular range covered by the tongue contours. The translation of the angular range means the rotation of the tongue contours in the plane of the image. Thus, relationships (2.1) fit the US tongue contour to the corresponding MRI frame. Applying the inverse of transformations (2.1), however, also the reverse conversion can be executed, i.e., by dint of the inverse operations</p><formula xml:id="formula_1">𝑟 = 𝑟 ′ 𝑅 , 𝜙 = 𝜙 ′ 𝐹 𝐼 , 𝜙 0 = 𝜙 ′ 0 − 𝐹 𝐼 0 ,<label>(2.2)</label></formula><p>the MRI tongue contour can be mapped onto the corresponding US frame. The parameter set {𝑅, 𝐹 𝐼, 𝐹 𝐼 0 } of the transformations performed in the directions US-MRI and MRI-US must necessarily be the same since, thereby, the maintaining of the relative scale ratio of the US and MRI environment can be ensured independently of the direction of the conversion. During the investigations, we fixed the value of factor 𝐹 𝐼 by 𝐹 𝐼 = 1, which means that the transformation is conformal. Transformations (2.1) and (2.2) become valid by the numerical determination of parameters 𝑅 and 𝐹 𝐼 0 , to which the optimisation of the values of the parameters offers a possible way. During the optimisation procedure, using an algorithm elaborated by us, we find the parameter set, in the case of that the distance between the transformed US tongue contour and the MRI tongue contour serving as a reference curve is minimal. The calculation of the distance is carried out for all of the possible pairs of points of the two curves, then the average of the smallest distances assigned to each point of the US tongue contour is minimised <ref type="bibr" target="#b11">[12]</ref>. To the successful transformation, however, not only the exact values of parameters 𝑅 and 𝐹 𝐼 0 are needed but also centre 𝐶 ′ designated in the MRI frame must be known, which is the image of centre 𝐶 of the US record. Beyond these, during the construction of the optimisation algorithm, also the peak of the epiglottis can serve as a good reference point demonstrated by Figure <ref type="figure" target="#fig_0">1</ref>, where the peaks of the epiglottis 𝐺 and 𝐺 ′ are marked by green circles in the US and MRI frames, and also centres 𝐶 and 𝐶 ′ are located by red cross signs. The meaning of the parameter set {𝑅, 𝐹 𝐼 0 , 𝐶 ′ , 𝐺, 𝐺 ′ } can be understood by the geometrical considerations of Figure <ref type="figure" target="#fig_1">2</ref>, where the left-side block depicts the points of the US frames (𝐶 and 𝐺), while the right-side block carries the points of the MRI frames (𝐶 ′ and 𝐺 ′ ) in agreement with Figure <ref type="figure" target="#fig_0">1</ref>. The radial distances 𝑅 1 and 𝑅 2 are measured between the centre of the images and the peaks of the epiglottis. The polar angles 𝜙 1 and 𝜙 2 are made by the central vertical axis of the image and radii 𝑅 1 and 𝑅 2 . Using these quantities, parameter 𝑅 is interpreted as a magnification factor by</p><formula xml:id="formula_2">𝑅 = 𝑅 2 𝑅 1 ,<label>(2.3)</label></formula><p>and parameter 𝐹 𝐼 0 is produced by</p><formula xml:id="formula_3">𝐹 𝐼 0 = 𝜙 1 + 𝜙 2 , (2.4)</formula><p>which is actually a difference, since 𝜙 1 is negative and 𝜙 2 is positive, as the polar angle is related to the vertical direction in both frames. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">The Optimisation Procedure for Minimising the Distance Between Tongue Contours</head><p>Proceeding along the geometrical features of Figure <ref type="figure" target="#fig_1">2</ref>, we created an optimisation algorithm using MATLAB via such mathematical formulas which enable the simultaneous optimisation of the parameters {𝑅, 𝐹 𝐼 0 , 𝐶 ′ , 𝐺, 𝐺 ′ }. Some details of the MATLAB scripts can be found in the appendix, where [𝑛1, 𝑛2], [𝑔1, 𝑔2], and [𝑔3, 𝑔4] are the coordinates of 𝐶 ′ , 𝐺, and 𝐺 ′ , respectively, while 𝐹 𝐼𝐾𝑂𝑅𝑅 stands for 𝐹 𝐼 0 . The construction of the mathematical formulas for 𝑅 1 , 𝑅 2 and 𝐹 𝐼1, 𝐹 𝐼2 follows the geometrical structure of Figure <ref type="figure" target="#fig_1">2</ref>. As explained in (2.3) and (2.4), the scale factor 𝑅 is the ratio of the radial distances between the centre and the peak of the epiglottis, measured in the US and MRI frames separately, and the angular displacement 𝐹 𝐼 0 is given as the sum of the signed polar angles of the US and MRI frames, made by a radial section and the central vertical axis of the image. In the MATLAB scripts, 𝑐𝑢1 denotes the actual US or MRI curve having points with complex coordinates to be transformed, 𝑢ℎ𝑢𝑠1 and 𝑚𝑟𝑖𝑠1 gives the US and MRI curves to be compared, and 𝑛𝑛𝑢1 = 𝑛𝑛𝑢𝑚1 provides the average of the smallest distances of all of the possible pairs of points of the two tongue contours.</p><p>In the appendix, the codes are written only for one of the contours belonging to the investigated speech sounds and labelled by index 1, but to have the complete script, these program blocks must be repeated with the same structure for the other examined sounds with indices 2, 3, 4, . . . , as well.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Results</head><p>We fulfilled the optimisation of the parameters {𝑅, 𝐹 𝐼 0 , 𝐶 ′ , 𝐺, 𝐺 ′ } for two speech sounds, k and t simultaneously, and we obtained the following numerical results: (2.5)</p><formula xml:id="formula_4">𝑅 = 0.</formula><p>Based on the values in (2.5), it can be observed in Figure <ref type="figure" target="#fig_3">3</ref> that the optimised positions of the peaks of the epiglottis are very close to the static position when the voice box is at rest without speaking. In addition, the centre of the MRI frame is located under the jaw of the speaker. Using the parameters of (2.5), the transformation of the US and MRI tongue contours can be implemented in a bidirectional way according to Figure <ref type="figure" target="#fig_4">4</ref>, where the tongue contours belonging to sound k can be seen. The green curves stand for the US, while the red curves represent the MRI tongue contours. In the MRI frame, also the contour of the palate is drawn by the yellow curve. The figures clearly show that the US and MRI tongue contours fit each other in an acceptable way, since it is demonstrated visually, as well, that the distance is minimised by the optimisation algorithm detailed in the appendix let the two curves mostly overlap with each other. At this level, it is a sufficient criterion for the acceptance of the results without any specified lower or upper limit for the distance between the two curves because we aimed to find the relative position of the two tongue contours, which corresponds to the minimal distance ensured by parameters (2.5). So, graphically, a realistic matching is obtained, and only this was expected. For instance, if the transformed tongue contour was out of the region of the oral cavity represented by the US or MRI image or placed visually at an unrealistic distance from the reference curve then the optimisation would be surely false.  In order to verify the results, we also checked the projection of such US-MRI pairs of contours onto each other, besides the setting of parameters (2.5) provided by the optimisation, which were not present in the set of sounds 𝑘 and 𝑡. Accordingly, Figure <ref type="figure" target="#fig_5">5</ref> exemplifies our results in the case of sound 𝑒. It can be stated that the matching of the US and MRI tongue contours is approximately as good as in the case of sound 𝑘.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Validation</head><p>Further validation of the results is in progress currently, as well. To gain more experience about the harmonisation of US and MRI geometry and improve the fitting of the tongue contours, we aim to develop our research work in several directions. We wish to extend the optimisation procedure to more than two sounds to understand the connection between the result of the optimisation and the number of speech sounds. We would also like to investigate the optimisation in the dependence of different sound contexts and speakers of different nationalities and genders. Furthermore, advancing to a large number of speech sounds, we plan to involve machine learning algorithms, as well.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. The peaks of the epiglottis 𝐺 and 𝐺 ′ , and the centres 𝐶 and 𝐶 ′ in the US (a.) and MRI (b.) frames.</figDesc><graphic coords="4,68.99,306.42,170.29,127.78" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. The graphical representation of the main parameters of the US (left) and MRI (right) frames.</figDesc><graphic coords="5,145.50,158.78,90.14,85.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>4348, 𝐹 𝐼 0 = 0.1293 rad, 𝐶 ′ = (250, 140), 𝐺 = (244, 230), 𝐺 ′ = (158, 198).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 .</head><label>3</label><figDesc>Figure 3. The results of the optimisation for the centre of the MRI frame (b.) and the positions of the peaks of the epiglottis in the US (a.) and MRI (b.) frames.</figDesc><graphic coords="6,68.99,268.00,170.29,127.78" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 .</head><label>4</label><figDesc>Figure 4. The results of the optimisation in the case of sound k by presenting the US (green) and MRI (red) tongue contours simultaneously. The contour of the palate is indicated by the yellow curve in the MRI frame.</figDesc><graphic coords="7,68.99,114.69,170.29,127.78" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 5 .</head><label>5</label><figDesc>Figure 5. The results of the optimisation in the case of sound e by presenting the US (green) and MRI (red) tongue contours simultaneously. The contour of the palate is indicated by the yellow curve in the MRI frame.</figDesc><graphic coords="7,68.99,353.81,170.29,127.78" type="bitmap" /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Acknowledgements. We would like to thank the MTA-ELTE Lendület Lingual Articulation Research Group for providing the recordings with the Micro system.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Appendix</head><p>Program code for the optimisation  </p><p>Program code for the calculation of the distances between US and MRI tongue contours K = 1; clear dc1 du1 for q = K : lh1 for qq = 1 : lm1 b1 = imag(uhus1(q)) − imag(mris1(qq)); b2 = real(uhus1(q)) − real(mris1(qq)); dc1(qq) = sqrt(b1 * b1 + b2 * b2); end du1(q) = min(dc1); end nnu1 = mean(du1(K : length(du1)));</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Multimodal fusion of electromagnetic, ultrasound and MRI data for building an articulatory model</title>
		<author>
			<persName><forename type="first">M</forename><surname>Aron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-O</forename><surname>Berger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kerrien</surname></persName>
		</author>
		<idno>ffinria-00326290f</idno>
		<ptr target="https://hal.inria.fr/inria-00326290/document" />
	</analytic>
	<monogr>
		<title level="m">8th International Seminar On Speech Production -ISSP&apos;08</title>
				<meeting><address><addrLine>Strasbourg, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008-12">Dec 2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Comparing articulatory images: An MRI/Ultrasound Tongue Image database</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cleland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Wrench</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Scobbie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Semple</surname></persName>
		</author>
		<idno>doi:</idno>
		<ptr target="https://eresearch.qmu.ac.uk/handle/20.500.12289/2477" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th International Seminar on Speech Production</title>
				<meeting>the 9th International Seminar on Speech Production</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="page" from="163" to="170" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Quantitative analysis of multimodal speech data</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">G</forename><surname>Danner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Barbosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Goldstein</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.wocn.2018.09.007</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.wocn.2018.09.007" />
	</analytic>
	<monogr>
		<title level="j">Journal of Phonetics</title>
		<imprint>
			<biblScope unit="volume">71</biblScope>
			<biblScope unit="page" from="268" to="283" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Speech synthesis from real time ultrasound images of the tongue</title>
		<author>
			<persName><forename type="first">B</forename><surname>Denby</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stone</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICASSP.2004.1326078</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1109/ICASSP.2004.1326078" />
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Acoustics, Speech, and Signal Processing</title>
				<imprint>
			<date type="published" when="2004">2004. 2004</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">685</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Geometry of the vocal tract and properties of phonation near threshold: calculations and measurements</title>
		<author>
			<persName><forename type="first">L</forename><surname>Fulcher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lodermeyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kahler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Becker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kniesburges</surname></persName>
		</author>
		<idno type="DOI">10.3390/app9132755</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.3390/app9132755" />
	</analytic>
	<monogr>
		<title level="j">Applied Sciences</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page">2755</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Automated segmentation of upper airways from MRI-vocal tract geometry extraction</title>
		<author>
			<persName><forename type="first">A</forename><surname>Ojalammi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Malinen</surname></persName>
		</author>
		<idno type="DOI">10.5220/0006138300770084</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.5220/0006138300770084" />
	</analytic>
	<monogr>
		<title level="j">International Conference on Bioimaging</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="77" to="84" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Speech MRI: morphology and function</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">D</forename><surname>Scott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wylezinska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Birch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Miquel</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ejmp.2014.05.001</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.ejmp.2014.05.001" />
	</analytic>
	<monogr>
		<title level="j">Physica Medica</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="604" to="618" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">span | speech production and articulation knowledge group: the rtMRI IPA chart (John Esling)</title>
		<ptr target="https://sail.usc.edu/span/rtmri_ipa/je_2015.html" />
		<imprint>
			<date type="published" when="2020-05-09">2020 May 9th</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A guide to analysing tongue motion from ultrasound images</title>
		<author>
			<persName><forename type="first">M</forename><surname>Stone</surname></persName>
		</author>
		<idno type="DOI">10.1080/02699200500113558</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1080/02699200500113558" />
	</analytic>
	<monogr>
		<title level="j">Clinical Linguistics and Phonetics</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="455" to="501" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A comparative study on the contour tracking algorithms in ultrasound tongue images with automatic re-initialization</title>
		<author>
			<persName><forename type="first">K</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">G</forename><surname>Csapó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Roussel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Denby</surname></persName>
		</author>
		<idno type="DOI">10.1121/1.4951024</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1121/1.4951024" />
	</analytic>
	<monogr>
		<title level="j">Journal of the Acoustical Society of America</title>
		<imprint>
			<biblScope unit="volume">139</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="L154" to="L160" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Automatic tracking of tongue contours in ultrasound records</title>
		<author>
			<persName><forename type="first">L</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Czap</surname></persName>
		</author>
		<idno type="DOI">10.15775/Beszkut.2019</idno>
		<idno>.331-343</idno>
		<ptr target="https://doi.org/10.15775/Beszkut.2019" />
	</analytic>
	<monogr>
		<title level="j">Beszédtudomány -Speech Science</title>
		<imprint>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="331" to="343" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Measuring lingual coarticulation from midsagittal tongue contours: Description and example calculations using English /t/ and /a</title>
		<author>
			<persName><forename type="first">N</forename><surname>Zharkova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hewlett</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.wocn.2008.10.005</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.wocn.2008.10.005" />
	</analytic>
	<monogr>
		<title level="j">Journal of Phonetics</title>
		<imprint>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="248" to="256" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
