<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Stravinsqi/De Montfort University at the MediaEval 2014 C@merata Task</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Tom</forename><surname>Collins</surname></persName>
							<email>tom.collins@dmu.ac.uk</email>
							<affiliation key="aff0">
								<orgName type="department">Faculty of Technology De</orgName>
								<orgName type="institution">Montfort University Leicester</orgName>
								<address>
									<postCode>+44 116 207 6192</postCode>
									<country key="GB">UK</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Stravinsqi/De Montfort University at the MediaEval 2014 C@merata Task</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">3307CEE8F9D3284BA62FA1647D3851CE</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T16:11+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>An overview is provided of the Stravinsqi-Jun2014 algorithm and its performance on the MediaEval 2014 C@merata Task. Stravinsqi stands for STaff Representation Analysed VIa Natural language String Query Input. The algorithm parses a symbolic representation of a piece of music as well as a query string consisting of a natural language expression, and identifies where event(s) specified by the query occur in the music. The output for any given query is a list of time windows corresponding to the locations of relevant events. To evaluate the algorithm, its output time windows are compared with those specified by music experts for the same query-piece combinations. In an evaluation consisting of twenty pieces and 200 questions, Stravinsqi-Jun2014 had recall .91 and precision .46 at the measure level, and recall .87 and precision .44 at the beat level. Important potential applications of this work in music-educational software and musicological research are discussed.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">INTRODUCTION</head><p>Given a natural language query and a piece of music in digital staff notation representation, the C@merata task <ref type="bibr" target="#b6">[6]</ref> evaluates an algorithm's ability to identify where one or more events specified by the query occur in the music. It is the latest example of a longstanding interest in querying music represented as (or derived from) staff notation. The C@merata task challenges researchers to extend current knowledge in two respects:</p><p>1. Accepting a music-analytic query in the form of a naturallanguage string, such as "perfect fifth followed by a D4";</p><p>2. Reliably retrieving instances of higher-level musictheoretic concepts from staff notation, such as functional harmonies (e.g., "Ib") or cadences (e.g., "interrupted cadence").</p><p>One application of an algorithm that performs well on the C@merata task would be within music notation software, so that students could query and hear/see results for the pieces with which they are working, in order to develop their understanding of various music-theoretic terms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">APPROACH 2.1 Overview</head><p>The Stravinsqi-Jun2014 algorithm that was entered in the C@merata task is embedded in a Common Lisp package called MCStylistic-Jun2014 (hereafter, MCStylistic), which has been under development since 2008 <ref type="bibr" target="#b1">[1]</ref>. <ref type="foot" target="#foot_0">1</ref> MCStylistic includes  implementations of algorithms from the fields of music information retrieval (MIR) and music psychology <ref type="bibr" target="#b2">[2]</ref><ref type="bibr" target="#b3">[3]</ref><ref type="bibr" target="#b4">[4]</ref><ref type="bibr">[5]</ref>.</p><p>From a natural language perspective, there are two types of queries: compound queries such as "a Bb followed a bar later by a C followed by a tonic triad", and ordinary queries such as "perfect cadence". Stravinsqi checks the query string for compound queries and splits it into N query elements if necessary, e.g., "a Bb" and "a bar later by a C" and "tonic triad".</p><p>The piece is converted from its MusicXML format to kern format using the xml2hum script. <ref type="foot" target="#foot_1">2</ref> The kern file is parsed by import functions in MCStylistic to give the following representations, which are referred to as point sets: (1) instrument/staff and clef names at the beginning of each staff; (2) bar numbers where time signatures are specified, together with the number of beats per bar, the type of beat, and the corresponding ontime (incrementing time in staff notation); (3) a point-set representation of the piece, where each point represents a note. The five-dimensional point consists of the ontime of the note, its MIDI note number, its morphetic pitch number <ref type="bibr" target="#b3">[3]</ref>, its duration in crotchet beats, and its numeric staff number; (4) a point-set representation of the piece with three extra dimensions, one each for articulation, dynamics, and lyrics information; (5) a point-set representation of the piece, where each point represents a notated rest.</p><p>Each query element is passed to several sub-functions (e.g., harmonic-interval-of-a, duration&amp;pitch-timeintervals, rest-duration-time-intervals, etc.), along with the appropriate point set(s). For example, the function rest-duration-time-intervals takes a query element, the point set of notated rests, and the point set of instrument/staff and clef names as its arguments, because these three information sources are sufficient for locating rests of specific duration. If a query string is ordinary (contains one element only), then the time windows in the first nonempty sub-function's output are passed to a final function that converts the time windows into the XML format required by the task. For compound queries, plausible sequences of time windows for the component query elements are merged before passing to the final syntax-conversion function.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Example Output for Three Sub-Functions</head><p>Figure <ref type="figure" target="#fig_1">1</ref> contains example output for the sub-functions harmonic-interval-of-a, (arrows indicate retrieved harmonic seconds), HarmAn-&gt;Roman (functional harmonic labels below each staff), and cadence-time-intervals (three perfect cadences surrounded by black boxes). All three functions involve implementing and extending MIR/musicpsychology algorithms to achieve promising results, especially for the higher-level music-theoretical concepts such as functional harmonies and cadences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">RESULTS AND DISCUSSION</head><p>Figure <ref type="figure" target="#fig_0">2</ref> shows recall and precision results for the Stravinsqi algorithm on the 2014 C@merata task. The measure metrics reward an algorithm's output if it is in the same bar/measure as a ground-truth item, whereas the beat metrics require an algorithm's output to be in the same bar and on the same beat as a groundtruth item. The mean category in Figure <ref type="figure">4</ref> shows the overall results, with Stravinsqi having recall .91 and precision .46 at the measure level, and recall .87 and precision .44 at the beat level. <ref type="foot" target="#foot_2">3</ref>Stravinsqi's strong performance on the first eight of twelve categories (pitch, duration,…, melodic interval) is encouraging, as is the small decrease in recall (.91 to .87) and precision (.46 to .44) with the change from measure-to beat-level granularity. The drop in precision for compound queries is due to over-lenient criteria used to select and combine time intervals for the different elements that comprise a compound query. This can be fixed in future work. For triad labelling, Stravinsqi suffered from an under-labelling issue in two instances, missing two first-inversion triads because the same triad in root position preceded them, and the two triads got one root-position label. The triad and texture categories are somewhat underrepresented in the training and test data, and so more attention ought to be given to these categories in future. Less-than-perfect performance on melodic and harmonic interval questions can be attributed to inconsistencies between the task description/training collection, and test collection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">CONCLUSION</head><p>Algorithms that perform strongly on the C@merata task open up new, interesting potential applications in music education and musicological research. The Stravinsqi algorithm described above is one such strong performer, and has effectively solved seven of the twelve C@merata task categories shown in Figure <ref type="figure">4</ref> (pitch, duration, pitch and duration, articulation, voice specific, lyrics, and melodic interval). As for the remaining five categories, future work will involve bug fixes, resolving task inconsistencies, and acquiring more data for cadence and texture query categories. It may also be helpful to have two experts provide annotations for the higher-level music-theoretic concepts. The addition of new, higher-level music-theoretic query categories would be welcome in future iterations of C@merata as well, in order to keep the task at the forefront of research in music computing.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 .</head><label>2</label><figDesc>Figure 2. Results of the Stravinsqi-Jun2014 algorithm on the MediaEval 2014 C@merata task. Overall results are indicated by the mean label, and followed by results for twelve question categories.</figDesc><graphic coords="1,319.05,501.80,239.60,179.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 .</head><label>1</label><figDesc>Figure 1. Prelude to Te Deum H146 by Marc-Antoine Charpentier (1643-1704), annotated with a bar number error (tick and cross), intervals of a harmonic second (arrows), functional harmonies below each staff, and three perfect cadences (black boxes).</figDesc><graphic coords="1,321.00,170.89,236.05,257.19" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://www.tomcollinsresearch.net</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://extras.humdrum.org/bin/osxintel64/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Stravinsqi is labelled DMUN03 in the task overview paper<ref type="bibr" target="#b6">[6]</ref>. The other runs DMUN01 and DMUN02 are not remarkable: incorrect bar numbers in four pieces (see, for instance, the cross and correction in Figure1) and xml2hum conversion caused issues in DMUN01 and DMUN02.</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title/>
		<author>
			<persName><surname>References</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Improved Methods for Pattern Discovery in Music, with Applications in Automated Stylistic Composition</title>
		<author>
			<persName><forename type="first">T</forename><surname>Collins</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
		<respStmt>
			<orgName>Faculty of Mathematics, Computing and Technology, The Open University</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Doctoral Thesis</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Cognitive Foundations of Musical Pitch</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L</forename><surname>Krumhansl</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1990">1990</date>
			<publisher>Oxford University Press</publisher>
			<pubPlace>New York, NY</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Algorithms for discovering repeated patterns in multidimensional representations of polyphonic music</title>
		<author>
			<persName><forename type="first">D</forename><surname>Meredith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lemström</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Wiggins</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. New Music Res</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="321" to="345" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Algorithms for chordal analysis</title>
		<author>
			<persName><forename type="first">B</forename><surname>Pardo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">P</forename><surname>Birmingham</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Comput. Music J</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="27" to="49" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Visual hierarchical key analysis</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">S</forename><surname>Sapp</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Computers in Entertainment</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="1" to="19" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">The C@merata task at MediaEval 2014: natural language queries on classical music scores</title>
		<author>
			<persName><forename type="first">R</forename><surname>Sutcliffe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Crawford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Fox</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Root</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hovy</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">MediaEval 2014 Workshop</title>
				<meeting><address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014-10-16">2014. October 16-17 2014</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
