<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Music Note Ontology?,??</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>The Music Note Ontology</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>FICLIT, University of Bologna</institution>
          ,
          <addr-line>Bologna</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>LILEC, University of Bologna</institution>
          ,
          <addr-line>Bologna</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In this paper we propose the Music Note Ontology, an ontology for modelling music notes and their realisation. The ontology addresses the relation between a note represented in a symbolic representation system, and its realisation, i.e. a musical performance. This work therefore aims to solve the modelling and representation issues that arise when analysing the relationships between abstract symbolic features and the corresponding physical features of an audio signal. The ontology is composed of three di erent Ontology Design Patterns (ODP), which model the structure of the score (Score Part Pattern), the note in the symbolic notation (Music Note Pattern) and its realisation (Musical Object Pattern).</p>
      </abstract>
      <kwd-group>
        <kwd>Ontology Design Patterns</kwd>
        <kwd>Computational Musicology</kwd>
        <kwd>Computer Music</kwd>
        <kwd>Music Information Retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        A music note is de ned as \the marks or signs by which music is put on paper.
Hence the word is used for the sounds represented by the notes" [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. However,
musical notation provides some general information on how to play a certain
note. These indications are then enriched in the context of a performance by a
large amount of information, for example, from the musician's sensitivity or the
conductor's instructions. Historically, music notation has been a crucial
innovation that allowed the study of music and hence the rise of musicology. Music
scores were initially introduced for the primary purpose of recording music by
giving other musicians the possibility of playing the same piece in turn,
reproducing it. However, musical notation by de nition entails expressive constraints
on the description of musical content.
      </p>
      <p>Representing music involves a number of issues that are closely related to
the complexity of the musical content. This can be found in every system of
? Copyright © 2021 for this paper by its authors. Use permitted under Creative</p>
      <p>
        Commons License Attribution 4.0 International (CC BY 4.0).
?? This project has received funding from the European Union's Horizon 2020 research
and innovation programme under grant agreement No 101004746.
? ? ? Corresponding author: andrea.poltronieri2@unibo.it
music representation, from music scores to their various forms of digital
encoding, usually referred to as symbolic representation systems. As Cook observed,
music scores symbolise rather than represent music, as people don't play
musical rhythms as written and often they don't play the pitches as written: that is
because the notation is only an approximation [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. To underline this concept, it
is useful to draw a distinction between a score and a musical object [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. While
the former can be thought of as a set of instructions that the musician or a
computer system uses to realise a piece of music, a musical object consists of the
realisation of this process. In other words, musical notation precedes realisation
or interpretation and de nes only partially the musical object [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>
        Music notation is not just a mechanical transformation of performance
information. Performance nuances are lost when moving from performance to
notation, and symbolic structure is partly lost in the translation from notation
to performance [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Moreover, in music notation some features of musical
content are completely overlooked. For example, timbre information is generally
ignored and is at best provided through generic information about the
instrument that is supposed to play the part. Many of these problems have not been
overcome by the numerous systems of symbolic representation proposed in recent
decades. Music representation systems (MSR) have been evaluated by Wiggins
et al. [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ] using a system based on two orthogonal dimensions. These two
dimensions have been named as expressive completeness and structural generality.
Expressive completeness is described as the extent to which the original musical
content may be retrieved and/or recreated from its representation [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
Structural generality, instead, refers to the range of high-level structures that can
be represented and manipulated [
        <xref ref-type="bibr" rid="ref26">26</xref>
        ]. For example, raw audio is very expressive
since it contains a great deal of information about a certain music performance;
but a MIDI representation contains much less information, e.g. timbral features
are not represented in this format. On the contrary, when evaluating structural
generality, raw audio performs poorly, due to the di culties in the extraction of
structured information (e.g., tempo, chords, notes, etc.) from such a format.
      </p>
      <p>
        Another problem in representing music is linked to the twofold nature of
musical content, which contains information that can be reduced to mathematical
functions, but also information related to the emotional spectrum and a wide
range of psychoacoustic nuances. In fact, music is distinguished by the presence
of many relationships that can be treated mathematically, including rhythm and
harmony. However, elements such as tension, expectancy, and emotion are less
prone to mathematical treatment [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        The limits of symbolic representations are also accentuated when the
performance is considered in relation to the concept of interpretation. For each
composition, several performances of the same one correspond to as many
interpretations of the musical piece. These interpretations can also vary greatly
from one another from agogic (note accentuation, [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]), timbral and dynamic
viewpoints.
      </p>
      <p>Therefore, this work aims to represent musical notes both symbolically and as
a realisation of the note itself. The alignment between these two representation
systems can be relevant on several grounds. For example, it can allow musical
analysis using a structured representation (i.e. a symbolic representation) while
simultaneously taking into account all the information that is only contained in
the signal (e.g. timbre information). This would allow a di erent level of analysis,
taking into account information that is usually overlooked in music scores and
symbolic representations. On the other side, this type of representation would
allow structured analysis of the music signal by having the signal information
encoded in strict relation with the music score and the multiple hierarchies that
the music score entails.</p>
      <p>Furthermore, this type of representation allows the analysis of di erent
realisations of the same note, also providing information on how a score is performed
di erently in di erent performances.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Physical Features and Perceptual Features</title>
      <p>The features that need to be taken into account when representing music can be
reduced to four main categories: tempo, pitch, timbre and dynamics. However,
all these concepts refer to music perception. Symbolic representations, however,
only represent abstractions of these features. In contrast, musical performance, as
analysed in recorded form (i.e. signal representation), contains information that
refers to physical characteristics. Through the analysis of these characteristics,
it is in turn possible to abstract the perceptual characteristics aroused by a
particular sound. However, this work aims not to analyse the perceptual aspects
of music, but rather to formalise a representation that aims to express a musical
note from the points of view of both scores and musical objects.</p>
      <p>
        As far as time is concerned, the main distinction that arises is that between
the quantised time of a symbolic representation, and real time, which is expressed
in seconds. A symbolic representation describes temporal information using
predetermined durations for each note (e.g., quarter notes, eighth notes, sixteenth
notes, etc.). These durations are executed in relation to a time signature,
indicated by a tempo marking (e.g., slow, moderate, allegro, vivace, etc.), or by
a metronome mark, which indicates the tempo measured in beats per minute
(bpm). Generally, symbolic notations (e.g. MusicXML) use the same
expedients as classical musical notation. However, some representation systems rely on
real-time, such as the MIDI standard [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        When considering frequency, the representation issue is more challenging. A
sound, e.g. a note in the form of a musical note, is usually associated with a
pitch. However, although pitch is associated with frequency [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], it is a
psychological percept that does not map straightforwardly onto physical properties of
sound [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. The American National Standards Institute (ANSI) de nes pitch as
the \auditory attribute of sound according to which sounds can be ordered on
a scale from low to high" [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. In fact, in many languages, the pitch is described
by using terms having a spatial connotation such as \high" and \low" [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
However, the spatial arrangement of sounds has been shown to be in uenced
by the listener's musical training [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ] and their ethnic group [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The relation
between pitch and frequency is evident in the case of pure tones [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. For
example, a sinusoid having a frequency of 440 Hz corresponds to the pitch A4.
However, real sounds are not composed of a simple pure note with a unique
and well-de ned frequency. Playing a single note on an instrument may result
in a complex sound that contains a mixture of di erent frequencies changing
over time. Intuitively, such a musical tone can be described as a superposition
of pure tones or sinusoids, each with its frequency of vibration, amplitude, and
phase. A partial is any of the sinusoids, by which a musical tone is described
[
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. The frequency of the lowest partial is de ned as the fundamental frequency.
Most instruments produce harmonic sounds, i.e. composed of frequencies close
to the partial harmonics. However, some instruments such as xylophones, bells,
and gongs produce inharmonic sounds, i.e. composed of frequencies that are not
multiples of the fundamental. As a result, the pitch perception of these sounds
is distorted for a listener who is unfamiliar with these instruments [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Several
approaches have been proposed to determine the perceived pitch of these sounds,
based both on the periodicity of the frequencies that compose the sound [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] and
on spectral clues [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. However, recent research suggests that pitch perception
involves learning to recognise a sound timbre over a range of frequencies, and then
associating changes in frequency with visuospatial and kinesthetic dimensions of
an instrument [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        On the other hand, timbre is a sound property even more di cult to grasp.
ANSI de nes the timbre as the \attribute of auditory sensation, in terms of
which a subject can judge that two sounds similarly presented and having the
same loudness and pitch are dissimilar" [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. However, even today, the de nition
of timbre is rather debated. For example, it has been de ned as a \misleadingly
simple and vague word encompassing a very complex set of auditory attributes,
as well as a plethora of psychological and musical issues" [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The main
problem in de ning timbre is that there is no precise correspondence with one or
more physical aspects. Timbre is also challenging to abstract, as it is usually
described using adjectives such as light, dark, bright, rough, violin-like, etc.
Numerous studies have shown that timbre depends on a large number of factors
and physical characteristics. Some research has tried to represent the perception
of timbre using multidimensional scaling (e.g. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]). Other studies in this area
focus on the grouping of timbre into a \Timbre Space" [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. While those studies
represented timbre in terms of perceptual dimensions, others have represented
timbre in terms of physical dimensions, such as the set of parameters needed for
a particular synthesis algorithm [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. These physical characteristics concern both
the time axis, such as attack time and vibrato variation and amplitude, spectral
features such as spectral centroid, and harmonic features such as odd-to-even
ratio and inharmonicity. An approximation of these features is provided by a
model called ADSR, which consists of the analysis of the envelope of the sound
wave and the measurement of attack (A), decay (D), sustain (S) and release (R).
The envelope can be de ned as a smooth curve outlining the extremes in the
amplitude of a waveform. The ADSR model, however, is a strong simpli cation
and only yields a meaningful approximation for amplitude envelopes of tones
that are generated by speci c instruments [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>
        The same distinction drawn between frequency and pitch can be made for
loudness and sound intensity. Loudness is a perceptual property whereby sounds
are ordered on a scale from quiet to loud. Sound intensity is generally measured
in dB, and expresses the sound power (expressed in Watts) in relation to the
physical space in which the sound propagates (expressed in square metres) [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
As a perceptual property, it is by its nature subjective and is closely related to
loudness. However, loudness also depends on other sound characteristics such as
duration or frequency [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Also, the sound duration in uences perception since
the human auditory system averages the e ect of sound intensity over an interval
up to a second.
      </p>
      <p>Formalizing the aforementioned perceptual concepts is beyond the scope of
this paper. Instead, the purpose of this contribution is to de ne (some of) the
attributes that can provide helpful information to describe sound material. The
importance of this analysis is to highlight the inner complexity of the sound
material, underlying the connections that can be drawn between physical aspects
and perceptual components, as well as the relationships among these aspects.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Related Work</title>
      <p>
        Several Semantic Web ontologies have been proposed for modelling musical data.
The Music Theory Ontology (MTO) [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] aims at modelling music theoretical
concepts. This ontology focuses on the rudiments that are necessary to
understand and analyse music. However, this ontology is limited to the description of
note attributes (dynamics, duration, articulation, etc.) at the level of detail of a
note set. This peculiarity considerably reduces the expressiveness of the model,
especially in relation to polyphonic and multi-voice music.
      </p>
      <p>
        The Music Score Ontology (Music OWL) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] aims to represent similar
concepts described in the Music Theory Ontology. However, it focuses on music
notation, and the classes it is composed of are all aimed at representing music
in a notated form, hence related to music sheets.
      </p>
      <p>
        The Music Notation Ontology [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is an ontology of music notation content,
focused on the core \semantic" information present in a score and subject to
the analytic process. Moreover, it models some high-level analytic concepts (e.g.
dissonances), and associates them with fragments of the score.
      </p>
      <p>However, all these ontologies only consider the characteristics of the musical
material from the point of view either symbolic representation or music score.
The aim of this paper is instead to propose a model that can be as general as
possible and that can therefore represent musical notation expressed in di erent
formats and considering at the same time scores and symbolic notations. This
would allow interoperability between di erent notation formats and the
standardisation of musical notation in a format that aims to be as expressive as
possible.</p>
      <p>Moreover, to the best of our knowledge there is no works that relates the
musical note (understood as part of the symbolic notation) to the corresponding
musical object (understood as the realisation of a musical note). In addition,
none of the proposed ontologies seem to model the physical information related
to the audio signal.</p>
      <p>Therefore, this paper proposes an ontology that can represent a note both
from a symbolic and a performance point of view. To do this, it is necessary
to take into account both the symbolic characteristics (the signs and attributes
typical of musical notation), and the physical characteristics of the note as
reproduced by a musical instrument. However, this work is not intended to go into
the perceptual characteristics of music.</p>
      <p>
        The problem of modelling information with respect to its realisation has
already been analysed in the past. For example, the Ontology Design Pattern
(ODP) Information Realization3 {extracted from DOLCE+DnS Ultralite
Ontology4 [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]{ proposes to help ontology designers to model this type of scenario.
This pattern allows designers to model information objects and their
realizations. This allows to reason about physical objects and the information they
realize, by keeping them distinguished.
      </p>
      <p>
        The same kind of relationship can be found in the FRBR5 [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] conceptual
entity-relationship model, which groups entities into four di erent levels (work,
expression, manifestation and item). The relationship between expression,
dened as the \speci c intellectual or artistic form that a work takes each time it
is realized" and manifestation, de ned as \the physical embodiment of an
expression of a work", can be traced back to the modelling problem discussed in
this paper.
      </p>
      <p>However, these examples do not solve the modelling problems that have been
advanced in the previous sections. In particular, the relationship between the
approximate features represented by the symbolic notation and their realisation
remains a challenge in the eld of ontology design. To achieve this, this research
proposes a new ontology to represent symbolic musical notation and its
realisation, considering the relationships between the abstraction of musical features
expressed in the symbolic notation and the physical characteristics of the audio
signal. Furthermore, this paper proposes a pattern-based approach, by proposing
an ontology composed of modular and reusable elements for the representation
of both audio and symbolic musical content.
4</p>
    </sec>
    <sec id="sec-4">
      <title>The Music Note Ontology</title>
      <p>The Music Note Ontology6 (see Figure 1) models a musical note both as a
constituent element of the score, and as a musical object, i.e. as a realisation of
3 http://www.ontologydesignpatterns.org/cp/owl/informationrealization.owl
4 http://ontologydesignpatterns.org/wiki/Ontology:DOLCE+DnS_Ultralite
5 https://www.ifla.org/publications/functional-requirements-for-bibliographic-records
6 The OWL le of the ontology can be found at: https://purl.org/
andreapoltronieri/notepattern
the event represented by a score or symbolic notation. To do this, the proposed
ontology represents both the elements and the typical hierarchies of a score,
describing all its attributes. In addition, the ontology models the realisation of
a note, describing its physical characteristics in a given performance.</p>
      <p>The proposed ontology is composed of three distinct Ontology Design
Patterns, which are presented in this paper and have been submitted to the ODP
portal at the following links::
{ The Score Part Pattern: http://ontologydesignpatterns.org/wiki/Submissions:</p>
      <p>Scorepart
{ The Musical Object Pattern: http://ontologydesignpatterns.org/wiki/</p>
      <p>Submissions:Musicalobject
{ The Music Note Pattern: http://ontologydesignpatterns.org/wiki/Submissions:</p>
      <p>Notepattern.</p>
      <p>ID
CQ1
CQ2
CQ3
CQ4
CQ5
CQ6
CQ7</p>
      <p>Competency Question
what is the name of a note?
what part of the score does a note belong to?
what are the dynamic indications referring to a note in the score?
what is the fundamental frequency of a note?
what are the di erent frequencies that make up the spectrum of a note?
what is the duration in seconds of a note, in a given performance?
how is the envelope of a note shaped?</p>
      <p>The note as represented in the score is described by the Music Note Pattern
(in green in the gure), which in turn imports the Score Part pattern (in orange
in the gure) and the Musical Object pattern (in yellow in the gure). The former
import describes the relationships between the di erent components of a score,
while the latter models the musical object from the point of view of its physical
features.</p>
      <p>A selection of competency questions that the ontology must be able to
answer are listed in Table 1. However, the complete list of compency questions in
available on the repository of the project7.</p>
      <p>Some of the classes and properties of the ontology have also been aligned with
currently available ontologies and the alignments are available in a separate le8.
Alignments with other ontologies were mainly made on the Score Part Pattern
and the Music Note Pattern. Although some of the available ontologies (see
Section 3) de ne some of the classes and properties present in these patterns, for
7 The project repository is available at: https://github.com/andreamust/music_
note_pattern/
8 Alignments to the Music Note Ontology are available at the URI: purl.org/
andreapoltronieri/notepattern-aligns
this we needed to re-model the score and symbolic elements. In fact, the objective
of this work is, among others, to abstract from a speci c representation system,
combining the attributes of the musical score with other elements of the MIDI
representation (as the most widely used symbolic representation) and, to the beat
of our knowledge, there are no ontologies available with this characteristics.
The hierarchy (i.e. the mereological structure) of a music score is described
by the Score Part Pattern9. The Part class describes the instrumental part of
a score, which refers to a speci c sta of the musical notation that is
associated with an instrument. The instrument playing the part is modelled by the
class Instrument, which is associated by means of a datatype property with
the MIDI program assigned to that instrument (hasMidiProgram). Each part of
the score is also divided into sections (class Section), similar music fragments
(class HomogeneousFragment) and voices (class Voice). A voice is de ned as
9 Score Part Pattern URI: https://purl.org/andreapoltronieri/scorepart
one of several parts that can be performed within the same sta (e.g. alto and
soprano). A section is instead a part of the music score de ned by the repetition
signs. A fragment, on the other hand, is a grouping of notes that share the same
metric, tempo and clef, described by the object properties hasMetric, hasTempo
and hasClef.
4.2</p>
      <sec id="sec-4-1">
        <title>The Musical Object Pattern</title>
        <p>The Musical Object Pattern10 models the physical features of a musical note
object, i.e. the execution of a note. Speci cally, this pattern models the
physical characteristics that can be extracted from the sound wave produced by an
instrument playing a musical note. The MusicalObject class is connected to
four classes that describe these physical characteristics, namely duration, sound
intensity, frequency and envelope. The MusicObjectDuration class expresses
the duration in seconds of the musical object, by means of the object
property hasDurationInSeconds. In the same way, the musical intensity is modelled
via the SoundIntensity class. Frequency is modelled by means of the class
Frequency and its sub-classes FundamentalFrequency and PartialFrequency.
For each expressed frequency, the magnitude of the frequency is also indicated
using the hasFrequencyMagnitude datatype property. Finally, the Envelope class
is connected to four datatype properties that describe the envelope of the
waveform according to the ADSR model, namely hasAttack, hasSustain, hasDecay
and hasRelease.
4.3</p>
      </sec>
      <sec id="sec-4-2">
        <title>The Music Note Pattern</title>
        <p>The Music Note Pattern11 models the symbolic note (as represented in a score
or in most of symbolic representation systems) and its attributes. Moreover, this
ODP aims to abstract over di erent music representation systems. This means
that it is possible to represent with this pattern music that was originally encoded
in other formats. To do so, the pattern proposed describes each symbolic notes
both by means of the standard music score notation and by using the MIDI
reference for each of the notes's attributes. This would allow the conversion
between
SymbolicNote is the central class representing the note as represented in the
score. This class has seven subclasses, one for each of the seven notes of the
tempered system (primitive notes), plus 35 subclasses which are de ned as the
combination between primitives and accidentals. The class Accidental (which has
in turn ve subclasses: Flat, Sharp, Natural, DoubleFlat, and DoubleSharp)
allows the representation of altered notes. This allows to represent every note
of the Western notation system, taking into account the function with respect
to the tonality of the piece (e.g. di erentiating between the notes A# and Bb).
10 Musical Object Pattern URI:</p>
        <p>musicalobject
11 Music Note Pattern URI: https://purl.org/andreapoltronieri/notepattern
https://purl.org/andreapoltronieri/
Notice that such di erentiation has an important cognitive function for the
music reader (distinguishing enharmonics according to the tonality), function that
is mostly irrelevant for a computer reading a midi, but very relevant for an AI
learning to generate music from notation). Furthermore, the SymbolicNote class
is linked to several classes and data properties, describing the note's attributes.
The class Position de nes the position of the note both in relation to the score
(datatype property hasToMeasure) and within the measure (datatype property
hasPositionInMeasure). The class NoteDuration describes the symbolic
duration of the note, as expressed in the music score. The NoteDynamic class,
instead, describes the dynamics of the note with reference to both musical
notation (datatype property hasLiteralDynamic, e.g crescendo, vibrato, etc.) and
the MIDI standard (data property hasMidiVelocity). Similarly, the NotePitch
class de nes both the octave in which the note is located (datatype property
isDefinedByOctave) and the MIDI pitch (datatype property hasMidiPitch).
In addition, the SymbolicNote class is linked to the other two patterns that
make up the Music Note Ontology. The SymbolicNote belongs in fact to a
Section rather than to a Voice, while it has realisation in a MusicalObject. Finally,
the relationships between the attributes of the SymbolicNote and those of the
MusicalObject are expressed.</p>
        <p>The NotePitch of a symbolic note is represented as the abstraction of the
musical object's frequency (expressed by the object property
hasFrequencyAbstraction), while NoteDynamic and NoteDuration are described
as the abstraction of SoundIntensity and MusicObjectDuration, respectively
(object properties hasDynamicAbstraction and hasDurationAbstraction).</p>
        <p>In order to test the ontology, a Knowledge Base (KB)12 containing a single
note and a single musical object was created. SPARQL queries were also created
for each of the competency questions de ned in Table 1. The KB can be tested
against the SPARQL queries, in order to verify the expressiveness of the ontology
with respect to the competency questions. The complete list of SPARQL queries
is available on the repository of the project.</p>
        <p>
          The ontology can also be used to enrich the description of music content that
is already annotated using other ontological models. For example, the classes
and properties describing the structure of musical notation can be aligned to the
widely used Music Ontology [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. Similarly, the same classes can be aligned with
the aforementioned Music OWL and Music Score Ontology.
5
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Future Perspectives</title>
      <p>In this paper we have proposed the Music Note Ontology, an ontology for
modelling musical content both from a symbolic point of view and in terms of its
realisation. To do this, three di erent Ontology Design Patterns have been
designed to model the internal relations of the score, the note from the symbolic
point of view, and the note as musical object. The relationships between the
12 https://purl.org/andreapoltronieri/notepatterndata
di erent musical notations have also been considered, in particular between the
symbolic abstraction and the features that can be extracted from the audio signal
of a performance.</p>
      <p>In our future work we aim at modelling the perception of the di erent
components of the audio signal, describing the di erent perceptual levels at which it is
possible to experience music. We also aim to formalise the musical perception of
di erent features from a subjective point of view, e.g. in individuals with di erent
musical skills or from di erent musical cultures. Furthermore, since the model
proposed at the moment can only express notes related to the traditional
notation system (and thus related to the temperate system of the Western musical
tradition) we plan to expand the ontology with notes from other temperaments
and musical traditions, as well as unpitched notes and microtonal variations.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Antovic</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Musical metaphors in serbian and romani children: An empirical study</article-title>
          .
          <source>Metaphor and Symbol</source>
          <volume>24</volume>
          (
          <issue>3</issue>
          ),
          <volume>184</volume>
          {202 (Jul
          <year>2009</year>
          ). https://doi.org/10.1080/10926480903028136, https://doi.org/10.1080/ 10926480903028136
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Association</surname>
            ,
            <given-names>M.M.:</given-names>
          </string-name>
          <article-title>The complete midi 1.0 detailed speci cation: Incorporating all recommended practices (</article-title>
          <year>1996</year>
          ), https://books.google.it/books?id= cPpcPQAACAAJ
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Brown</surname>
          </string-name>
          , C.:
          <article-title>Historical performance, metronome marks and tempo in Beethoven's symphonies</article-title>
          .
          <source>Early Music XIX(2)</source>
          ,
          <volume>247</volume>
          {
          <volume>260</volume>
          (05
          <year>1991</year>
          ). https://doi.org/10.1093/earlyj/XIX.2.247, https://doi.org/10.1093/earlyj/ XIX.2.
          <fpage>247</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bruneau</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scelo</surname>
          </string-name>
          , T.,
          <string-name>
            <surname>d'Acoustique</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          : Fundamentals of Acoustics. ISTE, Wiley (
          <year>2013</year>
          ), https://books.google.it/books?id=1\_y6gKuGSb0C
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Carlyon</surname>
            ,
            <given-names>R.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shackleton</surname>
          </string-name>
          , T.M.
          <article-title>: Comparing the fundamental frequencies of resolved and unresolved harmonics: Evidence for two pitch mechanisms?</article-title>
          <source>The Journal of the Acoustical Society of America</source>
          <volume>95</volume>
          (
          <issue>6</issue>
          ),
          <volume>3541</volume>
          {
          <fpage>3554</fpage>
          (
          <year>1994</year>
          ). https://doi.org/10.1121/1.409971, https://doi.org/10.1121/1.409971
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Cher</surname>
            ,
            <given-names>S.S.s.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guillotel</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hamdi</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rigaux</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Travers</surname>
          </string-name>
          , N.:
          <article-title>Ontologybased annotation of music scores</article-title>
          .
          <source>In: Proceedings of the Knowledge Capture Conference. K-CAP</source>
          <year>2017</year>
          ,
          <article-title>Association for Computing Machinery</article-title>
          , New York, NY, USA (
          <year>2017</year>
          ). https://doi.org/10.1145/3148011.3148038, https://doi.org/ 10.1145/3148011.3148038
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Cook</surname>
          </string-name>
          , N.:
          <article-title>Towards the complete musicologist</article-title>
          .
          <source>In: Proceedings of the 6th International Conference on Music Information Retrieval</source>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Dannenberg</surname>
          </string-name>
          , R.B.:
          <article-title>Music representation issues, techniques, and systems</article-title>
          .
          <source>Computer Music Journal</source>
          <volume>17</volume>
          (
          <issue>3</issue>
          ),
          <volume>20</volume>
          {
          <fpage>30</fpage>
          (
          <year>1993</year>
          ), http://www.jstor.org/stable/3680940
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Grey</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          :
          <article-title>Multidimensional perceptual scaling of musical timbres</article-title>
          .
          <source>The Journal of the Acoustical Society of America</source>
          <volume>61</volume>
          (
          <issue>5</issue>
          ),
          <volume>1270</volume>
          {
          <fpage>1277</fpage>
          (
          <year>1977</year>
          ). https://doi.org/10.1121/1.381428, https://doi.org/10.1121/1.381428
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Grove</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fuller-Maitland</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Grove's Dictionary of Music and Musicians. No. v. 4 in Grove's Dictionary of Music and Musicians</article-title>
          ,
          <string-name>
            <surname>Macmillan</surname>
          </string-name>
          (
          <year>1908</year>
          ), https://books. google.it/books?id=FBsPAAAAYAAJ
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Institute</surname>
            ,
            <given-names>A.N.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sonn</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , of America, A.S.
          <article-title>: American National Standard Psychoacoustical Terminology</article-title>
          . N.: ANSI, American National Standards Institute (
          <year>1973</year>
          ), https://books.google.it/books?id=5Fo0GQAACAAJ
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Jones</surname>
            , J., de Siqueira Braga,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tertuliano</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kauppinen</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Musicowl: The music score ontology</article-title>
          .
          <source>In: Proceedings of the International Conference on Web Intelligence</source>
          . p.
          <volume>1222</volume>
          {
          <fpage>1229</fpage>
          . WI '
          <volume>17</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA (
          <year>2017</year>
          ). https://doi.org/10.1145/3106426.3110325, https://doi. org/10.1145/3106426.3110325
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Mascardi</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cord</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>A comparison of upper ontologies</article-title>
          .
          <source>In: WOA</source>
          . pp.
          <volume>55</volume>
          {
          <issue>64</issue>
          (01
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>McAdams</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giordano</surname>
            ,
            <given-names>B.L.:</given-names>
          </string-name>
          <article-title>The perception of musical timbre</article-title>
          ., pp.
          <volume>113</volume>
          {
          <fpage>124</fpage>
          . Oxford library of psychology., Oxford University Press, New York, NY, US (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>McLachlan</surname>
            ,
            <given-names>N.M.</given-names>
          </string-name>
          :
          <string-name>
            <surname>Timbre</surname>
          </string-name>
          , Pitch, and
          <string-name>
            <surname>Music</surname>
          </string-name>
          . Oxford University Press (
          <year>Jun 2016</year>
          ). https://doi.org/10.1093/oxfordhb/9780199935345.013.44, https://doi. org/10.1093/oxfordhb/9780199935345.013.44
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Moore</surname>
            ,
            <given-names>B.C.J.:</given-names>
          </string-name>
          <article-title>An introduction to the psychology of hearing</article-title>
          , 5th ed.
          <article-title>An introduction to the psychology of hearing</article-title>
          , 5th ed., Academic Press, San Diego, CA, US (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17. Muller, M.: Fundamentals of Music Processing: Audio, Analysis, Algorithms, Applications. Springer Publishing Company, Incorporated, 1st edn. (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Nattiez</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dunsby</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          :
          <article-title>Fondements d'une semiologie de la musique</article-title>
          .
          <source>Perspectives of New Music</source>
          <volume>15</volume>
          (
          <issue>2</issue>
          ),
          <volume>226</volume>
          {
          <fpage>233</fpage>
          (
          <year>2021</year>
          /05/14/
          <year>1977</year>
          ). https://doi.org/10.2307/832821, http://www.jstor.org/stable/832821, full publication date: Spring - Summer,
          <year>1977</year>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Rashid</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>De Roure</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McGuinness</surname>
            ,
            <given-names>D.L.:</given-names>
          </string-name>
          <article-title>A music theory ontology</article-title>
          .
          <source>In: Proceedings of the 1st International Workshop on Semantic Applications for Audio and Music</source>
          . p.
          <volume>6</volume>
          {
          <fpage>14</fpage>
          . SAAM '
          <volume>18</volume>
          ,
          <string-name>
            <surname>Association</surname>
          </string-name>
          for Computing Machinery, New York, NY, USA (
          <year>2018</year>
          ). https://doi.org/10.1145/3243907.3243913, https: //doi.org/10.1145/3243907.3243913
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Raymond</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abdallah</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sandler</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giasson</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>The music ontology</article-title>
          .
          <source>In: Proceedings of the 8th International Conference on Music Information Retrieval (ISMIR</source>
          <year>2007</year>
          ). pp.
          <volume>417</volume>
          {
          <fpage>422</fpage>
          . Vienna, Austria (Sep
          <year>2007</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Rusconi</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kwan</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giorndano</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Umilta</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Butterworth</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          :
          <article-title>Spatial representation of pitch height: the SMARC e ect</article-title>
          .
          <source>Cognition</source>
          <volume>99</volume>
          (
          <issue>2</issue>
          ),
          <volume>113</volume>
          {129 (Mar
          <year>2006</year>
          ). https://doi.org/10.1016/j.cognition.
          <year>2005</year>
          .
          <volume>01</volume>
          .004, https://doi.org/ 10.1016/j.cognition.
          <year>2005</year>
          .
          <volume>01</volume>
          .004
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Schouten</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>The Perception of Subjective Tones</article-title>
          . Separaat / Laboratoria N. V. Philips Gloeilampenfabrieken, Verlag nicht ermittelbar (
          <year>1938</year>
          ), https://books. google.it/books?id=d9vjzQEACAAJ
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Stewart</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Walsh</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Frith</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          :
          <article-title>Reading music modi es spatial mapping in pianists</article-title>
          .
          <source>Perception &amp; Psychophysics</source>
          <volume>66</volume>
          (
          <issue>2</issue>
          ),
          <volume>183</volume>
          {195 (Feb
          <year>2004</year>
          ). https://doi.org/10.3758/bf03194871, https://doi.org/10.3758/bf03194871
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>P.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shadle</surname>
            ,
            <given-names>S.:</given-names>
          </string-name>
          <article-title>Frbr in the real world</article-title>
          .
          <source>The Serials Librarian</source>
          <volume>52</volume>
          (
          <issue>1-2</issue>
          ),
          <volume>93</volume>
          {
          <fpage>104</fpage>
          (
          <year>2007</year>
          ). https://doi.org/10.1300/J123v52n01 09, https://doi.org/10.1300/ J123v52n01_
          <fpage>09</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Wiggins</surname>
          </string-name>
          , G.:
          <article-title>Computer-representation of music in the research environment</article-title>
          .
          <source>Modern Methods for Musicology: Prospects</source>
          , Proposals, and Realities pp.
          <volume>7</volume>
          {
          <issue>22</issue>
          (
          <year>2009</year>
          ), https://doi.org/10.4324/9781315595894
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Wiggins</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Smaill</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Harris</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>A framework for the evaluation of music representation systems</article-title>
          .
          <source>Computer Music Journal</source>
          <volume>17</volume>
          (
          <issue>3</issue>
          ),
          <volume>31</volume>
          {
          <fpage>42</fpage>
          (
          <year>1993</year>
          ), http://www.jstor.org/stable/3680941
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>