<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>MusicWeb: music discovery with open linked semantic metadata</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alo Allik</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mariano Mora-Mcginity</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gyorgy Fazekas</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mark Sandler</string-name>
          <email>mark.sandlerg@qmul.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Queen Mary University of London</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>This demo presents MusicWeb, a novel platform for linking music artists within a web-based application for discovering associations between them. MusicWeb provides a browsing experience using connections that are either extra-musical or tangential to music, such as the artists' political a liation or social in uence, or intra-musical, such as the artists' main instrument or most favoured musical key. The platform integrates open linked semantic metadata from various Semantic Web, music recommendation and social media data sources. The connections are further supplemented by thematic analysis of journal articles, blog posts and content-based similarity measures focussing on high level musical categories.</p>
      </abstract>
      <kwd-group>
        <kwd>Semantic Web</kwd>
        <kwd>Linked Open Data</kwd>
        <kwd>music metadata</kwd>
        <kwd>semantic audio analysis</kwd>
        <kwd>music information retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        MusicWeb is a music discovery platform which o ers users the possibility of
exploring editorial, cultural and musical links between artists. It gathers, extracts
and manages metadata from many di erent sources. The connections between
artists are based on YAGO categories such as style, geographical location,
instrumentation, record label, but also more obscure links, for instance, artists
who have received the same award, have shared the same fate, or belonged
to the same organisation or religion. The connections are further enhanced by
thematic analysis of journal articles and blog posts, content-based music
information retrieval similarity metrics and proximity measures in a 2-dimensional
mood space. Information about artists is collated and processed from several
different web knowledge content resources and presented for the user to navigate
in a faceted manner [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. MusicWeb as a discovery platform is subtly di erent in
function from recommender systems like dbrec [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which suggests music obtained
from DBpedia by computing a measure of semantic distance as the number of
indirect and distinct links between resources in a graph. This demo intends to
show the methods of linking artists employed in the system and how these could
help overcome issues such as infrequent access of lesser known artists in large
music catalogues (the \long tail" problem) or the di culty of recommending
artists without user ratings in systems that employ collaborative ltering (\cold
start" problem) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
1 http://musicbrainz.org
2 http://dbpedia.org
3 http://sameas.org
4 http://last.fm
for linking artists.
the Music Ontology [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] provides main concepts and properties for describing
musical entities, including artists, on the Semantic Web.
      </p>
      <p>AcousticBrainz5 service, which gathers crowd-sourced acoustic information
about music, facilitates content-based similarity calculation.
3</p>
    </sec>
    <sec id="sec-2">
      <title>Artist similarity</title>
      <p>
        There are many ways in which artists can be considered related. MusicWeb
uses Semantic Web technologies and linked data to facilitate faceted searching
and displaying of information [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. This is done by modeling artist similarities in
four di erent domains: socio-cultural, research and journalistic literature,
crowdsourced tag statistics and content-based information retrieval.
      </p>
      <p>
        Socio-cultural connections between artists in MusicWeb are primarily derived
from YAGO categories that are incorporated into entities in DBpedia. Many
categories, in particular those that can be considered extra-musical or tangential to
music, stem from the particular methodology used to derive YAGO categories
from Wikipedia [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>Literature-based linking is achieved by data-mining research articles and
online publications using natural language processing. MusicWeb uses Mendeley6
and Elsevier7 databases for accessing research articles that are curated and
categorised by keywords, authors and disciplines. Online newspapers, music
magazines and blogs, on the other hand, constitute non-curated data. Relevant
information in this case must be extracted from the body of the text by Web-crawling
based on keywords or tags. The Alchemy API8 is then used for named entity
recognition and keyword extraction.</p>
      <p>
        Crowd-sourced tags enable modelling similarity based on projected mood.
This method involves using the Semantic Web version of ILM10K music mood
dataset that consists of over 4000 unique artists [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The dataset is based on
crowd-sourced mood tag statistics from Last.fm users, which have been
transformed to 2-dimensional coordinates re ecting energy and pleasantness. The
similarity between artists is measured by rst obtaining the average location of
each artist based on their track coordinates. The average locations then enable
computing distances between artists and using these as the similarity metric.
Content-based linking involves methodology of Music Information Retrieval
(MIR) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] which facilitate applications that rely on perceptual, statistical,
semantic or musical features derived from audio using digital signal processing
and machine learning methods. These features may include statistical
aggregates computed from time-frequency representations extracted over short time
windows. Higher-level musical features include keys, chords, tempo, rhythm, as
well as semantic features like genre or mood, with speci c algorithms to
extract this information from audio. To exploit di erent types of similarity, we
5 https://acousticbrainz.org/
6 http://dev.mendeley.com/
7 http://dev.elsevier.com/
8 http://www.alchemyapi.com
model each artist using three main categories of audio descriptors: rhythmic,
harmonic and timbral. The features are obtained from the AcousticBrainz Web
service which provides descriptors in each category of interest. For each artist
in our database, we retrieve features for a large collection of their tracks in the
above categories, including beats-per-minute and onset rate (rhythmic), chord
histograms (harmonic) and MFCC (timbral) features.
4
      </p>
    </sec>
    <sec id="sec-3">
      <title>Conclusion</title>
      <p>MusicWeb is an emerging application to explore the possibilities of linked
databased music discovery. It facilitates users to engage in interesting discovery paths
through the space of music artists. The aim is to gather in one application various
di erent approaches to music discovery and how they can bene t from linked
music metadata. The next steps are directed toward evaluating its potential
acceptance by end users, in particular, exploring which linking methods listeners
nd most appealing or interesting, and which they would use more often.
MusicWeb is accessible online: http://musicweb.eecs.qmul.ac.uk/</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>Mathieu</given-names>
            <surname>Barthet</surname>
          </string-name>
          , Gyorgy Fazekas, Alo Allik, and
          <string-name>
            <surname>Mark</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Sandler</surname>
          </string-name>
          .
          <article-title>Moodplay: an interactive mood-based musical experience</article-title>
          .
          <source>In George Kalliris and Charalampos Dimoulas</source>
          , editors,
          <source>Proceedings of the Audio Mostly 2015 on Interaction With Sound</source>
          ,
          <source>AM '15</source>
          ,
          <string-name>
            <surname>Thessaloniki</surname>
          </string-name>
          , Greece, October 7-
          <issue>9</issue>
          ,
          <year>2015</year>
          , pages
          <article-title>3:1{3:8</article-title>
          . ACM,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Michael</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Casey</surname>
            , Remco Veltkamp, Masataka Goto, Marc Leman, Christophe Rhodes, and
            <given-names>Malcolm</given-names>
          </string-name>
          <string-name>
            <surname>Slaney</surname>
          </string-name>
          .
          <article-title>Content-based music information retrieval: Current directions and future challenges</article-title>
          . volume
          <volume>96</volume>
          . IEEE Proceedings,
          <year>April 2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>O.</given-names>
            <surname>Celma</surname>
          </string-name>
          .
          <article-title>Music Recommendation and Discovery:The Long Tail, Long Fail, and Long Play in the Digital Music Space</article-title>
          . Springer Verlag,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>M.S.</given-names>
            <surname>Fabian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gjergji</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Gerhard</surname>
          </string-name>
          . YAGO:
          <article-title>A core of semantic knowledge unifying wordnet and wikipedia</article-title>
          .
          <source>In 16th International World Wide Web Conference, WWW</source>
          , pages
          <volume>697</volume>
          {
          <fpage>706</fpage>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>Gary</given-names>
            <surname>Marchionini</surname>
          </string-name>
          .
          <article-title>Exploratory search: from nding to understanding</article-title>
          .
          <source>Communications of the ACM</source>
          ,
          <volume>49</volume>
          (
          <issue>9</issue>
          ),
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>E.</given-names>
            <surname>Oren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Delbru</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Decker</surname>
          </string-name>
          .
          <article-title>Extending faceted navigation for RDF data</article-title>
          .
          <source>In ISWC</source>
          , pages
          <volume>559</volume>
          {
          <fpage>572</fpage>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>Alexandre</given-names>
            <surname>Passant</surname>
          </string-name>
          . dbrec
          <article-title>- music recommendations using DBpedia</article-title>
          . In Peter F.
          <string-name>
            <surname>Patel-Schneider</surname>
            , Yue Pan, Pascal Hitzler,
            <given-names>Peter</given-names>
          </string-name>
          <string-name>
            <surname>Mika</surname>
            , Lei Zhang, Je
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Pan</surname>
          </string-name>
          , Ian Horrocks, and Birte Glimm, editors,
          <source>The Semantic Web - ISWC 2010 - 9th International Semantic Web Conference, ISWC 2010</source>
          , Shanghai, China, November 7-
          <issue>11</issue>
          ,
          <year>2010</year>
          ,
          <string-name>
            <given-names>Revised</given-names>
            <surname>Selected</surname>
          </string-name>
          <string-name>
            <given-names>Papers</given-names>
            ,
            <surname>Part</surname>
          </string-name>
          <string-name>
            <surname>II</surname>
          </string-name>
          , volume
          <volume>6497</volume>
          of Lecture Notes in Computer Science, pages
          <volume>209</volume>
          {
          <fpage>224</fpage>
          . Springer,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>Yves</given-names>
            <surname>Raimond</surname>
          </string-name>
          , Samer A Abdallah,
          <article-title>Mark B Sandler,</article-title>
          and
          <string-name>
            <given-names>Frederick</given-names>
            <surname>Giasson</surname>
          </string-name>
          .
          <article-title>The music ontology</article-title>
          .
          <source>In ISMIR</source>
          , pages
          <volume>417</volume>
          {
          <fpage>422</fpage>
          .
          <string-name>
            <surname>Citeseer</surname>
          </string-name>
          ,
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>