<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>IUI'18 Workshop on Intelligent Music Interfaces for Listening and Creation (MILC)</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Peter Knees</string-name>
          <email>peter.knees@tuwien.ac.at</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Markus Schedl</string-name>
          <email>markus.schedl@jku.at</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rebecca Fiebrink</string-name>
          <email>ebrink@gold.ac.uk</email>
          <email>r.fiebrink@gold.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Goldsmiths, University of</institution>
          ,
          <addr-line>London</addr-line>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Johannes Kepler University</institution>
          ,
          <addr-line>Linz</addr-line>
          ,
          <country country="AT">Austria</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>TU Wien</institution>
          ,
          <addr-line>Vienna</addr-line>
          ,
          <country country="AT">Austria</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Digital music technology constitutes a key factor in the music ecosystem. Through intelligent user interfaces, music consumers and producers can effectively and intuitively access and create sound. The goal of the MILC workshop is to provide a forum for the latest developments and trends in intelligent interfaces for music listening and creation by bringing together researchers from areas such as interactive machine learning, music information retrieval, recommender systems, human computer interaction, and adaptive systems.</p>
      </abstract>
      <kwd-group>
        <kwd>Author Keywords music listening</kwd>
        <kwd>music creation</kwd>
        <kwd>sound synthesis</kwd>
        <kwd>music recommendation</kwd>
        <kwd>music information retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>MOTIVATION
Today’s music ecosystem is permeated by digital technology –
from recording to production to distribution to consumption.
Intelligent technologies and interfaces play a crucial role
during all these steps. On the creation side, tools and interfaces
like new sensor-based musical instruments or software like
digital audio workstations and sound and sample browsers
support creativity. Generative systems can support novice
and professional musicians by automatically synthesizing new
sounds or even new musical material. On the consumption
side, tools and interfaces such as recommender systems,
automatic radio stations, or active listening applications allow users
to navigate the virtually endless spaces of music repositories.
Both ends of the music market therefore heavily rely on and
benefit from intelligent approaches that enable users to
access sound and music in unprecedented manners. This
ongoing trend draws from manifold areas such as interactive
machine learning, music information retrieval (MIR) – in
particular content-based retrieval systems, recommender systems,
human-computer interaction, and adaptive systems, to name
but a few prominent examples. In this light, the MILC
workshop held in the context of IUI, fosters the convention of the
digital music creation and performance and MIR communities
©2018. Copyright for the individual papers remains with the authors.
Copying permitted for private and academic purposes.</p>
      <p>MILC ’18, March 11, 2018, Tokyo, Japan
with adaptive user interface experts and provides a forum for
the latest developments and trends in intelligent interfaces in
these areas.1
CONTRIBUTIONS
The contributions to the MILC workshop reflect the relevance
of intelligent interfaces on both ends of the spectrum.
Papers dealing with personalization address needs of both,
consumers and creators. In “How Automated Recommendations
Affect the Playlist Creation Behavior of Users,” Kamehkhosh
et al. analyze the influence playlist construction support tools
have on resulting playlists and user behavior. In “geMsearch:
Personalized Explorative Music Search,” Esswein et al. make
music collections accessible by facilitating approximate
querying and visualization through low-dimensional vector
representations learned via graph embedding. Shi and Mysore propose
“MedleyAssistant – A system for personalized music medley
creation” that enables also non-experts to create medleys, while
maintaining the possibility to express their individual style.
Interaction with intelligent music systems and user interfaces
presents another diverse area. Vigliensoni et al. propose an
interactive machine learning approach to optical music
recognition in “An environment for machine pedagogy: Learning
how to teach computers to read music” and show that
performance is continuously improved when humans can intervene
and correct, therefore teach, the machine. Lindh questions
and investigates usability, accessibility and intuitiveness of the
ubiquitous skeuomorphic design in music creation interfaces
in “Beyond a Skeuomorphic Representation of Subtractive
Synthesis”. In “Overviewing a Field of Self-Organising Music
Interfaces: Autonomous, Distributed, Environmentally Aware,
Feedback Systems,” Kollias identifies and surveys the area of
“self-organising music,” which denotes a field comprising of
various intelligent sound and music interfaces and systems.
Intelligent approaches to composition support music creators
and open up new perspectives. Roberts et al. introduce an
interface to explore complex note sequence, drum pattern, and
timbre spaces with intuitive controls by utilizing deep-learned
representations in “Learning Latent Representations of Music
to Generate Interactive Musical Palettes”. In “Lumanote:
A Real-Time Interactive Music Composition Assistant” by
Granger et al., songwriters are interactively supported with
real-time, scale-aware chord and note suggestions in the
process of composition.
1https://iui2018milc.github.io
WORKSHOP ORGANIZERS</p>
      <p>Peter Knees is an Assistant Professor of the Institute of
Information Systems Engineering of TU Wien. In the last
decade, he has been an active member of the Music
Information Retrieval research community, reaching out to the related
areas of multimedia, text IR, and recommender systems.
Webpage: https://www.ifs.tuwien.ac.at/~knees/</p>
      <p>Markus Schedl is an associate professor at the Department
of Computational Perception of the Johannes Kepler
University Linz. His main research interests include web and social
media mining, recommender systems, information retrieval,
multimedia, and music information research. He (co-)authored
more than 150 refereed conference papers and journal articles.
Webpage: http://www.cp.jku.at/people/schedl/</p>
      <p>Rebecca Fiebrink is a Senior Lecturer at Goldsmiths,
University of London. Much of her research focuses on designing
the use of machine learning as a creative tool. Fiebrink is the
developer of the Wekinator, open-source software for real-time
interactive machine learning whose current version has been
downloaded over 10,000 times. She is the creator of a MOOC
titled “Machine Learning for Artists and Musicians,” which
launched in 2016 on the Kadenze platform.</p>
      <p>Webpage: https://www.doc.gold.ac.uk/~mas01rf/
PROGRAM COMMITTEE</p>
      <p>Baptiste Caramiaux, IRCAM, France
Mark Cartwright, New York University, USA
Matthew Davies, INESC TEC Porto, Portugal
Christian Dittmar, International Audio Laboratories
Erlangen, Germany
Bruce Ferwerda, Jönköping University, Sweden
Ichiro Fujinaga, McGill University, Canada
Jason Hockman, Birmingham City University, UK
Masataka Goto, National Institute of Advanced Industrial
Science and Technology, Japan
Florian Grote, Native Instruments GmbH, Germany
Bogdan Ionescu, University Politehnica of Bucharest,
Romania
Vikas Kumar, University of Minnesota, USA
Cynthia Liem, Delft University of Technology, Netherlands
Matija Marolt, University of Ljubljana, Slovenia</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>