<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>June</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Personalization of Object-based Audio for Accessibility using Narrative Importance</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ben Shirley</string-name>
          <email>b.g.shirley@salford.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lauren Ward∗</string-name>
          <email>l.ward7@edu.salford.ac.uk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Emmanouil Theofanis</string-name>
          <email>e.t.chourdakis@qmul.ac.uk Queen Mary University London, UK BBC Research &amp; Development, UK</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Chourdakis∗</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Salford</institution>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>University of Salford, UK, BBC Research &amp; Development</institution>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <volume>0</volume>
      <fpage>3</fpage>
      <lpage>05</lpage>
      <abstract>
        <p>An increasing incidence of hearing impairment and of reported problems with broadcast audio is leading to an increased demand for personalized audio services. Previous research has treated these issues as a 'speech in noise' problem; sounds are viewed as speech (good) or as competing masker (bad). This binary approach to accessible audio disregards the important role of some non-speech sounds in facilitating understanding of broadcast programme narrative. This work, as part of the S3A project, has taken a more holistic approach to audio personalization using categories of narrative importance to provide complex manipulations of broadcast audio based on narrative comprehension, instead of simply intelligibility. A simple, intuitive user-interface allows the user to adjust the complexity of audio scenes based on their personal hearing needs, metadata is generated at production using plugins to generate appropriate metadata and audio previews of user-narrative importance settings. This paper outlines the concept of narrative importance, the production tools and the end-user interface designed to deliver it. Response to these tools from target users and production staf are discussed as well as ongoing work.</p>
      </abstract>
      <kwd-group>
        <kwd>accessibility</kwd>
        <kwd>broadcast audio</kwd>
        <kwd>hearing impaired</kwd>
        <kwd>object-based audio</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 INTRODUCTION</title>
      <p>
        Hearing impairment is estimated to afect around 1 in 6
people in the UK [
        <xref ref-type="bibr" rid="ref1">Action on Hearing Loss 2015</xref>
        ] with similar
numbers reflected throughout Europe and North America
[Agrawal et al. 2008; Shield 2006]. The majority of hearing
impaired people have mild or moderate hearing loss
        <xref ref-type="bibr" rid="ref19 ref3 ref6">(usually
defined as between 20dB and 70dB loss in their better
hearing ear [British Society of Audiology 2011])</xref>
        . Such listeners
still make use of the soundtrack while watching television
broadcasts and the majority consider that hearing well when
watching TV/video was ‘very important’ or ‘extremely
important’
        <xref ref-type="bibr" rid="ref5">(84% in a 2018 study)</xref>
        [Strelcyk and Singh 2018]. An
ageing demographic suggests that the proportion of people
with hearing loss is likely to rise significantly [Ofice fo
        <xref ref-type="bibr" rid="ref4">r
National Statistics 2015</xref>
        ; Roth et al. 2011].
      </p>
      <p>
        Media coverage of inaudible television speech in broadcast
has become commonplace over recent years [Fullerton 2017;
Plunke
        <xref ref-type="bibr" rid="ref27">tt 2016</xref>
        ], even being debated in the UK Parliament’s
upper
        <xref ref-type="bibr" rid="ref10">house [Hansard 2017</xref>
        ]. A survey by the BBC found
that 60% of respondents had dificulty hearing what was said
in broadcasts at some point
        <xref ref-type="bibr" rid="ref8">during a single evening [Cohen
2011</xref>
        ]. An earlier study by the Royal National Institute for the
Deaf reported that 87% of hard of hearing viewers struggled
to understand speech on television [Royal National
Institute for Deaf People 2008]. This coupled with rising rates
of hearing loss makes addressing the challenge of broadcast
accessibility increasingly important.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>PREVIOUS ACCESSIBLE BROADCAST AUDIO</title>
    </sec>
    <sec id="sec-3">
      <title>RESEARCH</title>
      <p>Early accessible audio research has largely approached the
challenge as a speech in noise problem; speech is seen as
desirable, non-speech is seen as ‘background’ that can mask
speech. In conventional (channel-based) broadcast original
‘clean’ speech, separate from the rest of the sound, is usually
unavailable so creating accessible broadcast audio has often
been tackled using speech enhancement techniques which
can add unwanted artifacts to the speech [Armstrong 2011].</p>
      <p>
        More recent object-based audio (OBA) formats that are
beginning to be used in broadcast have the potential to mitigate
this problem considerably. Using OBA it is possible to send
individual elements of the broadcast sound scene as
independent audio objects with accompanying descriptive metadata
which describes how the audio should be replayed. User
personalization of OBA has been proposed for personalization
of sports broa
        <xref ref-type="bibr" rid="ref8">dcast [Mann et al. 2013</xref>
        ; Mehta and Ziegler
2014; Meltzer et al. 2014] and alternate language p
        <xref ref-type="bibr" rid="ref4">rovision
[Bleidt et al. 2015</xref>
        ; Bleisteiner et al. 2018; Brun 2018] and now
a key driver for personalization is accessibility [Shirley and
Ward 2019; Ward and Shirley 2019; Ward et al. 2018]. Some
research has started utilizing OBA to present accessible
audio which includes the ability to adjust the balance between
‘foreground’ and ‘background’ sound [Bleisteiner et al. 2018;
Wal
        <xref ref-type="bibr" rid="ref27">ton et al. 2016</xref>
        , 2018] and an automated intelligibility
control of OBA using measurable objective models of speech
intelligibility has been proposed [Tang et al. 2018].
      </p>
    </sec>
    <sec id="sec-4">
      <title>Comprehension or Intelligibility?</title>
      <p>
        Research on accessible audio for broadcast has largely
focused on intelligibility of speech, usually defined in terms
of number of wo
        <xref ref-type="bibr" rid="ref4">rds recognized [Fontan et al. 2015</xref>
        ]. The
assumption has been that all non-speech sounds are
potentially maskers of speech and are not useful. However
understanding of broadcast programmes is dependent on more
than simple understanding of speech. Other sounds, such
as sound efects and music, can have important roles in
signalling, continuity and scene setting. Consider the case of
a live broadcast of a football game: speech (commentary) is
undoubtedly important however the sound of the referee
whistle is also important to understand the game’s narrative.
Imagine the movie ‘Jaws’ without its iconic musical score
all of the tension is lost and a scene becomes simply a person
swimming in the sea.
      </p>
      <p>
        Non-speech sound efects have also been shown to assist
intelligibility of speech. Recent researc
        <xref ref-type="bibr" rid="ref10">h [Ward et al. 2017</xref>
        ]
found that the inclusion of sound efects related to keywords
could improve word recognition rates in noise from 35.8%
without sound efects to 60.7% with sound efects for normal
hearing listeners. This was also shown to be the case for
some hard of hearing listeners although the efect was highly
dependent on hearing acuity. People with mild hearing loss
(20dB to 40dB loss in best ear) had results similar to people
with normal hearing.
      </p>
      <p>
        It can be argued that accessible audio should focus more
on comprehension of broadcast instead of measures of
intelligibility. This was the focus of research carried out by Shirley
et al in the FascinatE project which utilized a prototype OBA
format to allow user-control of levels of separate objects for
commentary, crowd and on-pitch sounds in a football game
[Shi
        <xref ref-type="bibr" rid="ref4">rley and Oldfield 2015</xref>
        ]. Further work using the MDA
(Multi-Dimensional Audio) OBA format presented drama
and sports media clips to hard of hearing participants with
separate controls for categories of speech, music, background
and foreground audio objects [S
        <xref ref-type="bibr" rid="ref10">hirley et al. 2017</xref>
        ]. Whereas
speech was always set to a higher level than other objects a
significant number of participants stated that the foreground
objects helped to understand the narrative of the media
content.
3
      </p>
    </sec>
    <sec id="sec-5">
      <title>NARRATIVE IMPORTANCE AND ITS</title>
    </sec>
    <sec id="sec-6">
      <title>IMPLEMENTATION</title>
      <p>Earlier OBA work showed promise in moving beyond a
simple binary (speech vs. non-speech) approach to accessible
audio. However the user interface could be considered overly
complex; few people would wish to adjust 5 diferent level
controls for each programme as was the case in the research
discussed earlier using the MDA audio format (1 gain control
for each of 4 categories and an overall volume). The research
presented in this paper utilizes hierarchical narrative
importance categorization and presents the user with a single
control which has diferent efects on the diferent categories
of sounds.</p>
      <p>Audio objects in object-based audio have accompanying
metadata. For personalization in accessibility, an additional
ifeld of metadata describing narrative importance has been
added for each object. Producers tag objects during
postproduction with appropriate levels of narrative importance
based on the importance of the sound in conveying the
narrative of the programme. The metadata is retained in the
OBA audio stream to the user. The user interface, a single
control, is then used to manipulate levels of objects within
each category based on user preference set by the control.
The remainder of this paper describes the work to implement
and evaluate tools which enable this approach to accessible
audio.</p>
    </sec>
    <sec id="sec-7">
      <title>User Interface</title>
      <p>The control acts as an audio complexity control for the media
and can be seen in Figure 1. At maximum setting the audio
mix is as the producer intended, fully immersive and with all
objects at default, producer-set levels. At minimal setting the
audio mix contains only elements that are essential or highly
important to understanding the narrative. Consequently the
mix is less complex. Each object category stem is
continuously variable throughout the full range of the control with
difering attenuation factors applied based on the narrative
importance of that object-category.</p>
    </sec>
    <sec id="sec-8">
      <title>Production Tools</title>
      <p>During post-production audio objects must be tagged by the
producer, dubbing mixer or sound designer with metadata
appropriate to the function of the audio object in
understanding the narrative. The producer-interface has been designed
for familiarity and simplicity, in order to avoid introducing
additional time, and therefore cost, to media production, and
so that new skills do not need to be learned. The process
of assigning an object, or an audio track, to a category is
therefore made analogous to assigning the audio track to a
mix bus or auxiliary track. As one dubbing mixer described
it,“instead of selecting L/R on the channel I just press 0, 1, 2 or
3 - it’s not going to add any more time to my work.”</p>
      <p>The efect has been implemented as a VST plugin for Linux
and Windows as well as an AAX Pro Tools plugin. It was
developed using the Faust [Orlarey et al. 2009] signal
processing programming language and faust2juce [Letz et al.
2015]. Figure 2 shows the DSP block diagram of the efect
and figure 3 a screen-grab of its operation. It consists of a
single dial control (as in the user interface) which ranges
from 0 to 100 (101 points) with 0 producing the mix with
narrative elements only, and 100 the full immersive mix. The
dial controls the gain levels for each of the 4 stereo input
tracks xi,c and mixes them to a single stereo output track
y. Additionally, values of the dial are smoothed using an
exponential envelope in order to avoid ‘scratches’ due to
processing rate being much higher than the control rate.</p>
    </sec>
    <sec id="sec-9">
      <title>User and Producer Assessment</title>
      <p>Demonstrations, focus groups and semi-structured
interviews have been carried out with hard of hearing people
to identify their response to the prototype user-interface
and to identify improvements that can be made [Ward et al.
2018]. Responses indicated a very positive impression of the
interface and use of narrative importance and enthusiasm for
the additional control it gave them over the audio. Most
participants considered that retaining some of the background
sounds, even in the least complex personalized mix, helped
to retain what they described as the content’s ‘depth’ and
‘colour’. Several participants expressed a wish for greater
control - for a greater range so that audio-objects in the less
important audio categories could be reduced still further
where desired. Mix-sessions were undertaken with 2
professional sound mixers to more clearly understand the process
by which narrative importance metadata could be generated
Complexity</p>
      <p>−0.01
and how choices were made. Both found the concept easy to
understand with one commenting, ‘that’s how I decide how
loud and element should be in the mix anyway - how important
it is to the story.’ The sessions produced useful pointers as
to how production workflows incorporating narrative
importance metadata could be made more eficient. Neither
expressed any wariness about their mix being altered by the
viewer at home although it is very likely that this sentiment
may not extend to premium movie production.
4</p>
    </sec>
    <sec id="sec-10">
      <title>DISCUSSION</title>
      <p>It is clear that for some hard of hearing people accessibility is
more than simply about speech enhancement and that other
narratively important elements of the soundtrack can also be
important to comprehension. Demonstrations of prototypes
of the user-control with production professionals have
indicated that utilizing narrative importance metadata in this
way need not significantly increase production workload.
The next challenge is to identify how narrative importance
metadata could be implemented in broadcast and to broadly
assess its impact on user experience. To this end some of the
current work is focusing on implementation in the Standard
Media Player (SMP) in collaboration with BBC R&amp;D. The
SMP is a responsive accessible player which delivers BBC
media content across News, Sport, Weather, iPlayer, Radio
and live events. Other ongoing work is looking at how
different producers interpret narrative importance of sounds in
TV broadcast and how consistent this interpretation is for
individual sound mixers.
5</p>
    </sec>
    <sec id="sec-11">
      <title>CONCLUSIONS</title>
      <p>This paper has introduced the idea of utilizing narrative
importance metadata into object-based accessible audio for
people with hearing impairments. It has documented
development of both user interface and production tools
consisting of VST plugins that can be used in many common
digital audio workstations and outlined next steps for further
development and implementation.
6</p>
    </sec>
    <sec id="sec-12">
      <title>ACKNOWLEDGMENTS</title>
      <p>This work was supported by the EPSRC Programme Grant
S3A: Future Spatial Audio for an Immersive Listener
Experience at Home (EP/L000539/1) and the BBC as part of the
BBC Audio Research Partnership. Lauren Ward is funded by
the General Sir John Monash Foundation.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <source>Action on Hearing Loss</source>
          .
          <year>2015</year>
          .
          <article-title>Hearing Matters Report</article-title>
          . https://www. actiononhearingloss.org.uk/how
          <article-title>-we-help/information-and-resources/ publications/research-reports/hearing-matters-report/</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Y.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Platz</surname>
          </string-name>
          , and
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Niparko</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Prevalence of hearing loss and diferences by demographic characteristics among US adults: data from the National Health and Nutrition Examination Survey,</article-title>
          <year>1999</year>
          -
          <fpage>2004</fpage>
          .
          <source>Archives of internal medicine 168</source>
          ,
          <issue>14</issue>
          (
          <year>2008</year>
          ),
          <fpage>1522</fpage>
          -
          <lpage>1530</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Mike</given-names>
            <surname>Armstrong</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Audio processing and speech intelligibility: a literature review</article-title>
          .
          <source>In BBC Research &amp; Development Whitepaper.</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>R.</given-names>
            <surname>Bleidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Borsum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Fuchs</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Weiss</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Object-Based Audio: Opportunities for Improved Listening Experience and Increased Listener Involvement</article-title>
          .
          <source>SMPTE Motion Imaging J</source>
          .
          <volume>124</volume>
          ,
          <issue>5</issue>
          (
          <year>July 2015</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          . https: //doi.org/10.5594/j18579
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>W.</given-names>
            <surname>Bleisteiner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Silzle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Liebl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Warusfel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ragot</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Epain</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>D5.6: Report on Audio subjective tests and User tests</article-title>
          .
          <source>Technical Report. The Orpheus Project</source>
          . https://orpheus-audio.eu/wpcontent/uploads/2018/07/orpheus-d5.
          <article-title>6_report-on-audio-subjectiveand-user-tests_v1.3</article-title>
          .pdf.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <source>British Society of Audiology</source>
          .
          <year>2011</year>
          .
          <article-title>Recommended procedure: Pure-tone air-conduction and bone-conduction threshold audiometry with and without masking</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Rupert</given-names>
            <surname>Brun</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Successful Demonstration of Interactive Audio Streaming Using MPEG-H Audio at Norwegian Broadcaster NRK</article-title>
          . /www.audioblog. iis.fraunhofer.com/mpegh-nrk/
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>D.</given-names>
            <surname>Cohen</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Sound Matters, BBC College of Production</article-title>
          . http://www. bbc.co.uk/academy/production/article/art20130702112136134
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Lionel</given-names>
            <surname>Fontan</surname>
          </string-name>
          , Julien Tardieu, Pascal Gaillard, Virginie Woisard, and
          <string-name>
            <given-names>Robert</given-names>
            <surname>Ruiz</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Relationship between speech intelligibility and speech comprehension in babble noise</article-title>
          .
          <source>J. Speech, Lang. Hear. Res</source>
          .
          <volume>58</volume>
          ,
          <issue>3</issue>
          (
          <year>2015</year>
          ),
          <fpage>977</fpage>
          -
          <lpage>986</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Fullerton</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>BBC drama SS-GB criticised for “mumbling" and bad sound quality in ifrst episode</article-title>
          . http://www.radiotimes.com/news/2017-02-26/ bbc-drama
          <article-title>-ss-gb-criticised-for-mumbling-and-bad-sound-quality-in-first-episode</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Hansard. 2017. Television</given-names>
            <surname>Broadcasts</surname>
          </string-name>
          : Audibility. https://hansard.parliament.uk/lords/2017-04-04/ debates/F84C55A0-3D8B
          <string-name>
            <surname>-</surname>
          </string-name>
          41F7
          <string-name>
            <surname>-</surname>
          </string-name>
          A19C-CC216F8C7B0B/ TelevisionBroadcastsAudibility
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Stephane</given-names>
            <surname>Letz</surname>
          </string-name>
          , Sarah Denoux, Yann Orlarey, and
          <string-name>
            <given-names>Dominique</given-names>
            <surname>Fober</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Faust audio DSP language in the Web</article-title>
          .
          <source>In Proceedings of the Linux Audio Conference (LAC-15)</source>
          , Mainz, Germany.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>M.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. W. P.</given-names>
            <surname>Churnside</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bonney</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Melchior</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Object-based audio applied to football broadcasts</article-title>
          .
          <source>In ACM international workshop on Immersive media experiences. ACM</source>
          ,
          <volume>13</volume>
          -
          <fpage>16</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>Sripal</given-names>
            <surname>Mehta</surname>
          </string-name>
          and
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Ziegler</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Personalized and immersive broadcast audio</article-title>
          .
          <source>In International Broadcast Convention. IET.</source>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <given-names>Stefan</given-names>
            <surname>Meltzer</surname>
          </string-name>
          , Max Neuendorf, Deep Sen, and
          <string-name>
            <given-names>Peter</given-names>
            <surname>Jax</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>MPEG-H 3D Audio-The Next Generation Audio System</article-title>
          . In International Broadcast Convention. IET.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Ofice</surname>
          </string-name>
          for National Statistics. Oct,
          <year>2015</year>
          .
          <article-title>National Population Projections: 2014-based Statistical Bulletin</article-title>
          . https://www.ons.gov. uk/peoplepopulationandcommunity/populationandmigration/ populationprojections/bulletins/nationalpopulationprojections/ 2015-10-29#older-people
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <given-names>Yann</given-names>
            <surname>Orlarey</surname>
          </string-name>
          , Dominique Fober, and
          <string-name>
            <given-names>Stéphane</given-names>
            <surname>Letz</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>FAUST: an eficient functional approach to DSP programming</article-title>
          .
          <source>New Computational Paradigms for Computer Music</source>
          <volume>290</volume>
          (
          <year>2009</year>
          ),
          <fpage>14</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>J.</given-names>
            <surname>Plunkett</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Heard this before? BBC chief speaks out over Happy Valley mumbling</article-title>
          . https://www.theguardian.com/media/2016/apr/08/ bbc-happy
          <article-title>-valley-mumbling-jamaica-inn-sarah-lancashire</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Niklaus</surname>
          </string-name>
          <string-name>
            <surname>Roth</surname>
          </string-name>
          , Dirk Hanebuth, and
          <string-name>
            <given-names>Rudolf</given-names>
            <surname>Probst</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Prevalence of age-related hearing loss in Europe: a review</article-title>
          .
          <source>Eur. Arch. Otorhinolaryngol</source>
          .
          <volume>268</volume>
          ,
          <issue>8</issue>
          (
          <year>2011</year>
          ),
          <fpage>1101</fpage>
          -
          <lpage>1107</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Royal</given-names>
            <surname>National Institute for Deaf People</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <source>Annual Survey Report</source>
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>Bridget</given-names>
            <surname>Shield</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Evaluation of the social and economic costs of hearing impairment</article-title>
          . https://www.hear-it.org/sites/default/files/hear-it% 20documents/Hear%20It%
          <fpage>20Report</fpage>
          %
          <fpage>20October</fpage>
          %
          <fpage>202006</fpage>
          .pdf
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <given-names>B.</given-names>
            <surname>Shirley</surname>
          </string-name>
          and
          <string-name>
            <given-names>R.</given-names>
            <surname>Oldfield</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Clean audio for TV broadcast: An objectbased approach for hearing-impaired viewers</article-title>
          .
          <source>J. Audio Eng Soc</source>
          .
          <volume>63</volume>
          ,
          <issue>4</issue>
          (
          <year>2015</year>
          ),
          <fpage>245</fpage>
          -
          <lpage>256</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <given-names>Ben</given-names>
            <surname>Shirley</surname>
          </string-name>
          and
          <string-name>
            <given-names>Lauren</given-names>
            <surname>Ward</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Intelligibility vs. Comprehension: Understanding Quality of Accessible Next-generation Audio Broadcast. Universal Access in the Information Society Special Issue on“Quality of Media Accessibility Products</article-title>
          and Services” (
          <year>2019</year>
          ). [Accepted; In press].
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <given-names>Ben</given-names>
            <surname>Guy</surname>
          </string-name>
          <string-name>
            <surname>Shirley</surname>
          </string-name>
          , Melissa Meadows, Fadi Malak, James Stephen Woodcock, and
          <string-name>
            <given-names>Ash</given-names>
            <surname>Tidball</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Personalized object-based audio for hearing impaired TV viewers</article-title>
          .
          <source>J. Audio Eng Soc</source>
          .
          <volume>65</volume>
          ,
          <issue>4</issue>
          (
          <year>2017</year>
          ),
          <fpage>293</fpage>
          -
          <lpage>303</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <given-names>Olaf</given-names>
            <surname>Strelcyk</surname>
          </string-name>
          and
          <string-name>
            <given-names>Gurjit</given-names>
            <surname>Singh</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>TV listening and hearing aids</article-title>
          .
          <source>PloS one 13</source>
          ,
          <issue>6</issue>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. M.</given-names>
            <surname>Fazenda</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T. J.</given-names>
            <surname>Cox</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Automatic speech-to-background ratio selection to maintain speech intelligibility in broadcasts using an objective intelligibility metric</article-title>
          .
          <source>Appl. Sci. 8</source>
          ,
          <issue>1</issue>
          (
          <year>2018</year>
          ),
          <fpage>59</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <given-names>T.</given-names>
            <surname>Walton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Evans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kirk</surname>
          </string-name>
          , and
          <string-name>
            <given-names>F.</given-names>
            <surname>Melchior</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Does Environmental Noise Influence Preference of Background-Foreground Audio Balance?</article-title>
          .
          <source>In 141st Audio Eng. Soc. Convention. AES</source>
          , Los Angeles, U.S.A.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <given-names>Tim</given-names>
            <surname>Walton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Michael</given-names>
            <surname>Evans</surname>
          </string-name>
          , David Kirk,
          <string-name>
            <given-names>and Frank</given-names>
            <surname>Melchior</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Exploring object-based content adaptation for mobile audio</article-title>
          .
          <source>Personal Ubiquitous Comput</source>
          .
          <article-title>(</article-title>
          <year>2018</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <given-names>Lauren</given-names>
            <surname>Ward</surname>
          </string-name>
          and
          <string-name>
            <given-names>Ben</given-names>
            <surname>Shirley</surname>
          </string-name>
          .
          <year>2019</year>
          .
          <article-title>Personalization in Object-based Audio for Accessibility: A Review of Advancements for Hearing Impaired Listeners</article-title>
          . Audio Engineering Society Journal Special Issue on Object-Based
          <string-name>
            <surname>Audio</surname>
          </string-name>
          (
          <year>2019</year>
          ). [Under Review].
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <given-names>Lauren</given-names>
            <surname>Ward</surname>
          </string-name>
          , Ben Shirley, and
          <string-name>
            <given-names>Jon</given-names>
            <surname>Francombe</surname>
          </string-name>
          .
          <year>2018</year>
          .
          <article-title>Accessible objectbased audio using hierarchical narrative importance metadata</article-title>
          .
          <source>In Audio Engineering Society Convention</source>
          <volume>145</volume>
          . Audio Engineering Society.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <given-names>Lauren</given-names>
            <surname>Ward</surname>
          </string-name>
          , Ben Shirley,
          <string-name>
            <given-names>Yan</given-names>
            <surname>Tang</surname>
          </string-name>
          , and
          <string-name>
            <given-names>William</given-names>
            <surname>Davies</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>The efect of situation-specific acoustic cues on speech intelligibility in noise</article-title>
          .
          <source>In Interspeech</source>
          <year>2017</year>
          . ISCA, Stockholm, Sweden,
          <fpage>2958</fpage>
          -
          <lpage>2962</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>