=Paper= {{Paper |id=Vol-3359/paper18 |storemode=property |title=The Impact of Salient Musical Features in a Hybrid Recommendation System for a Sound Library |pdfUrl=https://ceur-ws.org/Vol-3359/paper18.pdf |volume=Vol-3359 |authors=Jason Brent Smith,Ashvala Vinay,Jason Freeman |dblpUrl=https://dblp.org/rec/conf/iui/SmithV023 }} ==The Impact of Salient Musical Features in a Hybrid Recommendation System for a Sound Library== https://ceur-ws.org/Vol-3359/paper18.pdf
The Impact of Salient Musical Features in a Hybrid
Recommendation System for a Sound Library
Jason Brent Smith1 , Ashvala Vinay1 and Jason Freeman1
1
 Georgia Tech Center for Music Technology
840 McMillan Street NW, Atlanta, Georgia, USA, 30308


                        Abstract
                        EarSketch is an online learning environment that teaches coding and music concepts through the computational manipulation
                        of sounds selected from a large sound library. It features sound recommendations based on acoustic similarity and co-usage
                        with a user’s current sound selection in order to encourage exploration of the library. However, students have reported that
                        the recommended sounds do not complement their current projects in terms of two areas: musical key and rhythm. We
                        aim to improve the relevance of these recommendations through the inclusion of these two musically related features. This
                        paper describes the addition of key signature and beat extraction to the EarSketch sound recommendation model in order to
                        improve the musical compatibility of the recommendations with the sounds in a user’s project. Additionally, we present an
                        analysis of the effects of these new recommendation strategies on user exploration and usage of the recommended sounds.
                        The results of this analysis suggest that the addition of explicitly musically-relevant attributes increases the coverage of the
                        sound library among sound recommendations as well as the sounds selected by users. It reflects the importance of including
                        multiple musical attributes when building recommendation systems for creative and open-ended musical systems.


1. Introduction                                             filter by artist, genre, instrument, or key signature, and
                                                            mark them as favorites for future use (see Fig 1) and can
EarSketch [1] is a computational music remixing environ- preview or copy these sounds into their code as constants.
ment designed to teach music and computing concepts
through the process of writing code to creatively manip-
ulate audio loops. It is a web application that contains
a code editor for students to write Python or JavaScript
code using a custom API, and a Digital Audio Worksta-
tion for them to view and listen to the musical output
produced by their code.
   Previous analysis of EarSketch users revealed that a
sense of creative ownership and expression for their work
has been linked to intentions to persist in computer sci-
ence education [2]. To this end, EarSketch was designed
with the goal of being authentic to industry tools in terms
of music production, musical content, and computing lan-
guages. It achieves this with the design of its interface
and API as well as with the inclusion of a large sound
library for students to explore and find sounds that are
personally expressive and meaningful to them.
   EarSketch contains a library of over 4,500 sounds pro-
duced by professional artists such as sound designer
Richard Devine, hip-hop producer/DJ Young Guru, and
additional stems from popular musicians such as Alicia
Keys, Ciara, Common, Dakota Bear, Irizzary y Caraballo,
Jayli Wolf, Khalid, Milk + Sizz, Pharrell Williams, and
Samian. Users are able to search for sounds by name,
Joint Proceedings of the ACM IUI Workshops 2023, March 2023,                                      Figure 1: View of EarSketch Sound Browser interface (top),
Sydney, Australia                                                                                 with example recommendations (bottom).
$ jsmith775@gatech.edu (J. B. Smith); ashvala@gatech.edu
(A. Vinay); jason.freeman@gatech.edu (J. Freeman)
 0000-0002-7075-6132 (J. B. Smith); 0000-0002-2487-2052                                             A previous analysis of 20,000 user-created scripts showed
(A. Vinay); 0000-0003-3827-1060 (J. Freeman)
    © 2023 Copyright © 2023 for this paper by its authors. Use permitted under Creative Commons   that fewer than 200 library sounds were used in over 1%
    License Attribution 4.0 International (CC BY 4.0).
    CEUR Workshop Proceedings (CEUR-WS.org)                                                       of scripts and under 20 sounds were used in over 10% of
Figure 2: View of the EarSketch interface, with Sound Browser (left), Digital Audio Workstation (top), and Code Editor
(bottom), and Curriculum (right).



scripts. It was hypothesized that this was due to difficulty    a unique project than when matching sounds to others
in navigating the sound browser, as users reported that         out of the context of EarSketch [6].
it was hard to discover groups of sounds relevant to their         While the initial recommendation system, a hybrid
current work. In order to address this under-utilization of     model using collaborative filtering and content-based
the sound library and to promote compositional diversity        similarity metrics, improved the number of sounds ex-
among its users’ projects, a recommendation system was          plored by users, users have reported a lack of musical co-
added to EarSketch [3].                                         hesion between sounds after they have already included
   Diversity and coverage, measures of how different a          contrasting elements in a project, as well as a lack of
set of recommendations are from each other and how              sound suggestions that facilitated specific compositional
much of a set of available options is being recommended,        ideas such as creating a new section of a song. This work
are common design goal of recommendation systems                aims to improve the recommendation system’s impact on
[4]. Recommendation systems to present diverse com-             sound exploration and usage by adding two additional
positional material are prevalent in music production           musical features as inputs: key signature and beat sim-
platforms with which EarSketch aligns its design goals.         ilarity. These features are musically motivated in that,
The EarSketch sound recommendation system was de-               unlike the existing system’s use of Short-time Frequency
signed to assist in the process of navigating the sound         Transform, they use explicit human-understandable la-
library by presenting relevant, novel sounds for users          bels grounded in music theory. Although EarSketch does
to include in their code. By giving users more easily-          not include western music notation by design, each tonal
accessible sound options that match the content of their        sound was originally composed with a major or minor
in-progress compositions, the system aims to improve            key signature in mind. As such, by adding explicit key
the variety of sounds that users preview and copy into          labels [7] to sounds, the overall key of a user’s current
their scripts. It uses collaborative filtering [5] and acous-   project can be estimated and sounds with that key can
tic similarity metrics to minimize or maximize co-usage         have their recommendation scores increased. In addition
and similarity scores in various combinations to generate       to tonal similarity, the system can prioritize recommen-
recommendation scores, which can be used for different          dations that are rhythmically consistent with a user’s
recommendations such as “Songs that Fit Your Script”            project [8]. Beat detection is performed by generating a
or “Others Like You Used These”. Combining multiple             numerical vector representing the rhythm of each sound
recommendation strategies allowed for increased user            in the sound library, then computing the distance be-
exploration and sound usage and that users preferred            tween two sounds’ vectors and factoring it into their
different types of recommendations when freely creating         pairwise recommendation scores.
   By adding the above features, we aim to answer the          compute the key signatures for the dataset where key
following question:                                            signatures were appropriate.1
     • How does the addition of salient musical features          Beats were extracted using librosa’s [12] beat track
       in the EarSketch sound recommendation system            prediction method. The method takes an audio signal
       impact the diversity of sounds recommended and          and predicts its tempo and beat track. Details of the
       used in student projects?                               method can be found in Daniel Ellis’ paper [13], which
                                                               is the implementation used by librosa. The beat track
   The contributions of this work include the augmenta-        prediction provided by librosa is a series of timestamps
tion of a hybrid recommendation system, combining col-         indicating where a beat might be. We take these time
laborative filtering with multiple aspects of feature-based    stamps and construct an audio signal that is a click at
audio similarity, and the evaluation of sound recommen-        those time stamps. For the sake of computational and
dations in a creative, open-ended task. The rest of this       space efficiency when computing scores, we downsam-
paper will detail the process of adding the musically moti-    pled the signal from 44100 Hz to 100 Hz.
vated features of key signature and beat similarity to the        In the paper detailing the implementation of the beat
EarSketch recommendation system (the dataset, the key          tracker, it is shown that the dynamic programming ap-
signature and beat similarity extraction, and how they         proach is capable of achieving 93.4% accuracy on the
were incorporated into the recommender), followed by           MIREX beat tracking dataset [13]. Since we did not have
the methodology and analysis of an evaluation of these         ground truth annotations, we manually verified the beat
recommendations on aggregate statistics of users on the        predictions internally by using generated click tracks on
EarSketch website.                                             a random subset of the sound library with an informal
                                                               subjective evaluation. Using 5 sets of 16 sounds at a time,
                                                               testers from the EarSketch development team rated the
2. Implementation                                              implementation as appropriately matching their percep-
The EarSketch web client continuously monitors the             tion for each example.
sounds included in a user’s project as they edit their
code. Once a change is detected, the recommendation            2.2. Recommendations
system generates a set of recommendations using the
newly stored list of sounds as input [6]. The output is        The previous algorithm to recommend sounds to users
presented to users as a list of recommendations in the         was described in [3]. In short, sounds in the library were
sound browser (Figure 1). This section will discuss the        assigned a score 𝑆
implementations of key signature estimation and beat                              −1        −1
similarity calculation, as well as their addition to the ex-                 𝑆 = 𝒟𝑆𝑇 𝐹 𝑇 + 𝒟𝑀 𝐹 𝐶𝐶 + 𝒰                   (1)
isting EarSketch recommendation algorithm at the time          where 𝒟𝑆𝑇 𝐹 𝑇 and 𝒟𝑀 𝐹 𝐶𝐶 are acoustic feature dis-
of generation.                                              tances between a given sound and every sound in the
                                                            library and 𝒰 is the co-usage score, i.e, a score indicating
2.1. Key signature and beat extraction                      how often two sounds were used together.
                                                               In order to add key signatures into our algorithm, we
In order to extract key signatures for the clips in the compute the key signature of the project 𝒦proj as the
sound library, we used Essentia [9], a popular software most frequent key signature label across all sound clips in
package for music information retrieval. It implements a project. For a given sound clip S and its corresponding
several key-profiles to estimate the key signature for a key signature 𝒦S we compute it’s key signature score 𝒦
given sound such as “edmm” [10], a profile that is gener- as
ally suited to estimating key signatures from electronic             {︃
music and “braw” [11], a more general key signature                    1, if 𝒦S ∈ {(𝒦proj | relative(𝒦proj )}
estimation profile. In addition to the key signature, Es-      𝒦=                                                         (2)
                                                                       0 otherwise
sentia’s key signature estimator also produced a strength
score indicating how strong the presence of an annotated       where 𝒦 is set to 1 if the clip’s key signature matches
key signature is in the sample.                             the project or has a relative major/minor key relationship
   Identifying the best key signature profile in Essentia with the project’s key signature.
was done using the annotated subset of the library de-         To add a beat similarity score, we compute the Ham-
scribed in the section above. For each profile, we com- ming distance [14] between two given beat tracks. This
pared the predicted key signatures for the subset against is denoted as 𝒟−1 . We assume that users might select
                                                                             ℎ𝑎𝑚𝑚
the ground truth annotations. The “edmm” profile stood
out as the best profile since it predicted the largest num-     1
                                                                  We excluded purely percussive sounds and short, single shot
ber of correct annotations. Therefore, it was used to examples - for e.g, snare samples
a set of samples with varying attributes, for example,                                                        Distribution of unique recommendations per sound
                                                                                           0.0008
genre or instrumentation, but happen to have a consis-                                                                                                                          After update
                                                                                           0.0007                                                                               Before update
tent rhythmic structure. Hamming distances have been
                                                                                           0.0006
shown by Toussiant et al[? ] as a good measure of rhyth-




                                                                       Density of events
                                                                                           0.0005
mic similarity. Given that EarSketch time-stretches sam-
                                                                                           0.0004
ples to match a specified tempo, we wanted to choose
                                                                                           0.0003
a similarity measure that was tempo invariant and pri-
                                                                                           0.0002
marily focused on the difference in how the rhythms are
                                                                                           0.0001
performed directly.
                                                                                           0.0000
   Adding key and beat information to the system was                                                0        2000   4000        6000        8000         10000    12000    14000       16000
                                                                                                                            Frequency of unique recommendations
done as an addition to the score 𝑆 described in Equation 1:
                                                              Figure 3: The distribution of how frequently a sound was
                                                              recommended across the entire library before and after the
          −1        −1                −1
     𝑆 = 𝒟𝑆𝑇 𝐹 𝑇 + 𝒟𝑀 𝐹 𝐶𝐶 + 𝒰 + 𝒦 + 𝒟hamm             (3)    addition of key signature and beat similarity to the recommen-
                                                              dation system. The figure is scaled by the number of unique
   Like the co-usage and acoustic similarity scores in the    user sessions in the two time periods.
initial version of the recommendation system [6], the
key signature estimation and beat extraction processes                                                                Distribution of Unique Usages
are performed offline for the whole sound library. Their                          0.06                                                                                      After update
results are deployed to the EarSketch web client to be re-                        0.05
                                                                                                                                                                            Before update

trieved for individual sound-sound pairs and used in real-    Density of events   0.04
time recommendations. This is done to allow for faster
                                                                                  0.03
recommendations without the requirement for heavy
                                                                                  0.02
audio processing while users are editing a project.
                                                                                  0.01

                                                                                  0.00
                                                                                             0          50            100                150               200            250               300
3. Results                                                                                                                      Unique Usage Frequency



As with a previous evaluation of the recommendation sys-      Figure 4: The distribution of how recommended sounds
tem [6], the impact of this recommender was measured          were added to projects before and after the addition of key
through statistical analysis of the sounds recommended        signature and beat similarity to the recommendation system.
and added to projects by EarSketch users. The key signa-      The figure is scaled by the number of unique user sessions in
                                                              the two time periods.
ture and beat similarity recommendation changes were
added to the EarSketch website in October 2022. Using
an analytics engine, the actions of 103,828 users before
the update were recorded between July and September           recommended. The higher average and lower skew of
2022, and the actions of 133,349 users after the update       the distribution after the addition of key signature and
were recorded between October and December 2022.              beat similarity indicates that more sounds from the EarS-
   During each session for a given user, each unique rec-     ketch library are more likely to be recommended. In the
ommendation made is stored as a separate event rec-           period prior to the update, a given sound would be rec-
ommendation. A separate event, recommendationUsed,            ommended an average of 1790 times (1.72% of sessions).
is stored when the student uses a recommended sound           Comparatively, following the update a given sound would
in their project by copying it directly from the sound        be recommended 3240 times (2.42% of sessions). Using
browser interface or by writing the name of the sound         a two-sample t-test, we note that the difference in rec-
into their code. The density of sounds being recom-           ommendation frequency across the entire library was
mended is a function of recommendationUsed / recom-           statistically significant (𝑝 < 0.05).
mendation for each individual sound constant. We deter-          When comparing the usage of recommendations across
mine coverage of the sound library by the distribution        both periods, we measured the frequency distribution of
of unique recommendations made, as well as the rate           recommendationUsed events, or the unique instances of
at which these recommendations are used in student            previously recommended sounds being used in projects
projects.                                                     (see Fig 4). A two sample t-test shows that there is a sta-
   The analysis of this data suggests that the inclusion      tistically significant increase in recommendation usage
of musically driven features further improves the diver-      frequency following the update (𝑝 < 0.05). On aver-
sity of sounds suggested by our hybrid recommenda-            age, a recommended sound was used 0.94% of times it
tion system. Fig 3 depicts the frequency density of each      was recommended following the update compared to the
sound based on the number of times that sound was             average of 0.73% prior to the update.
   We also observed that the inclusion of rhythmic fea-      how recommendation behavior influences the song cre-
tures coincides with a noticeable uptick in the usage of     ation process for students. By rating the understanding
percussive loops with a larger proportion of used sounds     of and satisfaction with recommendations by users with
from a recommendation being percussive. We investi-          and without the musical features in a controlled setting,
gated the top 10 sounds being recommended and used           we can determine how effective these features are and
during both periods. We found that 6 of the 10 most          what visual design changes are necessary to enhance the
frequently used recommendations after the update are         effectiveness of musically-informed recommendations.
categorized as purely percussive sounds. In the period          In conclusion, this analysis of the impact that salient
prior to the update, there was no majority among the         musical features have on EarSketch users reveals multiple
instruments in the most used recommendations.                insights for the design of recommendation systems and
                                                             other creative systems. The use of recommendation den-
                                                             sity to compare groups shows how artifact analysis can
4. Discussion and Future Work                                represent trends in user interaction with a creative musi-
                                                             cal assistant, even at its most simple form. The significant
We implemented key signature and beat extraction in the
                                                             change in the density of unique sound recommendations
EarSketch sound recommendation system, to improve
                                                             shows the effectiveness of multimodal domain knowl-
the diversity and coverage of sounds that are being rec-
                                                             edge on recommendation generation. As the EarSketch
ommended to users and to make more musically relevant
                                                             recommendation system either minimizes or maximizes
suggestions for a student’s project. We analyzed two
                                                             co-usage scores as well as acoustic similarity [3], the ad-
periods of data above to identify trends in usage before
                                                             dition of features to multiple types of recommendations
and after the addition of these two musical features.
                                                             shows the importance of understanding task specifica-
   In our results, we were able to successfully demon-
                                                             tions when discussing recommendations for a creative
strate that the inclusion of these features improves the
                                                             system.
diversity and coverage of recommended sounds. By com-
paring the distributions of unique recommendations per
sound before and after the change, we found that the 5. Acknowledgments
number of recommended sounds was more evenly dis-
tributed across the sound library after the change. This This material is based upon work supported by the Na-
may be because the algorithm is able to pick up on more tional Science Foundation Award No. 1814083. Any opin-
sounds that are pertinent to a given user’s project more ions, findings, and conclusions or recommendations ex-
frequently. Additionally, there was a statistically sig- pressed in this material are those of the author(s) and
nificant increase in how often students elected to use a do not necessarily reflect the views of the National Sci-
recommendation. This could be attributed to the promi- ence Foundation. EarSketch is available online at https:
nence of beat similarity in the recommendation algorithm //earsketch.gatech.edu
providing sounds that stylistically match a user’s current
sounds and as such present more viable options to try in
a given project.                                             References
   We noticed that there has been a shift in the types
                                                               [1] B. Magerko, J. Freeman, T. Mcklin, M. Reilly, E. Liv-
of recommended sounds that are more frequently used
                                                                   ingston, S. Mccoid, A. Crews-Brown, Earsketch: A
across both periods. Following the introduction of our
                                                                   steam-based approach for underrepresented popu-
updated algorithm, we found that a majority of the most
                                                                   lations in high school computer science education
used recommendations were percussive or primarily rhyth-
                                                                   16 (2016) 1–25. doi:10.1145/2886418.
mic. We believe that this is an artifact of how the key sig-
                                                               [2] T. McKlin, B. Magerko, T. Lee, D. Wanzer, D. Ed-
natures and rhythmic similarities of sounds are weighted
                                                                   wards, J. Freeman, Authenticity and personal cre-
in the recommendation process. We speculate that stu-
                                                                   ativity: How EarSketch affects student persistence,
dents are largely seeking rhythmic sounds at the begin-
                                                                   in: Proceedings of the 49th ACM Technical Sym-
ning stages of the song-creating process. Given that the
                                                                   posium on Computer Science Education, 2018, pp.
weighting for pitched sounds necessitates the existence
                                                                   987–992. doi:10.1145/3159450.3159523.
of a key signature, the recommendation algorithm skews
                                                               [3] J. Smith, D. Weeks, M. Jacob, J. Freeman,
heavily towards rhythmic sounds at the start of a new
                                                                   B. Magerko, Towards a Hybrid Recommendation
project. Additionally, users with developed projects may
                                                                   System for a Sound Library, in: Joint Proceedings
prefer recommendations that do not clash with their cur-
                                                                   of the ACM IUI 2019 Workshops, CEUR-WS, 2019.
rent selections, such as the percussion samples without
                                                               [4] C. C. Aggarwal, Recommender Systems, Springer
a key signature. In order to understand this behavior
                                                                   International Publishing, 2016. URL: http://link.
better, we need a more in-depth user study to understand
     springer.com/10.1007/978-3-319-29659-3. doi:10.                P. Friesch, M. Vollrath, T. Kim, Librosa/Librosa:
     1007/978-3-319-29659-3.                                        0.9.2, 2022. URL: https://doi.org/10.5281/zenodo.
 [5] J. B. Schafer, D. Frankowski, J. Herlocker, S. Sen,            6759664. doi:10.5281/zenodo.6759664.
     Collaborative Filtering Recommender Systems, in:          [13] D. P. W. Ellis,          Beat tracking by dy-
     P. Brusilovsky, A. Kobsa, W. Nejdl (Eds.), The                 namic programming,             Journal of New
     Adaptive Web: Methods and Strategies of Web                    Music Research 36 (2007) 51–60. URL:
     Personalization, Lecture Notes in Computer Sci-                https://doi.org/10.1080/09298210701653344.
     ence, Springer, 2007, pp. 291–324. URL: https://doi.           doi:10.1080/09298210701653344.
     org/10.1007/978-3-540-72079-9_9. doi:10.1007/             [14] G. T. Toussaint, A comparison of rhythmic sim-
     978-3-540-72079-9_9.                                           ilarity measures., in: Proceedings of the 5th In-
 [6] J. Smith, M. Jacob, J. Freeman, B. Magerko, T. Mck-            ternational Conference on Music Information Re-
     lin, Combining collaborative and content filtering             trieval, ISMIR, Barcelona, Spain, 2004. URL: https:
     in a recommendation system for a web-based daw,                //doi.org/10.5281/zenodo.1416812. doi:10.5281/
     in: A. Xambó, S. R. Martín, G. Roma (Eds.), Proceed-           zenodo.1416812.
     ings of the International Web Audio Conference,
     WAC ’19, NTNU, Trondheim, Norway, 2019, pp.
     53–58.
 [7] M.-K. Shan, F.-F. Kuo, M.-F. Chiang, S.-Y. Lee,
     Emotion-based music recommendation by affin-
     ity discovery from film music 36 (2009) 7666–7674.
     doi:10.1016/j.eswa.2008.09.042.
 [8] X. Wang, Y. Wang, Improving content-based and
     hybrid music recommendation using deep learn-
     ing, in: Proceedings of the 22nd ACM Interna-
     tional Conference on Multimedia, 2014, pp. 627–636.
     doi:10.1145/2647868.2654940.
 [9] D. Bogdanov, N. Wack, E. Gómez, S. Gulati, P. Her-
     rera, O. Mayor, G. Roma, J. Salamon, J. Zapata,
     X. Serra, ESSENTIA: An open-source library for
     sound and music analysis, in: Proceedings of
     the 21st ACM International Conference on Mul-
     timedia, MM ’13, Association for Computing Ma-
     chinery, 2013, pp. 855–858. URL: https://doi.org/
     10.1145/2502081.2502229. doi:10.1145/2502081.
     2502229.
[10] Á. Faraldo, E. Gómez, S. Jordà, P. Herrera, Key Es-
     timation in Electronic Dance Music, in: N. Ferro,
     F. Crestani, M.-F. Moens, J. Mothe, F. Silvestri, G. M.
     Di Nunzio, C. Hauff, G. Silvello (Eds.), Advances in
     Information Retrieval, volume 9626 of Lecture Notes
     in Computer Science, Springer International Pub-
     lishing, 2016, pp. 335–347. URL: http://link.springer.
     com/10.1007/978-3-319-30671-1_25. doi:10.1007/
     978-3-319-30671-1_25.
[11] Á. Faraldo, S. Jordà, P. Herrera, A Multi-Profile
     Method for Key Estimation in EDM, in: Pro-
     ceedings of Conference on Semantic Audio, 2017,
     p. 7. URL: https://doi.org/10.5281/zenodo.3855499.
     doi:10.5281/zenodo.3855499.
[12] B. McFee, A. Metsai, M. McVicar, S. Balke, C. Thomé,
     C. Raffel, F. Zalkow, A. Malek, Dana, K. Lee, O. Ni-
     eto, D. Ellis, J. Mason, E. Battenberg, S. Seyfarth,
     R. Yamamoto, viktorandreevichmorozov, K. Choi,
     J. Moore, R. Bittner, S. Hidaka, Z. Wei, nullmighty-
     bofo, A. Weiss, D. Hereñú, F.-R. Stöter, L. Nickel,