<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Overview of the EEG Pilot Subtask at MediaEval 2021: Predicting Media Memorability</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Lorin Sweeney</string-name>
          <email>lorin.sweeney8@mail.dcu.ie</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ana Matran-Fernandez</string-name>
          <email>amatra@essex.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sebastian Halder</string-name>
          <email>s.halder@essex.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alba G. Seco de Herrera</string-name>
          <email>alba.garcia@essex.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alan Smeaton</string-name>
          <email>alan.smeaton@dcu.ie</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Graham Healy</string-name>
          <email>graham.healy@dcu.ie</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Computer Science and Electronic Engineering, University of Essex</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Computing, Dublin City University</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <fpage>13</fpage>
      <lpage>15</lpage>
      <abstract>
        <p>The aim of the Memorability-EEG pilot subtask at MediaEval'2021 is to promote interest in the use of neural signals-either alone or in combination with other data sources-in the context of predicting video memorability by highlighting the utility of EEG data. The dataset created consists of pre-extracted features from EEG recordings of subjects while watching a subset of videos from Predicting Media Memorability subtask 1. This demonstration pilot gives interested researchers a sense of how neural signals can be used without any prior domain knowledge, and enables them to do so in a future memorability task. The dataset can be used to support the exploration of novel machine learning and processing strategies for predicting video memorability, while potentially increasing interdisciplinary interest in the subject of memorability, and opening the door to new combined EEG-computer vision approaches.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>INTRODUCTION AND RELATED WORK</title>
      <p>Even though the nature and constitution of people’s memories
remains elusive, and our understanding of what makes one thing
more/less memorable than another is still nascent, combining
computational (e.g., machine learning) and neurophysiological (e.g.,
electroencephalography; EEG) tools to investigate the mechanisms
(formation and recall) of memory may ofer insights that would be
otherwise unobtainable. While EEG is not a tool that can directly
explain what makes a video more/less memorable, it can help us trim
the umbral undergrowth surrounding the subject, shedding light
and ofering a potential leap forward in our understanding of the
interplay between the mechanisms of memory and memorability.</p>
      <p>
        The purpose of this pilot study at MediaEval’2021 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] was to
collect enough EEG data for proof of concept and demonstration
purposes, showcasing what could be done in subsequent work on
predicting media memorability. The study involved the collection,
ifltering, and interpretation of neurophysiological data, and the
use and evaluation of machine learning methods to enable the
assessment of EEG data as a predictor of video memorability. The
study has culminated in a demonstration of the utility of EEG in
the context of video memorability, along with the public release of
processed EEG features for others to explore1. This study has the
potential to not only broaden the research horizons of computing
1Dataset and examples of use, as well as the code to replicate the results in this paper,
are available at https://osf.io/zt6n9/
researchers, allowing them to explore and leverage EEG features
without any of the requisite domain knowledge, but also increase
the interdisciplinary interest in the subject of memorability more
broadly.
      </p>
      <p>
        Applying EEG to the question of whether an experience will be
subsequently remembered or forgotten is a well researched area
[
        <xref ref-type="bibr" rid="ref12 ref15 ref7 ref9">7, 9, 12, 15</xref>
        ]. Memorability, however, has been shown to be distinct
from subsequent memory efects [
        <xref ref-type="bibr" rid="ref1 ref14">1, 14</xref>
        ], and received little
interdisciplinary attention. Additionally, even though the application
of machine learning to EEG is an active area of interest—allowing
for the automation or augmentation of neurological diagnostics
[
        <xref ref-type="bibr" rid="ref10 ref3 ref4 ref5">3–5, 10</xref>
        ], and the classification of emotional states [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], mental
tasks [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], and sleep stages [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]—the use of EEG to predict visual
memorability has yet to be firmly established, and was previously
limited to static content [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. To the best of our knowledge, this
paper outlines the first application of EEG to video memorability.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>EXPERIMENT DESIGN AND STRUCTURE</title>
      <p>
        The stimuli used in the study are a subset of the subtask 1 data
(i.e., the short-term video memorability prediction task) in
MediaEval’2021 [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and consists of 450 videos, 96 of which were
designated as targets and selected to reflect the bottom and top 50
memorable videos from the TRECVid dataset, 200 were selected to
reflect the next top and bottom 100, and 100 were selected to reflect
the middle 100 memorable videos (95 selected + 5 duplicates) from
the set of subtask videos. EEG data was collected from 11 subjects
while they completed a short-term memory experiment, which was
used to annotate the videos for memorability. EEG data acquisition2
was carried out in two separate locations using a shared
experimental procedure, and each location annotated the same set of videos.
Rather than being split into separate encoding and recognition
phases, the experiment was continuous in nature.
      </p>
      <p>Before the experiment was carried out, participants were given
a verbal description of the experiment procedure, presented with
a set of written instructions, and taken through a practice run of
3 videos to familiarise them with the experiment. The experiment
used a total of 450 videos, 192 of which were the target videos (96
targets, shown twice), and the remaining 258 videos were the fillers.
The experiment was broken into 9 blocks of 50 videos, where a
2Data collection for participants 1–5 was carried out at Dublin City University (DCU)
with approval from the university’s Research Ethics Committee (DCUREC / 2021 /
171), and for participants 6–11 at the University of Essex (UoE) with approval from
the Ethics Committee (ETH2122-0001). Data at DCU was collected using a 32-channel
ANT Neuro eego system with a sampling rate of 1000 Hz. Data at UoE was collected
using a 64-channel BioSemi ActiveTwo system at a sampling rate of 2048 Hz.
ifxation cross was displayed for 3–4.5s, followed by the video
presentation for its ~6 second duration, followed by a “get ready to
answer” prompt of 1–3 seconds, followed by a 3s period for
recognition response (repeated video or not). The time per block was
approximately 700 seconds (~12 minutes) without accounting for
30-second closed/open eye baselines and breaks, which occurred
between blocks. In order to account for recency efects, the first 50
videos presented did not include targets, but had 5 filler repeats,
and the presentation positions of targets between each of the
participants was pseudo-randomised, with the distances between target
and repeat videos roughly fitting a uniform distribution, and the
position of each block aside from block 1 being rotated by 1 for
each participant.
3</p>
    </sec>
    <sec id="sec-3">
      <title>ANALYSIS AND RESULTS</title>
      <p>EEG data from both locations were processed in the same way for
the 30 channels that were common across the two setups: data were
ifrst referenced using a common average and band-pass filtered
between 0.1–30 Hz using a symmetric linear-phase FIR filter.
Independent Component Analysis (ICA) was used to remove artifacts,
and trial rejection using subject-specific thresholds was applied.</p>
      <p>To establish a baseline using features extracted from the time
domain, the EEG was low-pass filtered with a cutof frequency of
15 Hz and downsampled to 30 Hz. We applied baseline correction
to the average of the 250-ms pre-stimulus interval and extracted
the data corresponding to the first second of each repeated clip,
from each of the 30 channels, and concatenated it to form a feature
vector. We term these the Event-Related Potential (ERP) features.</p>
      <p>A second set of features were extracted from the EEG, this time
from the time-frequency domain, which we refer to as ERSP
(EventRelated Spectral Perturbation) features. For this, we extracted
4second long epochs and computed a trial-by-trial time-frequency
representation using Morlet wavelets for frequencies between
230 Hz. For this set of features, we used data from only 4 channels,
namely Fz, Cz, Pz, and O1.</p>
      <p>
        Since there were very few forgotten clips, in this task we
differentiate between the first and the second viewing of clips that
were successfully remembered based only on EEG data. To establish
a baseline, we standardised the data to have mean zero and unit
standard deviation, and used scikit-learn’s Bayesian Ridge regressor
with default parameters. Results were obtained through 20-fold
cross-validation with a 20% train-test split, separately for ERP and
ERSP features. The individual classification results for each
participant are shown in Table 1, measured using Area Under the Receiver
Operating Characteristic Curve [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
4
      </p>
    </sec>
    <sec id="sec-4">
      <title>DISCUSSION AND OUTLOOK</title>
      <p>This was an exploratory pilot task to guide the development of
a future experimental protocol for capturing EEG signatures
relating to successful memory encoding and retrieval to be used in
predicting video memorability. While our experimental protocol
resulted in too little data to examine diferences between
successful and unsuccessful encoding, we show EEG-related diferences
exist between the encoding and recognition phases of previously
seen videos. These results indicate that EEG signatures relating to
memory processes for video are present, and thus suitable to be
0.522 ± 0.09
0.558 ± 0.07
0.532 ± 0.07
0.626 ± 0.09
0.649 ± 0.08
0.522 ± 0.10
0.525 ± 0.08
0.674 ± 0.08
0.489 ± 0.06
0.618 ± 0.09
0.611 ± 0.12
0.575 ± 0.06
collected with a revised experimental protocol and more
participants to support a future fully-fledged task for predicting video
memorability. The preprocessed EEG data captured is released to
the research community.</p>
    </sec>
    <sec id="sec-5">
      <title>ACKNOWLEDGMENTS</title>
      <p>This work was part-funded by NIST Award No. 60NANB19D155 and
by Science Foundation Ireland under grant number SFI/12/RC/2289_P2.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Wilma</surname>
            <given-names>A Bainbridge</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Daniel D Dilks</surname>
            , and
            <given-names>Aude</given-names>
          </string-name>
          <string-name>
            <surname>Oliva</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Memorability: A stimulus-driven perceptual neural signature distinctive from memory</article-title>
          .
          <source>NeuroImage</source>
          <volume>149</volume>
          (
          <year>2017</year>
          ),
          <fpage>141</fpage>
          -
          <lpage>152</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Farideh</given-names>
            <surname>Ebrahimi</surname>
          </string-name>
          , Mohammad Mikaeili, Edson Estrada, and
          <string-name>
            <given-names>Homer</given-names>
            <surname>Nazeran</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Automatic sleep stage classification based on EEG signals by using neural networks and wavelet packet coeficients</article-title>
          .
          <source>In 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society</source>
          . IEEE,
          <fpage>1151</fpage>
          -
          <lpage>1154</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Denis</surname>
            <given-names>A Engemann</given-names>
          </string-name>
          , Federico Raimondo,
          <string-name>
            <surname>Jean-Rémi</surname>
            <given-names>King</given-names>
          </string-name>
          , Benjamin Rohaut, Gilles Louppe, Frédéric Faugeras, Jitka Annen, Helena Cassol, Olivia Gosseries, Diego Fernandez-Slezak, and others.
          <source>2018</source>
          .
          <article-title>Robust EEG-based cross-site and cross-protocol classification of states of consciousness</article-title>
          .
          <source>Brain</source>
          <volume>141</volume>
          ,
          <issue>11</issue>
          (
          <year>2018</year>
          ),
          <fpage>3179</fpage>
          -
          <lpage>3192</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Behshad</given-names>
            <surname>Hosseinifard</surname>
          </string-name>
          , Mohammad Hassan Moradi, and
          <string-name>
            <given-names>Reza</given-names>
            <surname>Rostami</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Classifying depression patients and normal subjects using machine learning techniques and nonlinear features from EEG signal</article-title>
          .
          <source>Computer methods and programs in biomedicine 109</source>
          ,
          <issue>3</issue>
          (
          <year>2013</year>
          ),
          <fpage>339</fpage>
          -
          <lpage>345</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Cosimo</given-names>
            <surname>Ieracitano</surname>
          </string-name>
          , Nadia Mammone, Amir Hussain, and Francesco C Morabito.
          <year>2020</year>
          .
          <article-title>A novel multi-modal machine learning based approach for automatic classification of EEG recordings in dementia</article-title>
          .
          <source>Neural Networks</source>
          <volume>123</volume>
          (
          <year>2020</year>
          ),
          <fpage>176</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Sang-Yeong Jo</surname>
          </string-name>
          and
          <string-name>
            <surname>Jin-Woo Jeong</surname>
          </string-name>
          .
          <year>2020</year>
          .
          <article-title>Prediction of visual memorability with EEG signals: A comparative study</article-title>
          .
          <source>Sensors</source>
          <volume>20</volume>
          ,
          <issue>9</issue>
          (
          <year>2020</year>
          ),
          <fpage>2694</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Demetrios</given-names>
            <surname>Karis</surname>
          </string-name>
          , Monica Fabiani, and
          <string-name>
            <given-names>Emanuel</given-names>
            <surname>Donchin</surname>
          </string-name>
          .
          <year>1984</year>
          .
          <article-title>“P300” and memory: Individual diferences in the von Restorf efect</article-title>
          .
          <source>Cognitive Psychology 16</source>
          ,
          <issue>2</issue>
          (
          <year>1984</year>
          ),
          <fpage>177</fpage>
          -
          <lpage>216</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Rukiye</given-names>
            <surname>Savran</surname>
          </string-name>
          <string-name>
            <given-names>Kiziltepe</given-names>
            , Mihai Gabriel Constantin,
            <surname>Claire-Hélène</surname>
          </string-name>
          <string-name>
            <surname>Demarty</surname>
          </string-name>
          , Graham Healy, Camilo Fosco, Alba García Seco de Herrera, Sebastian Halder, Bogdan Ionescu, Ana Matran-Fernandez,
          <string-name>
            <given-names>Alan F.</given-names>
            <surname>Smeaton</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Lorin</given-names>
            <surname>Sweeney</surname>
          </string-name>
          .
          <year>2021</year>
          .
          <article-title>Overview of The MediaEval 2021 Predicting Media Memorability Task</article-title>
          .
          <source>In Working Notes Proceedings of the MediaEval 2021 Workshop.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Wolfgang</given-names>
            <surname>Klimesch</surname>
          </string-name>
          .
          <year>1999</year>
          .
          <article-title>EEG alpha and theta oscillations reflect cognitive and memory performance: a review and analysis</article-title>
          .
          <source>Brain research reviews 29</source>
          ,
          <fpage>2</fpage>
          -
          <lpage>3</lpage>
          (
          <year>1999</year>
          ),
          <fpage>169</fpage>
          -
          <lpage>195</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Christoph</surname>
            <given-names>Lehmann</given-names>
          </string-name>
          , Thomas Koenig, Vesna Jelic, Leslie Prichep, Roy E John,
          <string-name>
            <surname>Lars-Olof</surname>
            <given-names>Wahlund</given-names>
          </string-name>
          , Yadolah Dodge, and
          <string-name>
            <given-names>Thomas</given-names>
            <surname>Dierks</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Application and comparison of classification algorithms for recognition of Alzheimer's disease in electrical brain activity (EEG)</article-title>
          .
          <source>Journal of neuroscience methods 161</source>
          ,
          <issue>2</issue>
          (
          <year>2007</year>
          ),
          <fpage>342</fpage>
          -
          <lpage>350</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Nan-Ying</surname>
            <given-names>Liang</given-names>
          </string-name>
          , Paramasivan Saratchandran,
          <string-name>
            <surname>Guang-Bin Huang</surname>
            , and
            <given-names>Narasimhan</given-names>
          </string-name>
          <string-name>
            <surname>Sundararajan</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>Classification of mental tasks from EEG signals using extreme learning machine</article-title>
          .
          <source>International journal of neural systems 16, 01</source>
          (
          <year>2006</year>
          ),
          <fpage>29</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Eunho</surname>
            <given-names>Noh</given-names>
          </string-name>
          , Grit Herzmann, Tim Curran, and
          <string-name>
            <surname>Virginia R de Sa</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Using single-trial EEG to predict and analyze subsequent memory</article-title>
          .
          <source>NeuroImage</source>
          <volume>84</volume>
          (
          <year>2014</year>
          ),
          <fpage>712</fpage>
          -
          <lpage>723</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Fabian</surname>
            <given-names>Pedregosa</given-names>
          </string-name>
          , Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel,
          <string-name>
            <given-names>Peter</given-names>
            <surname>Prettenhofer</surname>
          </string-name>
          , Ron Weiss, Vincent Dubourg, and others.
          <source>2011</source>
          .
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>the Journal of machine Learning research 12</source>
          (
          <year>2011</year>
          ),
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Michael</surname>
            <given-names>D</given-names>
          </string-name>
          <string-name>
            <surname>Rugg</surname>
            and
            <given-names>Tim</given-names>
          </string-name>
          <string-name>
            <surname>Curran</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Event-related potentials and recognition memory</article-title>
          .
          <source>Trends in cognitive sciences 11</source>
          ,
          <issue>6</issue>
          (
          <year>2007</year>
          ),
          <fpage>251</fpage>
          -
          <lpage>257</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Thomas</surname>
            <given-names>F Sanquist</given-names>
          </string-name>
          , John W Rohrbaugh, Karl Syndulko, and Donald B Lindsley.
          <year>1980</year>
          .
          <article-title>Electrocortical signs of levels of processing: Perceptual analysis and recognition memory</article-title>
          .
          <source>Psychophysiology</source>
          <volume>17</volume>
          ,
          <issue>6</issue>
          (
          <year>1980</year>
          ),
          <fpage>568</fpage>
          -
          <lpage>576</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Xiao-Wei</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dan Nie</surname>
          </string-name>
          , and
          <string-name>
            <surname>Bao-Liang Lu</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Emotional state classification from EEG data using machine learning approach</article-title>
          .
          <source>Neurocomputing</source>
          <volume>129</volume>
          (
          <year>2014</year>
          ),
          <fpage>94</fpage>
          -
          <lpage>106</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>