=Paper= {{Paper |id=Vol-3287/paper6 |storemode=property |title=Grounding Words in Visual Perceptions: Experiments in Spoken Language Acquisition |pdfUrl=https://ceur-ws.org/Vol-3287/paper6.pdf |volume=Vol-3287 |authors=Fabio De Ponte,Sarah Rauchas |dblpUrl=https://dblp.org/rec/conf/aiia/PonteR22 }} ==Grounding Words in Visual Perceptions: Experiments in Spoken Language Acquisition== https://ceur-ws.org/Vol-3287/paper6.pdf
Grounding Words in Visual Perceptions: Experiments
in Spoken Language Acquisition
Fabio De Ponte, Sarah Rauchas
Department of Computing, Goldsmiths, University of London, Lewisham Way, New Cross, London SE14 6NW, UK


                                      Abstract
                                      In recent years, Natural Language Processing models have shown compelling progress in generating
                                      and translating text. Yet, the symbols that are manipulated by these models are not produced within
                                      the models themselves. On the contrary, they are externally given in the form of tokens. The models
                                      only measure the probability that a specific token comes after another (or a group of others) and allow
                                      the generation of a list of tokens, each of which has a certain probability of following the previous one.
                                      There is no connection to sensory perceptions, and the semantic interpretation of the outputs – as well
                                      as of the inputs – of these models is completely invisible to the system that produces them. Therefore,
                                      language cannot be used by the system to manipulate information about the perceived world. This
                                      is commonly referred to as the Symbol Grounding Problem and, as of today, there is not a generally
                                      accepted procedure to solve it. This paper explores a possible solution: a sequence-to-sequence model
                                      trained over videos characterised by visual elements which reliably predict the presence of acoustic
                                      co-occurring elements.
                                          A dataset was created ad-hoc, with videos that include 5 types of objects and 5 actions. Two research
                                      questions were considered: whether such a model could map video features onto audio features, in
                                      fact producing a categorization without labels, where the categories would emerge from the parallel,
                                      simultaneous generalization of both input and target; and whether the model would be able to combine
                                      learned information about objects and movements to correctly describe a new combination, shown in a
                                      video it was not exposed to during training, a process that is referred to as compositional semantics.
                                          The experiment showed that the model was able to generalize simultaneously over videos and over
                                      the utterances that were paired with them. Further, it produced sentences that were in some cases more
                                      accurate than the original ones, precisely because of the process of generalization.
                                          However, the results suggest also that the model did not develop the ability to combine information
                                      taken from different samples. In other words, while symbol grounding seems to have been achieved,
                                      compositional semantics does not. The experiment shows that sensory perceptions can be mapped onto
                                      one another with a sequence-to-sequence model trained over a dataset where elements coming from
                                      different sensory domains are paired. However, it is not sufficient to develop compositional semantics.

                                      Keywords
                                      Symbol grounding, Compositional semantics, NLP




1. Introduction
Recently, Natural Language Processing models like GPT3 and BERT have shown compelling
progress in interpreting and predicting text. Yet, while they help us understand the way

NL4AI 2022: Sixth Workshop on Natural Language for Artificial Intelligence, November 30, 2022, Udine, Italy [52]
$ fabio.deponte@gmail.com (F. De Ponte); s.rauchas@gold.ac.uk (S. Rauchas)
€ https://github.com/fabiodeponte/ (F. De Ponte)
 0000-0001-7139-3057 (F. De Ponte)
                                    © 2021 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
 CEUR
 Workshop
 Proceedings
               http://ceur-ws.org
               ISSN 1613-0073
                                    CEUR Workshop Proceedings (CEUR-WS.org)
languages work — offering at the same time valuable tools for different applications — they do
not address the question of the relationship between words and sensory perceptions and of how
language emerged in the first place. These questions date back centuries but have lately become
known within the field of artificial intelligence as “the symbol grounding problem,” since the
issue was clearly defined by Harnad [13, Abstract], who stated it in these terms: «How can the
semantic interpretation of a formal symbol system be made intrinsic to the system, rather than
just parasitic on the meanings in our heads?»
   Nowadays, even a model as simple as an n-gram conditional frequency model, trained over a
sufficiently large corpus of text, can produce apparently meaningful text. However, the symbols
that are manipulated by the system are not inherent to the system itself. Applying the model,
the machine only calculates the probability that a certain token comes after another (or a group
of others) and then produces a list of tokens, each of which has a certain probability of following
the previous one. In principle, this could be done by a human being on an unknown language.
As John R. Searle [34] famously argued in his paper “Minds, brains, and programs,” a person
could apply a similar method to a Chinese corpus and produce a Chinese text, without knowing
the meaning of a single Chinese word.
   Yet, recent advancements led a senior engineer at Google, Blake Lemoine, to claim that a
language model, LaMDA, short for Language Model for Dialogue Applications, a system for
building chatbots, was «sentient» [46]. The company quickly dismissed the claim and, in a
statement, a Google spokesperson said: «There was no evidence that LaMDA was sentient
(and lots of evidence against it).»[46]. In order to support his claim, Lemoine published a few
astonishing fragments of chats between himself and the system. However, as surprising as they
are, they do not seem to suggest the existence of a sentient mind. In fact, a simpler explanation
is possible. Language models capture language structures and, with them, a fraction of the
logic behind language itself. Mikolov, Yih and Zweig [20] were able to design a model such
that some seemingly logical operations between word vectors were possible, like “King - Man +
Woman”, resulting in a vector very close to “Queen”, corresponding to the analogy “a man is
to a king as a woman is to a queen.” Operations like this between embedding vectors do not
always work, even in more recent models. However, they show that sometimes it is possible to
convert logical operations into geometrical operations. More generally, they suggest that human
reasoning reflects in language and that a sufficiently complex language model can capture part
of that logic. This could explain the surprising performances of the most advanced language
models but leaves us with the unsolved problem of how a system can develop a meaningful
semantic interpretation of the symbols that logic is applied to. Therefore, the challenge is to
design a technique that would let the system define its own meaningful symbols. According to
Harnad, «there is really only one viable route from sense to symbols: from the ground up.» [13,
Conclusions]. In other words, symbols have to be grounded to perceptions. There have been
many attempts to do that, we will see some of them below.


2. Background Literature
Many attempts have been made at developing a method for grounding symbols to perceptions.
One of the earliest was proposed by Roy [29, section 1], who designed «a computational model
which learns from untranscribed multisensory input» where «acquired words are represented in
terms associations between acoustic and visual sensory experience.» The model was designed to
learn the same way it is claimed children do, by discovering «words by searching for segments
of speech which reliably predict the presence of visually co-occurring shapes.» The author
recorded a number of sessions of adults speaking to babies in a room. The adults were playing
with the babies with toys, one toy at a time. The system included a speech processor that
«converted spoken utterances into sequences of phoneme probabilities»; and a visual processor
that «extracted statistical representations of shapes and color from images of objects.» Phoneme
probabilities and statistical representations of co-occurring images were stored in a short-term
memory so that the model was able to predict the most probable next word, given an image.
While the experiment was a major step forward, it had an important drawback: not all utterances
contained the name of the toy. Sometimes, adults said only, for example, «Here it comes!»
referring to the toy car they were playing with. Therefore, the results were inconclusive.
   Another interesting experiment was proposed by Tuci et al. [48]. The authors set up a virtual
environment where subsequent generations of evolving robots were trained «to access linguistic
instructions and to execute them by indicating, touching or moving specific target objects.»
[48, Abstract]. In the course of the process, they were able to learn that the commands were
composed of different parts. For example, “touch the pen” is made of two parts, “touch” and “the
pen”, while “move the spoon” is composed of “move” and “the spoon.” Once they had learned
this, they were able to generalize to new compositions, e.g. “touch the spoon,” and execute them,
even though they had never seen that specific command before. This, according to the authors,
demonstrated «how the emergence of compositional semantics is affected by the presence of
behavioural regularities in the execution of different actions.» A drawback was that, while
this experiment effectively demonstrated the principle, it did not provide a method to enable a
system to acquire language, other than within the very limited scope of the experiment itself.
   A similar approach was proposed by Sugita and Tani [43]. They focused on the creation
of a geometrical n-dimensional space, where the geometric arrangements represented «the
underlying combinatoriality» among symbols. They defined 26 actions and recorded 120
corresponding algorithmically generated sensor-motor time series for each. They were then
able to create a model where «the composition of symbols is realized by summing up their
corresponding vectors,» a compositional semantics potentially much more powerful than the
one of Tuci et al. [48].
   More recently, other approaches have been proposed. One came from Gonzalez-Billandon
et al., who converted «the audio signal to an embedding space where embeddings for the same
words are closer than different words, regardless of speaker.» [11, Section II]. In order to achieve
this, they used a Vector Quantized-Variational Autoencoder network. Then, the embeddings
were associated with images. The project is still ongoing and the results have not been released
yet.
   Another method came from Wang et al. [51], who proposed MAXSAT, a SATNet layer that
can be integrated into neural network architectures that could successfully be used to learn
logical structures. Their technique was subsequently used by Topan, Rolnick and Si [47], to
map visual inputs to symbolic variables without explicit supervision, through a self-supervised
pre-training pipeline.
   A different approach was tried by Tan and Bansal [45]. They proposed LXMERT (Learning
Cross-Modality Encoder Representations from Transformers), a large-scale transformer model.
In essence, it is a framework designed to learn direct vision-to-language connections and it
represents a step forward in the task of describing the content of images in words. It outperforms
previous models on visual question-answering datasets. However, it does not include sound or
any perception other than vision. Therefore, it does not offer a direct contribution to tackle the
symbol grounding problem and it does not address the question of compositional semantics.
   A proposal in this direction was instead put forward by Liu, Li and Cheng [17] who replaced
text with audio. They proposed a new general-purpose neural sound synthesis network, based
on generative adversarial networks, that was able to generate sound directly from visual inputs.
In their work, the task was formulated as a regression problem to map a sequence of video
frames to a sequence of raw audio waveform. This model paves the way for the design of a
general method to associate visual to acoustic features and vice versa. It represents a significant
step in the direction of the solution of the symbol grounding problem. However, it was designed
to reproduce any kind of sound, and the authors had in mind mostly noise produced by specific
objects (e.g. cars, motorcycles, scrolling water, etc.). It was not designed for spoken language
and, again, it does not address the question of compositional semantics.


3. Methodology
3.1. Objectives of the research
Our work focused on the following research objectives:
   1. Verify whether it is possible for a system to learn words from an unlabeled dataset of
      spoken utterances and visual representations, through a sequence-to-sequence neural
      network.
   2. Verify if, once such words are learned, compositional semantic would emerge, that is
      whether the system would be able to combine them to compose sentences that were not
      present in the training dataset.

3.2. Dataset
The dataset is composed of 1,000 videos, each showing an object – either staying still or moving
– while a voice reads a sentence, for example “this is a pen,” that refers to what we see. There
are twenty voices in total: ten voices are natural – i.e. they have been recorded by real people –
while the other ten are artificial (produced through the services of the website www.murf.ai).
Each video is exactly 3 seconds long. It is 180x180 pixel sized and the audio is sampled to 16
KHZ. The dataset is available online 1 .
   Each voice recorded 50 sentences: the first 25 (Table 1) describe still objects, the next 25
(Table 2) refer to the moving of objects. A video was recorded for each sentence. Audio and
video files were then paired and saved into a single MP4 file. It is a balanced dataset: each object
is represented 5 times not moving and 5 times moving for each group of 50 videos.

    1
    It can be downloaded from https://www.kaggle.com/datasets/fabiodeponte/symbolgrounding and it is released
under licence Creative Commons Attribution-ShareAlike 4.0 Generic (CC BY-SA 4.0).
Table 1
Objects
                     Number       Spoken utterance         Video content           Times
                        1-5       “This is a pen”          A pen on a table           5
                       6-10       “This is a phone”        A phone on a table         5
                       11-15      “This is a spoon”        A spoon on a table         5
                       16-20      “This is a knife”        A knife on a table         5
                       21-25      “This is a fork”         A fork on a table          5


  The videos were augmented by five operations: flipping image, increasing brightness, de-
creasing brightness, increasing saturation and zooming in. Audios were augmented by five
operations as well: increasing and decreasing speed, increasing and decreasing pitch, and
increasing volume. Lastly, data was split into training, validation and test sets.
  After completing the data augmentation, each of the five augmented audio sets of 1,000
samples, plus the original one, was paired with each of the five augmented video sets of 1,000
samples, plus the original one. In total, 36,000 samples were generated.

3.3. Model architecture
Once the dataset was ready, a system was designed to map visual elements to spoken utterances.
Video files were processed by two pre-trained neural networks, namely Wav2vec (a model
introduced by Baevski et al. [2] at Facebook for self-supervised learning of representations
from raw audio, trained on LibriSpeech corpus) and CLIP (developed by Radford et al. [26] at
OpenAI, trained on a dataset of 400 million image-text pairs collected form a variety of publicly
available sources on the Internet), that extracted respectively acoustic and visual features. Then
a sequence-to-sequence neural network mapped the extracted features of the visual part onto
the extracted features of the acoustic part 2 .
   Wav2vec returns features vectors of 150 integer values, which in turn can be in turn decoded
to a written text by the same library. As we had a dataset composed of 36,000 videos, we got an
audio features matrix sized 36,000 x 150. CLIP returns vectors of 512 float values. However, it
does not provide a tool to convert the features back to a video or to any other directly readable
form. For this reason, while Wav2vec makes it relatively simple to evaluate the predictions of
a video-to-audio neural network, CLIP does not do the same for an audio-to-video network.
Therefore, we focused on the transformation of video features into audio features and not the
reverse, developing a sequence-to-sequence video-to-audio network.
   As a preliminary check, audio features were decoded to written text, through one of the
libraries made available by Facebook with Wav2vec. From this verification, it emerged that
not all features produced intelligible text. In particular, the sentences recorded by real people
contained many errors. The reason is probably that, as they are non-native speakers, their
pronunciation was not good enough for the library. Moreover, the artificial voices that had their
pitch multiplied by a factor of ten for data augmentation purposes resulted in unintelligible
    2
     The code of audio and video extractor and of sequence-to-sequence model is available at https://github.com/
fabiodeponte/symbol_grounding.
Table 2
Actions
          Number    Spoken utterance                Video content                 Times
            26      “Move the pen to the left”      A pen moving to the left      Once
            27      “Move the pen to the right”     A pen moving to the right     Once
            28      “Move the pen up”               A pen moving up               Once
            29      “Move the pen down”             A pen moving down             Once
            30      “Rotate the pen”                A pen rotating                Once
            31      “Move the phone to the left”    A phone moving to the left    Once
            32      “Move the phone to the right”   A phone moving to the right   Once
            33      “Move the phone up”             A phone moving up             Once
            34      “Move the phone down”           A phone moving down           Once
            35      “Rotate the phone”              A phone rotating              Once
            36      “Move the spoon to the left”    A spoon moving to the left    Once
            37      “Move the spoon to the right”   A spoon moving to the right   Once
            38      “Move the spoon up”             A spoon moving up             Once
            39      “Move the spoon down”           A spoon moving down           Once
            40      “Rotate the spoon”              A spoon rotating              Once
            41      “Move the knife to the left”    A knife moving to the left    Once
            42      “Move the knife to the right”   A knife moving to the right   Once
            43      “Move the knife up”             A knife moving up             Once
            44      “Move the knife down”           A knife moving down           Once
            45      “Rotate the knife”              A knife rotating              Once
            46      “Move the fork to the left”     A fork moving to the left     Once
            47      “Move the fork to the right”    A fork moving to the right    Once
            48      “Move the fork up”              A fork moving up              Once
            49      “Move the fork down”            A fork moving down            Once
            50      “Rotate the fork”               A fork rotating               Once


decoded text as well. Therefore, all real voices and some of the artificial ones were discarded,
along with the corresponding videos. The dataset was thus reduced in size and as a result was
composed of only 14,500 samples. Each object was represented 2,900 times: 1,450 times laying
still and 1,450 moving.
   It should be noted that this does not necessarily mean that the discarded samples are useless.
They could be used in a model that does not make use of Wav2vec. In fact, whether the features
are intelligible or not for a speech-to-text converter is not relevant for symbol grounding, as
long as the same utterances are pronounced consistently in similar ways.
   The sequence-to-sequence models are composed of two parts: an encoder and a decoder.
Each of them includes one or more Long Short-Term Memory (LSTM) layers. During training,
the encoder basically compresses the sequence into a vector, while the decoder learns a series
of conditional frequency distributions: for each value and each vector coming from the encoder,
it calculates the most likely value to come next. In terms of the translation process, this means
that the decoder calculates the most probable next word given the last predicted word and the
vector coming from the encoder, which summarizes information about the whole sentence to
be translated.
Figure 1: The structure of the model.


  In our case, we can consider each word of the input sentence one value of the video features
vector and each word of the target translated sentence a value of the audio features target vector.
Then, mapping video features onto audio features may be interpreted as a translation task: on
the input side, we have a sequence that represents the contents of a video; on the output side,
we expect a sequence that represents the same contents expressed in audio format.

3.4. The compositional semantics experiment
As we mentioned, while the first aim was to design a model that would be able to map videos onto
audio recordings, the second was to verify whether such a model could develop compositional
semantics. In order to do that, an experiment was set up. The videos containing a specific
moving object were removed from the dataset, while videos showing the same object laying still
were kept, along with all other objects, both moving and still. The model was trained on the
reduced dataset, and it was tested against the removed videos. The audio features predicted by
the model were converted into text and so it was possible to compare them with the intended
ones.
   For example, at one point all videos of moving pens were removed, so that the model was
trained only on videos showing pens laying still (paired with the utterance “this is a pen” read by
different voices) and other objects (spoons, forks, knives and phones) both staying still (paired
with utterances like “this is a spoon”) and moving (paired with utterances like “move the spoon
to the right”). No videos showing a pen moving to the right and the utterance “move the pen to
the right” were included in the training set. The experiment consisted of verifying whether the
model was able to compose the utterance “move the the pen to the right,” once such a video
was given to the network. In order to do that, the model should have been able to combine the
information coming from the shape of the pen and the information coming from the movement
of other objects. We repeated the experiment five times, with one object at a time.
Table 3
Original and predicted utterances.
             N.    ORIGINAL UTTERANCE                         PREDICTED UTTERANCE
             0     MOVE THE FORK TO THE RIGHT                 MOVE THE FORK TO THE RIGHT
             1     ROTATE THE KNIFE                           WROTATE THE KNIFE
             2     THIS IS A SPOON                            THIS IS A SPOON
             3     MOVE THE FORK TO THE RIGHT                 MOVE THE FORK TO THE RIGHT
             4     MOVE THE SPOON DOWN                        MOVE THE SPOON DOWN
             5     HIS IS A SPURN                             THIS IS A SPERN
             6     MOVED THE FORK TO THE LEFT                 MOVE THE FORK TO THE LEFT
             7     ROTATE THE SPOON                           ROTATE THE SPOON
             8     LUSA PHONAP                                MOVES THE PHONE UP
             9     THIS IS A SPOON                            THIS IS A SPOON
             10    MOVE THE PHONE TO THE RIGHT                MOVE THE PHONE TO THE RIGHT


4. Results and discussion
4.1. Symbol grounding
Twenty different configurations of the sequence-to-sequence model were tried3 , modifying the
size of the layers, the learning rate, the optmizer and the directionality. The best performing
one had an LSTM layer of size 1,024 for the encoder and an LSTM layer of the same size for the
decoder. It was not bidirectional, as bidirectionality made the training slower without offering
any gain in terms of performance. Categorical cross-entropy loss function and Adam optimizer
were adopted, with learning rate 0.001, and a batch size of 64. The video features values were
normalized to the range 0-1, multiplied by 100 and rounded to integer. Therefore, each value
was one-hot encoded in a vector sized 101. This model showed a test loss of 0.0336 and a test
accuracy of 0.9890. Along with loss and accuracy, another performance metric was adopted:
the cosine distance between the predicted features and the expected ones. This measure was
calculated comparing predicted and target vectors on 100 samples extracted from the train set
and 100 from the test set. The showed a test cosine distance of 50.12.
   Another important method of evaluation was the visual inspection. Due to space constraints,
we show here only the results of the best performing model4 . We can see in Table 3 a sample of
11 randomly picked sentences, as predicted by the model. The left column shows the sentence
originally paired with the video (as converted by Wav2vec from the original audio into text),
while the right column shows the predicted sentence. As we can see, the model was able to
produce sentences that in some cases were even more understandable than the original. In the
batch of 11 sentences that we randomly extracted from the test dataset:

    • The model correctly predicted “move the fork to the right” (sample number 0), “this is a
      spoon” (n. 2), “move the fork to the right” (n. 3), “move the spoon down” (n. 4), “rotate
      the spoon” (n. 7), “this is a spoon” (n. 9) and “move the phone to the right” (n.10).
    3
        The detailed description of this process can be found on https://github.com/fabiodeponte/symbol_grounding.
    4
        The results of the other models are described on the Github repository.
    • It predicted “wrotate the knife” (n. 1), a substantially correct output with the minor error
      of adding a “W” to the word “rotate”.
    • It predicted “This is a spern” (n. 5) for the video of a spoon staying still that was originally
      paired with a scarcely comprehensible utterance, namely “his is a spurn”, instead of
      “this is a spoon”. In fact, generalizing over similar videos and over the utterances that
      were paired with them, the model showed a result that turned out to be better than the
      original. This was again the case with the utterance (n. 6). The model returned features
      corresponding to “move the fork to the left”, whereas the original video had been paired
      with “moved the fork to the left”, with a “d” at the end of “move” that made the utterance
      slightly incorrect. And again (n. 8) the model returned “moves the phone up”, which
      was a strong improvement over the utterance originally paired with the video, that was
      interpreted by the Wav2vec library as “lusa phonap”.

4.2. Compositional semantics test
Once video features were mapped onto audio features, the experiment described in 3.4 was
performed. In order to do that, five subsequent tests were performed, one for each object: pen,
phone, spoon, knife and fork. For each of them, a reduced dataset was prepared, removing the
videos that showed the object moving to the left, to the right, up, down and rotating. The model
was trained on the reduced dataset, resulting in 13,050 samples, and then tested against the
1,450 videos that had been removed. The results are shown in Table 4.

Table 4
Model trained on reduced dataset and tested on removed moving object.

                              TRAIN AND TEST
                            ON REDUCED DATASET            TEST ON MOVING OBJECT
                          Train   Train    Test    Test                      Cosine
             OBJECT       loss    acc.     loss    acc.   Loss    Acc.       dist.
               PEN         0.02    0.99    0.04    0.99   0.73    0.88        59.81
              FORK         0.01    0.99    0.03    0.99    0.8    0.88        66.28
             PHONE         0.03    0.99    0.04    0.99   0.73    0.89        59.59
              KNIFE        0.03    0.99    0.04    0.98   0.86    0.87        64.59
             SPOON         0.02    0.99    0.04    0.99   0.97    0.87        61.47

   As we can see, loss increased dramatically when the model was tested over videos of moving
objects that it had not been exposed to during training. We can clearly see that it is not a
generalization problem, because the Test loss column shows that, when tested against videos
previously unseen but belonging to known groups, the model showed a loss ranging between
0.030 and 0.043. Yet, when exposed to new kinds of videos, its loss increased dramatically,
ranging between 0.73 and 0.97. The model was able to generalize when similar videos were
present during training. However, it was not able to combine information from the videos that
showed the objects staying still and information about the movement applied to other objects,
to form a sentence composed by “move” and the name of the object.
   In Table 5, ten predictions (out of 1,450) are shown. As we can see, the model did not predict
a sentence containing either “move the pen” or “rotate the pen” when exposed to a moving pen,
Table 5
Model tested on PEN videos removed from dataset during training.
            Number     ORIGINAL UTTERANCE                 PREDICTED UTTERANCE
            1          OVED THE PAN TO THE LEFT           THIS IS A PEN
            2          MOVE THE PEN TO THE RIGHT          THIS IS A PEN
            3          MOVE THE PEN UP                    THIS IS A PEN
            4          OVE THE PEN DOWN                   THIS IS A PEN
            5          ROTAKE THE PEN                     THIS IS A PEN
            6          MOVE THE PEN TO THE LEFT           OVE THE IFE TO THE LEFT
            7          MOVE THE PEN TO THE RIGHT          THIS IS A PEN
            8          MOVE THE PENNOP                    MOVE THE KNIFE DOWN
            9          MOVE THE PEN DOWN                  THIS IS A PEN
            10         ROTATHE THE PEN                    THIS IS A PEN


Table 6
Model tested on PHONE videos removed from dataset during training.
       Number    ORIGINAL UTTERANCE                      PREDICTED UTTERANCE
       1         MOVED THE PHONE TO THE LEFT             THIS IS A PHAWN
       2         MOVED THE PHONE TO THE RIGHT            MOTHE THE SPOON TO THE LEFT
       3         OOV THE PHON UP                         IS IS A FARN
       4         OVE THE FALM DOWN                       THIS IS A PHONE
       5         ROUTAKE THE PHONE                       THIS IS A PHAWN
       6         MOVE THE PHONE TO THE LEFT              THIS IS A PHONE
       7         MOVE THE PHONE TO THE RIGHT             THIS IS A PHONE
       8         MOVE THE PHONA                          THIS IS A PHONE
       9         MOVE THE FONE DOWN                      THIS IS A FOWN
       10        ROTATE THE PHONE                        THIS IS A PEN


nor was it able (Table 6) to predict “move the phone” or “rotate the phone” when exposed to
the moving phone videos. Similarly, it could not predict any of the correct sentences for the
other three objects. In the dataset of 14,500 videos, each object was represented 2,900 times:
1,450 times laying still and 1,450 moving. The model was trained five times, each time leaving
aside the videos of a particular object depicted while moving. Each time it was tested against
those videos. In total, it returned 7,250 audio features, converted then to text. Again, it cannot
be shown here due to space constraints, but a through inspection showed that the combination
never occurred5 .
   The model mostly favoured the shape of the object, predicting a sentence in the form “this
is...,” thus ignoring the movement. In a minority of cases, it recognised the movement, but
it failed to combine the information with the shape of the object and simply predicted the
utterance “move” accompanied by a different object, often the most similar. For example, “move
the pen” was frequently mistaken for “move the knife.”
   5
    See COMPOSITIONAL SEMANTICS RESULTS - FOR EACH OBJECT COMPARE ORIGINAL AUDIO FEATURES
AND PREDICTED AUDIO FEATURES.ipynb on Github repository.
5. Conclusions
5.1. Symbol grounding and compositional semantics
This project had two aims. The first was to design a system, composed of three neural networks,
that had to be able to directly map, without labels, visual elements onto spoken utterances, to
achieve symbol grounding. The second aim was to verify whether, once the mapping was done,
such a system would develop the ability to compose sentences that were not present in the
training dataset, combining information gathered from different samples, in order to achieve
compositional semantics.
    The results shown above suggest that the ability to directly produce a correct sentence when
exposed to a video has been achieved. The model was able to generalize on both sides of the
dataset and map videos onto audio recordings without labels. However, the results also suggest
that the model did not develop the ability to combine information taken from different samples.
    With this experiment, we tried to capture the power of categorization carried by words into
an artificial system. In fact, the co-occurring audio features associated with each video, once
generalized, play the role of categories. However, they are categories that the model itself
extracts from the sensory perceptions through a process of generalization, rather than from
externally given labels. The fact that words are intimately linked to categories is not surprising,
as with language comes indeed «the ability to generalize,» as Oliver Sacks [33, p. 42] pointed
out in “Seeing voices”, a book devoted to the relationships between sensory perceptions and
language. As Lupyan [19] argues, «merely perceiving an object does not require categorizing
it. In contrast, naming an object (whether to communicate to another individual or for your
own benefit) does require placing it into a category.» However, classification into categories
is not enough. Once things have names, the necessity of a grammar arises, in order to allow
the acquired categories to be manipulated. And, as Corballis [7, p. 37] illustrates, the origin of
grammar could very well have been compositional semantics: "The simplest events consist of
objects and actions, such as baby screams, snake approaches, or apple falls. Suppose, for example,
that an animal’s experience includes five meaningful objects and five meaningful actions. If
each object is associated with a single action, so that only babies scream or only apples fall,
then there are only five events to be signaled, and five event symbols will do the trick; the
objects do not need to be distinguished from the actions associated with them. But if all possible
combinations of objects and actions can occur, then it would be more economical to learn five
symbols for the objects and five for the actions, making ten in all, than to learn twenty-five
symbols to cover all their possible combinations. This might be the source of protolanguage,
leading eventually to grammar."

5.2. Further work
This experiment shows that it is possible to achieve symbol grounding through a sequence-to-
sequence model trained over a dataset with audio and video co-occurring elements. The work
could be expanded in several directions:

   1. Apply an attention layer as introduced by Bahdanau et al. [3] to the sequence-to-sequence
      model, something that might allow the model to achieve compositional semantics.
   2. Give up pre-trained features extraction networks. For real symbol grounding to be
      fully achieved, the system should not include networks that were trained over labels.
      They could be replaced by autoencoders, a solution that could offer the possibility of
      reconstructing the original audio and video files.
   3. Expand the dataset. The dataset developed for this project was relatively small, with one
      thousand original videos and only five objects represented. It could be expanded, with
      the addition of more objects, either staying still or moving, and new voices.
   4. Apply the model to the “something something” video database, that offers images of
      thousands of actions on objects, paired with a linguistic label, composed by a description
      of the action in the form “do  on ” (hence the name). That
      dataset seems perfectly suitable for purposes of compositional semantics. However, it
      lacks audio. In order to overcome the problem, captions – at least a fraction of them –
      could be converted to spoken utterances through artificial voices.


References
[1] Anderson, P., Wu, Q., Teney D., Bruce, J., Johnson, M., Sünderhauf, N., Reid,
    I., Gould, S., van den Hengel, A.: Vision-and-language navigation: Interpreting
    visually-grounded navigation instructions in real environments. In: Proceedings of the
    IEEE Conference on Computer Vision and Pattern Recognition 2018, pp. 3674–3683.
    (2018). Available at: https://openaccess.thecvf.com/content_cvpr_2018/html/Anderson_
    Vision-and-Language_Navigation_Interpreting_CVPR_2018_paper.html. Last Accessed 16
    September 2022.
[2] Baevski, A., Zhou, H., Mohamed, A.R., Auli, M.: Wav2vec 2.0: A Framework for Self-
    Supervised Learning of Speech Representations. In: 34th Conference on Neural Informa-
    tion Processing Systems (NeurIPS 2020), Vancouver, Canada (2020). Available at: https://
    proceedings.neurips.cc/paper/2020/file/92d1e1eb1cd6f9fba3227870bb6d7f07-Paper.pdf. Last
    Accessed 14 September 2022.
[3] Bahdanau, D., Cho, K., Bengio, Y.: Neural Machine Translation by Jointly Learning to
    Align and Translate. In: 3rd International Conference on Learning Representations (2015).
    Available at: https://arxiv.org/pdf/1409.0473.pdf. Last Accessed 6 September 2022.
[4] Cangelosi, A., Greco, A., Harnad., S.: From robotic toil to symbolic theft: Grounding transfer
    from entry-level to higher-level categories. Connection Science, Vol.12, No. 2, 143-162 (2000).
    Available at: https://doi.org/10.1080/09540090050129763. Last Accessed 8 May 2022.
[5] Cho, K., van Merrienboer, B., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., Bengio,
    Y.: Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine
    Translation. In: Proceedings of the 2014 Conference on Empirical Methods in Natural
    Language Processing (EMNLP), 1724-1734 (2014). Available at: https://aclanthology.org/
    D14-1179. Last Accessed 19 September 2022.
[6] Chollet, F.: Character-level recurrent sequence-to-sequence model, Keras, 29 September
    (2017). Available at: https://keras.io/examples/nlp/lstm_seq2seq. Last Accessed 5 September
    2022.
[7] Corballis, M.C.: From hand to mouth. The origins of language. Princeton University Press,
    Princeton and Oxford (2002).
[8] Fanello S.R., Ciliberto C., Natale L., Metta G.: Weakly supervised strategies for natural
    object recognition in robotics in IEEE International Conference on Robotics and Automation
    (ICRA), Karlsruhe, Germany, May 6-10 (2013). Available at: https://ieeexplore.ieee.org/
    document/6631174. Last Accessed 23 May 2022.
[9] Gehring, J., Auli, M., Grangier, D., Yarats, D., Dauphin, Y.N.: Convolutional Sequence
    to Sequence Learning. In: Proceedings of the 34th International Conference on Machine
    Learning - Volume 70 - 1243–1252 (2017). Available at: https://arxiv.org/abs/1705.03122.
    Last Accessed 6 September 2022.
[10] Gentner, D.: Some interesting differences between verbs and nouns. In: Cognition and
    Brain Theory, 1981, vol. 4, 161-178 (1981). Available at: https://groups.psych.northwestern.
    edu/gentner/papers/Gentner81c.pdf. Last Accessed 22 May 2022.
[11] Gonzalez-Billandon, J., Grasse, L., Sciutti, A., Tata, M., Rea, F.: Cognitive Architec-
    ture for Joint Attentional Learning of word-object mapping with a Humanoid Robot.
    In: IEEE/RSJ International Conference on Intelligent Robots and Systems (2019). Avail-
    able at: https://www.researchgate.net/publication/344432053_Cognitive_Architecture_for_
    Joint_Attentional_Learning_of_word-object_mapping_with_a_Humanoid_Robot. Last Ac-
    cessed 12 June 2022.
[12] Hannagan, T., Agrawal, A., Cohen, L., Dehaene S.: Emergence of a compositional neural
    code for written words: Recycling of a convolutional neural network for reading. In: Pnas,
    Vol. 118 - No. 46 (2021). Available at: https://www.pnas.org/doi/10.1073/pnas.2104779118.
    Last Accessed 6 September 2022.
[13] Harnad, S.: The Symbol Grounding Problem. Physica D 42: 335-346 (1999). Available at:
    https://arxiv.org/abs/cs/9906002v1. Last Accessed 7 May 2022.
[14] Harnad, S.: Symbol grounding and the origin of language, University of Southampton
    Institutional Research Repository (2002). Available at: https://eprints.soton.ac.uk/256471.
    Last Accessed 16 September 2022.
[15] Huang, H., Wong, R.K.: Attention-based Seq2seq Regularisation for Relation Extraction.
    In: International Joint Conference on Neural Networks (IJCNN) (2021). Available at: https:
    //ieeexplore.ieee.org/abstract/document/9533807. Last Accessed 6 September 2022.
[16] CLIP, Video Features Documentation, https://iashin.ai/video_features/models/clip. Last
    Accessed 5 September 2022.
[17] Liu, S., Li, S., Cheng, H.: Towards an End-to-End Visual-to-Raw-Audio Generation with
    GAN. In: IEEE Transaction on Circuits and Systems for Video Technology, Vol. 32, Is-
    sue 3 (2021). Available at: https://ieeexplore.ieee.org/document/9430540. Last Accessed 6
    September 2022.
[18] Luong, M.T., Pham, H., Manning, C.D.: Effective Approaches to Attention-based Neural
    Machine Translation. In: Proceedings of the 2015 Conference on Empirical Methods in
    Natural Language Processing (2015). Available at: https://arxiv.org/abs/1508.04025. Last
    Accessed 6 September 2022.
[19] Lupyan, G.: Carving nature at its joints and carving joints into nature: how labels augment
    category representations, Progress in Neural Processing, Vol. 16, 2005, 87-96 (2005). Available
    at: https://doi.org/10.1142/9789812701886_0008. Last Accessed 8 May 2022.
[20] Mikolov, T., Yih, W., Zweig, G.: Linguistic Regularities in Continuous Space Word Repre-
    sentations. In: Proceedings of the 2013 Conference of the North American Chapter of the
    Association for Computational Linguistics: Human Language Technologies. Association for
    Computational Linguistics, 746–751 (2013). Available at: https://aclanthology.org/N13-1090.
    Last Accessed 16 September 2022.
[21] Nolfi S.: “Emergence of communication and language in evolving robots” in Lefebvre, C.,
    Comrie, B. and Cohen H., New Perspectives on the Origins of Language, 533–554 (2013).
    Available at: https://doi.org/10.1075/slcs.144.20nol. Last Accessed 8 May 2022.
[22] Pasquale, G., Ciliberto, C., Odone, F., Rosasco, L., Natale, L.: Are we done with object
    recognition? The iCub robot’s perspective. In: Robotics and Autonomous Systems, Volume
    112, February 2019, 260-281 (2019). Available at: https://www.sciencedirect.com/science/
    article/pii/S0921889018300332. Last Accessed 12 June 2022.
[23] Pinker, S.: Language Learnability and Language Development (1984/1996). Cambridge,
    MA: Harvard University Press.
[24] Pinker, S.: How could a child use verb syntax to learn verb semantics?. In: Lingua Vol. 92,
    April 1994, 377-410 (1994). Available at: https://www.sciencedirect.com/science/article/pii/
    0024384194903476. Last Accessed 22 Maj 2022.
[25] Prickett, B., Traylor, A., Pater, J.: Seq2Seq Models with Dropout can Learn Generalizable
    Reduplication. In: Proceedings of the 15th SIGMORPHON Workshop on Computational
    Research in Phonetics, Phonology, and Morphology, 93–100, Brussels, Belgium, October
    31 (2018). Available at: https://aclanthology.org/W18-5810.pdf. Last Accessed 5 September
    2022.
[26] Radford, A., Kim, J.W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, A., Sastry, G., Askell,
    A., Mishkin, P., Clark, J., Krueger, G., Sutskever, I.: Learning Transferable Visual Models
    From Natural Language Supervision. In: Proceedings of the 38 th International Conference
    on Machine Learning, PMLR 139 (2021). Available at: http://proceedings.mlr.press/v139/
    radford21a/radford21a.pdf. Last Accessed 14 September 2022.
[27] Ramanathan, V., Tang, K., Mori, G., Fei-Fei, L.: Learning Temporal Embeddings for Complex
    Video Analysis in Proceedings of International Conference on Computer Vision (ICCV)
    (2015). Available at: https://arxiv.org/abs/1505.00315. Last Accessed 26 June 2022.
[28] Roger, E., Banjac, S., Thiebaut de Schotten, M., Baciua, M.: Missing links: The func-
    tional unification of language and memory in Neuroscience & Biobehavioral Reviews,
    Volume 133, February (2022). Available at: https://www.sciencedirect.com/science/article/
    pii/S0149763421005601. Last Accessed 26 June 2022.
[29] Roy, D.: Grounded speech communication. In: Proceedings of the 6th International Confer-
    ence on Spoken Language (ICSLP), vol. 4, 69-72 (2000). Available at: https://www.isca-speech.
    org/archive_v0/archive_papers/icslp_2000/i00_4069.pdf. Last Accessed 22 May 2022.
[30] Roy, D.: Learning Visually Grounded Words and Syntax of Natural Spoken Language,
    Computer Speech & Language, Volume 16, Issues 3–4, July–October 2002, 353-385 (2002).
    Available at: https://www.media.mit.edu/cogmac/publications/evol_comm_2002.pdf. Last
    Accessed 12 May 2022.
[31] Roy, D.: Grounded Spoken Language acquisition: Experiments in Word Learning. In: IEEE
    Transactions on Multimedia, Vol. 5, No. 2 (2003). Available at: https://www.media.mit.edu/
    cogmac/publications/ieee_multimedia_2003.pdf. Last Accessed 12 May 2022.
[32] Roy, D.: Grounding words in perception and action: computational insights. In: Trends in
    Cognitive Sciences, Vol.9, No.8 (2005). Available at: https://lsm.media.mit.edu/papers/Roy_
    TICS_2005.pdf. Last Accessed 22 May 2022.
[33] Sacks, O.: Seeing voices. 3rd edn. London: Picador books (2012).
[34] Searle, J.R.: Minds, brains, and programs. In: The behavioral and brain
    sciences,      417-457 (1980). Available at:               https://www.cambridge.org/core/
    journals/behavioral-and-brain-sciences/article/abs/minds-brains-and-programs/
    DC644B47A4299C637C89772FACC2706A. Last Accessed 12 May 2022.
[35] Selsam, D., Lamm, M., Bunz, B., Liang, P., de Moura, L., Dill, D. L.: Learning a sat solver
    from single-bit supervision. In: Proceedings of The International Conference on Learning
    Representations (ICLR) (2018). Available at: https://arxiv.org/abs/1802.03685. Last Accessed
    12 June 2022.
[36] Shao, L., Migimatsu, T., Zhang, Q., Yang, K., Bohg, J.: Concept2Robot: Learning ma-
    nipulation concepts from instructions and human demonstrations. In: The International
    Journal of Robotics Research, Vol. 40, Issue 12-14, October (2021). Available at: https:
    //journals.sagepub.com/doi/abs/10.1177/02783649211046285. Last Accessed 16 September
    2022.
[37] Siskind, J.M.: Naive Physics, Event Perception, Lexical Semantics, and Language Acquisi-
    tion. In: AI technical reports 1964-2004 (1993). Available at: https://dspace.mit.edu/handle/
    1721.1/6784. Last Accessed 22 Maj 2022.
[38] Siskind, J.M.: Grounding the Lexical Semantics of Verbs in Visual Perception using Force
    Dynamics and Event Logic. In: Journal Of Artificial Intelligence Research, Volume 15, 31-90
    (2001). Available at: https://arxiv.org/abs/1106.0256. Last Accessed 22 Maj 2022.
[39] Speech to Text with Wav2Vec 2.0, KdNuggets. https://www.kdnuggets.com/2021/03/
    speech-text-wav2vec.html. Last Accessed 5 September 2022.
[40] Steels, L.: The symbol grounding problem has been solved, so what’s next?. In: Manuel
    de Vega, Arthur Glenberg, and Arthur Graesser (eds), Symbols and Embodiment: Debates
    on meaning and cognition, Oxford, 2008. Online edn, Oxford Academic, 2012. Available
    at: https://doi.org/10.1093/acprof:oso/9780199217274.003.0012. Last Accessed 14 November
    2022.
[41] Steels, L., Van Eecke, P., Beuls, K.: Usage-based learning of grammatical categories. In:
    Belgian/Netherlands Artificial Intelligence Conference (BNAIC) 2018 Preproceedings, 253-
    264. Available at: https://arxiv.org/pdf/2204.10201.pdf. Last Accessed 14 November 2022.
[42] Steels, L., Verheyen, L., van Trijp, R.: An experiment in measuring understanding. In:
    Proceedings of the First International Conference on Hybrid Human-Artificial Intelligence
    2022, 241-242. Available at: https://ebooks.iospress.nl/pdf/doi/10.3233/FAIA220203. Last
    Accessed 14 November 2022.
[43] Sugita, Y., Tani, J.: A sub-symbolic process underlying the usage-based acquisition of a com-
    positional representation: Results of robotic learning experiments of goal-directed actions.
    In: 2008 7th IEEE International Conference on Development and Learning, ICDL (2008).
    Available at: https://ieeexplore.ieee.org/document/4640817. Last Accessed 16 September
    2022.
[44] Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to Sequence Learning with Neural Networks.
    In: Proceedings of the 27th International Conference on Neural Information Processing
    Systems - Volume 2 (2014). Available at: https://arxiv.org/abs/1409.3215. Last Accessed 6
    September 2022.
[45] Tan, H., Bansal, M.: LXMERT: Learning Cross-Modality Encoder Representations from
    Transformers. In: Proceedings of the 2019 Conference on Empirical Methods in Natural
    Language Processing, 5100–5111 (2019). Available at: https://aclanthology.org/D19-1514.
    Last Accessed 1 July 2022.
[46] Tiku, N.: ’The Google engineer who thinks the company’s AI has come to life’, The Wash-
    ington Post, 11 June (2022). Available at: https://www.washingtonpost.com/technology/
    2022/06/11/google-ai-lamda-blake-lemoine. Last Accessed 1 July 2022.
[47] Topan, S., Rolnick, D., Si, X.: Techniques for Symbol Grounding with SATNet. In: 35th Con-
    ference on Neural Information Processing Systems (2021). Available at: https://proceedings.
    neurips.cc/paper/2021/hash/ad7ed5d47b9baceb12045a929e7e2f66-Abstract.html. Last Ac-
    cessed 12 June 2022.
[48] Tuci, E., Ferrauto, T., Zeschel. A., Massera, G.: An Experiment on Behaviour Generalisation
    and the Emergence of Linguistic Compositionality in Evolving Robots, IEEE transactions
    on autonomous mental development, 176–189 (2011). Available at: https://dx.doi.org/10.
    1109/TAMD.2011.2114659. Last Accessed 8 May 2022.
[49] Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, L.,
    Polosukhin. I.: Attention Is All You Need. In: 31st Conference on Neural Information
    Processing Systems (NIPS 2017), Long Beach, CA, USA (2017). Available at: https://arxiv.
    org/abs/1706.03762. Last Accessed 6 September 2022.
[50] Wang P., Donti P.L., Wilder B., Kolter Z.: Bridging deep learning and logical reasoning
    using a differentiable satisfiability solver, Proceedings of the 36th International Conference
    on Machine Learning, Long Beach, California, PMLR 97 (2019). Available at: https://arxiv.
    org/abs/1905.12149. Last Accessed 12 June 2022.
[51] Wang, W., Wu, M., Pan, S.J.: Deep Weighted MaxSAT for Aspect-based Opinion Extrac-
    tion. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language
    Processing, 5618–5628, November 16–20 (2020). Available at: https://aclanthology.org/2020.
    emnlp-main.453.pdf. Last Accessed 6 September 2022.
[52] D. Nozza, L. Passaro, M. Polignano, Preface to the Sixth Workshop on Natural Language for
    Artificial Intelligence (NL4AI), in: D. Nozza, L. C. Passaro, M. Polignano (Eds.), Proceedings of
    the Sixth Workshop on Natural Language for Artificial Intelligence (NL4AI 2022) co-located
    with 21th International Conference of the Italian Association for Artificial Intelligence
    (AI*IA 2022), November 30, 2022, CEUR-WS.org, 2022.