=Paper= {{Paper |id=Vol-2717/paper14 |storemode=property |title=No room for hate: What research about hate speech taught us about collaboration? |pdfUrl=https://ceur-ws.org/Vol-2717/paper14.pdf |volume=Vol-2717 |authors=Ajda Šulc,Kristina Pahor de Maiti |dblpUrl=https://dblp.org/rec/conf/dhn/SulcM20 }} ==No room for hate: What research about hate speech taught us about collaboration?== https://ceur-ws.org/Vol-2717/paper14.pdf
                      No room for hate: What research about socially
                   unacceptable discourse taught us about collaboration?

                                         Ajda Šulc1 and Kristina Pahor de Maiti2
                                1 Faculty of Social Sciences, University of Ljubljana, Slovenia
                                      2 Faculty of Arts, University of Ljubljana, Slovenia

                                             ajda.sulc@gmail.com
                                      kristina.pahordemaiti@ff.uni-lj.si



                       Abstract. This paper offers insights into the collaboration process of a research
                       team that brought together social scientists, humanists and computer scientists on
                       the topic of socially unacceptable discourse online. What seemed as a straight-
                       forward problem, proved to be a complex phenomenon that required intense dis-
                       cussions and several iterations of solutions development in order to arrive at a
                       result that would satisfy the individual needs of the disciplines involved. More
                       specifically, we present the challenges faced before and during the creation of a
                       corpus of socially unacceptable Facebook comments. From a collaboration point
                       of view, we learned that it is crucial to set aside enough time for regular brain-
                       storming sessions and feedback throughout the project since this prevents possi-
                       bly fatal detours due to misunderstanding with regard to terminology or the scope
                       of research. Moreover, we saw how a lack of a common system for taking scru-
                       pulous notes on all interventions into common data resource can lead into multi-
                       ple iterations of simple tasks. Finally, the collaboration thought us that listening
                       is crucial in order to optimally combine and exchange knowledge and analytical
                       approaches among the disciplines, but also to rationally simplify tasks whenever
                       possible.

                       Keywords: Socially unacceptable discourse, Hate speech, Social media, Anno-
                       tation schema


               1       Introduction

               In the last two decades, rapid development and raising popularity of social media con-
               siderably changed our communication habits. This applies especially to written com-
               munication which is now predominantly digital. We can find a large portion of our
               everyday exchanges on social media, but despite all the positive aspects that this can
               have, many of these exchanges now reflect intolerant ideas and even encouragements
               to violent acts. Such utterances are frequently found in comments to posts from news
               media outlets that use social media platforms, such as Facebook, to disseminate their
               content. It has been shown that intolerant and abusive speech harms the targets as well
               as the society as a whole (Nielsen, 2002). To prevent these negative consequences, ef-


               Copyright 2020 for this paper by its authors. Use permitted under Creative Commons
               License Attribution 4.0 International (CC BY 4.0).




Twin Talks 2 and 3, 2020        Understanding and Facilitating Collaboration in Digital Humanities            132/143
               2


               forts have been made to develop automated detection of intolerant utterances, and re-
               searchers from various disciplines (e.g., media studies, law, psychology, computational
               linguistics, sociology) are studying the phenomenon of socially unacceptable discourse
               with the aim to gain better understanding of its dynamics and curb its proliferation.
                  In this paper we use the developments and outcomes of our research on socially
               unacceptable discourse as an example on which we base our report on the collaboration
               experience in an interdisciplinary team of researchers. In Section 2, we explain our
               research problem from three scientific perspectives and state collaboration opportuni-
               ties. In Sections 3 and 4, we present the collaboration challenges and solutions that lead
               to a creation of a language resource that meets the needs of social scientists, humanists
               and computer scientists. Section 5 concludes the paper with the main takeaways from
               our collaborative experience.


               2       Research problem and collaboration opportunities

               Socially unacceptable discourse (SUD), as we named it, is an umbrella term for com-
               munication practices that are openly or covertly harassing, provocative or insulting,
               incite to violence or express negative generalizations, stereotypical judgements, ob-
               scenities or incivilities (Vehovar et al., 2020). In addition to this broad spectrum of its
               possible manifestations, SUD is influenced by many contextual factors, such as the
               identity of the author/target, cultural setting, medium, language and so on (Schmidt &
               Wiegand, 2017). Due to its proliferation on social media in the last decade, SUD has
               become a trending topic in various scientific fields, but due to its complexity, the sci-
               entific community still struggles with a comprehensive description of the phenomenon.
               In order to contribute to the pool of insights into the nature of SUD and improve the
               understanding of SUD on social media, we joined forces of three scientific fields: So-
               ciology, Linguistics and Computational linguistics. Each of the three had its own re-
               search interests in the project.
                  SUD is primarily a concept of Social Sciences. For this reason, the research on it in
               these disciplines is rich and varied. Several studies have covered SUD or some of its
               forms in the fields of sociology (Dragoš, 2007), communication studies (Bajt, 2018),
               media studies (Vehovar et al., 2012) or journalism (Milosavljevič, 2012). In our project,
               broadly speaking, the sociologists were mainly interested in the impacts of SUD on
               ideological stances of users and public communication. Therefore, they wanted to study
               the scope and forms of SUD in the comments, the influence of contextual factors on the
               formation of SUD (e.g., the media post topic, media type, target, etc.) and network
               interconnectivity.
                  In Linguistics, SUD has not been so thoroughly researched, but since it is primarily
               realized through linguistic means, there exists a certain volume of research on SUD
               from different theoretical and analytical perspectives (e.g., sociolinguistics (Gorjanc,
               2005; McEnery, 2004), psycholinguistics (Kapoor, 2016; Pinker, 2008), pragmatics
               (Jay & Janschewitz, 2008; Pahor de Maiti & Fišer, 2020), foreign language learning
               (Horan, 2013), critical discourse analysis (Methven, 2017), etc.). The central linguistic
               research question was whether SUD is characterized by specific linguistic features, and




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities            133/143
                                                                                                        3


               if so, what are they. To this end, the analysis of SUD needed to be conducted on differ-
               ent levels of linguistic description. The researchers wanted to look at orthographic,
               grammatical and lexical dimensions of SUD as well as investigate the power relations
               that are being constructed or maintained through language use.
                   Given the negative influence of SUD on communication level and society as a
               whole, efforts have been dedicated in the last decade to the development of tools that
               would enable automatic detection and removal of online SUD. But due to the complex-
               ity of SUD, accurate and timely detection has yet to be achieved (ElSherief et al., 2018;
               Vidgen & Yasseri, 2019; Zhang & Luo, 2019). The problem is usually regarded as a
               machine learning classification task in which researchers develop algorithms or pro-
               duce descriptive statistics (Fortuna & Nunes, 2018). But related work shows that many
               challenges remain unsolved. They are mainly related to the lack of a common definition
               of the phenomenon, the absence of a commonly accepted benchmark corpus and a pre-
               dominant focus on English data (Schmidt & Wiegand, 2017). Furthermore, researchers
               usually develop their datasets based on project-specific annotation schemas and use
               various sets of features for detection purposes (ibid.). All this hinders comparative anal-
               ysis and consequently the generalization of findings. In our project, the computational
               linguistics group of researchers had two main research interests. The first was related
               to the development of a robust annotation schema that would be applicable across lan-
               guages and cultures, and the second was linked to the creation of a set of features that
               would prove most useful for detection tasks of Slovene SUD.
                   Following the research interests outlined above, we saw two main collaboration op-
               portunities: (1) annotation schema and dataset creation, and (2) the exchange of theo-
               retical knowledge and analytical approaches. In order to be able to address all the indi-
               vidual needs of the three disciplines, we needed a dataset that would be enriched with
               extensive metadata and several annotation layers. In this step, the main collaborative
               efforts were therefore put into defining the necessary categories of metadata and lin-
               guistic annotations while balancing these requirements with limitations imposed by pri-
               vacy regulation and computational possibilities. During the dataset creation, as well as
               in the following analytical phases of the project, the collaboration focused on the ex-
               change of theoretical knowledge and methodological approaches. This collaboration
               was crucial due to the complex nature of the studied phenomenon. Since almost all the
               aspects of SUD surpass single scientific domain, we understood that in order to provide
               a comprehensive and reliable interpretation of the results, we will need close interdis-
               ciplinary collaboration.


               3       The solution

               The main idea was to extract a suitable volume of online communication to be manually
               annotated and thus categorized according to previously designed annotation schema.
               We needed a clean dataset with enough relevant comments that could be used for quan-
               titative analyses, but at the same time manually annotated. Since this was our common
               goal, the solution seemed simple. Sociology and Linguistics knowledge contributed to




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities            134/143
               4


               the content selection, while Computational linguistics experts took care of technical
               aspects – mainly accessing and extracting the material.


               3.1     Defining and balancing the research goals
               For the purposes of all the three disciplines, we agreed that we want to analyze authentic
               communication, i.e. real-world discourse, written spontaneously by users on the web.
               We needed a public source since we did not want to (and were not allowed to) invade
               the privacy of individuals, but also a source that would provide us with a coherent and
               extensive discussion. Consequently, we decided to use user comments under public
               news posts on Facebook that were published by the country’s most read media outlets.
               We found that most of them are using Facebook to regularly share their own articles,
               and a number of followers are regularly commenting the content shared thus forming a
               connected string of discourse.
                  To be able to extract a sufficient amount of posts and associated comments, we chose
               the top three media outlets by their popularity according to the Alexa service 1 (i.e.,
               24ur.com, SIOL.net and Nova24TV), and extracted the news posts they shared on their
               official Facebook profile. At the time of the extraction, RTV Slovenia was also among
               the most popular media outlets in Slovenia, but their Facebook shares did not have
               enough comments to be used for the analysis so we did not include it (Ljubešič, Fišer
               and Erjavec, 2019). In the next step, we agreed that we need a relevant sample of com-
               ments. Since we planned to manually annotate the harvested comments, preferably each
               comment by several annotators to reduce the possibility of error and subjectivity, we
               could not afford to use random discourse, since we assumed that most of the discourse
               would be neutral and thus not relevant for our analysis. To ensure time and cost-effi-
               cient annotation process, we therefore choose to filter our data. Following Social Sci-
               ence experts’ experiences on typical hateful discourse triggers, we chose the news posts
               on then controversial topics on two minority groups: the LGBT community and mi-
               grants/refugees. Comments under these posts were recognized as the most relevant and
               therefore chosen to be extracted separately for annotation.
                  A combination of manual and automated classifying based on key words was per-
               formed in order to filter out the posts about LGBT and migrants (Ljubešič, Fišer and
               Erjavec, 2019). We extracted all of the posts that were published on these two topics
               on the official page of the media outlet from the time their Facebook profile was acti-
               vated until the time of the data collection (the end of 2017). For the Slovene data, the
               algorithm identified 93 posts and 4.571 comments about LGBT and 967 posts and
               43.000 comments about migrants. The latter were reduced to 30 most relevant posts
               with 6.545 comments for the annotation process in order to have similar and manually
               doable amount of comments for both minorities (Vehovar et al., 2020).




               1 https://www.alexa.com/topsites/countries




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities           135/143
                                                                                                         5


               4       The collaboration experiences

                  Following the agreement on what data was to be annotated, an annotation schema
               had to be designed, tested and used. It first seemed like a simple task of choosing the
               relevant categories of discourse, but we found that there were quite some dilemmas
               resulting from different understandings of the main concept and different needs of the
               three disciplines.


               4.1     Annotation schema

                   What are investigating?
                   The main question we had to answer was ‘What are we researching?’ Harmonizing
               the concepts between different disciplines required detailed discussion on our under-
               standing, definitions and possibilities to adjust to others’ needs. First, the idea was to
               research hate speech, but noticeable divergence occurred at this stage. From Sociolo-
               gists’ point of view, hate speech term is closely related to the social power concept and
               is taking into account the social position of the speaker and targets of such speech.
               European Commission against Racism and Intolerance (ECRI) defines hate speech as
               a speech that: “entails the advocacy, promotion or incitement of the denigration, hatred
               or vilification of a person or group of persons, as well any harassment, insult, negative
               stereotyping, stigmatization or threat of such person or persons and any justification of
               all these forms of expression – that is based on a non-exhaustive list of personal char-
               acteristics or status that includes “race”, color, language, religion or belief, nationality
               or national or ethnic origin, as well as descent, age, disability, sex, gender, gender iden-
               tity and sexual orientation” (European Commission against Racism and Intolerance,
               2016). The focus in Sociological sense is therefore on the background of the person or
               group that is a target of hate speech. Additionally, Social Sciences’ research of hate
               speech is usually in relation to its legal aspects considering the current legal practice in
               this field as an important criterion for categorization of hate speech. In Slovenia, Public
               incitement to hatred, violence or intolerance is a criminal offense under the Article 297
               of Criminal Code (KZ-1, 2008), but the conditions for prosecution are more specific
               than just general incitement, taking into account also how radical the speech is, how
               likely it will encourage a concrete hostile act, and previously mentioned social position
               of the target. According to the Supreme State Prosecutor's Office’s “Position on the
               prosecution of the criminal offense of Public Incitement to Hatred, Violence or Intoler-
               ance under Article 297 of Criminal Code" (2013), public incitement to hatred, violence
               or intolerance should generally be expressed towards disprivileged, vulnerable social
               groups, or minorities, that are deprived of political and social power in a certain society,
               and whose inequality is further deepened by such speech.
                   Accordingly, the categorization of hate speech from the Sociologists’ point of view
               is a very complex task that surpasses the sole content analysis. On the other hand, Lin-
               guistics and Computational linguistics experts needed a categorization that would sep-
               arate hateful speech from non-hateful one, using broader definition without a relation
               to social groups belonging and social relationship between the speaker and the target.




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities             136/143
               6


               For them, the focus was on a discourse that generally expresses discriminatory attitudes
               and hatred (Baider et al., 2017). Considering different approaches and definitions, we
               did not want to use the term ‘hate speech’, since no matter which discourse exactly we
               were about to cover with this term, it would not be accurate enough for at least one of
               the disciplines. This led us to introduce a new umbrella term – socially unacceptable
               discourse (SUD) which covers the broad definition of hateful discourse that we wanted
               to analyze.

                  Who is the target?
                  The question which targets are we interested in was closely related to the definition
               the individual disciplines used. Within that, Sociologists needed a distinction between
               the targets attacked because of their background and the other targets, either individuals
               or groups that are not socially protected or potentially disprivileged. They especially
               wanted to focus on chosen minorities (LGBT and migrants), so those had to be specif-
               ically labeled. Given the more general definition of SUD that the other two disciplines
               used, a distinction between other several target groups was also desired, but again had
               to be relevant according to the expected targets of hateful discourse online. The agree-
               ment on that was reached with a common expectation that the most usual targets, be-
               sides the subjects of the main article posted (in our case LGBT or migrants), were the
               media outlet or journalists and other commenters. As Hammod and Abdu-Rassul (2017)
               noticed, many commenters responding to other commenters’ comments are indeed us-
               ing some kind of aggression towards each other.

                  Should we consider the context?
                  Different understandings of the concept of discourse produced a dilemma of how
               much of the context of the individual comment we should consider during the manual
               annotation. For the Linguistic and Sociological analysis, the social, cultural and histor-
               ical context are a crucial part of each text, assuming that the content often cannot be
               properly understood without knowing the background of what is expressed. Even
               though for the machine learning process this was not preferable, given the importance
               of the context for the message delivered, we choose to consider it.
                  Our dataset enables looking into the textual context as well, since the annotators
               were able to read the title of the main article as well as other previous comments, giving
               them an insight into what the conversation was about. In the end, all three disciplines
               agreed that the context should be included due to its importance as influencing factor.

                   Do we include borderline cases?
                   As much as sociological definition of hate speech is narrowing down the concept
               regarding the targets, it is, on the other hand, quite broad when it comes to the interpre-
               tation of message that the text is delivering. Researching hatred, Sociologists are also
               interested in indirect hateful messages, oblique allegations, and negative stereotyping
               that are reflected as everyday discrimination or remarks directed towards a person
               solely based on his or her belonging to a specific social group. For a cooperation with
               experts from Linguistics, though, this was not entirely desirable since they wanted a
               clear distinction between different levels of hatred expressed in the comments. The




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities            137/143
                                                                                                        7


               agreement was that indirect messages can be considered unacceptable, but not when
               this would be too oblique to understand it as hateful. We also choose not to include the
               cases where the commenter only agreed with a hateful message, but did not (re)produce
               SUD in any form.

                  The solution
                  Considering all the needs and divergences described above, a complex two-level
               schema was designed that allowed grouping the annotated comments in a way to cater
               to the research needs of all the domains involved. On the first level, it distinguishes six
               types of speech according to the radicality of the content and according to why was the
               target assaulted:
                     Acceptable speech
                     Inappropriate speech
                     Background – offensive speech
                     Background – violence
                     Other – offensive speech
                     Other – violence

                  On the second level, one of the five different target groups needs to be chosen:
                    Migrants/LGBT
                    Related to migrants/LGBT (their supporters or alleged supporters)
                    Journalist or media
                    Commenter
                    Other


               4.2     Annotation process

                  Following the described annotation schema, 32 annotators, trained specially for the
               given task by our experts, started the annotation process for Slovene comments. They
               were working via online crowdsourcing tool PyBossa, which has its drawbacks, but is
               recognized as a useful tool for working with a large group of annotators. The main post
               text, published by the media outlet, and all the comments bellow it were displayed and
               annotators individually chose a type and a potential target for each of the comments.
               Their work was monitored by technical team, regularly extracting the information on
               their progress and agreement ratio, while a Social Sciences expert was analyzing the
               cases where the agreement was the lowest and giving the annotators advices and direc-
               tions on how to improve their work.
                  Only after working with annotators as a fourth group of participants, and after the
               analysis of a significant amount of actual cases, some new dilemmas arose. We found
               that a certain amount of subjectivity will always be present when deciding on the degree
               of hatefulness of the text, so more annotations for one comment has proven to be a good
               solution, enabling the researchers to use the modal category when analyzing the data
               later. Authentic communication is also unpredictable – sometimes it is hard to under-
               stand, since the context might not be available or it can abuse several targets. Some of




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities            138/143
               8


               the cases with the lowest agreement had to be additionally checked and annotated by
               experts.


               5       Main takeaways on interdisciplinary collaboration

               In this section, we discuss the main conclusions that will guide our collaboration efforts
               in our future projects. They are based on positive and negative experience from the
               project and are arranged into four categories which convey our main takeaways.

               5.1      Take the time
               Immerged into specific research questions and occasionally overwhelmed with admin-
               istrative work, we saw project group meetings often as an unpleasant necessity, rather
               than a beneficial opportunity. Looking back, we see that cutting back on the time for
               discussions (of the whole project group or its parts) leads into misunderstanding that
               could otherwise be prevented. Consequently, what seemed at the beginning as time-
               saving measures, proved at the end as time-consuming ones. Moreover, our experience
               shows that not only the regularity of meetings, but their structure is of equal importance.
               We saw that our project group worked best on semi-structured meetings where the time
               was divided between the presentation of progress, pre-prepared Q&A time and ample
               time for open discussion. This last part proved especially beneficial in the initial phases
               of the project when we needed to negotiate the scope of the research and best ap-
               proaches to dataset creation.
                   Being eager to start early, we immediately dived into work on annotation schema
               and started with a small sample of real-life data and some made-up examples. This is a
               perfectly suitable approach for certain phenomena, but it was soon clear that it is not
               the optimal approach for research on SUD. In our case, data collection and annotation
               has been a highly elaborated process since affective spontaneous discourse is highly
               unpredictable and often hard to understand even in the context. In the first phases, this
               process was even more complex since the guidelines accompanying the schema have
               been quite basic. In the later phases, we have added several special cases to the guide-
               lines with expert explanations of the most appropriate tag. In our future projects, in
               order to lower the complexity of the annotation process, we will try to work on a con-
               siderable amount of real-life data from the beginning and reserve more time for testing
               the schema and for brainstorming sessions in order to improve the schema before the
               official launch of the annotation campaign.
                   It is inevitable that an interdisciplinary team of researchers will have different ap-
               proaches to data management and different understanding of the importance of various
               interventions into the dataset. A rich dataset, such as ours, might not get properly used
               if its elements are not adequately recorded. When working on a common dataset, it is
               not only important to discuss any interventions beforehand, but it is also crucial to keep
               the notes on the interventions updated. We learned this by resolving the question how
               to deal with comments with two modal categories and how to mark them for later use.
               This question needed a lot of coordination between the individual research teams inside




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities            139/143
                                                                                                         9


               the project since we did not share the common view on the usefulness of such com-
               ments. What was understood as an important detail in the sociological field, was per-
               ceived mainly as noise for (computational) linguists.

               5.2      Listen to each other and stay open
               Complex research problems, such as SUD, that surpass the domain of single scientific
               field require interdisciplinary approach. In fact, for a comprehensive description of such
               phenomena, it might not be enough to stick only to one own standard research tech-
               niques, but it might be beneficial to adopt and adapt techniques and approaches from
               other fields. In our project we thus first tried to share among ourselves the more general
               aspects that represent the strong points of each domain, such as the strictness in meth-
               odology from sociology, focus on qualitative interpretation from linguistics and goal-
               orientation from computer linguistics. In addition, we exchanged analytical techniques
               between the disciplines, for example corpus linguistic techniques were adopted by so-
               ciologists, while sociological survey and inferential statistical methods were adopted
               by linguists.
                  If special care is given to listening to the research needs and hesitations of all the
               researchers involved, the whole team can greatly benefit from this as was the case in
               our project. On the one hand, through careful listening and discussions we learned why
               certain compromises cannot be accepted by all stakeholders despite being reasonable
               to all the others (e.g., in order to respect the established concepts in sociology, we opted
               for new term – SUD – instead of sticking to the well know but nonunanimously defined
               term of hate speech). On the other hand, we observed that it is only possible to correctly
               interpret the findings and appropriately process the data if we are informed of as many
               aspect of the phenomenon as possible (e.g., Social Sciences experts helped the whole
               team understand what are the sensitive aspects of the data and raised awareness regard-
               ing the legal and ethical considerations that need to be taken into account when working
               with SUD data like the need for anonymization, the limitations regarding subsequent
               related data collection or the need for psychological support for annotators).

               5.3     Make sure to have the terminology straight
               Despite being aware from the beginning that SUD is first and foremost a concept from
               Social Sciences, we needed quite some time to really set the terminology and definitions
               to be used in our project. The main difficulty probably originated from the fact that
               SUD is a phenomenon that all of us frequently come across in our everyday life and
               thus we unconsciously felt that we know what our research problem really encom-
               passes. However, experiencing something in everyday life is not the same as approach-
               ing it scientifically, and we can say that, at the beginning, we did not consider this
               aspect seriously enough. Initially, we wanted to stick to one of the existing terms in
               order not to introduce even more complexity into the already terminologically very
               varied field of research. But given the seeming familiarity with the studied phenomenon
               and the fact that scientific definition of hate speech does not correspond with its popular
               definition, we believe that coining a new umbrella term was a good choice in order to
               avoid confusion.




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities             140/143
               10


                  It is somewhat clear that in researching complex and not clear-cut phenomena, such
               as SUD, terminology and the scope of the research needs to be clearly defined in ad-
               vance, and we even observed that it is welcome to regularly refresh this knowledge with
               the entire research team throughout the project. However, in an interdisciplinary pro-
               ject, the attention should not only be payed to such special cases as is the definition of
               the core phenomenon. Despite being tedious, we saw how important it is to avoid using
               too much discipline-specific jargon in order to ease the understanding of the discussion
               for the colleagues from other disciplines. Respecting this simple rule had a very positive
               impact on our work, since the discussions became more inclusive which led to several
               useful suggestions for future steps in the analysis from different members of the re-
               search group.

               5.4     Simplify
               The work we did on SUD was in many ways a great collaboration experience and an
               encouraging learning opportunity. One important conclusion is that compromises are
               inevitable, but that constant negotiation needs to be undertaken in order not to settle for
               simplistic solutions. This can be seen in the development process of our annotation
               schema. Even though we initially wanted a simple annotation schema, it was soon clear
               that a dataset based on such schema would not provide enough information to research-
               ers. For this reason, we initially developed a highly complex schema that proved too
               complicated for efficient annotation process. This led us to a simplification phase in
               which we collected several rounds of feedback and use it to curb the schema. After
               many iterations, we can say that the final version of the annotation schema is simple
               enough to provide a solid framework for the annotators and a rich output in terms of
               metadata. It can be applied to different languages and cultures with slight modifications
               (e.g., with respect to the topic). Nonetheless, it can be further simplified and still remain
               useful. However, we believe that by better managing our expectations and dedicating
               more time to discussions and work on real data, we could arrive at such schema earlier.
                  Throughout the project we learned that simplifying is one of the keys to success, and
               especially so in interdisciplinary settings. We saw that the results of simplifying are
               nothing like the process that is needed to arrive to these results. Mainly, it takes a lot of
               time and we will try to consider this in our next project. In conclusion, we believe that
               interdisciplinary collaboration requires a step back in expectations of each individual
               discipline, and a step forward in looking for innovative research questions that inter-
               twine knowledge of the disciplines, rather than just adding findings one beside the
               other.


               Acknowledgement

               The work described in this paper was funded by the Slovenian Research Agency within
               the national research project »Resources, methods, and tools for the understanding,
               identification, and classification of various forms of socially unacceptable discourse in
               the information society« (J7-8280, 2017 – 2020).




Twin Talks 2 and 3, 2020       Understanding and Facilitating Collaboration in Digital Humanities              141/143
                                                                                                               11


               References
                1. Bajt, V.: Online hate speech and the »refugee crisis« in Slovenia. In I. Žagar & et al. (Eds.),
                   The disaster of European refugee policy: Perspectives from the »Balkan route«. pp. 133–
                   155. Cambridge Scholars Publishing. (2018).
                2. Dragoš, S.: Sovražni govor. Socialno Delo, 46(3), 135–144. (2007).
                3. ElSherief, M., Kulkarni, V., Nguyen, D., Wang, W. Y., & Belding, E.: Hate lingo: A target-
                   based linguistic analysis of hate speech in social media. Twelfth International AAAI Con-
                   ference on Web and Social Media. (2018).
                4. Fortuna, P., & Nunes, S.: A survey on automatic detection of hate speech in text. ACM
                   Computing Surveys (CSUR), 51(4), pp. 1–30. (2018).
                5. Gorjanc, V.: Diskurz slovenskih spletnih forumov–dokončen pokop strpnosti. Večkulturnost
                   v Slovenskem Jeziku, Literaturi in Kulturi, 41, pp. 20–29. (2005).
                6. Horan, G.: ‘You taught me language; and my profit on’t/Is, I know how to curse’: Cursing
                   and swearing in foreign language learning. Language and Intercultural Communication,
                   13(3), pp. 283–297. (2013).
                7. Jay, T., & Janschewitz, K.: The pragmatics of swearing. Journal of Politeness Research,
                   4(2), pp. 267–288. (2008).
                8. Kapoor, H.: Swears in context: The difference between casual and abusive swearing. Journal
                   of Psycholinguistic Research, 45(2), pp. 259–274. (2016).
                9. McEnery, T. Swearing in English: Bad language, purity and power from 1586 to the present.
                   Routledge. (2004).
               10. Methven, E. P.: Dirty talk: A critical discourse analysis of offensive language crimes. Uni-
                   versity of Technology Sydney. (2017).
               11. Milosavljevič, M.: Regulacija in percepcija sovražnega govora: Analiza dokumentov in od-
                   nosa urednikov spletnih portalov. Teorija in Praksa, 49(1), pp. 112–130. (2012).
               12. Nielsen, L. B.: Subtle, pervasive, harmful: Racist and sexist remarks in public as hate speech.
                   Journal of Social Issues, 58(2), pp. 265–280. (2002).
               13. Pahor de Maiti, K., & Fišer, D.: Analiza kazalnih zaimkov v družbeno nesprejemljivih splet-
                   nih komentarjih. In Slovenščina – diskurzi, zvrsti in jeziki med identiteto in funkcijo.
                   Znanstvena založba Filozofske fakultete. (2020).
               14. Pinker, S.: Freedom’s curse. The Atlantic Monthly, 302, pp. 28–29. (2008).
               15. Schmidt, A., & Wiegand, M.: A survey on hate speech detection using natural language
                   processing. 1–10. (2017).
               16. Vehovar, V., Motl, A., Mihelič, L., Berčič, B., & Petrovčič, A.: Zaznava sovražnega govora
                   na slovenskem spletu. Teorija in Praksa, 49(1), pp. 171–189. (2012).
               17. Vehovar, V., Povž, B., Fišer, D., Ljubešić, N., Šulc, A., & Jontes, D.: Družbeno nesprejem-
                   ljivi diskurz na facebookovih straneh novičarskih portalov. Teorija in Praksa, 2, pp. 622–
                   645. (2020).
               18. Vidgen, B., & Yasseri, T.: Detecting weak and strong Islamophobic hate speech on social
                   media. Journal of Information Technology & Politics, pp. 1–13. (2019).
               19. Zhang, Z., & Luo, L.: Hate speech detection: A solved problem? The challenging case of
                   long tail on twitter. Semantic Web, 10(5), pp. 925–945. (2019).
               20. Ljubešić, N., Fišer, D. in Erjavec, T.: The FRENK Datasets of Socially Unacceptable Dis-
                   course in Slovene and English. (2019). Available at https://arxiv.org/pdf/1906.02045.pdf
               21. European Commission against Racism and Intolerance: ECRI General Policy Recommen-
                   dation no. 15 on Combating Hate Speech. (2019). Available at https://rm.coe.int/ecri-gen-
                   eral-policy-recommendation-no-15-on-combating-hate-speech/16808b5b01




Twin Talks 2 and 3, 2020        Understanding and Facilitating Collaboration in Digital Humanities                   142/143
               12


               22. Kazenski Zakonik (KZ-1) [Criminal Code of the Republic of Slovenia]: Uradni list RS,
                   50/12. (2008).
               23. Vrhovno državno tožilstvo Republike Slovenije [Supreme State Prosecutor's Office]: Pravno
                   stališče o pregonu kaznivega dejanja Javnega spodbujanja sovraštva, nasilja ali nestrpnosti
                   po 297. Členu KZ-1 [Position on the prosecution of the criminal offense of Public Incitement
                   to Hatred, Violence or Intolerance under Article 297 of Criminal Code]. (2013).
               24. Baider, F. H., Assimakopoulos, S., & Millar, S. L.: Hate speech in the EU and the
                   CONTACT project. In Online Hate Speech in the European Union: A Discourse-Analytic
                   Perspective (pp. 1–6). Springer. SpringerBriefs in Linguistics: https://doi.org/10.1007/978-
                   3-319-72604- 5_1. (2017).
               25. Hammod, N. M. in Abdul-Rassul, A.: Impoliteness Strategies in English and Arabic Face-
                   book Comments. International Journal of Linguistics, 9(5), 97–112. (2017).




Twin Talks 2 and 3, 2020        Understanding and Facilitating Collaboration in Digital Humanities                143/143