<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Protecting the Researcher in Digital Contexts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Coppélie Cocq</string-name>
          <email>coppelie.cocq@umu.se</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Evelina Liliequist</string-name>
          <email>evelina.liliequist@umu.se</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lacey Okonski</string-name>
          <email>lacey.okonski@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Umeå University</institution>
          ,
          <addr-line>901 87 Umeå</addr-line>
          ,
          <country country="SE">Sweden</country>
        </aff>
      </contrib-group>
      <fpage>195</fpage>
      <lpage>202</lpage>
      <abstract>
        <p>In recent years, a growing need for protecting researchers has become necessary as online risks such as death threats and “doxing” are more frequent risks in relation to an increased digital landscape of anti-gender, far right extremists, and anti-science movements. This paper suggests resources and strategies for preventing threats and protecting researchers. By improving safety and support, entities such as universities, departments, and research groups can avoid the negative impact of online harassment on researchers' reputation and health, on academic research and for democracy. risky research; online abuse; research ethics</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In recent years, a growing need for protecting researchers has become necessary as greater levels of
risk are posed to academics through online settings. In relation to an increased digital landscape of
antigender, far right extremists, and anti-science movements, risks such as death threats and “doxing”, or
sharing information publicly for the purposes of harassment and intimidation, have become more
frequent. Especially targeted are researchers whose work, and/or whose public identity is norm breaking
- e.g., ethnicity, minority identity, sexual identity, political activism, etc. [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ] or challenges white
male supremacy, colonialism, heteronormativity, and/or in other ways critically studies power
structures. Furthermore, Massanari (2018) observes how “[i]ndividuals working within the humanities
and social sciences are particularly at risk, given the ontological and epistemological rationales for their
research” [5, p. 2].
      </p>
      <p>The growing need for developing resources to protect researchers has been emphasized by e.g. the
Association of Internet Researchers, stating for instance that an “essential measure is that institutions
develop policy detailing support procedures for researchers experiencing online threats or harassment
related to their work” [6, p. 11]. Other texts and statements published by scientific societies encourage
universities to develop policies and strategies for protecting researchers conducting risky research [see
for instance 7].
safety reasons.</p>
      <p>In our research, we are taking a first step toward identifying and understanding unsafe research
situations primarily in a Swedish context. This paper suggests resources and strategies for preventing
threats and protecting researchers in the Humanities. By improving safety and support, entities such as
universities, departments, and research groups can avoid the negative impact of online harassment on
researchers’ reputation and health, and ensure that researchers do not drop funded lines of research for</p>
      <p>2022 Copyright for this paper by its authors.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Understanding researchers at risk</title>
      <p>Today, individuals and groups face increased visibility on social media. As new tools for seeing and
being seen have been introduced in the era of social media the act of sharing has become a central, and
often socially expected, aspect of participation in online communities [cf. 8, p. 1052]. For researchers
in particular, visibility is often constructed as something positive to strive for: we are encouraged (and
to some extent expected) by our employers and funding institutions to inform others about our research
by having an online presence. Visibility is also necessary as a means for networking and for societal
impact – we want people to read our publications, to know about our results, and to extend our findings
beyond the ivory tower.</p>
      <p>However, visibility and online presence have a dark side. In the context of online academic
communication, researchers face the risk of becoming products, where focus is not only on what they
do and on their findings, but also on who they are, something that McMillan Cottom (2015) describes
as the construction of “microcelebrity''. This contributes to the development of a form of academic
capitalism, i.e. “the ways in which knowledge production increasingly embeds universities in the new
economy” [1, see also 9]. Moreover, one problematic aspect is the fact that researchers as professionals
are expected to share information that sometimes touches upon their lives as private persons, in contexts
where they do not have much control. The internet can function as a mechanism of harassment that
shields harassers under “free speech” while offering limited protection to those exposed to risks and
targeted. Social media sites function through a logic that encourages amplification of messages that can
cause harmful content [e.g. 5].</p>
      <p>
        In interdisciplinary contexts such as gender, minority, and environmental research, for instance, the
exposure of researchers is an issue that has been addressed more recently in international research [
        <xref ref-type="bibr" rid="ref10 ref3">10,
3</xref>
        ]. The conducted studies indicate that safety considerations can restrict research and lead to termination
of funded high-quality research in progress. Previous research [
        <xref ref-type="bibr" rid="ref2 ref3 ref5">2, 3, 5</xref>
        ] has also shown that being part
of a marginalised group (e.g. female, Indigenous, black, lgbtq) may increase the level/risk exposure.
Also, the implications of risk and the level of harm is often bigger in these groups than if you are part
of a privileged community (e.g white, cisgender, heterosexual, male). For example, Yelin &amp; Clancy
(2021) argue that threats and abusive comments are often gendered and that derogatory gender-based
terms are used with the purpose of making women feel uncomfortable speaking out [
        <xref ref-type="bibr" rid="ref11 ref3">3,11</xref>
        ]. Vera-Gray
(2017) also points out the added labour of ‘safety work’ as a problematic aspect as it “forms an invisible
backdrop to the methodological decisions of many feminist researchers” [12, p. 62]. Implications of
risks may also manifest as self-censorship as limiting oneself can be a strategy to avoid being put at
risk.
      </p>
      <p>
        Research about risky research in Swedish academic contexts is scarce, and addressed mostly in “grey
literature” (reports, etc.). Surveys indicate that about 50% of respondents have experienced hate speech
and/or toxic speech online [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ]. The Swedish Defence Research Agency recently published two
brief reports addressing the issue of toxic language online. The first report [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] investigates toxic speech
(hate speech and dangerous speech) in Swedish online forums and platforms, where "societal
institutions'' (a category that would include researchers) is mentioned as one category that receives these
types of comments. The second report [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] observes that forms of derogatory comments vary between
men and women. While men are exposed to devaluing comments about lack of competence or
performance in their profession or in general, women are to a greater extent exposed to derogatory
comments about appearance including allegations that a person is ugly or unattractive [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. These reports
underscore the scope of the problem of toxic language use in Swedish online contexts. Studies have
focused on specific professional groups, i.e. journalists, politicians, influencers, comedians and artists
[14; 17]. Exposure in academic contexts and towards researchers is less researched.
      </p>
      <p>
        Another survey in a Swedish context found that people involved in societal issues are particularly
exposed to threats and hatred [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. This applies, for example, to elected representatives, journalists,
artists, opinion leaders, researchers, and representatives of civil society. Academics and journalists who
write about feminism and anti-racism often suffer from online hate in an organized form, something
described in the report as "continuously opposed by more or less organized hatred"[13, p. 24]. These
findings indicate that the democratic conversation is limited and important voices risk being silenced.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Risky research and its implications</title>
      <p>
        While surveillance from the alt-right can intimidate researchers [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], giving researchers the resources
they need to safely conduct research on topics such as gender, race, climate, and politics has far reaching
implications for societal and political spheres and for research ethics and methodologies more generally.
For example, if researchers do not have proper protections in place and avoid researching topics that
put them at risk they may miss opportunities to conduct work that informs public debate and challenges
harmful narratives/paradigms [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Alternatively, if researchers are enabled to publish risky projects
within a research collective and with the institutional support of the university they may be able to
spend more time contributing to research and less time mitigating risk factors.
      </p>
      <p>Potentially, any researcher can be exposed, but recent examples in a Swedish context might help us
to grasp the scope of the issue and, in relation to other international studies, discern some patterns.</p>
      <p>
        In 2018, an employee at the Swedish National Secretariat for Gender Research found a suspicious
object in a bag by the entrance. It later was found to be a dummy device that resembled a dangerous
object, nonetheless we have good reasons to assume that fear and harm was inflicted, both emotionally
and psychologically, to the staff, as well as having implications for gender scholars in general. This
should not be regarded as an isolated event. Gender scholars in Swedish contexts report recurrent
threats, often in conjunction with being published, as being part of everyday life for gender researchers
[
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. This can also be understood as part of the increased threats towards gender research(ers)
worldwide [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Other highly politicised fields and areas of research where dangerous or hate speech
occur include minority studies and research concerned with marginalised groups. In the case of
Indigenous and Sámi research, for instance, the existence of such fields of research in themselves is
questioned, the identity, colonial heritage etc of these groups are questioned [
        <xref ref-type="bibr" rid="ref19 ref20 ref21">19, 20, 21</xref>
        ], and
researchers are in the front line. These examples highlight the increased risk for threats towards certain
groups in relation to gender and race.
      </p>
      <p>
        Gendered and racial violence can also be noted in relation to the phenomena Zoombombing – a
practice of crashing Zoom meetings and posting distressing and/or abusive comments, pictures or
videos during the meeting. This is an increasing problem noted especially during the Covid-19
pandemic when meetings, lectures and seminars were transferred to exclusively online events [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. As
stated by Ling et al. 2021 “Online meeting tools like Zoom and Google Meet have become central to
our professional, educational, and personal lives. This has opened up new opportunities for large scale
harassment”. In Racist Zoombombing [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ], the phenomenon is examined and explained as a form of
racial violence in forms of racist harassment and hate speech. This malicious practice risks disturbing
important academic work and events and, as Nakumara et al. (2021) show, racial abuse during zoom
meetings can cause lasting trauma, anxiety, and anger [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ].
      </p>
      <p>
        Recently, scholars engaged in research related to the Covid-19 pandemics have also witnessed how
they have been the target of massive threats and harassment [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ]. Concerns have been raised by
employers and scientific boards about the implications of such threats. Such a situation not only has
consequences on the researchers’ work environment, but also severe implications for the willingness of
researchers to engage in research of crucial importance for our societies. One example experienced by
colleagues at another department at Umeå University (Sweden) occurred when students were
conducting an online study about perceptions of climate change which revealed their email within the
consent form as a part of standard practice in online research. They subsequently suffered email
harassment from anti-climate activists causing the department to be hesitant to tackle such topics and
nervous about protocols for conducting research which is extremely important to understanding
attitudes towards climate and mitigation strategies. If these experiences continue, issues such as
Covid19 and climate would be less likely to be addressed in the social sciences and humanities literature and
more vulnerable to disinformation from malicious actors.
      </p>
      <p>
        Online harassment can take multiple forms. The threat is not always explicit and a researcher might
experience a feeling of discomfort that may already have an impact on their well-being. For instance,
Massanari (2018) brings attention to the phenomena of sea lioning, a type of trolling or harassment that
consists of pursuing people with persistent requests for evidence or repeated questions, while
maintaining a pretence of civility and sincerity [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Doxing – sharing information publicly for the
purposes of harassment and intimidation - is a more explicit form of threat that has been emphasised
and addressed as an increasing problem, for instance, by the Association of Internet Researchers [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
AoIR also points out potential risk for threats and retaliation should researchers’ identities become
known, in relation to field work, e.g. in research that investigates discussions or phenomena in toxic
environments, for instance 8chan and to some extent Flashback, and/or in studies on anti-gender and
ALT-right groups or other violent online and offline political extremists on social media. Related risks
worth mentioning, which are also highlighted by AoIR [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], are the risks for psychological consequences
involved with research that addresses directly sensitive topics - such as violence, pornography, religious
fanaticism, etc. Although this constitutes another form of risk than the one illustrated above, it is
important to consider what support researchers might need while doing fieldwork in potentially toxic
environments and/or researching potentially heavy and heartbreaking subjects, and also afterwards
while handling such data.
      </p>
      <p>Online threat, harassment, hate, and dangerous speech can obviously have serious implications for
individuals in terms of well-being and work environment. In addition, there are implications for the
employer as well. A university may not be doxed but they may risk losing important research grants,
having bad PR, or a decline in public support (with the rise of anti-intellectual sentiments). They risk
not reaching their potential in research excellence if their employees are being harassed and face this
additional emotional labour, intellectual labour and sometimes even legal battles. Highlighting the
financial and PR risks the university could suffer may be key to get support at the university level.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Strategies</title>
      <p>Based on literature and discussions with peers in international academic contexts, we have identified,
and suggest below, a set of strategies for proactively and actively supporting and protecting scholars
doing risky research, or being exposed for other reasons. This list of suggestions is far from being
exhaustive, but can hopefully spark inspiration and discussion within research environments in need of
addressing issues similar to those addressed in this paper. We also think and hope that these suggestions
can be helpful for supervisors of students at different levels, as exposure and risky research can also be
of immediate interest for students writing essays and theses on sensitive topics and/or doing fieldwork
in toxic environments.</p>
      <p>In all of these suggested strategies, we start from the idea of solidarity, with inspiration from ethics
of care, which starts with “the real experience of being embedded in relationships with uneven power
relations” [25, p. 67; see also 26].</p>
    </sec>
    <sec id="sec-5">
      <title>4.1 Need for intersectional awareness</title>
      <p>
        We see intersectionality as key for understanding how aspects of a person’s identities combined create
different modes of discrimination and privilege in relation to interlocking systems of power [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ].
Intersectional perspectives highlight the complexity of the issue of researchers being increasingly at
risk. It is an approach for understanding who, in what ways, and to what extent researchers, in relation
to gender and gender identity, race, age, functionality, sexuality and so on, can be at risk in digital
contexts with different implications.
      </p>
      <p>
        We also want to highlight the potential different implications of risk in relation to employment status
and academic career level, as junior, non-permanent, or not yet tenured scholars may have more to lose
if their work is put under intense scrutiny in public forums or if they drop an important line of research.
As stated by Massanari (2018) visibility can be profoundly damaging for researchers in precarious
professional situations (eg adjuncts, graduate students, or job applicants) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>Intersectional awareness can help to be proactive. Also, an intersectional awareness helps to address
the problem in a context-based way, rather than a one solution-fits all approach which may be excluding
and/or fail to grasp the whole problem and potential long term implications, especially on an individual
level.</p>
    </sec>
    <sec id="sec-6">
      <title>4.2 Managing Visibility</title>
      <p>While visibility is often central in our academic work, visibility can also be problematic, as we have
stated in this paper. In this section, we will highlight some examples of situations where protecting
one’s identity might be needed, and suggestions for possible strategies.</p>
      <p>Researchers often need (and want) to be part of the digital context they study, which involves
immersion in the context: “social justice–oriented research on “alt-right” adjacent groups, such
entanglements are a form of risk—to students, researchers, their families, and their institutions” [10, p.
320]. Rambukkana (2019) suggests “engaging in covert invisible non-participatory observation” as a
precaution to avoid being put at risk [see also 28].</p>
      <p>
        When conducting online fieldwork, using computational tools etc, a technology awareness is
necessary. VPN, ad blockers (against malicious code), and firewalls can offer some protection [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ]. To
establish support and consultation from an IT security team that can recommend software and hardware
precautions is thus suggested for researchers and students prior to engaging with risky research and
risky fieldwork situations.
      </p>
      <p>Talking in a collective voice can serve as a good strategy in order to avoid individual researchers to
stand on the front line, for instance when presenting sensitive results or in toxic contexts. A research
group, a lab, or a department can be the primary contact when presenting a study or sharing results, or
in publications. This can be especially useful for younger researchers. As senior researchers are more
likely to have better support networks among their peers, as well as experience, and a less vulnerable
employment situation, their engagement in communication of research as a collective voice is
important. Again, due to the increased intersectional risk factors of being targeted, marginalised
researchers can also benefit from being supported and/or represented through their peers. We suggest
research groups early in the research phase discuss potential risk and plan strategies for dealing with
potential unwanted attention.</p>
      <p>
        Another strategy related to the collective voice is to publish anonymously or under a pseudonym
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Publishing anonymously in academic research is not a common practice. Transparency,
responsibility, and accountability are core principles in research that stand in contradiction with
anonymity in publication. However, there have been cases when journals have allowed authors to
publish anonymously, for instance in the case of threat to personal safety [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ]. Such a strategy needs to
be used only upon careful consideration, as responsibility in authorship should be prioritised as much
as possible. Also, there are implications for research impact when a scholar publishes anonymously or
under a pseudonym.
      </p>
      <p>An awareness of potential risks for researchers is also important for personnel working with
communication in universities, at conferences and academic journals. News related to publications are
likely to be spread in external communication including social media to gain positive attention to a
department, journal, or author. Visibility can also become problematic. One way to avoid attracting
unwanted attention is to inform the author of the article beforehand and give them the choice to be
tagged or not, and on which social media platform. Tagging on social media is a great way to attract
attention but it can also attract unwanted attention.</p>
    </sec>
    <sec id="sec-7">
      <title>4.3 Planning online events</title>
      <p>During the Covid-19 pandemic, our research lab relocated all academic and public events online. In
this shift, we have experienced the balance and potential conflict between a need to promote events
while still ensuring the safety of our invited speakers. The issue of safety can be illustrated with this
(slightly edited) quote from one of our speakers in an email conversation about how and where to
market the event: “I have been doing public talks about LGBTQ+ issues for 20 years. It feels really
odd to be worried about visibility online, but this kind of security has been a major issue.” Having a
dialogue beforehand with the invited speaker made it possible for us to decide together on a marketing
plan that fit both parties' need for visibility, but in a manner that provided control and was based on
consent.</p>
      <p>
        Another example of how to ensure security during our events – and more specifically for avoiding
the risk of “Zoombombing”- is to always require pre-sign up to events and not sharing the zoom link
until a day or two before the event. Although anyone can sign up, this is a strategy to make it harder
to get access to the meeting, as it requires more steps to the zoom-link [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. This strategy increases
our control as event organisers. Further, Douglas suggests that “A general principle is that online
activities should be mediated by an IT Security team with appropriate special responsibility and
training” [29, p. 78]. For each of our events we have at least one technician present in the meeting,
ready to act, with attention to any unwanted behaviour in the audience. In events deemed risky, we set
up limits for how the audience can act during the meeting (e.g not turn on mics and cameras without
the hosts’ permission). Having established a strategy for how to plan, promote, and conduct our
events, we as organisers have better control and are better prepared. Hopefully seemingly small
actions provide improved security and protection for our invited speakers and participants.
      </p>
    </sec>
    <sec id="sec-8">
      <title>4.4 The responsibility of institutions</title>
      <p>
        While many universities have established policies to address and deal with explicit harassment,
threats, or violence, the vulnerability experienced by researchers in online environments is often
addressed to a lesser extent. Based on previous research, we argue that individual researchers should
not be left to fend for themselves, as work-related risks should be viewed as a work-related problem
and the responsibility must be placed on an institutional level. Thus, protocols and strategies for
ensuring the security and safety of all employees must be installed and continuously updated to
include strategies for handling online risks and dealing with potential harm and hatred posed in and
through digital contexts. As Yelin &amp; Clancy (2021) state: “Universities have a duty of care to all
researchers, not just during the media work, but before and afterwards. It is important for universities
to take responsibility for the wellbeing of all researchers engaging in impact work which will benefit
the academy” [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. While many universities in a Swedish context have established policies to address
and deal with explicit harassment, threats or violence, there is still a problematic absence of relevant
university policies at many universities as many of these protocols (if existing at all) fail to include
online risks and implications of such risks. They also fail to grasp the complex nature of how both
risks and its implications can differ (and thus may need different solutions) depending on who, what,
and where risks are posed. This is especially important in the case of researchers who are
disadvantageously affected by intersectional implications and converging risk factors. Finally, there is
a need for research to better understand where and why risks may arise, types of risk, and
implications. We strongly recommend research groups and research leaders develop policies and
preparedness for risky research. Resources listed in the next section can be a helpful start.
      </p>
    </sec>
    <sec id="sec-9">
      <title>5. Resources</title>
      <p>
        Already mentioned is the set of “Best practices for conducting risky research and protecting yourself
from online harassment” by Marwick et al (2016) published by Data &amp; Society [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. This resource
addresses primarily young researchers but is relevant for all academics. It also suggests a way to
approach university administrations and suggest modes of action for providing support to their
researchers. A list of additional valuable resources can also be found in this document.
      </p>
      <p>
        Friedman et al. (2016) have developed resources “especially designed for women, people of color,
trans and genderqueer people, and everyone else whose existing oppressions are made worse by digital
violence”[
        <xref ref-type="bibr" rid="ref31">31</xref>
        ]. These are not specific for researchers, but include useful advice about security practices
for online behaviour, documentation of misconduct, and addressing emotional impacts of digital
violence. Other resources include websites specifically addressing online abuse, for instance Crash
Override, a crisis helpline, advocacy group and resource centre that offers helpful tools, educational
materials and DIY security guides [
        <xref ref-type="bibr" rid="ref32">32</xref>
        ]. The Swedish Crime Victim Authority provides information and
advice through their campaign “Do not fall silent” against online hatred, abuse and threats [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ]. Among
other things, the website includes support and guidance on how to file a police rapport.
      </p>
      <p>National, regional or local review boards are also resources toward which researchers can turn for
support when preparing a risky project. Employers’ policies and their security services often include
support in case of threats or violence. Even when those do not directly address the risks, threats, and/or
harassments that researchers meet in online environments, these structures provide a first contact when
in need of support or protection. Occupational health services might also provide support.</p>
    </sec>
    <sec id="sec-10">
      <title>6. Conclusions</title>
      <p>This paper aimed to provide an intersectional account of how researchers experience online risks,
the types of risks they face, who is typically targeted, and how this risk exposure can be a detriment to
research, to research institutions, and to the health of democracy. We argue that intersectional awareness
needs to be the starting point for developing protocols and strategies for avoiding and dealing with risk.</p>
      <p>A second aim in this paper was to identify already available resources and strategies that
researchers and research groups can take to enable an increased intersectional awareness, manage
visibility, and better plan for secure online events to mitigate the costly demands placed on at-risk
researchers in online environments so that their resources can be better allocated to scholarly activities.
We conclude by arguing that institutions themselves should bear a greater responsibility to care for
researchers and take a more active approach in mitigating risk factors. Universities themselves stand to
lose well-funded and high-profile projects if they do not act. It is unknown how many projects to date
have been abandoned by researchers or have not been funded due to inadequate preparedness for online
research. Online methodologies and online risks are increasing at an unprecedented rate. Meanwhile,
online environments are increasingly the context where societal issues and misinformation are
discussed, exacerbated, and propagated. Universities are not immune from these digital threats and, as
online risk increases, universities will need to keep up with technology and develop clear guidelines
and avenues of support to researchers. It is our most fervent hope that Swedish institutions develop
these protocols with an openly intersectional approach. In this era of fake news, tweetstorms, and tik
tok tidbits, democracy itself depends upon academic institutions protecting open dialogue in a
responsible way.</p>
    </sec>
    <sec id="sec-11">
      <title>7. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>McMillan</given-names>
            <surname>Cottom</surname>
          </string-name>
          , T. “'Who Do You Think You Are?': When Marginality Meets Academic Microcelebrity”.
          <source>Ada: A Journal of Gender</source>
          , New Media, and
          <string-name>
            <surname>Technology</surname>
          </string-name>
          ,
          <source>No.7</source>
          . (
          <year>2015</year>
          ). doi:
          <volume>10</volume>
          .7264/N3319T5T.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Vera-Gray</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          “
          <article-title>'talk about a cunt with too much idle time': trolling feminist research”</article-title>
          .
          <source>Feminist review</source>
          .
          <volume>115</volume>
          (
          <issue>1</issue>
          ):
          <fpage>61</fpage>
          -
          <lpage>78</lpage>
          , (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .1057/s41305-017-0038-y
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Yelin</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Clancy</surname>
          </string-name>
          , L. “
          <article-title>'Doing impact work while female: Hate tweets, 'hot potatoes' and having 'enough of experts'”</article-title>
          ,
          <source>European Journal of Women's Studies</source>
          ,
          <volume>28</volume>
          (
          <issue>2</issue>
          ):
          <fpage>175</fpage>
          -
          <lpage>193</lpage>
          , (
          <year>2021</year>
          ).
          <source>doi: 10.1177/1350506820910194</source>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Engebretsen</surname>
            ,
            <given-names>E.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Scientizing</surname>
            <given-names>Gender</given-names>
          </string-name>
          ?
          <article-title>An Examination of Anti-Gender Campaigns on Social Media, Norway</article-title>
          , in: Eslen-Ziya,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Giorgi</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . (eds) Populism and Science in Europe .
          <source>Palgrave Studies in European Political Sociology. Palgrave Macmillan</source>
          ,
          <string-name>
            <surname>Cham</surname>
          </string-name>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Massanari</surname>
            ,
            <given-names>A. L.</given-names>
          </string-name>
          “
          <article-title>Rethinking Research Ethics, Power, and the Risk of Visibility in the Era of the 'Alt-Right' Gaze”</article-title>
          .
          <source>Social media + society. 4</source>
          (
          <issue>2</issue>
          ), (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .1177/2056305118768302.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>[6] franzke, a</article-title>
          . s.,
          <string-name>
            <surname>Bechmann</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zimmer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ess</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          &amp;
          <article-title>the Association of Internet Researchers</article-title>
          .
          <source>Internet Research: Ethical Guidelines 3.0</source>
          , (
          <year>2020</year>
          ). URL: https://aoir.org/reports/ethics3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Marwick</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blackwell</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Lo</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          .
          <article-title>Best Practices for Conducting Risky Research and Protecting Yourself from Online Harassment (Data &amp; Society Guide)</article-title>
          . New York: Data &amp; Society Research Institute (
          <year>2016</year>
          ). URL: https://datasociety.net/library/best
          <article-title>-practices-for-conductingrisky-research/.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Marwick</surname>
            ,
            <given-names>A. E.</given-names>
          </string-name>
          &amp; boyd, d. “
          <article-title>Networked privacy: How teenagers negotiate context in social media”</article-title>
          ,
          <source>New Media &amp; Society</source>
          ,
          <volume>16</volume>
          (
          <issue>7</issue>
          ):
          <fpage>1051</fpage>
          -
          <lpage>1067</lpage>
          (
          <year>2014</year>
          , p.
          <fpage>1052</fpage>
          ). doi:
          <volume>10</volume>
          .1177/1461444814543995.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Berman</surname>
            ,
            <given-names>E. P. Creating</given-names>
          </string-name>
          <article-title>the market university: How academic science became an economic engine</article-title>
          . Princeton University Press,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Rambukkana</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          “
          <article-title>The Politics of Gray Data: Digital Methods, Intimate Proximity, and Research Ethics for Work on the 'Alt-Right'”. Qualitative inquiry</article-title>
          .
          <volume>25</volume>
          (
          <issue>3</issue>
          ):
          <fpage>312</fpage>
          -
          <lpage>323</lpage>
          (
          <year>2019</year>
          ). doi:
          <volume>10</volume>
          .1177/1077800418806601.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Sobieraj</surname>
            <given-names>S.</given-names>
          </string-name>
          “
          <article-title>Bitch, slut, skank, cunt: Patterned resistance to women's visibility in digital publics”</article-title>
          .
          <source>Information, Communication and Society</source>
          <volume>21</volume>
          (
          <issue>11</issue>
          ):
          <fpage>1700</fpage>
          -
          <lpage>1714</lpage>
          (
          <year>2018</year>
          ). doi:
          <volume>10</volume>
          .1080/1369118X.
          <year>2017</year>
          .
          <volume>1348535</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Vera-Gray</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          “
          <article-title>'talk about a cunt with too much idle time': trolling feminist research”</article-title>
          .
          <source>Feminist review</source>
          .
          <volume>115</volume>
          (
          <issue>1</issue>
          ):
          <fpage>61</fpage>
          -
          <lpage>78</lpage>
          , (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Svensson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , Bjö rkenfeldt,
          <string-name>
            <surname>O.</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          ̊ strö m,
          <string-name>
            <given-names>F.</given-names>
            , &amp;
            <surname>Dahlstrand</surname>
          </string-name>
          ,
          <string-name>
            <surname>K.</surname>
          </string-name>
          <article-title>Nä that och demokratiskt deltagande - en kunskapsö versikt</article-title>
          .
          <source>Brottsoffermyndigheten</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Fernquist</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaati</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pelzer</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lindberg</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akrami</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cohen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Pollack Sarnecki</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <article-title>Det digitala hatets karaktä r: En studie av hat mot kvinnor och mä n i utsatta yrkesgrupper</article-title>
          .
          <source>FOI Memo 7429</source>
          ,
          <string-name>
            <surname>Totalfo</surname>
          </string-name>
          <article-title>̈ rsvarets forskningsinstitut (FOI) (</article-title>
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Kaati</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pelzer</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Asplund Cohen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wallgren</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Akrami</surname>
            ,
            <given-names>N</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Yourstone</surname>
            <given-names>J.</given-names>
          </string-name>
          <article-title>Toxiskt språk i svenska digitala miljöer</article-title>
          .
          <source>FOI Memo 7740</source>
          ,
          <string-name>
            <surname>Totalfo</surname>
          </string-name>
          <article-title>̈ rsvarets forskningsinstitut (FOI) (</article-title>
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Kaati</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Asplund Cohen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pelzer</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wallgren</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Akrami</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Yourstone</surname>
            ,
            <given-names>J. Könsskillnader</given-names>
          </string-name>
          <article-title>i utsatthet för toxiskt språk online</article-title>
          .
          <source>Memo</source>
          <volume>7741</volume>
          ,
          <string-name>
            <surname>Totalfo</surname>
          </string-name>
          <article-title>̈ rsvarets forskningsinstitut (FOI) (</article-title>
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <article-title>Ministry of Culture. “Till det fria ordets fö rsvar - å tgä rder mot utsatthet fö r hot och hat bland journalister, fö rtroendevalda och konstnä rer”</article-title>
          .
          <source>Regeringskansliet</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Lofrup</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          “
          <article-title>Gender researcher: threats and hate are part of everyday life” (</article-title>
          <year>2019</year>
          ). URL: https://www.staff.lu.se/article/gender-researcher
          <article-title>-threats-and-hate-are-part-everyday-life.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Carlson</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Fazer R. Indigenous</surname>
          </string-name>
          Digital Life:
          <source>The Practice and Politics of Being Indigenous on Social Media</source>
          . Springer International Publishing AG (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Katz</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keeley</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spears</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bates</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Swirski</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Taddeo</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <article-title>Research on youth exposure to, and management of, cyberbullying incidents in</article-title>
          <source>Australia: Synthesis Report</source>
          ,
          <year>2014</year>
          . URL: https://www.communications.gov.au/publications/publications/research-youth
          <article-title>-exposure-andmanagement-cyber-bullying-incidents-australia-synthesis-</article-title>
          <string-name>
            <surname>report-</surname>
          </string-name>
          june-2014.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Mobin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>C. X.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Neudorf</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          “
          <article-title>Cybervictimization among pre-adolescents in a community-based sample in Canada: Prevalence and predictors”</article-title>
          .
          <source>Canadian Journal of Public Health</source>
          ,
          <volume>108</volume>
          (
          <issue>5-6</issue>
          ):
          <fpage>e475</fpage>
          -
          <lpage>e481</lpage>
          (
          <year>2017</year>
          ). doi:
          <volume>10</volume>
          .17269/cjph.108.5878
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Ling</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Balci</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blackburn</surname>
            <given-names>J.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Stringhini</surname>
            ,
            <given-names>G. A First</given-names>
          </string-name>
          <string-name>
            <surname>Look at Zoombombing</surname>
          </string-name>
          ,
          <source>2021 IEEE Symposium on Security and Privacy (SP)</source>
          (
          <year>2021</year>
          , pp.
          <fpage>1452</fpage>
          -
          <lpage>1467</lpage>
          ), doi: 10.1109/SP40001.
          <year>2021</year>
          .
          <volume>00061</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Nakamura</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stiverson</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Lindsey</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Racist Zoombombing</surname>
          </string-name>
          , 1st ed., New York: Routledge (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Torkelsson</surname>
            ,
            <given-names>A-C.</given-names>
          </string-name>
          “
          <article-title>Forskare vittnar om hot efter att ha pratat om covid-19 med media”</article-title>
          .
          <source>Läkartidningen</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <article-title>[25] franzke, a</article-title>
          . s. ”Feminist Research Ethics”,
          <source>IRE 3</source>
          .
          <issue>0 Companion 6</issue>
          .3, Association of Internet Researchers (
          <year>2020</year>
          ). URL: https://aoir.org/reports/ethics3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Suomela</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chee</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Berendt</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Rockwell</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <article-title>Applying an Ethics of Care to Internet Research: Gamergate and Digital Humanities. Digital Studies/Le champ numé rique, 9(1), 4</article-title>
          . (
          <year>2019</year>
          , p.
          <fpage>2</fpage>
          ). https://www.digitalstudies.org/articles/10.16995/dscn.302/.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Crenshaw</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          “
          <article-title>Mapping the Margins: Intersectionality, Identity Politics, and Violence against Women of Color.” Stanford law review 43(6</article-title>
          ), (
          <year>1991</year>
          :
          <fpage>1241</fpage>
          -
          <lpage>1299</lpage>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Pollock</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          “
          <article-title>Researching white supremacists online: Methodological concerns of researching hate online”</article-title>
          .
          <source>Internet Journal of Criminology</source>
          , (
          <year>2009</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Douglas</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          “Operational Security: Central Considerations”,
          <source>IRE 3</source>
          .
          <issue>0 Appendices 7</issue>
          .1, Association of Internet Researchers (
          <year>2020</year>
          ). URL: https://aoir.org/reports/ethics3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30] Council of Science Editors,
          <source>Editorial Policy Committee, Authorship and Authorship Responsibilities</source>
          (
          <year>2020</year>
          ). URL: https://www.councilscienceeditors.org/resource-library/editorialpolicies/white
          <article-title>-paper-on-publication-ethics/2-2-authorship-and-authorship-responsibilities/</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <surname>Friedman</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sarkeesian</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Sherman</surname>
            ,
            <given-names>R. B. “Speak</given-names>
          </string-name>
          <string-name>
            <surname>Up</surname>
          </string-name>
          &amp;
          <article-title>Stay Safe(r): A Guide to Protecting Yourself From Online Harassment” (</article-title>
          <year>2016</year>
          ). URL: https://onlinesafety.feministfrequency.com/en/#about-us.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>[32] http://www.crashoverridenetwork.com/resources.html</mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>[33] https://www.tystnainte.se/en/</mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>