<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Normalization for Swiss German</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pius von Däniken</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manuela Hürlimann</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mark Cieliebak</string-name>
          <email>ciel@zhaw.ch</email>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Lugano, Switzerland</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Centre for Artificial Intelligence, Zurich University of Applied Sciences (ZHAW)</institution>
          ,
          <addr-line>Winterthur</addr-line>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Written Swiss German is not standardized and varies across authors and their dialects and its use is almost exclusively constrained to communication on social media or via text messaging. Many corpora will therefore contain many distinct surface forms for the same word which can make their analysis challenging. It is therefore desirable to be able to normalize them to a single common surface form. We collected Swiss German utterances from social media and two annotators mapped every token to a corresponding form in Standard German. The task is to build models that can perform such a mapping automatically. This is diferent from translation since the resulting normalized utterance will in general not be grammatically correct Standard German as word order is preserved. A similar efort has previously been undertaken for text messages by the SMS4Science project. There is also a recent related shared task on lexical normalization of other languages at WNUT2021 workshop. During the shared task session, we presented the shared task dataset, how it was created, and gave an overview of ∗Corresponding author.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>the annotation tool. Since there were no participants this year, we presented the results of a naive baseline on the task dataset.</p>
    </sec>
  </body>
  <back>
    <ref-list />
  </back>
</article>