<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <article-id pub-id-type="doi">10.1145/93385.93442</article-id>
      <title-group>
        <article-title>RNN Generalization to Omega-Regular Languages</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Charles Pert</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dalal Alrajeh</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alessandra Russo</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Imperial College London</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>4963</volume>
      <fpage>0009</fpage>
      <lpage>0009</lpage>
      <abstract>
        <p>Büchi automata (BAs) recognize -regular languages defined by formal specifications like linear temporal logic (LTL) and are commonly used in the verification of reactive systems. However, BAs face scalability challenges when handling and manipulating complex system behaviors. As neural networks are increasingly used to address these scalability challenges in areas like model checking, investigating their ability to generalize beyond training data becomes necessary. This work presents the first study investigating whether recurrent neural networks (RNNs) can generalize to -regular languages derived from LTL formulas. We train RNNs on ultimately periodic -word sequences to replicate target BA behavior and evaluate how well they generalize to out-of-distribution sequences. Through experiments on LTL formulas corresponding to deterministic automata of varying structural complexity, from 3 to over 100 states, we show that RNNs achieve high accuracy on their target -regular languages when evaluated on sequences up to 8× longer than training examples, with 92.6% of tasks achieving perfect or near-perfect generalization. These results establish the feasibility of neural approaches for learning complex -regular languages, suggesting their potential as components in neurosymbolic verification methods.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;recurrent neural networks</kwd>
        <kwd>omega-regular languages</kwd>
        <kwd>linear temporal logic</kwd>
        <kwd>büchi automata</kwd>
        <kwd>length generalization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Linear Temporal Logic (LTL) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] formulas specify properties of system execution traces through
regular languages, which are composed of infinitely long sequences called -words. While regular
languages are recognized by finite automata, -regular languages require Büchi automata (BAs) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
See Figure 1 for an example of a deterministic Büchi automaton (DBA). For a detailed introduction to
these concepts, we refer the reader to [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
      </p>
      <p>
        While BAs provide exact solutions, they can become computationally expensive to handle and
manipulate when representing complex behaviors. Recently, neurosymbolic methods have been used in
verification contexts such as model checking [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], areas that traditionally rely on BAs. Characterizing
whether neural networks can recognize -regular languages is a step toward enabling the development
of additional approaches. Recent studies have shown that recurrent neural networks (RNNs) have the
ability to generalize to the recognition of regular languages [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8">5, 6, 7, 8</xref>
        ], but this has not yet been shown
specifically for -regular languages. Related works [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ] have used graph neural networks to analyze
BA properties like emptiness checking and [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] demonstrated that Transformers [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] can generate
satisfying -words for LTL formulas, yet the problem of generalization to the recognition of -regular
languages remains open.
      </p>
      <p>
        Extending the existing work on regular language generalization to -regular languages requires
handling two practical challenges: (1) encoding: representing infinite sequences with finite-length
sequences; (2) labeling sequences: computing acceptance labels for large batches of sequences. While
ifnite automata accept a word when it terminates in an accepting state, BAs require -words to traverse
accepting states infinitely often, meaning RNNs must learn to recognize eventually periodic behavior
instead of just reaching an accepting state. We address challenge (1) by using ultimately periodic (UP)
-words, which uniquely characterize their -regular languages [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. This representation enables us to
investigate whether RNNs can approximate the symbolic acceptance mechanisms of BAs.
0
a ∧ b
¬a
¬a ∨ b
1
3
a ∧ ¬b
a ∧ ¬b
b
⊤
¬b
2
      </p>
      <p>
        While existing model checking tools like Spot [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] can compute acceptance labels for specific
sequences, this approach quickly becomes impractical for the large datasets required for neural network
training and does not address sequence generation. Instead, we use Spot to construct a BA from each
LTL formula, generate sequences using the BA representation and simulate acceptance by directly
processing the generated sequences through the BA. We restrict our approach to DBAs to simplify this
acceptance check, limiting our study to recurrence properties [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. While this means that this study
does not cover persistence properties, recurrence properties cover a large number of properties and
are commonly used in verification. Using BAs as data generators provides some control over sampling
diversity, which is essential since random sampling can induce data imbalance problems when accepted
(or rejected) traces are rare.
      </p>
      <p>
        This work establishes the feasibility of RNNs generalizing to the recognition of -regular languages
while identifying challenges for future research. Our investigation complements ongoing
neurosymbolic advances in verification, for example, neural certificates for model checking [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], neural circuit
synthesis [16], as well as neural specification mining, learning finite automata or LTL  [17] from
traces [18, 19, 20, 21].
      </p>
      <p>The main contributions of this work are:
• We present the first empirical evidence that RNNs achieve high accuracy on -regular languages
when evaluated at lengths up to 8× longer than the sequences in their training distribution.1
• We provide an analysis of RNN generalization, showing that generalization is not limited to toy
-regular languages and is robust across DBAs with over 100 states. We also show that the model
complexity is correlated with the complexity of the BA recognizing the -regular language.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Method</title>
      <p>
        Practical training of RNNs on -regular languages requires a finite representation of -words and an
eficient method for computing acceptance labels for generated sequences given the infinite acceptance
condition of BAs. Our approach resolves the finite encoding problem by only using UP -words of an
-regular language, , where  is a finite prefix and  is an infinitely repeating sufix; these -words
uniquely characterize the language [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. We encode  as $ [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The alphabet size of a DBA is
2| | + 1, where | | is the number of propositions present in the LTL formula used to construct the
DBA. Each symbol in the alphabet represents either an assignment to all propositions or the separator
symbol $.
      </p>
      <p>
        Importantly, the encoding $ establishes a bijection between UP -words in the target -regular
language and words in a derived regular language. The DBA can be algorithmically reconstructed from
1Code available at: https://github.com/pertcj/omega-generalization
the finite automaton recognizing this regular language [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Therefore, while our RNNs are learning
regular languages (and leveraging their established ability [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">6, 7, 5</xref>
        ]), they are learning canonical regular
representations of the target DBA that preserve all structural information for reconstruction.
      </p>
      <p>To determine the label of a $ sequence, we simulate DBA behavior. Simulating the finite prefix 
yields the state in which  terminates. For the sufix , we compute the state transition matrix induced
by reading  and use matrix exponentiation to determine reachability. The sequence is accepted if
repeated application of  can reach a cycle containing an accepting state.</p>
      <p>We illustrate how we sample $ sequences with fixed length  from a DBA. We first sample the
position  of $ uniformly between 1 and  − 1 and then sample  with length  − 1 and  with length
 − . We sample  and  by uniformly sampling valid paths in the DBA. Many DBAs exhibit strong
acceptance biases that skew random sampling toward rejection or acceptance. For example, the presence
of accepting or rejecting sink states (states that can only transition back to themselves) can dominate
uniform sampling. We address this through targeted sampling strategies to improve the balance of our
sampled sequences: (1) we oversample sequences, determine their labels, and selectively filter them,
targeting a balanced class distribution; (2) when sampling accepted sequences, we exclude transitions
to rejecting sink states during sampling (and vice versa for rejected sequences); (3) when sampling
rejected sequences, we prevent transitions to accepting states within the sufix  because if  contains a
state transition to an accepting state, the resulting -word is likely to be accepted. It is still possible to
sample equivalent rejected sequences as this constraint is not applied to . By controlling the sampling
of accepted and rejected sequences separately, the resulting dataset is more balanced compared to
uniform sampling of sequences.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Experiments</title>
      <p>Our experiments aim to answer the following research questions:</p>
      <p>RQ1: Can RNNs generalize to the recognition of sequences from -regular languages when trained
only on short UP -words?
RQ2: Do structural properties of the underlying DBAs influence (a) generalization performance and
(b) learned model complexity?</p>
      <p>To answer these questions, we use the -regular languages associated with two well-known LTL
benchmarks. The first benchmark ( alaska_lift) [22] consists of two encodings of safe lift behaviors
(we use the variants with bug-fixes [ 23]) parameterized by the number of floors: encoding (a) uses a
linear number of propositions per floor and encoding (b) uses a logarithmic number of propositions per
lfoor. The second benchmark ( acacia_example) [24] consists of 25 formulas specifying the behavior
of arbiters and trafic light controllers [ 25]. For this benchmark, we use the negated versions of the
formulas, which in our setting only flip the labels of the generated sequences.</p>
      <p>
        For each LTL formula, we generate its corresponding DBA using Spot [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. During training, we
generate sequences on-the-fly with uniformly sampled lengths between 2 and 64. Once sampled, we
transform sequences into one-hot encoded symbol embeddings for RNN processing. Our test data
consists of 512 sampled sequences for each length between 2 and 512. We apply a 10-minute timeout for
the construction of each DBA. We excluded automata with more than 200 states due to computationally
expensive sequence generation, not RNN training. This constraint caused the exclusion of lift formulas
for 4 or more floors and 2 formulas from acacia_example. The test data was balanced for the lift
formulas and for 20 of the acacia_example formulas. However, the training data was not balanced for
the lift formulas and the 3 acacia_example formulas with unbalanced test data.
      </p>
      <p>
        Each experiment trains a single-layer vanilla RNN [26] with a hidden dimension of 256 (consistent
with [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]), batch size 256 for 1 × 105 steps. We use linear warmup for the learning rate [27] from 1 × 10− 8
to 1 × 10− 3 at 20% of the training steps, the AMSGrad optimizer [28], and 2 regularization with weight
5 × 10− 4. We minimize cross-entropy loss. Results are from a single seed due to computational budget.
All experiments were conducted using an NVIDIA RTX 6000 Ada GPU. We measure in-distribution (ID)
accuracy, the mean accuracy of the test data in the training length range, and out-of-distribution (OOD)
accuracy, the mean accuracy of the test data outside of the training length range.
      </p>
      <p>To address RQ1, we assess whether RNNs can learn to classify -words by measuring performance on
the test data, reporting ID accuracy (lengths 2-64) and OOD accuracy (lengths 65-512) in a summary table.
To address RQ2a, we plot the relationship between the number of states in the DBAs and generalization
performance. To address RQ2b, we plot the number of states against the parameter norm (2 norm)
of the trained models (post-training). We report the Pearson correlation coeficients [ 29] and their
statistical significance between the number of states and both OOD accuracy and the trained models’
parameter norms to assess their linear correlations.
3.1. Results
In Table 1, we present the results answering RQ1. The majority of tasks (92.6%) had perfect or
nearperfect generalization. Two tasks had OOD accuracies of 81.5% and 76.8%.</p>
      <p>In Figure 2, we present the plots comparing the number of states with OOD accuracy (Figure 2a) and
the models’ parameter norms (Figure 2b). Figure 2a reveals that generalization performance remains
consistently high across automata of varying complexity, with 85.2% of tasks achieving perfect accuracy
regardless of state count, indicating that RNN capacity is robust to the structural complexity of the
underlying BA. OOD accuracy showed no significant correlation with the number of states in the DBA
( = 0.115,  = 0.567), suggesting that generalization performance does not necessarily degrade with
increasing DBA complexity. Figure 2b shows a strong positive correlation between the number of states
in the DBA and the parameter norm of the trained models ( = 0.880,  &lt; 0.001), indicating that model
complexity aligns with the complexity of the target -regular languages.</p>
      <p>)100
(%95
.
cc 90
AD 85
O 80
O
75
0
25 50 75</p>
      <sec id="sec-3-1">
        <title>Number of States</title>
        <p>100
25 50 75</p>
      </sec>
      <sec id="sec-3-2">
        <title>Number of States</title>
        <p>(a) OOD accuracy (acc.) remains consistently high (b) Parameter (param.) norm of the trained RNNs
across DBAs with large numbers of states. shows a strong positive correlation with the
number of states in the DBAs.</p>
        <p>We now examine the two tasks that exhibited poor generalization to understand their failure modes.
During training, both tasks demonstrate unstable validation patterns that predict their generalization</p>
      </sec>
      <sec id="sec-3-3">
        <title>Lift</title>
      </sec>
      <sec id="sec-3-4">
        <title>Acacia 100</title>
      </sec>
      <sec id="sec-3-5">
        <title>Lift</title>
      </sec>
      <sec id="sec-3-6">
        <title>Acacia</title>
        <p>r40
m
o
N
.
m30
a
r
a
P
0
failures (see Figure 3a). Acacia 13 achieves 100% validation accuracy initially but then degrades to
random chance (50%), while Acacia 22 oscillates between 100% and 50% throughout training, indicating
that neither model converges to a stable representation of the ground truth -regular language. This
instability corresponds to poor length generalization, as shown in Figure 3b, where both tasks achieve
perfect accuracy on shorter OOD sequences before experiencing sharp degradation to 50% at longer
lengths (degradation begins at 345 and 300 for Acacia 13 and Acacia 22, respectively). This degradation
suggests that these models converged to local minima, achieving perfect in-distribution accuracy
without learning the ground truth -regular language. The DBAs of both failure cases contained
accepting sink states. If a sequence reaches such a state, the RNN must encode this information when
processing all subsequent symbols. Further investigation is needed to confirm this hypothesis.
1.0
y
c
a
rcu0.8
c
A
l.a0.6
V</p>
      </sec>
      <sec id="sec-3-7">
        <title>Acacia 13</title>
      </sec>
      <sec id="sec-3-8">
        <title>Acacia 22</title>
        <p>0.0
0.5
Train Step
1.0
×105</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion and Future Work</title>
      <p>Our experiments demonstrate that RNNs generalize to the recognition of -regular languages from
short UP -words. Across 27 tasks with a diverse range of system behaviors, we achieved perfect or
near-perfect generalization in 92.6% of cases when testing on sequences up to 8 times longer than
training examples. This finding remained consistent across automata ranging from 3 to 105 states,
demonstrating that the approach scales to realistic verification problems. These results provide a
foundation for developing diferentiable Büchi automata , components within neurosymbolic systems.
These components might behave as monitors in reinforcement learning or enable gradient-based search
in model checking.</p>
      <p>
        This work has several limitations that future research should address. Sampling sequences from
complex automata (&gt;200 states) proved impractical despite our attempts to speed up the process.
Our experiments were also restricted to DBAs as a first step; notably, DBAs cannot represent all
-regular languages [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Further investigation of the failure cases is necessary to develop improved
training methods. Despite these limitations, this work provides strong evidence that neural approaches
could further enhance neurosymbolic methods by ofering more scalable, diferentiable alternatives
to traditional automata-theoretic methods. Exploring interpretability by adapting existing automata
extraction techniques (e.g. [30, 31]) may enable verification of the learned representations required for
safety-critical applications.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work was supported by the UK EPSRC grant 2760033. The authors would like to thank Frederik
Kelbel for reading the paper and the reviewers for their constructive feedback.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Claude Sonnet 4 in order to: Paraphrase and
reword. After using this tool, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Pnueli</surname>
          </string-name>
          ,
          <article-title>The Temporal Logic of Programs</article-title>
          ,
          <source>in: 18th Annual Symposium on Foundations of Computer Science</source>
          (sfcs
          <year>1977</year>
          ),
          <year>1977</year>
          , pp.
          <fpage>46</fpage>
          -
          <lpage>57</lpage>
          . doi:
          <volume>10</volume>
          .1109/SFCS.
          <year>1977</year>
          .
          <volume>32</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J. R.</given-names>
            <surname>Büchi</surname>
          </string-name>
          ,
          <article-title>On a Decision Method in Restricted Second Order Arithmetic</article-title>
          ,
          <year>1990</year>
          , pp.
          <fpage>425</fpage>
          -
          <lpage>435</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-1-
          <fpage>4613</fpage>
          -8928-6\_
          <fpage>23</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Baier</surname>
          </string-name>
          , Christel and Katoen,
          <string-name>
            <surname>Joost-Pieter</surname>
          </string-name>
          ,
          <source>Principles of Model Checking</source>
          ,
          <year>2008</year>
          . URL: https://mitpress. mit.edu/9780262026499/principles-of-model-checking/.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Mirco</given-names>
            <surname>Giacobbe</surname>
          </string-name>
          and
          <string-name>
            <given-names>Daniel</given-names>
            <surname>Kroening</surname>
          </string-name>
          and
          <string-name>
            <given-names>Abhinandan</given-names>
            <surname>Pal</surname>
          </string-name>
          and
          <string-name>
            <given-names>Michael</given-names>
            <surname>Tautschnig</surname>
          </string-name>
          ,
          <source>Neural Model Checking, in: The Thirty-eighth Annual Conference on Neural Information Processing Systems</source>
          ,
          <year>2024</year>
          . URL: https://openreview.net/forum?id=dJ9KzkQ0oH.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Alexandra</given-names>
            <surname>Butoi</surname>
          </string-name>
          and
          <article-title>Ghazal Khalighinejad and Anej Svete and Josef Valvoda and Ryan Cotterell and Brian DuSell, Training Neural Networks as Recognizers of Formal Languages</article-title>
          ,
          <source>in: The Thirteenth International Conference on Learning Representations</source>
          ,
          <year>2025</year>
          . URL: https://openreview. net/forum?id=aWLQTbfFgV.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Delétang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ruoss</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Grau-Moya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Genewein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. K.</given-names>
            <surname>Wenliang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Catt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cundy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hutter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Legg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Veness</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Ortega</surname>
          </string-name>
          ,
          <article-title>Neural Networks and the Chomsky Hierarchy</article-title>
          ,
          <source>in: 11th International Conference on Learning Representations</source>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Svete</surname>
          </string-name>
          ,
          <article-title>Anej and Chan, Robin and Cotterell, Ryan, On Eficiently Representing Regular Languages as RNNs, in: Findings of the Association for Computational Linguistics: ACL 2024, Association for Computational Linguistics</article-title>
          ,
          <year>2024</year>
          , pp.
          <fpage>4118</fpage>
          -
          <lpage>4135</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .findings-acl.
          <volume>244</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .findings-acl.
          <volume>244</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Merrill</surname>
            , William and Weiss, Gail and Goldberg, Yoav and Schwartz, Roy and Smith,
            <given-names>Noah A.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Yahav</surname>
          </string-name>
          ,
          <string-name>
            <surname>Eran</surname>
            ,
            <given-names>A Formal</given-names>
          </string-name>
          <article-title>Hierarchy of RNN Architectures, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</article-title>
          , Association for Computational Linguistics,
          <year>2020</year>
          , pp.
          <fpage>443</fpage>
          -
          <lpage>459</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .acl-main.
          <volume>43</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          . acl-main.
          <volume>43</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Stammet</surname>
          </string-name>
          ,
          <article-title>Christophe and Dotti, Prisca and Ultes-Nitsche, Ulrich and Fischer, Andreas, Analyzing Büchi Automata with Graph Neural Networks</article-title>
          ,
          <source>arXiv preprint arXiv:2206.09619</source>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Stammet</surname>
          </string-name>
          ,
          <article-title>Christophe and Ultes-Nitsche, Ulrich and Fischer, Andreas, Universality of Büchi Automata: Analysis with Graph Neural Networks</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Hahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Schmitt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. U.</given-names>
            <surname>Kreber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. N.</given-names>
            <surname>Rabe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Finkbeiner</surname>
          </string-name>
          ,
          <article-title>Teaching Temporal Logics to Neural Networks</article-title>
          , in: International Conference on Learning Representations,
          <year>2021</year>
          . URL: https://openreview.net/forum?id=dOcQK-f4byz.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , L. u. Kaiser,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          , Attention Is All You Need,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>30</volume>
          ,
          <year>2017</year>
          . URL: https://proceedings.neurips.cc/paper_files/paper/2017/file/ 3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Calbrix</surname>
          </string-name>
          , Hugues and Nivat, Maurice and Podelski, Andreas, Ultimately Periodic Words of Rational -Languages, in
          <source>: Mathematical Foundations of Programming Semantics</source>
          ,
          <year>1994</year>
          , pp.
          <fpage>554</fpage>
          -
          <lpage>566</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Duret-Lutz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Renault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Colange</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Renkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. G.</given-names>
            <surname>Aisse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Schlehuber-Caissier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Medioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Martin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dubois</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gillard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lauko</surname>
          </string-name>
          ,
          <source>From Spot 2</source>
          .0 to Spot 2.10:
          <article-title>What's New?</article-title>
          ,
          <source>in: Proceedings of the 34th International Conference on Computer Aided Verification (CAV'22)</source>
          , volume
          <volume>13372</volume>
          of Lecture Notes in Computer Science,
          <year>2022</year>
          , pp.
          <fpage>174</fpage>
          -
          <lpage>187</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -13188-2\_9.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Manna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pnueli</surname>
          </string-name>
          ,
          <article-title>A Hierarchy of Temporal Properties (invited paper,</article-title>
          <year>1989</year>
          ), in: Proceedings of
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>