=Paper=
{{Paper
|id=Vol-3749/genesy-03
|storemode=property
|title=Modular Design Patterns for Generative Neuro-Symbolic Systems
|pdfUrl=https://ceur-ws.org/Vol-3749/genesy-03.pdf
|volume=Vol-3749
|authors=Maaike H. T. de Boer,Quirine S. Smit,Michael van Bekkum,André Meyer-Vitali,Thomas Schmid
|dblpUrl=https://dblp.org/rec/conf/esws/BoerSBMS24
}}
==Modular Design Patterns for Generative Neuro-Symbolic Systems==
Modular Design Patterns for Generative
Neuro-Symbolic Systems
Maaike H. T. de Boer1,∗ , Quirine S. Smit1,∗ , Michael van Bekkum1 ,
André Meyer-Vitali2 and Thomas Schmid3,4,5
1
TNO, dep. Data Science, The Hague, The Netherlands
2
Deutsches Forschungszentrum für Künstliche Intelligenz GmbH (DFKI), Saarbrücken, Germany
3
Martin Luther University Halle-Wittenberg, Halle (Saale), Germany
4
Lancaster University in Leipzig, Leipzig, Germany
5
Leipzig University, Leipzig, Germany
Abstract
Developing systems that are able to generate novel outputs is one of the dominating trends in current
Artificial Intelligence (AI) research. Both capabilities and availability of such generative systems, in
particular of so-called Large Language Models (LLMs), have been exploding in recent years. While
Neuro-Symbolic generative models offer advantages over purely statistical generative models, it is
currently difficult to compare the different ways in which the training, fine-tuning and usage of the
growing variety of such approaches is carried out. In this work, we use the modular design patterns and
Boxology language of van Bekkum et al for this purpose and extend those to enable the representation of
generative models, specifically LLMs. These patterns provide a general language to describe, compare and
understand the different architectures and methods used. Our main aim is to support better understanding
of generative models as well as to support engineering of LLM-based systems. In order to demonstrate
the usefulness of this approach, we explore generative Neuro-Symbolic architectures and approaches as
use cases for these generative design patterns.
Keywords
design patterns, neuro-symbolic AI, generative models, Large Language Models
1. Introduction
Recently, Artificial Intelligence (AI) has taken a leap in the form of generative models. Promi-
nently, multimodal statistical models, such as DALL-E [1] and Stable Diffusion [2] have changed
the world of image generation, and with the release of OpenAI’s ChatGPT system1 , the world
of text generation has changed forever. Targeting text generation tasks in particular, both the
development and the number of Large Language Models (LLMs) has increased enormously.
Currently, many different generative models are popping up, both open-source and proprietary
[3]. Moreover, due to open challenges of LLMs, such as hallucination [4], explainability [5] and
trustworthiness, novel Neuro-Symbolic generative approaches have emerged [6, 7].
GeNeSy’24: First International Workshop on Generative Neuro-Symbolic AI, May 2024, Hersonissos, Crete, Greece
∗
Corresponding author; both authors contributed equally.
Envelope-Open maaike.deboer@tno.nl (M. H. T. d. Boer); quirine.smit@tno.nl (Q. S. Smit); andre.meyer-vitali@dfki.de
(A. Meyer-Vitali); thomas.schmid@medizin.uni-halle.de (T. Schmid)
Orcid 0000-0002-2775-8351 (M. H. T. d. Boer); 0000-0002-5242-1443 (A. Meyer-Vitali)
© 2024 Copyright (c) 2024 for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
1
https://chat.openai.com/chat
CEUR
ceur-ws.org
Workshop ISSN 1613-0073
Proceedings
Not only several LLMs, but also a large number of so-called foundation models dealing with
various input and output modalities have entered the scene in recent years. Due to the quantity
and diversity of emerging generative techniques, it becomes more and more challenging to keep
track of the ever-growing variety of models with different architectures and capabilities. One
of the solutions to tackle this issue is to create a high-level conceptual framework to discuss,
compare, configure and combine different models is using a Boxology. The Boxology started in
the field of Neuro-Symbolic systems, by Van Harmelen and Ten Teije [8] in 2019. This work is
extended in 2021 by van Bekkum et al. [9] by providing a taxonomically organised vocabulary
to describe both processes and data structures used in hybrid systems.
Here, we propose to use and extend the Boxology to gain insights in a variety of generative
models, specifically on LLMs. To this end, we test validity and usefulness of the Boxology in
this field on example architectures and applications, such as ChatGPT, KnowGL, GENOME and
Logic-LM. Our modular approach supports new architectures and engineering approaches to
systems based on generative AI models. Our pattern extensions promote transparency and
trustworthiness in system design, by providing interpretable, high-level component descriptions
of generative AI models.
The rest of the paper is organized as follows. In the next section, we give a more detailed
overview of the Boxology. In the third section, we propose to extend the Boxology by three
novel patterns in order to be able to handle generative models. In section 4, we dive into specific
applications and tasks in which generative models, specifically in Neuro-Symbolic systems, are
used. We conclude with summarizing our key findings and outlining future work.
2. Related Work on the Boxology
We will base our paper on the paper by van Bekkum et al. [9], in which the authors provide
a taxonomically organised vocabulary to describe both processes and data structures used in
hybrid systems. The highest level of this taxonomy contains instances, models, processes and
actors, which may be described as follows.
Instances: The two main classes of instances are data and symbols. Symbols are defined
as to have a designation to an object, class or a relation in the world, which can be
either atomic or complex, and when a new symbol is created from another symbol and
a system of operations, it should have a designation. Examples of symbols are labels
(short descriptions), relations (connections between data items, such as triples) and traces
(records of data and events). Data is defined as not symbolic. Examples are numbers,
texts, tensors or streams.
Models: Models are descriptions of entities and their relationships, which can be statistical
or semantic. Statistical models represent dependencies between statistical variables,
such as LLMs or Bayesian Networks. Semantic models specify concepts, attributes and
relationships to represent the implicit meaning of symbols, such as ontologies, taxonomies,
knowledge graphs or rule bases.
Processes: Processes are operations instances and models. Three types of processes are defined:
generation, transformation and inference. Generation can be done using, for example, the
training of a model or by knowledge engineering. Transformation is the transformation
of data, for example from knowledge graph to vector space. Inference can be inductive or
deductive, in which induction generalises instances and deduction reaches conclusions
on specific instances, such as with classification.
Actors: Actors can be humans, (software) agents or robots (physically embedded agents).
Meyer-Vitali et al. [10] extended the original paper with a definition of teams of actors in
the Boxology.
Besides the vocabulary, the visual language is defined in van Bekkum et al. [9], as an extension
on Van Harmelen and Ten Teije [8]. The visual language consists of rectangular boxes (instances),
hexagonal boxes (models), ovals (processes) and triangles (actors) and unspecified arrows
between them. Within the boxes the concept will be noted by each level in the vocabulary
using colon-separation from most generic to most-specific, for example a neural network will
be model:stat:NN.
van Bekkum et al. [9] present elementary patterns, which can then be combined into more
complex patterns. Patterns 1a and 2a from Figure 1, for example, can be combined into a pattern
which is named 3a in the paper (depicted in Figure 2). Whereas 1a describes the pattern of
training a model based on data (data generates a model), 2a describes the usage of the model in
deducing a symbol (data and model deduce a symbol), such as a prediction. The combination
in 3a describes a basic structure for a (statistical) Machine Learning (ML) model depicting the
training (creating the model) and testing or application phase (applying the model on new data).
In the past years, the Boxology has been used and extended in different ways. Three of the
most influential papers are the formalisation of the notions from the Boxology and implemen-
tation in the heterogeneous tool set (Hets) [11], the extension of the Boxology for (teams of)
actors [10] and the systematic study of nearly 500 papers published in the past decade in the
area of Semantic Web Machine Learning [12].
3. Design Patterns for Generative Models
While Generative AI originates in the realm of data-driven AI, it has demonstrated capabilities
that exceed classical machine learning tasks like classification and regression by far. In particular,
such generative systems specialise in the generation of content, such as images [1, 2], videos
[13], or text [14, 15, 16]. In the original, purely statistical setting, these capabilities are acquired
during a so-called (pre-)training phase [17] where a representation of a large data body is
learned and in a second phase used to process input to output that has not explicitly been
specified but follows the characteristics of the data body (application phase).
However, specific arrangements for both (pre-)training and representation usage in down-
stream tasks vary for different approaches and systems [18]. In order to allow for a coherent
description of the generative paradigm, we propose to extend the elementary patterns of van
Bekkum et al. [9] that describe the generic pattern for instances, models, processes and actors
(Figure 1 1a-1d and 2a-d). Please note that while patterns 1e and 1f are required for certain
aspects of the generative paradigm, their usage is not limited to this. Data generation and
labelling by humans may also be employed work with any statistical approach.
Figure 1: All elementary design patterns, including proposed additions 1e, 1f and 2e
In particular, when describing classical machine learning systems, mostly pattern 2a is used,
where the output is a symbol, such as a classification or a label. However, the key concept in
generative models is that the output is not a symbol, but data; this can be an image, video or
text, depending on the model. Additionally, actors play an important role in Generative AI, by
creating prompts or label data. To this end, we here propose three new elementary patterns:
pattern 1e, in which an actor can generate data, pattern 1f, in which an actor labels data, and 2e,
in which a model can deduce data from data. In the remainder of this section we mainly focus
on Large Language Models (LLMs). Please note, however, that the patterns proposed in this
section are transferable to other data types, for example to vision transformers, which follow a
similar architecture paradigm as transformers but operate on image data .
3.1. Transformer Models
The key technology behind basically all current LLMs is the so-called transformer architecture.
The original transformer paper by Vaswani et al. [19] proposed to use two interacting models,
Figure 2: Compositional design patterns, including proposed addition 3c made by combining elementary
pattern 1a and 2e.
an encoder and a decoder. In the transformer family, some models, however, only use the
encoder or the decoder part [18]. Figure 3A shows the architecture of a transformer model as a
design pattern. Transformers are made up of two parts, an encoder and a decoder. These are
usually trained end-to-end (such as flan-T5 [20]), but can also be used separately as encoder-
only (Figure 3B) or decoder-only (Figure 3C) models. In the following sections, we focus on an
encoder-only and a decoder-only family. Other sections focus on instructions and prompting of
different models and the interaction with actors.
3.1.1. Encoder only: BERT (base)
Some systems are encoder-only. These systems are specialised in contextual encoding, often
named a base model. They can ‘understand’ and encode input sentences. An encoder model is
trained using data, pattern 1a. It is often connected to other systems, such as a classification
system, pattern 3a (see Figure 3B), to be useful for tasks other than the encoding input sentences.
An example of this is BERT [21]. Encoders are transformer models, but not generative models.
3.1.2. Decoder only: GPT
Other transformer based systems have decoder-only architectures. This approach is comple-
mentary to the encoder-only paradigm, but structurally different [18]: an encoder processes the
input data (in these cases text) and transforms it into a different, machine interpretable, repre-
sentation, often a vector representation. A decoder-only system, on the other hand, decodes the
input data directly, without being transformed into a higher, more abstract representation, to
the desired representation (text or images). Examples of this are generative models from the
GPT family [14].
In the Boxology, both encoders and decoders have a similar representation. For generative
models from the GPT family, we suggest pattern 3c, (see Figure 2), which is a combination of 1a
and 2e, as presented in Figure 1: data is used to train a decoder model, which does not use an
Figure 3: Three uses of transformers. A shows the traditional encoder-decoder architecture. B shows
an encoder-only model applied to a classification task. C shows a decoder-only architecture.
encoder as input as well, such as with other transformers. This decoder model can be used to
deduce output data from input data directly.
Decoder-only architectures may be further divided into causal decoder architectures and
prefix decoder architectures. Causal decoder architectures, such as GPT [22, 14] and BLOOMZ
[23], use only unidirectional attention to the input sequence by using a specific mask. Prefix
decoder architectures, such as PaLM [24], uses the bidirectional attention for tokens in the prefix
while maintaining unidirectional attention for generating subsequent tokens. Both architectures
follow the elementary pattern 2e.
3.1.3. Prompts and Instructions
One of the main differences between current LLMs and earlier BERT or other transformer
models is that the model is fine-tuned on instructions [18]. Multi-task fine-tuning or instruction
tuning, is currently often done using a collection of datasets phrased as instructions, to improve
model performance and generalisation to unseen tasks [20]. The original model is often referred
to as foundation model [25], whereas the fine-tuned model is an adjusted model. In the Boxology,
we define this adjusted model as another model as we did with the encoder and decoder model
in Figure 3, but then stacking two decoder models. This instruction tuning also follows pattern
1a, but this data is different as it also contains instructions.
Next to instruction learning LLMs can also be tweaked by in-context learning. Here examples
are used as part of the prompt to give context for the answers to the instructions. In this case
the model weights are not changed. This optimizes the performance of models on different
tasks [26], but does not need as much training data as training a model from scratch. These
prompts can include a few (training) examples of the input and output (few-shot) or no examples
(zero-shot). These few-shot examples do not train the foundational or instruction model, and
therefore we model them as input data that is used to deduce data (text), which is pattern 2e.
Assistants or GPTs could, however, be seen as a new model, especially if they perform other
tasks, such as Retrieval Augmented Generation (RAG).
3.2. Actor Interaction
Actors play a large role in the current generative models. In the original paper by van Bekkum
et al. [9], patterns using actors are underspecified. On the one hand, actors often create data,
not only in the interaction with an agent that uses generative models, but also in common
Machine Learning approaches. Many of the created textual datasets are written, pre-processed
and labelled by actors. A first proposed pattern is pattern 1e, in which an actor creates data.
The second proposed pattern is pattern 1f, in which an actor generates a label, or annotates
data. Both patterns are depicted in Figure 1.
Generative models are often not used only once. With the current chat functions, actors are
interacting with the model multiple times. The main difference with other Machine Learning
models, where also more data is inputted and symbols are outputted, the data inputted is
often not dependent on the output of the previous data point. However, with conversational
generative models, prompts can be related to the previous response. Currently, recurrent or
iterative behaviour is not yet part of the pattern concepts.
4. Design Patterns for Generative Neuro-Symbolic AI
In this section, we describe and explore several papers that use generative models in a Neuro-
Symbolic system. The selected papers are chosen, as they represent a diverse set of possibilities
to use a generative model, at the start of the system, in the middle and at the end, but also to
act as a fluent language interface or a formal language interface. We also included ChatGPT,
which is the most famous generative AI system, and although mainly data driven, includes a
symbolic component in the reward modelling part of the training phase.
4.1. (Training of) ChatGPT
ChatGPT is an application of the foundational model GPT3 [14], and later GPT4 [27]. It is
trained further to be of aid in the setting of an assistant. The architecture of the training phases
is represented in Figure 4. The foundational model GPT3 is used as a basis for further training
Figure 4: Training phase of ChatGPT
Figure 5: Boxology representation of KnowGL [28]
(1a). Instructions and answers are used to train what will become ChatGPT. Then, based on
new prompts the model generates a response (3c).
To further train ChatGPT to give the desired responses the reward model is added. The reward
model is a separate model, which can judge if a response is a good one, given the instructions.
The reward model is trained by people annotating the multiple answers to instructions. To train
the reward model, the model trained on instructions is asked to output multiple answers. These
answers are then ranked by annotators to generate a training set for the reward model (1f). The
reward model is trained to compare answers of ChatGPT and return their score (3a). This is
then used in a loop with the ChatGPT to improve the instruction answering process. As one
can view, we have adapted Boxology patterns to be able to accept multiple inputs.
When applying ChatGPT in a pipeline, it suffices to show only pattern 3c, the block containing
ChatGPT and 1e to show the user writing the prompt.
4.2. KnowGL
Figure 5 shows KnowGL Parser [28], a NeSy system combining a generative module and symbolic
methods. The KnowGL Parser can be used to automatically extract knowledge graphs from
collections of documents. It is based on BART-large, which has an encoder-decoder architecture.
Figure 6: Boxology representation of KnowBERT [29]
The encoder receives a sentence (1a) and the decoder generates a list of ‘subject, relation, object’
(3c). These are then parsed (transformed) in preparation of the next step, fact ranking (1d). Here
a ranked list is created of distinct facts and their scores (2b). In the final step the generated
facts are linked to Wikidata. This is done using a mapping of labels to Wikidata IDs (2b). In the
case that the generative model has created a new entity, type or relation label that are not in
Wikidata it returns ‘null’.
4.3. KnowBERT
While knowledge is mostly injected to statistical generative models either during the input
or during the output stage, also approaches to inject knowledge inside the model have been
proposed. A prominent example is KnowBERT, a modified variant of the transformer architec-
ture BERT [29]. Although not a generative model, it stands out for its fusion of contextual and
graph representations, attention-enhanced entity spanned knowledge infusion, and flexibility in
injecting multiple Knowledge Graphs at various model levels. By integrating so-called Knowl-
edge Attention and Recontextualization (KAR) layers [30], graph entity embeddings are utilized
that are processed through an attention mechanism to enhance entity span embeddings. This
happens in later layers of the model to stabilize training but may potentially also used to inject
knowledge at earlier stages [6]. The Boxology pattern for KnowBERT is depicted in Figure 6.
4.4. Mathematical Conjecturing and LLMs
The system proposed by Johansson and Smallbone [31] assigns the generative task of discovery
of mathematical conjectures to a LLM (3c), while the results can be checked afterwards using
a symbolic theorem prover or counter-example finder (2b). The system is prompted with a
formal theory (e.g. a sort function), and has the LLM generate lemmas from the theory. These
generated lemmas are transformed from data to symbol and can then be used by the semantic
model(s). The pattern is depicted in Figure 7. The approach taken in Yang et al. [32] is also
captured by this pattern. The system proposed uses a LLM component to produce Prolog code
Figure 7: Boxology representation of using LLMs for discovery of mathematical conjectures [31]
Figure 8: Boxology representation of GENOME [33]
(3c) and a symbolic inference engine to produce answers and reasoning traces by executing the
aforementioned code (1d, 2b).
4.5. GENOME
Generative Neuro-Symbolic Visual Reasoning by Growing and Reusing Modules (GENOME)
[33] focuses on the task of generative software module learning, based on a LLM generating
signatures (input/output) and reasoning steps, then have a LLM create the software module
based on those and evaluate the module on test cases.
The system consists of three stages: module initialization, module generation, and module
execution. The design pattern is depicted in Figure 8. First a LLM assesses a visual-language
question and outputs new module signatures and operation steps as a response to the query
(3c), if current modules cannot provide an adequate response. In the next step, the LLM creates
a module (software code) based on the signature/test case (3c). Finally the module is executed
by passing it a visual query (2a).
Figure 9: Boxology representation of Logic-LM [34]
4.6. Logic-LM
Logic-LM [34] integrates LLMs with symbolic solvers to improve logical problem-solving. The
pattern is depicted in Figure 9: the system utilizes LLMs to translate a problem stated in natural
language problem into a symbolic formulation (3c). In the next step, a symbolic reasoner
performs logical inference on the formulated problem (1d, 2b, 1d). Finally, an LLM interprets
the results and outputs natural language (3c). The LLM thus functions as a fluent language
interface (both on input and output) to a symbolic reasoner component.
5. Conclusion and Future Work
Generative AI is currently a major technology with many applications and combining data-
driven approaches with knowledge-based techniques is a promising development to this end. In
this paper, we propose new design patterns for modular generative Neuro-Symbolic systems to
be included into the design pattern approach for Neuro-Symbolic systems as proposed by van
Bekkum et al. [9]. We show how the composition of elementary patterns can be used to describe
generative models, and we explore several specific generative models, such as ChatGPT, as well
as several generative NeSy papers, such as KnowGL, GENOME and Logic-LM.
We acknowledge that this is only the first step in a more elaborate exploration on generative
design patterns and the description of generative Neuro-Symbolic architectures. In future work,
we would like to validate our proposals for extending the Boxology, by applying them to more
examples from additional papers. In addition, we expect to further extend and deepen the
Boxology itself. In this paper, it became clear that the temporal or iterative aspect is not yet
visualised well, as well as the naming and formalisation of the Boxology, including the do’s
and don’ts: which pattern combinations are allowed and which are not? The importance of
modelling datasets for generative AI may be taken into account in future specifications of
particular subtypes of Instances and Models in the taxonomy. Additionally, the use of graphical
tools for software development is well-known from the Unified Modelling Language (UML) and
visual programming tools, such as LabView or Scratch. We are mostly concerned with graphical
representations of design patterns for system design and documentation, but the promise of
templates, low-code or no-code development is appealing for the future.
Acknowledgements
We would like to thank the TNO project GRAIL for their financial support, as well as Frank van
Harmelen and Annette ten Teije for their feedback. We would also like to thank Daan Di Scala
for his contribution to the KnowGL pattern.
References
[1] J. Betker, G. Goh, L. Jing, T. Brooks, J. Wang, L. Li, L. Ouyang, J. Zhuang, J. Lee, Y. Guo,
et al., Improving image generation with better captions, Computer Science 2 (2023) 8.
[2] R. Rombach, A. Blattmann, D. Lorenz, P. Esser, B. Ommer, High-resolution image synthesis
with latent diffusion models, 2021. arXiv:2112.10752 .
[3] H. Chen, F. Jiao, X. Li, C. Qin, M. Ravaut, R. Zhao, C. Xiong, S. Joty, Chatgpt’s one-year
anniversary: Are open-source large language models catching up?, arXiv:2311.16989
(2023).
[4] Z. Ji, N. Lee, R. Frieske, T. Yu, D. Su, Y. Xu, E. Ishii, Y. J. Bang, A. Madotto, P. Fung, Survey
of hallucination in natural language generation, ACM Comput. Surv. 55 (2023).
[5] H. Zhao, H. Chen, F. Yang, N. Liu, H. Deng, H. Cai, S. Wang, D. Yin, M. Du, Explainability
for large language models: A survey, ACM Transactions on Intelligent Systems and
Technology 15 (2024) 1–38.
[6] P. Colon-Hernandez, C. Havasi, J. Alonso, M. Huggins, C. Breazeal, Combining pre-trained
language models and structured knowledge, arXiv preprint arXiv:2101.12294 (2021).
[7] X. Wei, S. Wang, D. Zhang, P. Bhatia, A. Arnold, Knowledge enhanced pretrained language
models: A compreshensive survey, arXiv:2110.08455 (2021).
[8] F. Van Harmelen, A. Ten Teije, A boxology of design patterns for hybrid learning and
reasoning systems, Journal of Web Engineering 18 (2019) 97–123.
[9] M. van Bekkum, M. de Boer, F. van Harmelen, A. Meyer-Vitali, A. t. Teije, Modular design
patterns for hybrid learning and reasoning systems: a taxonomy, patterns and use cases,
Applied Intelligence 51 (2021) 6528–6546.
[10] A. Meyer-Vitali, W. Mulder, M. H. T. de Boer, Modular design patterns for hybrid actors,
2021. arXiv:2109.09331 .
[11] T. Mossakowski, Modular design patterns for neural-symbolic integration: refinement
and combination, arXiv:2206.04724 (2022).
[12] A. Breit, L. Waltersdorfer, F. J. Ekaputra, M. Sabou, A. Ekelhart, A. Iana, H. Paulheim,
J. Portisch, A. Revenko, A. t. Teije, et al., Combining machine learning and semantic web:
A systematic mapping study, ACM Computing Surveys (2023).
[13] Y. Liu, K. Zhang, Y. Li, Z. Yan, C. Gao, R. Chen, Z. Yuan, Y. Huang, H. Sun, J. Gao, et al.,
Sora: A review on background, technology, limitations, and opportunities of large vision
models, arXiv:2402.17177 (2024).
[14] T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan,
P. Shyam, G. Sastry, A. Askell, et al., Language models are few-shot learners, Advances in
neural information processing systems 33 (2020) 1877–1901.
[15] H. Touvron, T. Lavril, G. Izacard, X. Martinet, M.-A. Lachaux, T. Lacroix, B. Rozière,
N. Goyal, E. Hambro, F. Azhar, A. Rodriguez, A. Joulin, E. Grave, G. Lample, Llama: Open
and efficient foundation language models, 2023. arXiv:2302.13971 .
[16] S. Pichai, D. Hassabis, Introducing gemini: our largest and most capable ai model, 2023.
[17] D. Erhan, A. Courville, Y. Bengio, P. Vincent, Why does unsupervised pre-training help
deep learning?, in: Proceedings of the thirteenth international conference on artificial
intelligence and statistics, JMLR Workshop and Conference Proceedings, 2010, pp. 201–208.
[18] B. Min, H. Ross, E. Sulem, A. P. B. Veyseh, T. H. Nguyen, O. Sainz, E. Agirre, I. Heintz,
D. Roth, Recent advances in natural language processing via large pre-trained language
models: A survey, ACM Computing Surveys 56 (2023) 1–40.
[19] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, I. Polo-
sukhin, Attention is all you need, Advances in neural information processing systems 30
(2017).
[20] H. W. Chung, L. Hou, S. Longpre, B. Zoph, Y. Tay, W. Fedus, Y. Li, X. Wang, M. Dehghani,
S. Brahma, et al., Scaling instruction-finetuned language models, arXiv:2210.11416 (2022).
[21] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional
transformers for language understanding, 2019. arXiv:1810.04805 .
[22] A. Radford, J. Wu, R. Child, D. Luan, D. Amodei, I. Sutskever, et al., Language models are
unsupervised multitask learners, OpenAI blog 1 (2019) 9.
[23] N. Muennighoff, T. Wang, L. Sutawika, A. Roberts, S. Biderman, T. L. Scao, M. S. Bari,
S. Shen, Z.-X. Yong, H. Schoelkopf, et al., Crosslingual generalization through multitask
finetuning, arXiv:2211.01786 (2022).
[24] A. Chowdhery, S. Narang, J. Devlin, M. Bosma, G. Mishra, A. Roberts, P. Barham, H. W.
Chung, C. Sutton, S. Gehrmann, et al., Palm: Scaling language modeling with pathways,
Journal of Machine Learning Research 24 (2023) 1–113.
[25] R. Bommasani, D. A. Hudson, E. Adeli, R. Altman, S. Arora, S. von Arx, M. S. Bernstein,
J. Bohg, A. Bosselut, E. Brunskill, et al., On the opportunities and risks of foundation
models, arXiv:2108.07258 (2021).
[26] Y. Liu, H. He, T. Han, X. Zhang, M. Liu, J. Tian, Y. Zhang, J. Wang, X. Gao, T. Zhong, et al., Un-
derstanding llms: A comprehensive overview from training to inference, arXiv:2401.02038
(2024).
[27] T. Wu, S. He, J. Liu, S. Sun, K. Liu, Q.-L. Han, Y. Tang, A brief overview of chatgpt: The
history, status quo and potential future development, 2023.
[28] G. Rossiello, M. F. M. Chowdhury, N. Mihindukulasooriya, O. Cornec, A. M. Gliozzo,
Knowgl: Knowledge generation and linking from text, in: AAAI, 2023, pp. 16476–16478.
[29] M. E. Peters, M. Neumann, R. L. Logan IV, R. Schwartz, V. Joshi, S. Singh, N. A. Smith,
Knowledge enhanced contextual word representations, arXiv:1909.04164 (2019).
[30] I. Balažević, C. Allen, T. M. Hospedales, Tucker: Tensor factorization for knowledge graph
completion, arXiv:1901.09590 (2019).
[31] M. Johansson, N. Smallbone, Exploring mathematical conjecturing with large language
models, Proceedings of NeSy (2023).
[32] S. Yang, X. Li, L. Cui, L. Bing, W. Lam, Neuro-symbolic integration brings causal and
reliable reasoning proofs, 2023. arXiv:2311.09802 .
[33] Z. Chen, R. Sun, W. Liu, Y. Hong, C. Gan, Genome: Generative neuro-symbolic visual
reasoning by growing and reusing modules, 2023. arXiv:2311.04901 .
[34] L. Pan, A. Albalak, X. Wang, W. Y. Wang, Logic-lm: Empowering large language models
with symbolic solvers for faithful logical reasoning, arXiv:2305.12295 (2023).