<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Using ChatGPT to Refine Draft Conceptual Schemata in Supply-Driven Design of Multidimensional Cubes</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Stefano Rizzi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>DISI - University of Bologna</institution>
          ,
          <addr-line>Viale Risorgimento, 2, Bologna, 40136</addr-line>
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Refinement is a critical step in supply-driven conceptual design of multidimensional cubes because it can hardly be automated. In fact, it relies on the end-users' requirements on the one hand, and on the semantics of measures, dimensions, and attributes on the other. As a consequence, it is normally carried out manually by designers in close collaboration with end-users. The goal of this work is to check whether LLMs can act as facilitators for the refinement task, so as to let it be carried out entirely -or mostly- by end-users. The Dimensional Fact Model is the target formalism for our study; as a representative LLM, we use ChatGPT's model GPT-4o. To achieve our goal, we formulate two research questions aimed at understanding the basic competences of ChatGPT in refinement and investigating if they can be improved via prompt engineering. The results of our experiments show that, indeed, a careful prompt engineering can significantly improve the accuracy of refinement, and that the residual errors can quickly be fixed via one additional prompt. However, we conclude that, at present, some involvement of designers in refinement is still necessary to ensure the validity of the refined schemata.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Conceptual design</kwd>
        <kwd>Multidimensional model</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>ChatGPT</kwd>
        <kwd>Refinement</kwd>
        <kwd>Supply-driven design</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Conceptual design is a key step in the development of data
warehouse (DW) systems and multidimensional databases,
since it determines their information content and, ultimately,
the set of queries they can answer. The goal is to create an
implementation-independent representation of one or more
cubes structured according to the multidimensional model,
i.e., described in terms of measures, dimensions, and
attribute hierarchies. A lot of research has been done over the
last couple of decades on conceptual design of cubes, mainly
distinguishing between supply-driven approaches, where the
conceptual schema is determined starting from the schema
of a source databases, and demand-driven approaches, where
it is created based on the end-users’ requirements.</p>
      <p>
        An advantage of supply-driven design over
demanddriven design is that a draft conceptual schema can be
obtained from the source schema in automatic fashion, by
applying an algorithm that essentially chases the functional
dependencies coded in the source schema and uses them to
arrange hierarchies [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Although this significantly speeds
up design, the draft schema must then be refined in the light
of the end-users’ requirements. Refinement mainly implies
the following activities [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]:
• Removing attributes that are deemed not interesting for
analyses.
• Finding descriptive attributes, i.e., attributes that should
not be used for aggregation while being useful for
analyses (e.g., the name of a customer).
• Discretizing attributes with dense domains to make them
usable for aggregation (e.g., the weight of a product).
• Finding optional attributes, i.e., attributes that are
undeifned for some instances of the hierarchy (e.g., the State
attribute in a geographical hierarchy that also includes
non-US nations).
• Labeling measures based on whether the SUM operator
can be used or not to aggregate them (e.g., the exchange
rate of dollars to euros, which cannot be summed).
Unfortunately, these activities can hardly be automated by
an algorithm because they rely on the end-users’
requirements on the one hand, and on the semantics of measures,
dimensions, and attributes as expressed by their names on
the other. Then, they must be carried out manually by
designers in collaboration with end-users.
      </p>
      <p>
        This is a typical situation in software engineering where
Large Language Models (LLMs) may come to the rescue.
LLMs have proven to be a great tool for mimicking
human linguistic abilities because of their capacity to learn
from large corpora, which has had a disruptive efect in a
number of fields, and more specifically in software
engineering [
        <xref ref-type="bibr" rid="ref3 ref4 ref5">3, 4, 5</xref>
        ]. In particular, the experiments on using
LLMs for conceptual design [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ] showed that they can help
designers with this task by producing draft solutions in a
timely manner —although some human intervention is still
necessary to guarantee the accuracy of the outcomes.
      </p>
      <p>
        The goal of this work is to check whether LLMs can act
as facilitators for the refinement of conceptual schemata
of multidimensional cubes, so as to relieve designers from
their role or even, if possible, let refinement be carried out
entirely by end-users. The Dimensional Fact Model (DFM
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]) is the target formalism for our study; as a typical LLM,
we use ChatGPT’s model GPT-4o [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], which has gained
popularity for its smooth user interface and natural language
generating capabilities [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. To achieve our goal, we
formulate two research questions aimed at (i) understanding the
basic competences of ChatGPT in the refinement of a draft
DFM schema and (ii) investigating if the latter can be
improved via prompt engineering. An extended version of this
work is available in [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        An experiment to use LLMs for creating specifications from
requirements documents in the realm of smart devices is
described in [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. The authors contend that the fundamental
skill of conceptual design is still lacking, but they
acknowledge that LLMs are very useful in later phases of the
development process, like creating class diagrams and generating
source code. Additional experiments with ChatGPT for
conceptual modeling are discussed in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. The authors note that
ChatGPT can rapidly produce an initial draft diagram from
a natural language description; nevertheless, considerable
modeling expertise is still needed to improve and verify the
outcomes. The authors of [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] describe an experiment they
conducted using ChatGPT and come to the conclusion that
while adding LLMs to human-driven conceptual design does
not dramatically afect outcomes, it does greatly reduce the
time required to complete the design by requiring fewer
design steps. In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], many conceptual schemata produced
by an LLM are contrasted with a baseline of crowdsourced
solutions. On average, it is shown that crowdsourced ideas
are more innovative, whereas LLM-generated solutions are
more practical. In [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], the benefits of utilizing LLMs to
improve morphological analysis in conceptual design are
examined. The tests demonstrate how LLMs give designers
access to interdisciplinary knowledge; for optimal outcomes,
LLMs and designers should work closely together and use
smart prompt engineering. With relation to use case and
domain modeling, [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] examines how users engage with
LLMs during conceptual modeling. The primary
conclusions speak to the necessity of particular prompt templates
to assist users.
      </p>
      <p>
        As to multidimensional conceptual design, the main types
of methods in the literature are supply-driven (or
datadriven), demand-driven (or requirement-driven), mixed, and
query-driven. Supply-driven methods begin by designing
conceptual schemata from the schemata of the data sources
(such as relational schemata); end-user requirements
inlfuence design by enabling the designer to choose which
data are important for making decisions and by figuring out
how to structure them using the multidimensional model
[
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Demand-driven techniques begin with identifying
end-users’ business requirements, and only then do they
look into how to map these requirements onto the available
data sources [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]. Mixed techniques integrate
requirementsdriven and data-driven methods; here, both end-user
requirements and data source schemata are used
simultaneously [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. The set of OLAP queries that end-users are
willing to formulate is the starting point for the creation
of a multidimensional schema in query-driven approaches.
These queries can be specified using SQL statements [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ],
MDX expressions [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ], pivot tables [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ], or query trees [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ].
Multidimensional modeling techniques are reviewed in [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ],
and their cost-benefit analysis is provided in [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. The investigation process</title>
      <p>As stated in the Introduction, our goal in this work is to
assess the performance of ChatGPT in the refinement of a
draft DFM schema obtained by supply-driven design
starting from a source relational schema. We take as a reference
an advanced form of the DFM including, besides the basic
constructs of fact, measure, dimension, and attribute, the
advanced constructs of descriptive attributes, optional
attributes, and additivity. In this form, a DFM schema is a
graph whose root is the fact (represented as a box with the
fact name —e.g., SALES— followed by a list of measures —
e.g., Amount), whose other nodes are attributes —e.g.,
Product— represented as circles and connected by arcs
representing many-to-one roll-up relationships, i.e., functional
dependencies (FDs, for instance, Product → Category).
Descriptive attributes are represented without a circle;
optional attributes are dashed; a non-additive measures is
represented by adding its aggregation operator to its name
(e.g., ExchangeRate (AVG)).</p>
      <sec id="sec-3-1">
        <title>3.1. Research questions</title>
        <sec id="sec-3-1-1">
          <title>We formulate the following research questions:</title>
          <p>RQ.1: Is ChatGPT capable of refining a draft DFM schema
by (i) making attribute names more intuitive for
endusers, (ii) showing additivity, (iii) finding descriptive
attributes or discretizing them, (iv) finding optional
attributes, (v) completing time hierarchies, and (vi)
removing uninteresting attributes?
RQ.2: Can the performance of ChatGPT in refining a draft</p>
          <p>DFM schema be improved via prompt engineering?</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Experiment design</title>
        <p>
          Our experiment relies on five cornerstones, described in the
following subsections.
3.2.1. Base criteria and technology
The criteria we follow for our experiment are listed below:
• Learning. For learning we adopt a prompt-based
learning method, which is often used as an alternative to
finetuning [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ]. Specifically, for RQ.1 we adopt 0-shot
learning (which operates with no labeled examples); for RQ.2
we adopt few-shot learning and provide two training
examples [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. To further improve learning, for RQ.2 we
also employ the chain-of-thought technique [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ], which
includes a list of reasoning steps in the examples.
• Reproducibility. The lack of reproducibility of the tests
is a significant challenge when working with LLMs
because of their non-deterministic nature. The level of
“creativity” of ChatGPT is ruled by its temperature parameter;
in principle, no creativity is required for refining draft
schemata, so we set the temperature to 0 for every chat.
• Domain. The issue domain is acknowledged to be crucial
for LLMs; the more domain knowledge an LLM has, the
better the model it generates [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ]. Every example we
present describes actual domains, some of which are
wellknown (like purchases) and others are less common (like
crossfit workouts).
• Conversation-awareness. The answers obtained from
ChatGPT may depend heavily on the previous questions
asked during a conversation. Thus, as also suggested in
[
          <xref ref-type="bibr" rid="ref26">26</xref>
          ], we start a new chat for each case.
• Iteration. In all our tests, the first answer obtained is
considered. However, keeping in mind that the
refinement process is inherently iterative, in RQ.2 we tried to
improve the first answer by further prompting ChatGPT
with suggestions.
        </p>
        <p>As to the technological environment, experiments have
been carried out on the ChatGPT-4o model.
3.2.2. Input/output format
A draft DFM schema must be provided in input for each
of our research questions, and a refined one must be
provided as output. We employ YAML1, a human-readable data</p>
        <sec id="sec-3-2-1">
          <title>1https://yaml.org/spec/history/2001-08-01.html</title>
          <p>
            serialization language that is frequently used for
configuration files and in applications where data is being saved
or communicated, as a format to express DFM schemata.
Since ChatGPT is familiar with YAML, it does not need any
further instruction on the syntax of the language;
nevertheless, it needs to be taught about the particular tags we
added to denote multidimensional concepts (e.g., measures
to introduce the list of measures).
3.2.3. Prompt templates
Following the suggestions given in [
            <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
            ], the prompts we
adopt during our chats are structured according to the
following templates:
• Instruction prompts. These are used in RQ.1 and RQ.2
to assign ChatGPT a task and explain how to execute
it. Their structure includes: (1) ROLE: the specific roles
assigned to ChatGPT and to the user to provide a context
for the task; (2) FORMAT: how the input and output (i.e.,
the draft and refined DFM schemata) should be coded;
(3) TASK: the task assigned; (4) PROCEDURE (optional):
the method suggested to perform the task; (5) EXAMPLE
(optional): an example of some test cases, an explanation
of the procedure suggested to solve it (according to the
chain-of-thought principle), and the expected output.
• Case prompts. These are used in RQ.1 and RQ.2 to assign
a specific task to ChatGPT. Their structure includes: (1)
INPUT: the input of the task (a draft DFM schema coded
in YAML); (2) TASK: the task assigned; (3) OUTPUT: the
output required (a refined DFM schema coded in YAML).
3.2.4. Test cases
We created a set of five test cases with increasing dificulties,
based on some exercises in supply-driven design assigned
to the students of a master course in Business Intelligence.
Each exercise provided a source relational schema; from this
schema, a draft DFM schema was created in supply-driven
mode using the FD-chasing algorithm in [
            <xref ref-type="bibr" rid="ref1">1</xref>
            ]. The number
of dimensions and measures in the test cases ranges from 3
to 5 and from 0 to 5, respectively, while the overall number
of attributes (dimensions plus hierarchy levels) ranges from
10 to 34. Two of the test cases include shared hierarchies
(i.e., nodes entered by two or more arcs, as often is the case
with temporal hierarchies).
3.2.5. Evaluation of the results
Refinement is, to some extent, a subjective process because it
largely depends on the end-user requirements. For instance,
given a ProductWeight attribute, both making it descriptive
and discretizing it into WeightRanges are reasonable
refinements. As a consequence, creating a single ground truth
for each test case is hardly feasible. So we had to proceed
manually, by first identifying a set of feasible refinements
for each part of each draft DFM schema, and then
counting an error in the solution proposed by ChatGPT for each
deviation from this set of feasible refinements.
          </p>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Answer to RQ.1: Refinement</title>
        <p>To put ChatGPT to the test on refinement, we fed it with our
ifve test cases. In order to enable a more precise evaluation
of the abilities of ChatGPT, separate prompts are
submitted for each refinement step entailed by RQ.1. Thus, for
each test case we adopt a simple instruction prompt that
explains the DFM constructs followed by a request to carry
out a list of refinement steps; no PROCEDURE and
EXAMPLE components are present that suggest ChatGPT how to
operate. Then we formulate a sequence of case prompts
that (i) specify a draft DFM schema in input, (ii) assign as a
task one single refinement step, and (iii) require a refined
DFM schema in output. In the following we will separately
consider each step and briefly review its result.
(i) Make names intuitive. The attribute and measure
names in the draft DFM schema have the form
RELATION_NAME.attributeName, being RELATION_NAME
a table of the source relational schema used for deriving
the draft schema and attributeName one of its attributes
(see Figure 1). Making these names more intuitive for
end-users is mostly done correctly even if no specific
procedure is suggested. In some cases the relation name was
simply dropped, in others it was prefixed to the attribute
name (e.g., SUPPLIER.name became SupplierName). In
case C5, the most complex one, the shared hierarchy was
mistaken and the direction of some FDs was inverted.
(ii) Label measures. ChatGPT is quite good at dealing with
additivity. This is surprising, considering that this task
is often not easy even for end-users. The main errors we
found were syntactical: the measure was renamed in the
YAML code under the “measure” tag but not under the
“dependencies” tag, resulting in additional fake nodes.
(iii) Find descriptive attributes. ChatGPT performs poorly
in this task, with an average of almost four errors per test
case. On the one hand, it does not know under which
conditions an attribute should be made descriptive or
discretized; on the other, it does not use the correct syntax as
stated in the FORMAT section of the instruction prompt.
(iv) Find optional attributes. The identification of optional
attributes is strictly related to the end-user requirements.
Thus, for this refinement step the prompt simulates an
end-user statement; for instance, “Not all regions have
a state”. As in the previous case, ChatGPT always fails
in the syntax used (although it correctly identifies the
optional attribute).
(v) Complete time hierarchies. Here, ChatGPT correctly
adds Month → Year hierarchies to Date attributes.
However, it always fails in recognizing and managing shared
hierarchies (in C4 and C5).
(vi) Remove attributes. For the last refinement step, an
indication from end-users about which attributes are deemed
uninteresting for their analyses is required. Thus, like</p>
        <p>C3</p>
        <p>RQ.2
C1</p>
        <p>C2</p>
        <p>C4</p>
        <p>C5</p>
        <p>C1 C2 C3 C4 C5</p>
        <p>Renaming Additivity Descriptive Optional Time hier. Removal</p>
        <p>The number of errors made at each step for each test
case is shown in Figure 2 (top). As an example, Figure 3
(top) shows the final DFM schema for test case C2; note
that descriptive and optional attributes are not shown as
such because the visualizer does not recognize the wrong
YAML syntax for them, and that the graph is non-connected
because some FDs were dropped. Overall, the performance
are not very good but acceptable, with an average number
of total refinement errors per test case equal to 9. The errors
clearly tend to increase with the complexity of the draft
schema; the main problems are due to the YAML syntax and
to the presence of shared hierarchies. The most critical
reifnement steps appear to be the identification of descriptive
attributes and the removal of uninteresting attributes.
To answer RQ.2 we incrementally crafted an instruction
prompt by first trying to address the main issues emerged
in RQ.1, then progressively adding specific sentences to try
to fix the residual (or new) errors. The ROLE, FORMAT, and
TASK components are exactly like in RQ.1. However, for
each refinement step, we added a PROCEDURE component
to suggest ChatGPT how to operate and an EXAMPLE
component with two examples. The case prompts are exactly
the same used for RQ.1.</p>
        <p>No errors in additivity and optional attributes are made
when the improved prompt is used. A single renaming error
is made in C5, due to an unrecognized shared hierarchy. A
few errors are made on time hierarchies, again due to shared
hierarchies. Overall, the main causes of errors are related
to descriptive/discretized attributes (in some cases, a few of
them are not identified) and to the removal of uninteresting
attributes (sometimes, arcs are not correctly repositioned).
Noticeably, all these errors could be fixed in a single iteration
via specific prompts, e.g., “Merge drop-of date and pick-up
date into a single date node” to fix a shared hierarchy. In
some cases, even generic prompts were used successfully to
ifx errors, e.g., “Some arcs are missing, please try again” to
ifx the FDs after removal. Figure 3 (bottom) shows the final
DFM schema obtained for C2 after correcting two errors in
descriptive attributes via an iteration prompt.</p>
        <p>The results, in terms of number of errors made at each
step, are summarized in Figure 2 (bottom). It appears that
prompt engineering can significantly improve the accuracy
of refinement, with the average number of total refinement
errors per test case decreasing from 9 to 4. The main residual
errors are related to the recognition of shared hierarchies
and of descriptive/discretized attributes, as well as to the
removal of uninteresting attributes. In our tests, all these
errors could be fixed via an additional prompt that either
explains exactly how to proceed, or simply suggests to try
again paying more attention to some specific aspect.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>In this work we have investigated the capabilities of
ChatGPT to cope with a specific task in conceptual design,
namely, the refinement of draft DFM schemata obtained
by supply-driven conceptual design of multidimensional
data cubes —a task that is normally carried out manually
by designers and end-users in close collaboration. It turned
out that, although ChatGPT tends to mix the conceptual
level (DFM) with the logical level (star/snowflake schemata),
it can provide some acceptable results on test cases with
diferent degrees of complexity using simple prompts.
Noticeably, our tests show that, when prompts are enhanced
with detailed instructions and examples, the results
produced significantly improve in all cases. Indeed, when using
an improved prompt the average number of errors per
multidimensional concept across all test cases decreases from 0.5
to 0.2. In practice, the residual errors are still too many to
state that no involvement of designers is necessary and that
end-users can carry out refinement by directly interacting
with an LLM. However, we can conclude that LLMs can
significantly support designers in refinement, even considering
that all residual errors in our tests could quickly be fixed
via a simple additional prompt.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Golfarelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rizzi</surname>
          </string-name>
          ,
          <article-title>Data warehouse design: Modern principles and methodologies</article-title>
          ,
          <source>McGraw-Hill</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Antonelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bimonte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rizzi</surname>
          </string-name>
          ,
          <article-title>Multidimensional modeling driven from a domain language</article-title>
          ,
          <source>Autom. Softw. Eng</source>
          .
          <volume>30</volume>
          (
          <year>2023</year>
          )
          <article-title>6</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>W.</given-names>
            <surname>Ma</surname>
          </string-name>
          , S. Liu,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , L. Nie, Y. Liu,
          <string-name>
            <surname>LLMs:</surname>
          </string-name>
          <article-title>Understanding code syntax and semantics for code analysis</article-title>
          ,
          <source>CoRR abs/2305</source>
          .12138 (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>White</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Hays</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Fu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Spencer-Smith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. C.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <article-title>ChatGPT prompt patterns for improving code quality, refactoring, requirements elicitation, and software design</article-title>
          ,
          <source>CoRR abs/2303</source>
          .07839 (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>H.</given-names>
            <surname>Fill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Cabot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Maass</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. van Sinderen</surname>
          </string-name>
          ,
          <article-title>AI-driven software engineering - the role of conceptual modeling</article-title>
          ,
          <source>Enterp. Model. Inf. Syst. Archit. Int. J. Concept. Model</source>
          .
          <volume>19</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Fill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fettke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Köpke</surname>
          </string-name>
          ,
          <article-title>Conceptual modeling and large language models: Impressions from first experiments with ChatGPT, Enterp</article-title>
          .
          <source>Model. Inf. Syst. Archit. Int. J. Concept. Model</source>
          .
          <volume>18</volume>
          (
          <year>2023</year>
          )
          <article-title>3</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Lutze</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Waldhör</surname>
          </string-name>
          ,
          <article-title>Generating specifications from requirements documents for smart devices using large language models (LLMs)</article-title>
          ,
          <source>in: Proc. HCI</source>
          , Washington, DC, USA,
          <year>2024</year>
          , pp.
          <fpage>94</fpage>
          -
          <lpage>108</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>W.</given-names>
            <surname>Hariri</surname>
          </string-name>
          ,
          <article-title>Unlocking the potential of ChatGPT: A comprehensive exploration of its applications, advantages, limitations, and future directions in natural language processing</article-title>
          ,
          <source>CoRR abs/2304</source>
          .
          <year>02017</year>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Duh</surname>
          </string-name>
          ,
          <article-title>Examining how the large language models impact the conceptual design with human designers: A comparative case study</article-title>
          ,
          <source>Int. J. Hum. Comput. Interact</source>
          . (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Rizzi</surname>
          </string-name>
          ,
          <article-title>Using ChatGPT to refine draft conceptual schemata in supply-driven design of multidimensional cubes</article-title>
          , arXiv
          <volume>2502</volume>
          .02238v1 (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Grandi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>McComb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Goucher-Lambert</surname>
          </string-name>
          ,
          <article-title>Conceptual design generation using large language models</article-title>
          ,
          <source>CoRR abs/2306</source>
          .01779 (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>L.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Tsang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Jing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>A LLM-augmented morphological analysis approach for conceptual design</article-title>
          ,
          <source>in: Proc. DRS</source>
          , Boston, USA,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Ali</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Reinhartz-Berger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bork</surname>
          </string-name>
          ,
          <article-title>How are LLMs used for conceptual modeling? An exploratory study on interaction behavior and user perception</article-title>
          ,
          <source>in: Proc. ER</source>
          , Pittsburgh, USA,
          <year>2024</year>
          , pp.
          <fpage>257</fpage>
          -
          <lpage>275</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>O.</given-names>
            <surname>Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abelló</surname>
          </string-name>
          ,
          <article-title>Data-driven multidimensional design for OLAP</article-title>
          ,
          <source>in: Proc. SSDBM</source>
          , Portland,
          <string-name>
            <surname>OR</surname>
          </string-name>
          , USA,
          <year>2011</year>
          , pp.
          <fpage>594</fpage>
          -
          <lpage>595</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Jovanovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Simitsis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abelló</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Mayorova</surname>
          </string-name>
          ,
          <article-title>A requirement-driven approach to the design and evolution of data warehouses</article-title>
          ,
          <source>Inf. Syst</source>
          .
          <volume>44</volume>
          (
          <year>2014</year>
          )
          <fpage>94</fpage>
          -
          <lpage>119</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>F. D.</given-names>
            <surname>Tria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Lefons</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Tangorra</surname>
          </string-name>
          ,
          <article-title>Hybrid methodology for data warehouse conceptual design by UML schemas, Inf</article-title>
          . Softw. Technol.
          <volume>54</volume>
          (
          <year>2012</year>
          )
          <fpage>360</fpage>
          -
          <lpage>379</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>O.</given-names>
            <surname>Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abelló</surname>
          </string-name>
          ,
          <article-title>Automatic validation of requirements to support multidimensional design, Data Knowl</article-title>
          .
          <source>Eng</source>
          .
          <volume>69</volume>
          (
          <year>2010</year>
          )
          <fpage>917</fpage>
          -
          <lpage>942</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>T.</given-names>
            <surname>Niemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Nummenmaa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Thanisch</surname>
          </string-name>
          ,
          <article-title>Constructing OLAP cubes based on queries</article-title>
          ,
          <source>in: Proc. DOLAP</source>
          , Atlanta, Georgia, USA,
          <year>2001</year>
          , pp.
          <fpage>9</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bimonte</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Antonelli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Rizzi</surname>
          </string-name>
          ,
          <article-title>Requirements-driven data warehouse design based on enhanced pivot tables</article-title>
          ,
          <source>Req. Eng</source>
          .
          <volume>26</volume>
          (
          <year>2021</year>
          )
          <fpage>43</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wilson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Srinivasan</surname>
          </string-name>
          ,
          <article-title>A conceptual querydriven design framework for data warehouse</article-title>
          ,
          <source>Int. Jour. of Computer and Information Engineering</source>
          <volume>1</volume>
          (
          <year>2007</year>
          )
          <fpage>62</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>O.</given-names>
            <surname>Romero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abelló</surname>
          </string-name>
          ,
          <article-title>A survey of multidimensional modeling methodologies</article-title>
          ,
          <source>Int. J. Data Warehous. Min</source>
          .
          <volume>5</volume>
          (
          <issue>2009</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>F. D.</given-names>
            <surname>Tria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Lefons</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Tangorra</surname>
          </string-name>
          ,
          <article-title>Cost-benefit analysis of data warehouse design methodologies</article-title>
          ,
          <source>Inf. Syst</source>
          .
          <volume>63</volume>
          (
          <year>2017</year>
          )
          <fpage>47</fpage>
          -
          <lpage>62</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. A. H.</given-names>
            <surname>López</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Mussbacher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Varró</surname>
          </string-name>
          ,
          <article-title>Automated domain modeling with large language models: A comparative study</article-title>
          ,
          <source>in: Proc. MODELS</source>
          , Västerås, Sweden,
          <year>2023</year>
          , pp.
          <fpage>162</fpage>
          -
          <lpage>172</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>T. B. Brown</surname>
          </string-name>
          , et al.,
          <article-title>Language models are few-shot learners</article-title>
          ,
          <source>in: Proc. NeurIPS</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wei</surname>
          </string-name>
          , et al.,
          <article-title>Chain-of-thought prompting elicits reasoning in large language models</article-title>
          , in: S. Koyejo,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Belgrave</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Cho</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Oh (Eds.),
          <source>Proc. NeurIPS</source>
          , New Orleans, LA, USA,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>J.</given-names>
            <surname>Cámara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Troya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Burgueño</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vallecillo</surname>
          </string-name>
          ,
          <article-title>On the assessment of generative AI in modeling tasks: an experience report with ChatGPT and UML</article-title>
          , Softw.
          <source>Syst. Model</source>
          .
          <volume>22</volume>
          (
          <year>2023</year>
          )
          <fpage>781</fpage>
          -
          <lpage>793</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>