<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>DPS, and TTC. Koblenz,
Germany, June</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>AIPyCraft: AI-assisted software development lifecycle for 6G blockchain oracle validation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Antonio M. Alberti</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexis V. de A. Leal</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ariel Galante Dalla-Costa</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cristiano Bonato Both</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Graduate Program in Applied Computing, University of Vale do Rio dos Sinos (UNISINOS)</institution>
          ,
          <country country="BR">Brazil</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>School of Computer Science, University of Leeds</institution>
          ,
          <country country="UK">United Kingdom</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>1</volume>
      <fpage>0</fpage>
      <lpage>13</lpage>
      <abstract>
        <p>The growing interest in applying Artificial Intelligence (AI) to software engineering has accelerated since the release of ChatGPT 3.5 in 2022. This paper investigates how Large Language Models (LLMs) support applications' modular development and testing. We introduce AIPyCraft, a novel AI-assisted framework that facilitates the end-to-end lifecycle of software projects. Our approach leverages Google Gemini 2.5 Pro model to generate, correct, and manage software components within a semi-automated and incremental workflow. AIPyCraft enables project creation, environment setup, error correction, and feature evolution in an integrated manner. We develop and test a blockchain-based Oracle component designed for 6G wireless network environments, i.e., a complex, real-world scenario that demands secure data integration and modular extensibility. Preliminary experiments demonstrate AIPyCraft's potential to accelerate small-scale software project development through an “understand-by-building” methodology. Our findings show that using an LLM to generate efective TOML jobs for Of-chain 6G functions is feasible, with an average of 1.05 iterations to correct the TOML code and mean experiment time of 27.8 seconds.</p>
      </abstract>
      <kwd-group>
        <kwd>AI-assisted coding</kwd>
        <kwd>software engineering</kwd>
        <kwd>automated code generation</kwd>
        <kwd>modular software development</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The release of ChatGPT 3.5 on November 30, 2022, marked a significant milestone in the
evolution of Large Language Models (LLMs) for code generation tasks [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The performance and
accessibility introduced in version 3.5 sparked widespread interest across industry, academia,
and the broader developer community [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Initial experiments commonly involved short
interactions to assess the model’s ability to generate and manipulate source code. However,
more in-depth exploration revealed practical limitations: (i) iteratively prompting the model,
(ii) copying generated code, and (iii) manually testing outputs can be time-consuming and
cognitively demanding. Moreover, as projects grow in complexity, i.e., requiring multiple files,
interdependent modules, and isolated virtual environments, developers frequently encounter
challenges related to environment configuration, dependency management, and scalability.
These barriers have highlighted the need for more structured and automated approaches to
support the development of modular, testable, and maintainable software systems using LLMs.
      </p>
      <p>
        Given these challenges, a more fitness approach would involve using autonomous agents
capable of directly generating, testing, and refining code, eliminating the need for manual
copy-and-paste operations and enabling project-level validation beyond isolated files. For such
solutions to be practically applicable, it is essential that new features can be incrementally
added and seamlessly integrated into previously validated components [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. An incremental
development strategy plays a key role in maintaining manageable levels of complexity while
allowing iterative refinement. In real-world scenarios, developers begin with a general idea of
the desired functionality and benefit from evolving the system through progressive modifications
rather than attempting to define and fully implement the entire solution up front. Intelligent
software systems that support human-in-the-loop development, combining the generative
capabilities of Artificial Intelligence (AI) with human guidance, ofer considerable value [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Recent studies in AI-assisted software development have explored the integration of LLMs
into stages of the engineering workflow, including code generation [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ], error correction
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], configuration validation [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and test automation [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. While these eforts have yielded
promising results, most solutions are limited in scope, e.g., focusing on narrow tasks rather
than supporting the complete software lifecycle. Typical limitations include the absence of
modular project design, lack of orchestration across components, and insuficient handling
of environment setup or integration with virtual environments. Moreover, tools that embed
LLMs into developer environments lack structured control over incremental project evolution
and rely heavily on manual oversight [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Although techniques can be applied to emerging
domains such as blockchain-enabled Sixth Generation Networks (6G) components, such as
decentralized oracles and smart contract-based network automation, no comprehensive solution
in the literature addresses these challenges in this context. These gaps highlight the need
for holistic approaches that enable coherent, end-to-end development workflows, supporting
iterative refinement, automation, and scalability.
      </p>
      <p>This paper explores the subject of the AI-assisted coding lifecycle from a practical perspective,
trying to answer the question: Could LLMs help with the modular development and testing of 6G
components? Our central hypothesis is that LLMs can assist in the incremental and modular
development of projects with virtual environments. Moreover, LLMs can help automate tasks
such as code generation, virtual environment preparation, debugging, incremental feature
addition, and project structuring. AI-powered tools can ofer new paradigms to help developers
with the growing complexity of software development while testing new ideas "understanding
by building" prototypes experimenting with new features and directions on new developments.
We present the AIPyCraft1, an open-source collaborative project development assisted by LLMs,
integrating rapid prototyping with a virtual environment for automated code running and
checking. AIPyCraft is a first-look proposal for human-machine collaboration in software
development, exploring how AI Application Programming Interfaces (APIs) are integrated to
generate, manage, correct, and improve code iteratively. Moreover, AIPyCraft creates an entirely
1Source code available at: https://github.com/antonioalberti/AIPyCraft
new project, prepares its virtual environment, runs the project from its main program, collects
the obtained results, corrects the components according to errors, and adds new features to a
created project. The main contributions of this paper are:
• Open-source development automation: the proposal uses LLM APIs to generate
code, manage environments, iteratively correct errors, and integrate versions, ofering a
single-person end-to-end solution.
• A software component correction tool: corrects code based on existing running errors.</p>
      <p>The prompts focus on preserving existing functionality while correcting a failing module.
• AI-assisted development and testing: the analysis is performed on a blockchain-based</p>
      <p>Oracle component designed for 6G wireless network environments.</p>
      <p>Our findings indicate that using an LLM to generate efective TOML jobs in the domain of
Of-chain 6G functions, is feasible with a satisfactory number of interactions (1.05 average
interactions) and a reasonable time duration (27.80 seconds). The remainder of this paper
is structured as follows. Section II presents the fundamental background, including our 6G
development application. Section III presents the related work. Section IV describes our
proposed agent architecture. Section V presents a proof of concept, outlining how AIPyCraft
was evaluated to support the central hypothesis of the paper. Section VII concludes the study.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Background</title>
      <p>
        Deep Learning has emerged as a dominant reference in various subfields of AI, leveraging
artificial neural networks to model complex data distributions and achieving significant
breakthroughs in applications such as speech recognition, computer vision, and Natural Language
Processing (NLP). A pivotal advancement in NLP was Vaswani et al.’s [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] introduction of the
Transformer architecture in the seminal work "Attention Is All You Need". This architecture
innovation laid the groundwork for the development of LLMs [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], such as Google Gemini2,
OpenAI ChatGPT-4o3, DeepSeek4, and Anthropic Claude5, which are trained on extensive texts
and demonstrate remarkable proficiency in generating, interpreting, and manipulating natural
language in a human-like fashion. Adopting LLMs has influenced various domains, including
software engineering, education, and content generation. This adoption underscores their
potential to augment productivity and enable novel forms of human-AI collaboration.
      </p>
      <p>
        The growing complexity of software development has driven the adoption of technologies
to enhance the eficiency and quality of the development lifecycle [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. LLMs have attracted
attention for their remarkable ability to understand and generate source code. While initially
applied to machine translation, text summarization, and image generation tasks, LLMs have been
integrated into software engineering workflows [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Their capabilities extend to synthesis,
autocompletion of code snippets, and support for testing, debugging, and documentation,
positioning them as tools in development environments. The integration of LLMs into software
development workflows is facilitated through APIs, ofering a programmatic interface for
2Google Gemini. Available at: https://gemini.google.com
3OpenAI ChatGPT-4o. Available at: https://openai.com/chatgpt
4DeepSeek. Available at: https://www.deepseek.com
5Anthropic Claude. Available at: https://www.anthropic.com/index/claude
accessing the models’ capabilities [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. These APIs allow developers to send prompts or input
queries to the LLM and receive generated text responses, which can be parsed, transformed into
source code, and incorporated into local projects within an Integrated Development Environment
(IDE). Through this mechanism, LLMs assist in various stages of the software engineering
process, including automated code generation, unit test creation, and debugging support, thereby
contributing to increased development productivity and reduced manual efort [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>Prompt engineering has become a fundamental interface for human-AI interaction,
particularly in LLM-assisted software development. To fully harness the potential of LLMs, it is crucial
to devise efective strategies for designing and optimizing prompts. High-quality prompt design
is vital in improving generated content’s accuracy, relevance, and utility, enhancing the eficiency
and reliability of AI-assisted development workflows. By carefully structuring input queries
and layering them with contextual instructions, developers can guide the model’s behavior to
achieve more predictable and purposeful outputs. Moreover, managing virtual environments
represents another key aspect of LLM-assisted software development. A virtual environment
encapsulates a dedicated interpreter, project-specific libraries, and binaries, ensuring isolation
from environments and system-wide installations. This isolation is essential for maintaining
reproducibility, avoiding dependency conflicts, and supporting modular development
practices. Integrating LLMs with virtual environments enables robust and streamlined development
workflows, as the models can assist in code generation, environment setups, and configuration.
In more advanced scenarios, LLMs may be employed to identify and solve package conflicts,
enhancing the automation and reliability of the software development process.</p>
      <p>DLT (On-Chain)
Operator 1</p>
      <p>Node
Operator 3</p>
      <p>Node</p>
      <p>Operator 2</p>
      <p>Node
Operator N</p>
      <p>Node</p>
      <p>Off-Chain Jobs</p>
      <p>Oracle</p>
      <p>External External</p>
      <p>Data Data
External External</p>
      <p>Data Data
External External</p>
      <p>Data Data</p>
      <p>
        Adopting LLMs across domains such as software engineering becomes relevant to
emerging domains such as 6G network architectures, which demand programmable, resilient, and
autonomous systems. One application scenario involves blockchain-based oracle components
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], which secure bridges between Of-chain data sources and smart contracts deployed in
6G-enabled environments [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. These oracles enable mission-critical applications, such as
cognitive spectrum management, smart cities, and industrial, to access verified real-time
information while addressing key challenges such as scalability, interoperability, and latency.
Combining LLM-assisted development with 6G-oriented architectures represents a promising
research frontier, enabling the rapid prototyping of modular and trustworthy components
for next-generation wireless systems. Oracle components in the context of 6G networks act
as secure middleware between Of-chain data sources and smart contracts operating within
blockchain-based systems. These oracles are essential for applications that rely on real-world
data, such as dynamic spectrum management, decentralized authentication, and trustworthy
telemetry. They provide an interface for data acquisition, verification, and delivery of smart
contracts, ensuring the integrity and reliability of network services. Figure 1 illustrates a
conceptual oracle architecture integrating a 6G network with a blockchain layer. The architecture
is composed of three core components: (i) a Smart Contract Interface operating on a Distributed
Ledger Technology (DLT), defining and enforcing On-chain validation and storage rules [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ],
(ii) an Of-chain Oracle, which manages job execution, data processing, and the coordination of
trust policies [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ], and (iii) an External Data Layer [18], consisting of distributed data sources
that provide real-time information to be retrieved, verified, and delivered by the oracle.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>Recent advances in AI-assisted software development have increasingly focused on integrating
LLMs into stages of the software engineering lifecycle. Works have emerged targeting tasks
such as code generation, testing, error correction, and configuration management. This section
discusses the contributions of these works and highlights how each relates to AIPyCraft.</p>
      <p>
        Liu et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] proposed PromptV, a collaborative multi-agent framework for Verilog code
generation, where LLMs specialize in hardware description. Although the work emphasizes
coordination among agents, it lacks mechanisms for managing execution environments or
addressing the broader development lifecycle regarding testing and integration. PromptV focuses
only on isolated code-generation tasks. In contrast, AIPyCraft supports iterative construction
and orchestration of the lifecycle beyond the code synthesis. A complementary direction is
explored in LLMSecConfig [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], applying retrieval-augmented generation techniques to detect
and remediate misconfigurations in containerized environments. This solution incorporates
an automated correction pipeline and highlights the potential of LLMs in infrastructure-level
validation tasks. Nonetheless, its scope focuses on configuration-level errors and does not
encompass modular software design or lifecycle integration. In this context, AIPyCraft expands
this mode by integrating code generation with modular assembly and tracking artifacts.
      </p>
      <p>
        Plein et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] investigate test automation based on natural language input, employing
ChatGPT and CodeGPT to generate test cases from bug reports. Their results reinforce the feasibility
of leveraging LLMs for targeted development tasks. This investigation shows the viability of
generating tests from natural language, but it lacks integration with modular codebases or
version-controlled pipelines, which AIPyCraft enables through structured orchestration. In a
similar context, Nettur et al. introduced Cypress Copilot [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], ofering an AI assistant for
generating end-to-end test scripts in Web applications using Behavior Driven Development (BDD)
techniques. The tool generates structured and runnable code snippets by adopting few-shot
prompting with GPT-4o. While it efectively guides code creation, it does not address component
orchestration or project-level modularity. In contrast, AIPyCraft extends beyond isolated test
generation, ofering modular construction and integration across the development workflow.
      </p>
      <p>
        Cline [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is a project initiative integrating Claude 3.5/3.7 into the VSCode IDE, facilitating
autonomous, agent-based development workflows. Cline showcases how LLMs can be
emArticles/Projects
Liu et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
Ye et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
Plein et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
Nettur et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
Cline [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]
AIPyCraft (this work)
✗
✓
✗
✗
✓
✓
bedded into the developer’s IDE to execute commands, edit files, and manage context with
human oversight through manual approval mechanisms. Unlike AIPyCraft, which emphasizes
modularity and full-lifecycle project construction, Cline operates within existing codebases,
ofering flexibility but lacking structural guidance. Moreover, Cline does not target 6G network
code generation projects or deployment contexts. Table 1 reinforces the position of AIPyCraft as
a framework that unifies prompt engineering, modular orchestration, and lifecycle support. The
comparison includes key dimensions such as support for the development lifecycle, execution
within virtual environments, automated error correction, component-level structuring, modular
architecture, and relevance to 6G networks. While existing approaches tend to cover some of
these aspects in isolation, only AIPyCraft integrates them as a system.
4. AIPyCraft
      </p>
    </sec>
    <sec id="sec-4">
      <title>Architecture</title>
      <p>Leveraging LLM capabilities, AIPyCraft supports the creation and management of software
solutions in an automated and modular fashion, using an LLM API to help users generate, manage,
and run code solutions. The interaction with the tool occurs through a simple menu of options.
Moreover, AIPyCraft manages multiple stages of the open-source project lifecycle, including
creating, loading, running, and updating code. Figure 2 depicts our AI agent. AIPyCraft contains
three main abstractions: (i) Solution, (ii) Components, and (iii) Dispatcher. A Solution is an
abstraction representing a simple open-source project with multiple Components. Components
represent individual code files within a Solution Folder. An abstraction of a component
contains methods to execute the component and convert it to and from a dictionary format. The
Dispatcher holds the main menu of AIPyCraft, managing the lifecycle of solutions, including
creating, saving, loading, preparing virtual environments, and running them. Through the
Dispatcher, human users interact with the tool, selecting the actions and giving inputs that
complement pre-defined tuned prompts.</p>
      <p>One of the main characteristics of AIPyCraft is the solution management of LLMs using
APIs. In this context, AIPyCraft has: (i) AIConnector to analyze each solution component
remotely by employing an LLM, (ii) SolutionCreator to create new solutions and their
components using LLM assistance, (iii) SolutionLoader to load existing solutions from folders,
(iv) SolutionImporter to import external folders as solutions, and (v) SolutionImporter
Dispatcher
AIConnector</p>
      <p>Solution
Importer
Solution
Display
Component
Corrector</p>
      <p>…</p>
      <p>Component
to import repositories as a solution inside the tool. After creating a solution, AIPyCraft
offers: (vi) InstallationScriptGenerator function to install solution packages according
to their dependencies and avoid conflict, (vii) SolutionRunner to run solution calls and
automatically collects possible errors, (viii) SolutionDisplay to show solution details and
code, (ix) SolutionCorrecting (x), SolutionUpdater (xi), ComponentCorrector (xii),
and SolutionFeatureAdding to correct and improve a solution, and (xiii) AICodeParser
to extract, save, and detect the language of code blocks from AI responses.</p>
      <p>Figure 3 illustrates the workflow of AIPyCraft usage, showing the main steps while
collaboratively developing a new open-source project. SolutionCreator manages to create solutions
with multiple components/files and includes human approval in the feedback loop. This process
creates isolated environments within a solution’s directory, ensuring each solution has its
contained development environment with necessary dependencies installed. The user inputs a
solution name and semantic description. After, LLM generates a plan for components needed,
such as classes, accessory files, text, and main program. Each component is created individually
with AI assistance. Solutions are saved in the specified directory and rejected solutions are
stored so users can adjust the creative process in partnership with the LLM.</p>
      <p>The process begins by prompting the user to describe a desired new feature. AIPyCraft
constructs a detailed prompt for the LLM API, including the solution’s name, component
details, and the new feature description. AIPyCraft checks for a valid solution and locates
the required main program component. Moreover, AIPyCraft has an error correction system
that automatically fixes code/scripts when solutions fail during execution. In this context,
AIPyCraft works by (i) generating targeted prompts for AI code correction, (ii) analyzing each</p>
      <p>Start
solution component remotely by employing an LLM, (iii) handling code updates with proper
ifle I/O operations when required, (iv) trying to maintain naming conventions and parameter
consistency via prompt engineering, and (v) ensures proper imports to the main program.
Therefore, AIPyCraft writes the new code to the appropriate component file if updates are
found, ensuring its name remains unchanged. Finally, the process of code construction concludes
when no additional execution errors are detected or the user declines further feature additions.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Proof-of-Concept</title>
      <p>
        We test AIPyCraft with Google Gemini 2.5 Pro to correct a TOML script named config.toml.
This TOML script creates and runs the Of-chain Oracle using the Chainlink tool that requests
jobs for 6G based on the D6G architecture [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. We automated the execution of multiple
independent test runs (trials) for each Chainlink direct request job specification tested, as
illustrated in Figure 4. An initialization script is performed for each trial, passing the target
solution name and base path, to ensure that a TOML Script Tester starts from a known consistent
state. The Chainlink direct request job specification script is cleared before each test.
      </p>
      <p>Our multi-trial test script runs a Python-developed tester.py, providing it with the
maximum number of internal correction loops allowed (-LoopsValue), a unique identifier for the
trial (–run-id), the target solution name (–solution-name), the base path (–solutions-base-path),
and the specific correction instructions (–correction-prompt) to be used by the AI during that
trial’s correction phase. A total of 20 trials have been performed for each evaluated Chainlink
direct request job specification script. Table 2 shows our test configuration parameters. In each
trial, tester.py invokes AIPyCraft to load a TOML script tester containing the config.toml
Chainlink job specification and two human-developed testing programs. Two scenarios have
been evaluated: Scenario 1, which only verifies the syntax of the TOML script using the tomli
Python library, and Scenario 2, which provides successful Job script testing. In Scenario 1,
success means a call to tomli.load(f) returned without syntax errors. In Scenario 2, the
TOML script tester sends a new config.toml version created by the LLM to a Chainlink node
running on a virtual machine. A Job is created with the TOML script received and then run. In</p>
      <p>Run
experiment
1 Scenario 1
2 Scenario 2</p>
      <p>Chainlink</p>
      <p>Node
Job
run_tester_multiple</p>
      <p>20 trials
TOML Script Tester
this case, success means that the LLM generated a TOML script that enabled the creation of a
Job in the VM Chainlink node and ran without errors.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>We have integrated a new AI-assisted development tool called AIPyCraft with Gemini 2.5 Pro
to successfully correct and run TOML scripts (Jobs) in a local Chainlink node. The Jobs created
were able to listen to Of-chain Oracle requests within a real Chainlink implementation. Such
Mean Iterations (95% CI)
Mean Duration (95% CI)
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20</p>
      <p>Trial Number
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20</p>
      <p>Trial Number
(a) Number of interactions required.
(b) Time required.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20</p>
      <p>Trial Number</p>
      <p>AI-generated Jobs are essential to a disruptive 6G (D6G) wireless mobile network proposal being
developed by the authors. The Jobs will connect On-chain public smart contracts with Of-chain
6G service instances. Our experimental tests showed that the mean number of correction
interactions required to run a Job created by an LLM successfully was 1.05 attempts of 20
allowed. The mean time to achieve this result was 37.61 s, which is acceptable in deploying
new 6G services connected to On-chain smart contracts. In conclusion, these results prove
that our solution can be successfully applied in the modular development and testing of 6G
components with acceptable performance. Future work includes comparison among diferent
LLMs, evaluation of additional AI models for multi-agent collaboration, diferent AI-assisted
decision-making, expanding to other programming languages and environments, improving
security for AI-assisted workflows, and evaluation of diferent purpose Chainlink Jobs for 6G.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used Chat-GPT-4 to support the improvement
of grammar and spelling checking. After using these tool(s)/service(s), the author(s) reviewed
and edited the content as needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This work was conducted with partial financial support from CNPq, MCTIC/CGI.br/FAPESP
through PORVIR-5G project n 2020/05182-3, and RNP with resources from MCTIC Grant n
01245.020548/2021 − 07, under the Brazil 6G project.
and Potential Solutions, IEEE Network 37 (2023) 8–15.
[18] Y. Zuo, et al., A Survey of Blockchain and Artificial Intelligence for 6G Wireless
Communications, IEEE Communications Surveys &amp; Tutorials 55 (2023) 1–39.
[19] T. B. Brown, et al., Language Models are Few-Shot Learners, arXiv preprint
arXiv:2005.14165 (2020).
[20] A. Pasdar, et al., Connect API with Blockchain: A Survey on Blockchain Oracle
Implementation, ACM Computing Surveys 55 (2023) 1–39.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>V.</given-names>
            <surname>Venkatasubramanian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chakraborty</surname>
          </string-name>
          , Quo Vadis ChatGPT?
          <article-title>From large language models to Large Knowledge Models</article-title>
          ,
          <source>Computers &amp; Chemical Engineering</source>
          <volume>192</volume>
          (
          <year>2025</year>
          )
          <fpage>108895</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Buscemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A Comparative</given-names>
            <surname>Study</surname>
          </string-name>
          <article-title>of Code Generation using ChatGPT 3.5 across 10 Programming Languages</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2308</volume>
          .
          <fpage>04477</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Beyazit</surname>
          </string-name>
          , et al.,
          <article-title>Incremental testing in software product lines-an event based approach</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>2384</fpage>
          -
          <lpage>2395</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Anthropic</surname>
          </string-name>
          ,
          <source>Cline: Agentic Development with Claude 3.5/3</source>
          .7, https://github.com/anthropics/ cline,
          <year>2025</year>
          . Accessed: March 31,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>X.</given-names>
            <surname>Gu</surname>
          </string-name>
          , et al.,
          <article-title>On the Efectiveness of Large Language Models in Domain-Specific Code Generation</article-title>
          ,
          <source>ACM Trans. Softw. Eng. Methodol</source>
          .
          <volume>34</volume>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          , et al.,
          <article-title>PromptV: Large Language Model based Multi-agent Framework for Verilog Code Generation</article-title>
          ,
          <source>in: Conference on Artificial Intelligence</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plein</surname>
          </string-name>
          , et al.,
          <article-title>Automatic Generation of Test Cases based on Bug Reports: a Feasibility Study with Large Language Models</article-title>
          , in: International Conference on Software Engineering: Companion Proceedings,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Ye</surname>
          </string-name>
          , et al.,
          <source>LLMSecConfig: An LLM-Based Approach for Fixing Software Container Misconfigurations, arXiv preprint arXiv:2502</source>
          .
          <year>02009</year>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S. B.</given-names>
            <surname>Nettur</surname>
          </string-name>
          , et al.,
          <string-name>
            <surname>Cypress</surname>
            <given-names>Copilot</given-names>
          </string-name>
          :
          <article-title>Development of an AI Assistant for Boosting Productivity and Transforming Web Application Testing</article-title>
          ,
          <source>IEEE Access 13</source>
          (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          , et al.,
          <source>Attention Is All You Need, in: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>6000</fpage>
          -
          <lpage>6010</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhao</surname>
          </string-name>
          , et al.,
          <source>Explainability for Large Language Models: A Survey</source>
          ,
          <source>ACM Trans. Intell. Syst. Technol</source>
          .
          <volume>15</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>U.</given-names>
            <surname>Durrani</surname>
          </string-name>
          , et al.,
          <article-title>A decade of progress: A systematic literature review on the integration of ai in software engineering phases and activities (</article-title>
          <year>2013</year>
          -2023),
          <source>IEEE Access 12</source>
          (
          <year>2024</year>
          )
          <fpage>171185</fpage>
          -
          <lpage>171204</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Debbah</surname>
          </string-name>
          ,
          <article-title>Large language models for telecom</article-title>
          , in: 2023 Eighth International Conference on Fog and
          <source>Mobile Edge Computing (FMEC)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>3</fpage>
          -
          <lpage>4</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          , et al.,
          <source>Software Testing With Large Language Models: Survey</source>
          , Landscape, and
          <string-name>
            <surname>Vision</surname>
          </string-name>
          ,
          <source>IEEE Transactions on Software Engineering</source>
          <volume>50</volume>
          (
          <year>2024</year>
          )
          <fpage>911</fpage>
          -
          <lpage>936</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Cuellar</surname>
          </string-name>
          , et al., BSM-6G:
          <article-title>Blockchain-Based Dynamic Spectrum Management for 6G Networks: Addressing Interoperability and Scalability</article-title>
          ,
          <source>IEEE Access 12</source>
          (
          <year>2024</year>
          )
          <fpage>59464</fpage>
          -
          <lpage>59480</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>A. M. Alberti</surname>
          </string-name>
          , et al.,
          <article-title>Disruptive 6G architecture: Software-centric, AI-driven, and digital market-based mobile networks</article-title>
          ,
          <source>Computer Networks</source>
          <volume>252</volume>
          (
          <year>2024</year>
          )
          <fpage>110682</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Du</surname>
          </string-name>
          , et al.,
          <string-name>
            <given-names>A Novel</given-names>
            <surname>Oracle-Aided Industrial IoT Blockchain: Architecture</surname>
          </string-name>
          , Challenges,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>