<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>TLIMB - A Transfer Learning Framework for IMage Analysis of the Brain</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marc-André Schulz</string-name>
          <email>marc-andre.schulz@charite.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jan Philipp Albrecht</string-name>
          <email>jan-philipp.albrecht@mdc-berlin.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alpay Yilmaz</string-name>
          <email>alpay.yilmaz@student.hu-berlin.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexander Koch</string-name>
          <email>alexander.koch@charite.de</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dagmar Kainmüller</string-name>
          <email>Dagmar.Kainmueller@mdc-berlin.de</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ulf Leser</string-name>
          <email>leser@informatik.hu-berlin</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kerstin Ritter</string-name>
          <email>kerstin.ritter@charite.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Bernstein Center for Computational Neuroscience</institution>
          ,
          <addr-line>Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, Humboldt-Universität zu Berlin</institution>
          ,
          <addr-line>Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Department of Psychiatry and Neurosciences</institution>
          ,
          <addr-line>Charité - Universitätsmedizin Berlin, Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Max-Delbrueck-Center for Molecular Medicine</institution>
          ,
          <addr-line>Berlin</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Biomedical image analysis plays a pivotal role in advancing our understanding of the human body's functioning across diferent scales, usually based on deep learning-based methods. However, deep learning methods are notoriously data hungry, which poses a problem in ifelds where data is dificult to obtain such as in neuroscience. Transfer learning (TL) has become a popular and successful approach to cope with this issue, but is dificult to apply in practise due the many parameters it requires to set properly. Here, we present TLIMB, a novel python-based framework for easy development of optimized and scalable TL-based image analysis pipelines in the neurosciences. TLIMB allows for an intuitive configuration of source / target data sets, specific TL-approach and deep learning-architecture, and hyperparameter optimization method for a given data analysis pipeline and compiles these into a nextflow workflow for seamless execution over diferent infrastructures, ranging from multicore servers to large compute clusters. Our evaluation using a pipeline for analysing 10.000 MRI images of the human brain from the UK Biobank shows that TLIMB is easy to use, incurs negligible overhead and can scale across diferent cluster sizes.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;framework</kwd>
        <kwd>transfer learning</kwd>
        <kwd>biomedical image analysis</kwd>
        <kwd>nextflow</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Biomedical imaging, especially in neuroscience, is crucial
for understanding the complexities of the central nervous
system [15]. It allows for non-invasive examination of brain
structure and function, enabling clinical applications like
diagnosing and monitoring neurological and psychiatric
diseases [2]. Deep learning, with techniques such as
convolutional neural networks (CNNs) and transformer-based
architectures, show great promise in this domain [8]. Their
effectiveness in tasks such as lesion segmentation and disease
classification has been demonstrated [18, 20, 8]. However,
the success of these advanced architectures often hinges
on the availability of large and homogeneous datasets, a
challenge in biomedical settings due to their scarcity.</p>
      <p>Transfer learning (TL) has recently become popular for
overcoming the constraints of small and heterogeneous
datasets. In a nutshell, it allows leveraging a model trained
on a given source dataset for improving model performance
on a diferent target dataset [21]. However, applying TL in
neuroimaging practise has proven dificult, as it requires
the careful selection of a multitude of diferent yet close
interacting parameters, including the base image
analysis architecture (e.g. ResNet, diferent flavors of CNNs or
transformers), the concrete TL-method (e.g. fine-tuning,
multitask-learning), the concrete objective function, and the
source dataset to be used. Determining these parameters
manually in a framework like PyTorch is time-consuming
and error-prone, as it requires source code manipulation and
extensive experimentation to find optimal configurations.
These experimentations can be computationally extremely
time-consuming unless adequate parallel and/or distributed
infrastructures are available, which, however, makes
programming the analysis pipeline even more involved.</p>
      <p>In this work, we present TLIMB, a Transfer-Learning
Framework for Image Analysis of the Brain. TLIMB is
programmed in the widely-used general-purpose language
Python and based on PyTorch Lightning1. With TLIMB,
users specify their TL-pipeline in the form of simple and
intuitive configuration files, which are then compiled into a
concrete image analysis workflow in Nextflow [6], a popular
and powerful workflow engine than can execute an analysis
over a wide range of infrastructures, ranging from single
servers to large compute clusters. With TLIMB, researchers
thus are able to easily assess the efectiveness of diferent
TL setups across diverse datasets and environments.</p>
      <p>
        We specifically designed TLIMB as a framework and not
as a proper domain specific language (i.e., a programming
language tailored to a particular problem; DSL) because
of the advantages of this approach in terms of flexibility,
ease of creation, extensibility, and seamless integration with
existing tools [
        <xref ref-type="bibr" rid="ref11">13, 1</xref>
        ]. A Python-based framework, in
particular, provides a familiar environment for data scientists,
capitalizing on the language’s popularity and compatibility
with established machine and deep learning frameworks
(like PyTorch).
      </p>
      <p>Through a series of experiments following the
"brainage" paradigm [4], a widely-used method for assessing brain
health through neuroimaging data, we validated the
platform’s capability to create a diverse landscape of TL-based
pipelines and to execute them seamlessly over any
infrastructure supported by Nextflow.</p>
      <sec id="sec-1-1">
        <title>1https://pypi.org/project/pytorch-lightning/</title>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        There has been a number of eforts to develop DSLs as well
as frameworks for machine learning-based (image) data
analysis [9]. OptiMl, a DSL tailored for machine learning
tasks, seeks to provide an implicitly parallel, expressive, and
high-performance alternative to MATLAB and C++ [17].
However, it does not address TL and is agnostic to the data
types analysed and thus requires some efort for using it in
image analysis. Extending it with TL abilities would be
nontrivial due to its design as a DSL. P-Hydra employs Transfer
Learning and Multitask learning for image analysis in cancer
detection, aiming to validate its algorithmic efectiveness
and establish a baseline for other methods [
        <xref ref-type="bibr" rid="ref1">11</xref>
        ]. In contrast
to TLIMB, the method is implemented in a single pipeline
and not designed as configurable framework. Furthermore,
our approach supports multiple heads per model, enabling
a broader spectrum of TL-methods. Ilastik, designed as an
interactive tool for machine-learning-based (bio)image
analysis, addresses challenges associated with manual image
analysis by providing pre-defined workflows for
segmentation, object classification, counting, and tracking, with a
user-friendly interface emphasizing accessibility for
nonprogrammers [3]. In contrast to ilastik, our framework
concentrates on training neural networks for TL-based analysis.
Finally, SimpleITK, is a software package designed for
image analysis that provides a simplified interface for flexible
and reproducible computational workflows [24], aligning
closely with the goals of our framework. While SimpleITK
streamlines image analysis through Jupyter Notebooks and
introduces various abstractions, our framework adopts a
diferent approach, allowing users to initiate analysis by
starting with our framework components and building upon
them as needed.
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methods</title>
      <sec id="sec-3-1">
        <title>Architecture of TLIMB</title>
        <p>The core of the framework is constructed around three
primary components: the DataModule, Architecture,
and ObjectiveFunction. These are orchestrated within
a Scenario to create a comprehensive TL pipeline. Users
can execute diferent configurations of these Scenarios, such
as for hyperparameter tuning or model comparisons, by
automatically generating a Nextflow workflow from their
definitions. An overview of TLIMB’s architecture is shown
in Figure 1.</p>
        <p>The Scenario component is responsible for the training
logic: it orchestrates training, validation, and testing by
sourcing data from the DataModule, processing it through
the specified deep learning Architecture, calculating
losses using the ObjectiveFunction, and executing the
optimization step. This abstraction level facilitates not only
the conventional sequential pretrain-finetune TL workflows,
but also enables the implementation of workflows that
require simultaneous processing of both pretraining and
finetuning data, such as semi-supervised learning algorithms
[19]. Scenarios are designed to be data-operation agnostic,
i.e., independent of the specific deep learning architecture
and objective function, thereby enhancing the modularity
of the design.</p>
        <p>The Architecture component relates to the adaptable
configuration of network layers and nodes, providing users
with the versatility to select from predefined architectures
or to incorporate their own custom designs by referencing
them in the configuration file. To facilitate eficient TL,
architectures are decomposed into two main elements: the
encoder, which is often repurposed from the source task,
and the head, which is specific to and replaceable for the
target task. This modular structure supports a variety of TL
strategies, ensuring adaptability to methodologies such as
the core train-fine-tune paradigm and multitask learning.</p>
        <p>Objective Functions embody the core logic of TL,
composed of a primary objective (such as cross-entropy for
classification) and an auxiliary objective (such as an elastic
penalty on weights during fine-tuning or reconstruction
losses in semi-supervised training). During the training
phase, this class considers batches and the architecture from
the scenario class to computes the loss and performance
metrics. In pursuit of greater modularity, Objective
Functions have been architectured to remain decoupled from
other framework components. For instance, employing an
Objective Function designed for multitask learning does not
require predefined knowledge of the number of heads within
the configuration. This design choice facilitates transitions
of the objective function, enhancing the user experience
and adaptability within the TL workflow.</p>
        <p>Our DataModule defines the handling of diverse data
types, ranging from 3D brain MRI data to 1D fMRI time
series. It encompasses data-specific loading,
preprocessing, and data augmentation routines. It ensures that batch
preparation conforms to a defined structure and assembles
DataLoaders. In contrast to the standard PyTorch
Lightning (see below), our method imposes constraints on
DataLoader instantiation, mandates a uniform Dataset structure,
and centralizes all data-related augmentations and
transformations within the DataModule itself. Such a
separationof-concerns supports simple substitutability of Datasets
and DataModules. This module also inherits several
features from the PyTorch Lightning DataModule, such as
the on_after_batch_transfer and on_before_batch_transfer
hooks. These hooks grant users the capability to refine batch
post-retrieval but prior to their delivery to the Scenario,
enabling, for instance, the ofloading of resource-intensive data
augmentation strategies to a GPU. This design promotes
user-driven adaptability in our framework, ensuring the
lfexibility to customize components while preserving the
integrity of essential operations.</p>
        <p>Datasets, pivotal elements within the DataModule, are
tasked with providing the necessary data and associated
labels. The DataModule delineates the procedures for
processing a certain category of data, whereas the Dataset is
explicit about the specific input files to utilize, their
locations within the file system, and the particular labels to
retrieve (for instance, selecting the participant’s sex for a
pretraining task, and later using the same dataset to return
the participant’s age, thus facilitating straightforward label
specification). A DataModule can include multiple Datasets,
accommodating various TL strategies that incorporate data
from diverse sources. Each Dataset implements a custom
getitem method to ensure the standardized conveyance of
images and labels to the DataModule. This getitem method
invariably produces a tuple, which includes an image paired
with a list of labels, thereby adapting to the diverse labeling
demands posed by diferent Objective Functions. Varied
learning paradigms such as Multitask, Pre-train Fine-tune,
and Unsupervised Domain Adaptation require unique label
arrangements.</p>
        <p>The Configuration component serves as an important
tool for managing configuration within our framework,
offering users the ability to customize every aspect of their
workflow. Unlike PyTorch Lightning, which primarily
focuses on non-structural hyperparameters, our
Configuration empowers users to tailor Scenarios, Architectures,
Objective-functions, Datasets, DataModules, trainers, and
optimizer parameters, ensuring high configurability and
modularity. This user-centric approach minimizes coding
eforts, allowing users to predominantly interact with the
Configuration instead. The framework seamlessly
integrates with PyTorch Lightning components, enabling the
utilization of features like early stopping and automatic
optimizers, efortlessly configurable through the provided
configuration file. To ensure reproducibility, each
worklfow is associated with a defined configuration, facilitating
run reproduction. The configuration provides essential
details such as splits, which define the distribution of images
across training, validation, and testing sets. It also includes
adjustable seeds to guarantee consistent runs, except when
randomness is introduced by the user. Users are not
conifned to predefined components; instead, our framework
provides interfaces for Objective-functions, Architectures,
DataModules, Datasets, and Scenarios, making it easy to
implement specialized versions of these components, such
as a new Objective-function.
1–6</p>
      </sec>
      <sec id="sec-3-2">
        <title>Integrated models and implementation</title>
        <p>Our framework, implemented in Python, provides a
seamless integration of PyTorch and incorporates PyTorch
Lightning components. This integration ofers multiple benefits,
such as support for distributed training, compatibility with
multi-GPU setups, and optimized performance for various
machine learning tasks. The framework’s alignment with
Python and PyTorch’s popularity in the research
community simplifies the learning curve, making it a user-friendly
and accessible option for TL projects. However, it also ofers
signficant additional functionalities compared to PyTorch
Lightning. For instance, our TL Command-Line Interface
(CLI) distinguishes itself from the PyTorch Lightning CLI by
facilitating the passage of parameters from the DataModule
to the Scenario during initialization. This enables users to
customize various aspects, such as output size and input
size.</p>
        <p>The framework is used mainly via configuration files.
Users specify key components such as a particular
’DataModule’ for input data specifications, a ’Dataset’ for data
ifle and label paths, ’Architecture’ for neural network
structure, ’Objective’ for the transfer learning strategy, and
’Scenario’ for training details. The framework supports class
path parsing, allowing users to define parameters via class
references, which can be particularly useful for complex
configurations. To facilitate hyperparameter tuning,
multiple variants of these parameters can be provided. The
framework’s workflow generator leverages this information
to create Nextflow workflows, which orchestrate the
execution of tasks across the computational infrastructure. This
streamlined approach enables systematic exploration and
eficient optimization of model parameters.</p>
        <p>TLIMB comes with a number of readily available models
and configurations for its diferent components. Regarding
architectures, it currently ofers 3d-ResNets of diferent
depths [22], the 3d Simple-Fully-Convolutional-Network
[14], as well as vision and swin transformers , three highly
popular imaging architectures. ResNet utilizes shortcut
connections to enhance training performance, while SFCN is a
lightweight 3D convolutional neural network specifically
tailored for 3D neuroimaging data. Transformers are fully
connected deep encoder-decoder stacks with self-attention.
Several customization options, such as filter and kernel sizes
and addition of dropout layers are available for each. The
framework also integrates pre-processing, neuroimaging
domain specific data augmentation, and data transformation
techniques.</p>
        <p>Regarding TL algorithms, our framework encompasses
ifve methods: Pre-train-fine-tune, multitask learning,
selfsupervised semi-supervised learning, elastic penalty, and
unsupervised domain adaptation. Pre-train-fine-tune
involves using a pre-trained network for a target task, while
elastic penalty introduces an L² penalty to preserve learned
features during fine-tuning. Multitask learning optimizes
models by sharing representations between related tasks.
Self-supervised semi-supervised learning leverages both
labeled and unlabeled data. Unsupervised domain
adaptation allows training deep models using labeled data from
a source domain and unlabeled data from a target domain.
On overview of these techniques can be found in [10].</p>
        <p>TLIMB’s objective functions mirror PyTorch Lightning
training/validation/testing steps. Hyperparameter
optimization is facilitated through grid search and random
search. TLIMB comes with three directly usable
DataModules, namely BaseDataModule, CropCenterDataModule,
and BioImageDataModule. Additionally, several pre-defined
Datasets from the Human Connectome Project are
readily available, but researchers can efortlessly incorporate
any image analysis dataset of their choice by utilizing the
provided interface.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Nextflow as workflow manager</title>
        <p>Nextflow is a mature and popular scientific workflow
engine [7]. Workflows in Nextflow are written in a proper
workflow language based on Groovy and are executed by a
workflow engine which controls data dependencies,
maximises parallelism in task executions, and supports
reproducibility by a sophisticated logging mechanism. Workflows
can either be executed locally (non distributed) by the
system itself, or passed on to popular resource managers, such
as Slurm or Kubernetes [25], for scheduling on arbitrarily
large clusters. In TLIMB, we utilize Nextflow to assemble TL
workflows from user-provided configurations into a
worklfow script. This script can then be executed in parallel and
distributed across all supported infrastructures, significantly
accelerating the processing speed.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>For the evaluation of the TLIMB framework, we used
T1weighted brain images from the UK Biobank [12]. To
streamline the evaluation process, we processed images by
applying linear registration and extracting the central axial slices.
This reduced the dimensionality of the data, allowing us
to expedite the training process. We created three subsets
of randomly sampled images: 10,000 for pre-training, 500
for fine-tuning, and 1,000 for the test set. Models were
pre-trained on age regression, and fine-tuned on sex
classiifcation.</p>
      <p>To assess usability improvements, we specified a
simpliifed search space comprising two diferent neural network
architectures (ResNet-18 and a Vision Transformer), three
diferent learning rates ( 10−4, 10−3, 10−2), and an optional
elastic penalty loss [23] as an advanced fine-tuning
technique. Both models were pre-trained for 10 epochs and
ifne-tuned for one epoch. Such limited training time would
be insuficient for real world applications, but our aim here is
to investigate the framework’s usability and computational
overhead rather than achieving state-of-the-art accuracy.
Each variant was implemented in two ways: manually using
PyTorch with PyTorch Lightning, and through our TLIMB
framework compiled into a Nextflow workflow. These
implementations were then run in three diferent scenarios:
manually without the framework, with the framework
sequentially, and with the framework in parallel. The primary
metrics for evaluation were the execution times and the
lines of code required for each scenario. Execution times are
documented in Table 1, illustrating the comparison between
running the processes with and without the framework,
both sequentially and in parallel.</p>
      <p>Additionally, we conducted a minimal set of experiments
illustrating how TLIMB may be used in practice. On the
same data set and using the ResNet-18 architecture, we
compared fine-tuning efectiveness for diferent numbers of
frozen layers in the pre-trained model. Freezing lower
layers of a pre-trained model reduces the number of trainable
parameters and thus reduces the risk of overfitting during
the fine-tuning process. Metrics for pre-training and
finetuning performance are shown in Table 3 and 4 respectively.
TLIMB achieved expected levels of accuracy, in line with
other studies [5, 16].</p>
      <p>Execution Time: The execution times indicated
minimal to no computational overhead when using the TLIMB
framework. The parallel execution with nextflow
significantly reduced the time compared to the sequential runs,
showcasing the framework’s scalability (see Table 1).</p>
      <p>Lines of Code: A notable reduction in lines of code
was observed when using TLIMB, emphasizing the ease of
use and time savings in coding. The framework abstracted
many of the repetitive tasks, such as setting up data loaders,
model configurations, and hyperparameter tuning, which
contributed to a more streamlined development process.</p>
      <p>Prediction Performance: Although no exact replication
of literature results was attempted at the time of writing,
our preliminary results are compatible with literature
expectations.</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion and outlook</title>
      <p>In this study, we introduce our innovative solution – a
tailored implementation and evaluation platform for TL
techniques in biomedical imaging applications. Guided by
specific requirements, we opted for a comprehensive
framework over a DSL. Our framework comprises two key
components: firstly, a Python framework built upon PyTorch
Lightning, facilitating diverse user-defined TL tasks.
Secondly, a workflow generator and executor ensuring
scalability. We provide in-depth descriptions of both components,
highlighting their functionalities and capabilities. To
ascertain the efectiveness and utility of our framework, we
applied it to the "brain-age" paradigm. In this context, the
assessment of brain-age deviations from chronological age
serves as a metric for evaluating brain health. Our
framework demonstrates minimal or no computational overhead,
while significantly reducing the number of lines of code
required. In the pursuit of refining our framework, we
propose several avenues for future development. Firstly, we
recommend the establishment of a standardized template
to streamline the evaluation of TL methods. This template
would simplify result and methodology comparisons among
researchers, fostering a more cohesive and eficient research
environment. Moreover, to enhance the eficiency of model
tuning, we advocate for the implementation of additional
hyperparameter optimization methods within our
framework. Specifically, techniques like Bayesian Optimization
can be incorporated to further optimize model performance.
Furthermore, to minimize manual intervention and improve
user experience, we suggest enhancing the workflow
manager. This enhancement includes the addition of automatic
ranking capabilities, which will facilitate a more eficient
comparison and selection of the best-performing models,
guided by predefined evaluation metrics.</p>
      <p>Alexander Alexandrov et al. “Implicit parallelism
through deep language embedding”. In: SIGMOD.
2015, pp. 47–61.</p>
      <sec id="sec-5-1">
        <title>Sergi Valverde et al. “Improving automated multiple</title>
        <p>sclerosis lesion segmentation with a cascaded 3D
convolutional neural network approach”. In: NeuroImage
155 (2017).</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements and Funding</title>
      <p>This work was funded by FONDA (DFG; SFB 1404; Project
ID: 414984028).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Rohit</given-names>
            <surname>Bakshi</surname>
          </string-name>
          et al. “
          <article-title>MRI in multiple sclerosis: current status and future prospects”</article-title>
          .
          <source>In: The Lancet Neurology 7.7</source>
          (
          <issue>2008</issue>
          ), pp.
          <fpage>615</fpage>
          -
          <lpage>625</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Stuart</given-names>
            <surname>Berg</surname>
          </string-name>
          et al. “
          <article-title>Ilastik: interactive machine learning for (bio) image analysis”</article-title>
          .
          <source>In: Nature methods 16</source>
          .12 (
          <year>2019</year>
          ), pp.
          <fpage>1226</fpage>
          -
          <lpage>1232</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>James H Cole</surname>
          </string-name>
          et al. “
          <article-title>Predicting brain age with deep learning from raw imaging data results in a reliable and heritable biomarker”</article-title>
          .
          <source>In: NeuroImage</source>
          <volume>163</volume>
          (
          <year>2017</year>
          ), pp.
          <fpage>115</fpage>
          -
          <lpage>124</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>James H.</given-names>
            <surname>Cole</surname>
          </string-name>
          . “
          <article-title>Multimodality neuroimaging brainage in UK biobank: relationship to biomedical, lifestyle, and cognitive factors”</article-title>
          .
          <source>In: Neurobiology of Aging 92 (Aug</source>
          .
          <year>2020</year>
          ), pp.
          <fpage>34</fpage>
          -
          <lpage>42</lpage>
          . issn:
          <fpage>0197</fpage>
          -
          <lpage>4580</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.neurobiolaging.
          <year>2020</year>
          .
          <volume>03</volume>
          .014. url: https:// www.ncbi.nlm.nih.gov/pmc/articles/PMC7280786/ (visited on 03/04/
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>P.</given-names>
            <surname>Di Tommaso</surname>
          </string-name>
          et al. “
          <article-title>Nextflow enables reproducible computational workflows”</article-title>
          .
          <source>In: Nat Biotechnol 35.4</source>
          (
          <issue>2017</issue>
          ), pp.
          <fpage>316</fpage>
          -
          <lpage>319</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Paolo</given-names>
            <surname>Di</surname>
          </string-name>
          Tommaso et al. “
          <article-title>Nextflow enables reproducible computational workflows”</article-title>
          .
          <source>In: Nature biotechnology 35.4</source>
          (
          <issue>2017</issue>
          ), pp.
          <fpage>316</fpage>
          -
          <lpage>319</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Fabian</given-names>
            <surname>Eitel</surname>
          </string-name>
          et al. “
          <article-title>Promises and pitfalls of deep neural networks in neuroimaging-based psychiatric research”</article-title>
          .
          <source>In: Experimental Neurology</source>
          <volume>339</volume>
          (
          <year>2021</year>
          ), p.
          <fpage>113608</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <given-names>Joan</given-names>
            <surname>Giner-Miguelez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Abel</given-names>
            <surname>Gómez</surname>
          </string-name>
          , and Jordi Cabot.
          <article-title>“A domain-specific language for describing machine learning datasets”</article-title>
          .
          <source>In: Journal of Computer Languages</source>
          <volume>76</volume>
          (
          <year>2023</year>
          ), p.
          <fpage>101209</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <given-names>Padmavathi</given-names>
            <surname>Kora</surname>
          </string-name>
          et al. “
          <article-title>Transfer learning techniques for medical image analysis: A review”</article-title>
          .
          <source>In: Biocybernetics and Biomedical Engineering</source>
          <volume>42</volume>
          .1 (
          <issue>2022</issue>
          ), pp.
          <fpage>79</fpage>
          -
          <lpage>107</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <given-names>Jiyoung</given-names>
            <surname>Lee. “P-Hydra</surname>
          </string-name>
          :
          <article-title>Bridging Transfer Learning And Multitask Learning”</article-title>
          .
          <source>In: Master Thesis</source>
          , University of Friburg (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <source>[13] [15] [17] [19] [21] [23]</source>
          [25]
          <string-name>
            <surname>Thomas J Littlejohns</surname>
          </string-name>
          et al. “
          <article-title>The UK Biobank imaging enhancement of 100,000 participants: rationale, data collection, management and future directions”</article-title>
          .
          <source>In: Nature communications 11.1</source>
          (
          <issue>2020</issue>
          ), p.
          <fpage>2624</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Marjan</given-names>
            <surname>Mernik</surname>
          </string-name>
          , Jan Heering, and
          <string-name>
            <surname>Anthony</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Sloane</surname>
          </string-name>
          . “
          <article-title>When and How to Develop Domain-Specific Languages”</article-title>
          .
          <source>In: ACM Comput. Surv</source>
          .
          <volume>37</volume>
          .4 (
          <issue>Dec</issue>
          .
          <year>2005</year>
          ), pp.
          <fpage>316</fpage>
          -
          <lpage>344</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Han</given-names>
            <surname>Peng</surname>
          </string-name>
          et al. “
          <article-title>Accurate brain age prediction with lightweight deep neural networks”</article-title>
          .
          <source>In: Medical image analysis 68</source>
          (
          <year>2021</year>
          ), p.
          <fpage>101871</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Rangaraj M Rangayyan.</surname>
          </string-name>
          <article-title>Biomedical image analysis</article-title>
          .
          <source>CRC press</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Marc-Andre Schulz</surname>
          </string-name>
          et al. “
          <article-title>Performance reserves in brain-imaging-based phenotype prediction”</article-title>
          .
          <source>In: Cell Reports 43.1</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Arvind</given-names>
            <surname>Sujeeth</surname>
          </string-name>
          et al. “
          <article-title>OptiML: an implicitly parallel domain-specific language for machine learning”</article-title>
          . In: ICML.
          <year>2011</year>
          , pp.
          <fpage>609</fpage>
          -
          <lpage>616</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Jesper E Van Engelen</surname>
          </string-name>
          and
          <string-name>
            <surname>Holger H Hoos</surname>
          </string-name>
          .
          <article-title>“A survey on semi-supervised learning”</article-title>
          .
          <source>In: Machine learning 109.2</source>
          (
          <issue>2020</issue>
          ), pp.
          <fpage>373</fpage>
          -
          <lpage>440</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Sandra</given-names>
            <surname>Vieira</surname>
          </string-name>
          , Walter HL Pinaya, and Andrea Mechelli. “
          <article-title>Using deep learning to investigate the neuroimaging correlates of psychiatric and neurological disorders: Methods and applications”</article-title>
          .
          <source>In: Neuroscience &amp; Biobehavioral Reviews</source>
          <volume>74</volume>
          (
          <year>2017</year>
          ), pp.
          <fpage>58</fpage>
          -
          <lpage>75</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Karl</given-names>
            <surname>Weiss</surname>
          </string-name>
          ,
          <article-title>Taghi M Khoshgoftaar,</article-title>
          and DingDing Wang.
          <article-title>“A survey of transfer learning”</article-title>
          .
          <source>In: Journal of Big data 3.1</source>
          (
          <issue>2016</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>40</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Wanni</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <surname>You-Lei Fu</surname>
          </string-name>
          , and Dongmei Zhu.
          <article-title>“ResNet and Its Application to Medical Image Processing: Research Progress and Challenges”</article-title>
          . In: Computer Methods and Programs in Biomedicine (
          <year>2023</year>
          ), p.
          <fpage>107660</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>LI</given-names>
            <surname>Xuhong</surname>
          </string-name>
          ,
          <string-name>
            <surname>Yves Grandvalet</surname>
          </string-name>
          , and Franck Davoine. “
          <article-title>Explicit inductive bias for transfer learning with convolutional networks”</article-title>
          .
          <source>In: International Conference on Machine Learning. PMLR</source>
          .
          <year>2018</year>
          , pp.
          <fpage>2825</fpage>
          -
          <lpage>2834</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <given-names>Ziv</given-names>
            <surname>Yaniv</surname>
          </string-name>
          et al. “
          <article-title>SimpleITK image-analysis notebooks: a collaborative environment for education and reproducible research”</article-title>
          .
          <source>In: Journal of digital imaging 31.3</source>
          (
          <issue>2018</issue>
          ), pp.
          <fpage>290</fpage>
          -
          <lpage>303</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <given-names>Naweiluo</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <surname>Huan Zhou</surname>
            , and
            <given-names>Dennis</given-names>
          </string-name>
          <string-name>
            <surname>Hoppe</surname>
          </string-name>
          . “
          <article-title>Containerization for High Performance Computing Systems: Survey and Prospects”</article-title>
          .
          <source>In: IEEE Transactions on Software Engineering 49.4</source>
          (
          <issue>2022</issue>
          ), pp.
          <fpage>2722</fpage>
          -
          <lpage>2740</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>