<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Paris, France
∗Corresponding author.
£ sarah.lang@uni-graz.at(S. A. Lang); bernhard.liebl@uni-leipzig.d(Be. Liebl);
burghardt@informatik.uni-leipzig.d(Me. Burghardt)
ç https://sarahalang.com(S. A. Lang);
https://www.mathcs.uni-leipzig.de/ifi/forschung/computational-humanitie(sB/. Liebl);
https://www.mathcs.uni-leipzig.de/personenprofil/mitarbeiter/juniorprof-dr-manuel-burgh(Mar. dBturghardt)
ȉ</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Toward a Computational Historiography of Alchemy: Challenges and Obstacles of Object Detection for Historical Illustrations of Mining, Metallurgy and Distillation in 16th-17th Century Print</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sarah A. Lang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>BernhardLiebl</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ManuelBurghard</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computational Humanities Research Group, University of Leipzig</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department Centre for Information Modelling (ZIM), University of Graz</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>This study explores the use of modern computer vision methods for object detection in historical images extracted from 16th-17th century printed books containing illustrations of distillation, mining, metallurgy, and alchemical apparatus. We found that the transfer of knowledge from contemporary photographic data to historical etchings proves less e昀ective than anticipated, revealing limitations in current methods like visual feature descriptors, pixel segmentation, representation learning, and object detection with YOLOv8. These 昀椀ndings highlight the stylistic disparities between modern images and early print illustrations, suggesting new research directions for historical image analysis.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;computer vision</kwd>
        <kwd>object detection</kwd>
        <kwd>alchemy</kwd>
        <kwd>chymistry</kwd>
        <kwd>early-modern print</kwd>
        <kwd>metallurgy</kwd>
        <kwd>mining</kwd>
        <kwd>distillation</kwd>
        <kwd>annotation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        the norm in the kitchens and makeshi昀琀 laboratories of the past has been seen by Smith as the
precursor to the natural sciences and chemistry as we know them toda5y5,[p. 292]. Morris
argues that chemical laboratories in the modern sense emerged with the replacement of
multipurpose or make-shi昀琀 spaces, which were not speci昀椀cally designed for carrying out chemical
operations, with professionalized work environments for performing chemical and
metallurgical operations 4[1, p. 19–20]. He further states that this rise of chemical laboratories coincides
with the boom of a genre of metallurgical technical treatis4e0s][. Empirical evidence of these
椀昀rst laboratories remains scarce, with only a handful of alchemical laboratories discovered so
far [
        <xref ref-type="bibr" rid="ref35">34</xref>
        ]. This is where early modern handbooks on distillation, metallurgy and mining, rich
with illustrations, become invaluable. These texts provide unmatched insight into the
laboratories, processes, and practices in thaertes technicae at the time, illustrating the underpinnings
of the era’s chemistry and technology. Despite their signi昀椀cance for the history of technology
and the Chemical Humanities 4[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], these books remain relatively understudied to this day.
      </p>
    </sec>
    <sec id="sec-2">
      <title>1.1. Depicting mining, metallurgy, and distillation</title>
      <p>
        During the proto-industrial revolution, mining and metallurgy 昀氀ourished, leading to the
emergence of encyclopedic compendia of technological apparatus and processes. These include
works such as Georgius Agricola’Dse re metallica libri XII [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], Vannoccio Biringucci’sDe la
pirotechnia Libri X [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], Lazarus Ercker’sAula subterranea [
        <xref ref-type="bibr" rid="ref24">23</xref>
        ], and Giambattista della Porta’s
De distillatione libri IX [
        <xref ref-type="bibr" rid="ref45">44</xref>
        ]. Metallurgical technical treatises began to become a staple in the
genre of didactic manuals and were frequently accompanied by technical illustrations.
Beginning with smaller treatises, grander montanist works started appearing by the mid-16th
century, such as Vannoccio Biringuccio’Dse la Pirotechnia (1540) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and Georg Agricola’Dse
Re Metallica (1556) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. This knowledge, always accessible in books, as Michael Giesecke has
emphasized, was so attractive because it replaced the exchange with experts and, thus, o昀琀en
made expensive and time-consuming journeys unnecessary2[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Consequently, ease of 昀椀nding
relevant passages, through 昀椀tting illustrations or knowledge organization tools such as indices,
was pivotal to their success. Besides metallurgical-focused works, distillation treatises also
became popular in the 16th century3[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Particularly in昀氀uential was Hieronymus Brunschwig’s
Liber der Arte Distillandi (Straßburg 1512) [11] or Walther Hermann Ry昀’s Distillation Book
(Frankfurt am Main 1545) 5[0]. Brunschwig’s treatises have been published in a bewildering
variety of versions, translations, and re-edition3s5[, p. 284–287].1
      </p>
    </sec>
    <sec id="sec-3">
      <title>1.2. Research agenda and the case for automatic object detection</title>
      <p>
        Since book illustration was expensive, early modern printers opportunistically reused
illustrations from woodcuts and copper plates, thereby separating the images from their original
contexts. Thus, illustrations would be commissioned for one speci昀椀c publication, rendering
lots of detail and providing an alternative communication medium for the message expressed
in the text of that particular book, and then reused in other contexts where they 昀椀tted more
1The Strasbourg doctor and pharmacist 昀椀rst published hiLsiber de Arte Distillandi De Simplicibus in 1500. This is
referred to by research as the ‘small distillation book’. Twelve years later, the author followed up with a more
voluminousLiber de Arte Distillandi De Compositis, known as the ‘large distillation book1’1[].
or less well, much like modern stock photograph2y8][. However, this means that not every
image used in early modern print was made speci昀椀cally to illustrate the exact matter discussed
in a text passage. Medical books, herbiaries and distillation books are a medium particularly
rich in illustration, for which even legal battles for ‘copyright’ are not unheard of. Especially
later richly illustrated encyclopedic works could only be 昀椀nanced due to their reuse of earlier
image material. What does this mean for pragmatic literature though? Do the images
faithfully represent the processes being described and the equipment needed to carry them out?
We know, for example, that Lazarus Ercker’Asula Subterranea [
        <xref ref-type="bibr" rid="ref23">22</xref>
        ] (or ‘Bergwercksarten’) is a
true handbook, in the sense that it is detailed enough so that one can replicate the processes
described. But can this be true for all other books from that genre as well, given what we know
about the practices of illustration reuse in historical print?
      </p>
      <p>It is in this context that we propose to apply computer vision techniques to automatically
detect the illustrations in these books. Being able to detect relevant objects in digitised book pages
is a crucial 昀椀rst step for a quantitative Distant Viewing 5[] analysis of such apparatus within
early modern chymical and pragmatic literature. In this short paper, we discuss challenges and
obstacles we encountered during a 昀椀rst series of experiments in annotating a sample of such
illustrations and training di昀erent approaches for object detection for historical illustrations
of mining, metallurgy, and distillation in 16th–17th century print.</p>
      <sec id="sec-3-1">
        <title>2. Detecting alchemical apparatus</title>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>2.1. Related work</title>
      <p>
        We presume that a computational analysis of illustration practices can yield answers to the
questions outlined above. As for related work, there is one branch of works that uses
computer vision methods on illustrations in 15th/16th century pri3n7t,[
        <xref ref-type="bibr" rid="ref22 ref29">21, 28</xref>
        ]. However, these
approaches are less concerned with the recognition of individual objects and more focused
on identifying illustrations as a whole, particularly their reuse in di昀erent books. Cormier et
al. [
        <xref ref-type="bibr" rid="ref18">17</xref>
        ] use machine learning approaches to classify illustrations as either woodcut or
copperplate engravings. An interactive Visual Analytics System (VeCHArt) for comparing copies or
di昀erent states of a print is proposed by P昀氀üger et al. [
        <xref ref-type="bibr" rid="ref43">42</xref>
        ]. Valleriani et al.5[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] present an
empirical study on the visual similarity of early modern scienti昀椀c illustrations on cosmology
while Kaoua et al. 3[0] provide insights from a large-scale study on image collation, as they try
to match di昀erent illustrations in a large corpus of manuscripts.
      </p>
      <p>What all these approaches have in common is their emphasis on studying illustrations as
complete entities, analyzing their style, similarities, or reuse. However, for our speci昀椀c use
case of detecting alchemical apparatus, we require an approach that is able to detect singular
objects in a complex scene depicted in an illustration. Since we could not 昀椀nd any existing
methods for object detection in 16th/17th century book illustrations, we conducted a series of
experiments using various approaches on our own.</p>
    </sec>
    <sec id="sec-5">
      <title>2.2. First experiments with existing methods</title>
      <p>First of all, we experimented with out-of-the-box methods, such as the Distant Viewing Toolkit
(Figure1), Segment Anything (segment-anything.com/) (Figure2) and image querying, using
OWL-ViT (Figure3). While revealing some successes at 昀椀rst glance, a昀琀er some more testing
these algorithms have proven largely inadequate in di昀erentiating speci昀椀c objects of interest
from early modern prints. This medium is rich in visually similar etchings and contains typical
alchemical objects that algorithms trained on modern data may simply not be familiar with.</p>
    </sec>
    <sec id="sec-6">
      <title>2.3. Training and evaluation corpus</title>
      <p>
        Because of the shortcomings of the above approaches, we proceeded to compile some training
data, to provide a representative sample for the book genre de昀椀ned above, containing books
primarily concerned with mining, metallurgy or distillation. Some of them represent di昀erent
issues or print runs of the same book for standard works such as Hieronymus Brunschwig’s
De Arte Distillandi [11], Georgius Agricola’sDe re metallica [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] or Vannoccio Biringuccio’s
Pirotechnia [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], in which illustrations frequently di昀er in between di昀erent editions or print
shops. Unlike the training corpus, the evaluation corpus was constructed to contain books not
only concerned with mining, metallurgy or distillation. This allows us to verify if the algorithm
actually learned anything and is able to distinguish illustrations not related to our subject (such
as workshop scenes not related to alchemy, metallurgy, mining or distillation) from the objects
we wish to detect. Accordingly, we 昀椀rst evaluate the ability to detect illustrations of laboratory
equipment in early modern book pages, and then look at the performance for classifying
speci昀椀c objects. Our training corpus, thus, only contains books that we know contain a su昀케cient
number of relevant illustrations from the contexts of mining, metallurgy and distillation from
the 16th-17th centuries [
        <xref ref-type="bibr" rid="ref1 ref10 ref2 ref21 ref23 ref25 ref3 ref32 ref51 ref58 ref6 ref7 ref8 ref9">11, 12, 50, 51, 10, 13, 24, 57, 31, 20, 22, 3, 1, 2, 6, 7, 9, 8</xref>
        ], while the
evaluation corpus contains books from other alchemy-related areas and encycloped5i6a,s32[
        <xref ref-type="bibr" rid="ref19 ref20 ref34 ref4 ref49">,
33, 19, 18, 48, 4</xref>
        ]. The challenges of annotating the training corpus are described in the next
section.
      </p>
      <sec id="sec-6-1">
        <title>3. The alchemy of annotation</title>
        <p>The next step involved the semi-automatic annotation of images using the Supervisely
platform (https://supervise.ly), whereby each component of alchemical apparatus was labeled
individually in the hopes of providing the most useful form of annotations to improve model
training. This process resulted in the creation of pixel-level labels.</p>
        <p>
          We based our annotation on previous work done at the Herzog August Bibliothek
Wolfenbüttel [
          <xref ref-type="bibr" rid="ref27">26</xref>
          ].2 In Frietsch’s classi昀椀cation [
          <xref ref-type="bibr" rid="ref27">26</xref>
          ], ‘Alchemistic equipment’ (49E393,https://iconcl
ass.org/49E393) is a subclass of ‘Alchemy’ (49E39) in IconClass and organized as illustrated
in Figure4. As the annotation table (Table1) shows, we did not incorporate all of the
IconClass categories as labels. The classes to be used were selected by the relative frequency of
related images in our corpus and depending on whether it made sense to keep subclasses or
2Adhering to the alchemy IconClass classi昀椀cation and vocabulary created by Ute Frietsch, which includes most
alchemical apparatus, would not only keep a successful object detection model coming out of this work interoperable,
but it also provides us with 1,800 tagged images we may re-use for creating ground truth in future work.
not (as many of them are not that visually distinctive nor frequent enough in our corpus to
be e昀ective to annotate). The goal was to keep the number of necessary annotation labels
(and classes) as low as possible for our initial experiments. On the other hand, we introduced
a class forambices (singularambix, a distillation helmet), which are frequently depicted, yet
were lacking from Frietsch’s classi昀椀cation of alchemical equipmen3 tT.his approach represents
a compromise between keeping the number of classes as low as possible while still including
a su昀케cient number for making meaningful interpretations later. Had we annotated both the
non-explicitly alchemical and the explicitly alchemical tools the same way, we would probably
train our algorithm to simply detect tools, regardless of the label assigned to them coming from
the IconClass alchemy category.
3As we initially had planned not to include composite devices in the hopes of thus providing better training data
for the algorithms, some classes very visually distinctive for alchemy were not included, such as alembics and
moor’s heads (Figure5). Notably, within the category of ‘pots’o(llae), some objects exhibit visually distinct
alchemical characteristics, like triangular crucibles (for examples see T1a)b, lwehile others only can be interpreted
as alchemical within a guaranteed alchemical context such as cupels, which visually look like simple pots or cups.
We further opted to unite a range of furnace types under a single label.
49E3931 alchemistic vessels
49E39311 bottles (ampullae)
• philosophical egg (ovum philosophicum)
• pelican
• phial
• receiver (receptaculum)
49E39312 flasks (cucurbitae)
• alembic
• Moor’s head
• operculum
• retort
• rosenhut
49E39313 pots, jars (ollae)
• aludel
• chalice
• crucible
• cupel
49E3932 alchemistic furnace
• assay furnace
• athanor
• carburizing furnace
• ‘slow Harry’ (piger henricus)
• reverberatory furnace
• smelting furnace
49E3933 alchemistic bath (balneum)
• balneum arenae
• balneum Mariae
49E3939 other alchemistic equipment
        </p>
      </sec>
      <sec id="sec-6-2">
        <title>4. Prelimary results</title>
        <p>
          In the rapidly evolving Digital Humanities (DH) sub-昀椀eld oDfistant Viewing [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ], the application
of computer vision techniques in diverse research areas has been met with enthusiasm. But
despite this enthusiasm, our study reveals that these models may not yet readily adapt to
specialized tasks in the DH. We have encountered substantial challenges in deploying these models
for object detection in early modern depictions of chemical apparatus. The likely culprits were
not solely the unique visual style of these etchings but also the models’ unfamiliarity with the
nuances of early modern alchemical equipment and associated terminology. It is apparent that
these models, adept at interpreting modern visual styles and contexts, are confounded by the
distinct visual style of early modern etchings. In the following subsections, we present
preliminary results for the detection of alchemical objects in early modern illustrations that were
achieved with a range of di昀erent supervised and unsupervised computer vision approaches.
        </p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>4.1. Visual feature descriptors</title>
      <p>
        First, we experimented with an unsupervised clustering approach for visual feature description,
namely the ORB (Oriented FAST and Rotated BRIEF4[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]) method. This approach is tailored for
exact image reproduction (cf. the work done on woodcut reuse in chapbooks with VIS2E1][).
This did not involve any training or the usage of our annotations and was meant to discern
whether some intrinsic structures within the data could be utilized. Unfortunately, ORB failed
to demonstrate such patterns in our data set.
      </p>
    </sec>
    <sec id="sec-8">
      <title>4.2. Pixel Segmentation</title>
      <p>
        Next, we decided to try pixel segmentation approaches, which allow us to perform object
detection by dividing an image into segments and labeling each pixel, trying to map it to an object
class. We 昀椀rst deployed approaches, where models classify each pixel individually, namely
UNet [
        <xref ref-type="bibr" rid="ref48">47</xref>
        ] (with a ResNet-34 backbone) and the newer SegFormer5[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Despite being unable to
recognize several elements (notably, animals), the U-Net/ResNet deep learning model detected,
i.e. segmented, some plants correctly. Overall, however, the classi昀椀cation still proved to be
erroneous. With the ResNet-based pixel segmentation, we reached an overall accuracy of 33.0%
a昀琀er 昀椀ne tuning for 50 epochs. A similar story unfolded when using the SegFormer B15[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
deep learning model, which occasionally managed to identify the rough area of an object but
again without determining the correct category.
      </p>
    </sec>
    <sec id="sec-9">
      <title>4.3. Representation learning</title>
      <p>
        Furthermore, we continued assessing the e昀케cacy of non-supervised models, which operate
without annotation to discover structures in the data and thus are supposed to identify similar
objects. We employed SimSiam (Simple Siamese Representation Learning1)6[] and SimCLR
(Siamese Contrastive Representation Learning)1[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] for unsupervised clustering using Siamese
networks. Siamese networks are used in unsupervised visual representation learning to
maximize similarity between image augmentations. SimCLR (Contrastive Learning of Visual
Representations) performs unsupervised representation learning from unlabeled images, which
leverages data augmentation for contrasting di昀erent visual representations. SimCLR and
SimSiam perform well on ImageNet. Yet the methods yielded equally discouraging results on our
historical data.
      </p>
    </sec>
    <sec id="sec-10">
      <title>4.4. YOLO object detection</title>
      <p>Finally, we turned to the state-of-the-art object detection framework YOLYOou( Only Look
Once) version 8 mode4l, because a predecessor (YOLOv5) had been previously reported as being
suitable for detecting images in historical prin1t4[]. Unfortunately, the performance of YOLO –
like the previous approaches – fell short of our expectations. As YOLO is a popular framework
and widely known in the Computational Humanities community, we will discuss it in more
detail. We based our quality assessment on the model’s ability to correctly detect objects and
accurately label them.</p>
      <p>YOLO training was performed using about 50% of each class for training and the rest for
validation (昀椀gure 10). We initially experimented with 3-fold cross-validation, however, due to
the scarceness of our training data, we 昀椀nally opted for the single train-validation split.</p>
      <p>
        As each image usually contained various labels with di昀erent classes (see Figu6r)e,
producing such a strati昀椀ed sampling was unfortunately not straightforward, as one image must
be either assigned to the training set or to the validation set with all its containing labels to
prevent data leakage. Sometimes a large proportion of all available labels was on one or two
images. A further complication was the presence of sometimes partially overlapping label
annotations, such as distillation helmets being part of furnace setups. These create potential
sources of confusion for both training and validation (昀椀gur6e). We found no solution for
昀椀xing the overlapping labels, but we partitioned image regions with non-overlapping labels into
isolated (non-overlapping and non-leaking) sub-images, thereby producing a larger number
of possible assignments to training or validation sets (and through this a lower strati昀椀cation
error). The image partitioning was performed using a custom plane sweep algorithm that
produced a hierarchy of either horizontal or vertical axes that subdivided images without cutting
across label bounding boxes. To compute the actual training-validation split, we generated
10,000 random splits and picked the one that yielded a label distribution with the lowest mean
error in its test-val ratio over all classes. For future studies, we plan to resort to more robust
approaches [
        <xref ref-type="bibr" rid="ref54">53</xref>
        ]. Still, except for the furnace class, our approach produced a good strati昀椀cation
for all classes (昀椀gure10).
      </p>
      <p>We now report some of the training results. Trainingyoalo8n model with default
parameters yielded a model with a mAP@0.5 of 0.3. Switching toyaolov8s model with a resolution
of 1280 pixels (instead of the 640-pixel default) improved this score to 0.37 (discussed below
and shown in 昀椀gure 7). As the confusion matrix (昀椀gure 8) and the precision-recall curves
(昀椀gure 7) show, the classes that were best detected are ‘plants’, ‘ollae’ and ‘animals’. ‘Furnaces’,
‘other-equipment’, ‘cucurbitae-retorte’, ‘cucurbitae-rosenhut’ and ‘ampullae’ are detected
considerably less well, having both issues with precision and recall. The classes ‘human’,
‘mineralmetal’ and ‘cucurbitae’ showed very low overall precision. The detection of ‘cucurbitae-ambix’
did not seem to work at all. We also experimented with other resolutions (up to 1,600 pixels),
as well as adding augmentation througmhixup and various image transformations, as well as
tuning the mosaic setting and the box_loss gain. However, we found no improvements in
overall performance. Looking at the training curves, it turned out that for all tested YOLO
models, resolutions, and settings, from smalleysotlov8n model to the largeryolov8s model,
generalization for object localization did not work well, while generalization for object
classi椀昀cation seemed to present no issues at all: while the classi昀椀cation losscls_loss was reduced
rather symmetrically for both training and validation sets, and bthoxe_loss for the training
data showed a nearly perfect training curve in all regimes, tbhoex_loss for the validation
set turned out to be highly unstable and erratic in all cases, implying at least partial
over昀椀tting. Upon analyzing whyollae was recognized better than other classes, we noted that the
characteristic rounding could potentially account for a somewhat better model performance
in this category. Across other label classes, visual variance was higher, which is illustrated in
椀昀gure 10. For example, the depictions of objects in thaempullae category varied considerably
(e.g., the jugs with handles in the ‘training’ set lacked larger openings at the top).</p>
      <p>The overall lack of success was probably due to the ratio of large ‘variance in the data’ to
small ‘number of annotations’. The latter pales in comparison to the recommended 昀椀gure of
1,500 images (and 10,000 labels) per categor5y.The daunting prospect of manually annotating
such a volume of images, however, was contrary to our objectives of automating the task.
Annotating 1,500 objects per category would not only be very laborious and potentially
nonsensical for our task, this amount of examples per class also simply may not exist per object
type in our historical data.</p>
      <p>In preliminary experiments we observed that out-of-the-box YOLO models, pre-trained on
COCO, showed no advantage in terms of transfer learning for the task at ha6ndN. ot only are
COCO images modern, but their di昀erences also tend to be much bigger than amongst di昀erent
types of early modern alchemical laboratory apparatus. Thus, the model probably cannot adapt
5https://docs.ultralytics.com/yolov5/tutorials/tips_for_best_training_results/
6The COCO dataset consists of 80 distinct object classes (from a modern context) like cats, zebras, or baseball bats.
easily to grappling with historical data or distinguish in such a lot of details types of objects it
has never seen before and does not know what to call.</p>
      <sec id="sec-10-1">
        <title>5. Conclusions and future work</title>
        <p>As part of our endeavour to utilize computer vision techniques for detecting early modern
depictions of chemical apparatus, we initially embarked on experimental runs using readily
available toolkits. These preliminary e昀orts yielded encouraging results, suggesting the
viability of digging deeper into the intricacies of this interdisciplinary task. Encouraged by these
early indications, we decided to extend our exploration, leveraging custom annotations to
昀椀netune a model. However, as the previous sections have detailed, these subsequent e昀orts were
met with considerable obstacles and ultimately did not live up to the promise suggested by our
initial forays. This, in turn, strongly suggests that further in-depth investigation is required in
this area.</p>
        <p>Attempts to harness state-of-the-art computer vision models revealed a distinct lack of
generalisability to the idiosyncratic nature of early modern etchings. These unsuccessful attempts
underscore the unique challenges presented by these unconventional early modern images.
The ‘rendering’ or hatchings, i.e. the style of our images, could be what thwarts the algorithms.
They may also have issues with granularity due to the etchings’ visual similarity because all
the objects to be analyzed are early modern book illustrations characterized by cross-hatching
and strong black lines. The model may simply recognize them all as parts of books or book
pages but does not realize that it is the di昀erence between those particular illustrations that
we are interested in7.</p>
        <p>
          Going forward, we propose to explore one or few-shot approaches, although such methods
are not extensively supported for object detection. We might try reducing our object detection
problem (which is more complex than classi昀椀cation and for which there are also fewer
readily available frameworks) into a classi昀椀cation problem by working with cropped images. The
complex nature of the image data at hand suggests a need for more comprehensive annotation
or potentially attempting to leverage style transfer to enhance our outcomes. We had initially
tested this approach with theInstructPix2Pix model, which can convert hatching into
real7This may be because the training data it was trained on probably did not have lots of images like ours and when it
did, these simply may have been labelled as ‘book’ or ‘book page’ by annotators who, unlike us, were not interested
in their particular details. At least indications for this were witnessed when we were 昀椀rst trying out the Distant
Viewing Toolkit 5[] as an out-of-the-box tool (cf. Figur1e).
istic shading, but unfortunately, this led to the loss of crucial visually distinctive details in the
images and was ultimately unhelpful for our object detection task (Fig9u)r. eLeveraging the
classi昀椀cation capabilities of large Vision-Language Models (VLMs) such as BLIP-236[] would
be very interesting as well, however, the object localization issue needs to solved 昀椀rst, maybe
by using OWL-ViT [
          <xref ref-type="bibr" rid="ref40">39</xref>
          ] or SegmentAnything (Figure2) only for bounding box estimation but
not for classi昀椀cation [
          <xref ref-type="bibr" rid="ref53">52</xref>
          ].
        </p>
        <p>In conclusion, despite the growing enthusiasm foDristant Viewing in the DH, the application
of recent computer vision methods in the context of early modern print illustrations requires
more nuanced approaches. The models’ failure to recognize and classify early modern etchings
of chemical apparatus serves as a sobering reminder of the gap that still exists between the
outof-the-box availability of state-of-the-art technology and the challenges in its DH application
on historical data.</p>
        <p>IconClass DescriptionVisual Representation
other</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>G.</given-names>
            <surname>AgricolaD.e Re</surname>
          </string-name>
          <article-title>Metallica</article-title>
          . Basileae / Basel: In O昀케cina Frobeniana, Per Hier.
          <source>Frobenivm Et Nic. Episcopivm</source>
          ,
          <volume>1561</volume>
          . url:http://data.onb.ac.at/rep/10B29699.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>G.</given-names>
            <surname>Agricola.De Re Metallica</surname>
          </string-name>
          . Basileae / Basel: König, Emanuel,
          <volume>1657</volume>
          . urlh: ttp://data.onb .ac.at/rep/10B29680.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>G. Agricola.Vom</given-names>
            <surname>Bergkwerck</surname>
          </string-name>
          . Basel / Basel: Jeronymus Froben and Niclausen Bischo昀,
          <volume>1557</volume>
          . url: http://data.onb.ac.at/rep/108C988B.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>U.</given-names>
            <surname>AldrovandiM.usaeum Metallicum</surname>
          </string-name>
          . Bononiae / Bologna: Joan. Bapt. Ferronij,
          <volume>1648</volume>
          . url: http://data.onb.ac.at/rep/10B29761.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T. B.</given-names>
            <surname>Arnold</surname>
          </string-name>
          and
          <string-name>
            <given-names>L. Tilton. “Distant</given-names>
            <surname>Viewing: Analyzing Large Visual Corpora”D.Iingi</surname>
          </string-name>
          <article-title>:tal Scholarship in the Humanities (</article-title>
          <year>2019</year>
          ). doi: http://dx.doi.org/10.1093/digitalsh/fqz01.3
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>V.</given-names>
            <surname>Biringuccio</surname>
          </string-name>
          .Pirotechnia. Vinegia / Venice: Giouan Padoano,
          <volume>1550</volume>
          . url:http://data.on b.
          <source>ac.at/rep/108CB82E.</source>
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>V.</given-names>
            <surname>Biringuccio</surname>
          </string-name>
          .Pirotechnia. Venetia / Venice: Gironimo Giglio,
          <volume>1559</volume>
          . urlh:ttp://data.on b.
          <source>ac.at/rep/1085DDED.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V.</given-names>
            <surname>Biringuccio</surname>
          </string-name>
          .Pirotechnia. Vinegia / Venice: Comin da Trino di Monferrato : Navo, Curzio Troiano de,
          <volume>1559</volume>
          . urlh: ttp://data.onb.ac.at/rep/10727526.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Biringuccio</surname>
          </string-name>
          .Pirotechnia. Bologna / Bologna: Longhi,
          <volume>1678</volume>
          . urlh: ttp://data.onb.ac.at/r ep/106F01F6.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10] [11] [12] [13]
          <string-name>
            <given-names>H.</given-names>
            <surname>Brunschwig</surname>
          </string-name>
          .Distilierbuch der rechten Kunst.
          <source>Franckfurt am Mayn / Frankfurt am Main: Weygand Hanen Erben</source>
          ,
          <volume>1565</volume>
          . url:http://data.onb.ac.at/rep/1084C642.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Brunschwig</surname>
          </string-name>
          .Liber de arte distillandi. Straßburg / Strasbourg: Johann Grüninger,
          <volume>1512</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>url: http://data.onb.ac.at/rep/109559D1.</mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Brunschwig</surname>
          </string-name>
          .
          <article-title>The vertuose boke of distyllacyon of the waters of all maner of herbes</article-title>
          .
          <source>London: Laurens Andrewe</source>
          ,
          <volume>1527</volume>
          . urlh:ttps://archive.org/details/mobot3175300081606 3/page/n30/mode/thumb.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <given-names>H.</given-names>
            <surname>Brunschwig</surname>
          </string-name>
          and
          <string-name>
            <given-names>W. H.</given-names>
            <surname>Ry昀</surname>
          </string-name>
          . New Vollkommen Distillierbuch.
          <source>Franckfurt a. M. / Frankfurt am Main: Christian Egenol昀s Erben</source>
          ,
          <volume>1597</volume>
          . url:http://data.onb.ac.at/rep/10A5871B.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Büttner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martinetz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>El-Hajj</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Valleriani</surname>
          </string-name>
          . “
          <article-title>CorDeep and the Sacrobosco Dataset: Detection of Visual Elements in Historical Documents”</article-title>
          .
          <source>IJno:urnal of Imaging</source>
          <volume>8</volume>
          .10 (Oct.
          <volume>15</volume>
          ,
          <year>2022</year>
          ), p.
          <fpage>285</fpage>
          . doi:
          <volume>10</volume>
          .3390/jimaging8100285.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kornblith</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Norouzi</surname>
          </string-name>
          , and
          <string-name>
            <given-names>G.</given-names>
            <surname>HintonA. Simple</surname>
          </string-name>
          <article-title>Framework for Contrastive Learning of Visual Representations</article-title>
          .
          <source>June 30</source>
          ,
          <year>2020</year>
          . url: http://arxiv.org/abs/
          <year>2002</year>
          .0570 9.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          and
          <string-name>
            <given-names>K.</given-names>
            <surname>He</surname>
          </string-name>
          .
          <source>Exploring Simple Siamese Representation Learning. Nov. 20</source>
          ,
          <year>2020</year>
          . url: http://arxiv.org/abs/
          <year>2011</year>
          .1056 6.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Cormier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Park</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Beck</surname>
          </string-name>
          . “
          <article-title>Automatic Classi昀椀cation of Woodcuts and Copperplate Engravings”</article-title>
          .
          <source>In:2020 17th Conference on Computer and Robot Vision</source>
          (CRV) (
          <year>2020</year>
          ), pp.
          <fpage>85</fpage>
          -
          <lpage>92</lpage>
          . url: https://api.semanticscholar.org/CorpusID:21954848.
          <fpage>7</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>J. v.</given-names>
            <surname>Cuba</surname>
          </string-name>
          . Gart der Gesuntheit. Straßburg / Strasbourg: Mathia Apiario,
          <volume>1536</volume>
          . urhl:ttp: //data.onb.ac.at/rep/108C96EA.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>J. v. Cuba. Ortus</given-names>
            <surname>Sanitatis</surname>
          </string-name>
          . Venetiis / Venice: Joannes de Cereto de Tridino,
          <volume>1538</volume>
          . url: http://data.onb.ac.at/rep/109E61DA.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>G.</given-names>
            <surname>Della PortaD.e Distillatione</surname>
          </string-name>
          . Romae / Rome: Reu. Camerae Apostolicae,
          <volume>1608</volume>
          . urlh:tt p://data.onb.ac.at/rep/1058CD41.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dutta</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Bergel, and</article-title>
          <string-name>
            <given-names>A.</given-names>
            <surname>Zisserman</surname>
          </string-name>
          . “
          <article-title>Visual Analysis of Chapbooks Printed in Scotland”</article-title>
          .
          <source>In:The 6th International Workshop on Historical Document Imaging and Processing (HIP '21), September 5-6</source>
          ,
          <year>2021</year>
          , Lausanne, Switzerland (
          <year>2021</year>
          ). doi: https://doi.org/10.1145 /3476887.3476893.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>L. Ercker. Aula</given-names>
            <surname>Subterranea</surname>
          </string-name>
          .
          <source>Franckfurt am Mayn / Frankfurt am Main: Johan Feyerabendt</source>
          ,
          <volume>1598</volume>
          . url: http://data.onb.ac.at/rep/104A2CBD.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ercker</surname>
          </string-name>
          .
          <article-title>Aula subterranea</article-title>
          .
          <source>Frankfurt (Main): Zunner</source>
          ,
          <volume>1672</volume>
          . urlh:ttps://www.deutsche stextarchiv.de/book/show/ercker%5C
          <source>%5Faula01%5C%5F167.2</source>
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ficinus</surname>
          </string-name>
          . Liber de arte distulandi. Straßburg / Strasbourg: Johan Grüniger,
          <volume>1509</volume>
          . url: http://data.onb.ac.at/rep/10A81A42.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>H.</given-names>
            <surname>Fors</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. M.</given-names>
            <surname>Principe</surname>
          </string-name>
          , and
          <string-name>
            <given-names>H. O.</given-names>
            <surname>Sibum</surname>
          </string-name>
          . “
          <article-title>From the Library to the Laboratory and Back Again: Experiment as a Tool for Historians of Science”</article-title>
          .
          <source>IAn:mbix 63/2</source>
          (
          <year>2016</year>
          ), pp.
          <fpage>85</fpage>
          -
          <lpage>97</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>U.</given-names>
            <surname>Frietsch</surname>
          </string-name>
          .
          <article-title>IconClass alchemy notation</article-title>
          .
          <source>Herzog August Bibliothek Wolfenbüttel</source>
          ,
          <year>2017</year>
          . url: http://alchemie.hab.de/bilde.r
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>M.</given-names>
            <surname>Giesecke</surname>
          </string-name>
          . Der Buchdruck in der frühen Neuzeit.
          <article-title>Eine historische Fallstudie über die Durchsetzung neuer Informations-</article-title>
          und
          <string-name>
            <surname>Kommunikationstechnologien</surname>
          </string-name>
          .
          <source>Frankfurt am Main</source>
          ,
          <year>1991</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [28]
          <string-name>
            <surname>G. Götzelmann.</surname>
          </string-name>
          “Bilderschätze,
          <article-title>Bildersuchen: Digitale Auswertung von Illustrationswiederverwendungen im Buchdruck des 16</article-title>
          . Jahrhunderts”. InW: issen und Buchgestalt. Ed. by
          <string-name>
            <given-names>P.</given-names>
            <surname>Hegel</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Krewet</surname>
          </string-name>
          . Wiesbaden: Harrassowitz Verlag,
          <year>2022</year>
          , pp.
          <fpage>323</fpage>
          -
          <lpage>340</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [29]
          <string-name>
            <surname>M. M.</surname>
          </string-name>
          <article-title>A</article-title>
          . Hendriksen. “
          <article-title>Rethinking Performative Methods in the History of Science”</article-title>
          .
          <source>In: Berichte zur Wissenscha昀琀sgeschichte</source>
          <volume>43</volume>
          (
          <year>2020</year>
          ), pp.
          <fpage>313</fpage>
          -
          <lpage>322</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>R.</given-names>
            <surname>Kaoua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Durr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lazaris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Picard</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Aubry</surname>
          </string-name>
          . “Image Collation:
          <article-title>Matching illustrations in manuscripts”</article-title>
          .
          <source>InC:oRR abs/2108</source>
          .08109 (
          <year>2021</year>
          ). url: https://arxiv.org /abs/2108.08109.
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kerzenmacher</surname>
          </string-name>
          .Alle Farben Wasser. München / Munich, Frankfurt am Mayn / Frankfurt am Main: Egenol昀, 1589. url: https://digital.deutsches-museum.de/de/digital-catalogue /library-object/BV020856683/%5C#
          <fpage>1</fpage>
          .
          <fpage>2</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kircher</surname>
          </string-name>
          .
          <article-title>Mundus subterraneus 1</article-title>
          . Amstelodami / Amsterdam: Janssonius &amp; Weyerstraten, 1665. url: http://data.onb.ac.at/rep/1047DA01.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kircher</surname>
          </string-name>
          .
          <article-title>Mundus subterraneus 2</article-title>
          . Amstelodami / Amsterdam: Janssonius &amp; Weyerstraten, 1665. url: http://data.onb.ac.at/rep/1047D9FF.
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>S.</given-names>
            <surname>Lang</surname>
          </string-name>
          . “Alchemical Laboratories: Texts, Practices,
          <string-name>
            <given-names>Material</given-names>
            <surname>Relics</surname>
          </string-name>
          .
          <article-title>An Introduction”</article-title>
          . In: Alchemistische Labore. Praktiken, Texte und materielle Hinterlassenscha昀琀en / Alchemical Laboratories. Practices, texts, material relics. Ed. by
          <string-name>
            <given-names>S.</given-names>
            <surname>Lang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröstl</surname>
          </string-name>
          , and
          <string-name>
            <given-names>P.</given-names>
            <surname>Fiska</surname>
          </string-name>
          . Graz: Grazer Universitätsverlag,
          <year>2023</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>30</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>S.</given-names>
            <surname>Laube</surname>
          </string-name>
          . “
          <source>Am Anfang ist Gestaltung. Bemerkungen zu Titelblättern bei Destilliertraktaten des 16</source>
          .
          <article-title>Jahrhunderts”</article-title>
          . In:Wissen und Buchgestalt. Ed. by
          <string-name>
            <given-names>P.</given-names>
            <surname>Hegel</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Krewet</surname>
          </string-name>
          . Wiesbaden: Harrassowitz Verlag,
          <year>2022</year>
          , pp.
          <fpage>275</fpage>
          -
          <lpage>298</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          et al. “
          <article-title>BLIP-2: Bootstrapping Language-Image Pre-Training with Frozen Image Encoders and Large Language Models”</article-title>
          .
          <source>Ina:rXiv preprint arXiv:2301.12597</source>
          (
          <year>2023</year>
          ). url: htt p://arxiv.org/abs/2301.12597.
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>M.</given-names>
            <surname>Malaspina</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhong</surname>
          </string-name>
          . “
          <article-title>Image-matching technology applied to Fi昀琀eenth-century printed book illustration”</article-title>
          .
          <source>InLe:tt Mat Int</source>
          <volume>5</volume>
          (
          <year>2017</year>
          ), pp.
          <fpage>287</fpage>
          -
          <lpage>292</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>M.</given-names>
            <surname>Martinón-Torres</surname>
          </string-name>
          .
          <article-title>“Some recent developments in the historiography of alchemy”</article-title>
          .
          <source>In: Ambix 58/3</source>
          (
          <year>2011</year>
          ), pp.
          <fpage>215</fpage>
          -
          <lpage>37</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>M.</given-names>
            <surname>Minderer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gritsenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Neumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Weissenborn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dosovitskiy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mahendran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Arnab</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dehghani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Kipf</surname>
          </string-name>
          , and
          <string-name>
            <given-names>N.</given-names>
            <surname>Houlsby</surname>
          </string-name>
          . “
          <article-title>Simple Open-Vocabulary Object Detection with Vision Transformers”</article-title>
          .
          <source>IEnc:</source>
          cv
          <year>2022</year>
          .
          <year>2022</year>
          . doi: https://doi.org/10.48550/arXiv.2205.0623.0
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>P. J. T.</given-names>
            <surname>Morris</surname>
          </string-name>
          . “
          <article-title>The history of chemical laboratories: a thematic approach”C.Ihne:mTexts 7/21 (</article-title>
          <year>2021</year>
          ). doi: https://doi.org/10.1007/s40828-021-00146- x.
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          [41]
          <string-name>
            <given-names>P. J. T.</given-names>
            <surname>Morris</surname>
          </string-name>
          .
          <article-title>The Matter Factory: A History of the Chemistry Laboratory</article-title>
          . London: Reaktion Books,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          [42]
          <string-name>
            <given-names>H.</given-names>
            <surname>P昀氀üger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Thom</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schütz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bohde</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>Ertl</surname>
          </string-name>
          . “
          <article-title>VeCHArt: Visually Enhanced Comparison of Historic Art Using an Automated Line-Based Synchronization Technique”</article-title>
          .
          <source>In: IEEE Transactions on Visualization and Computer Graphics</source>
          <volume>26</volume>
          .10 (
          <year>2020</year>
          ), pp.
          <fpage>3063</fpage>
          -
          <lpage>3076</lpage>
          . doi:
          <volume>10</volume>
          .1109/tvcg.
          <year>2019</year>
          .
          <volume>2908166</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          [43]
          <string-name>
            <given-names>M.</given-names>
            <surname>Piorko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hendriksen</surname>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Werrett</surname>
          </string-name>
          . “Alchemical Practice:
          <article-title>Looking Towards the Chemical Humanities”</article-title>
          .
          <source>In:Ambix</source>
          <volume>69</volume>
          :
          <issue>1</issue>
          (
          <issue>2022</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>18</lpage>
          . doi:
          <volume>10</volume>
          .1080/00026980.
          <year>2022</year>
          .
          <volume>203</volume>
          5572.
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          [44]
          <string-name>
            <given-names>G. D.</given-names>
            <surname>Porta</surname>
          </string-name>
          .De Distillationibus libri IX.
          <source>Strasbourg: Zetzner</source>
          ,
          <fpage>1609</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref46">
        <mixed-citation>
          [45]
          <string-name>
            <given-names>L. M.</given-names>
            <surname>Principe</surname>
          </string-name>
          and
          <string-name>
            <given-names>W. R.</given-names>
            <surname>Newman</surname>
          </string-name>
          .
          <article-title>“Some Problems with the Historiography of Alchemy”</article-title>
          . In:
          <article-title>Secrets of Nature: Astrology and Alchemy in Early Modern Europe</article-title>
          . Ed. by
          <string-name>
            <given-names>W. R.</given-names>
            <surname>Newman</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Gra昀琀on</surname>
          </string-name>
          . Cambridge/Massachusetts: MIT Press,
          <year>2001</year>
          , pp.
          <fpage>385</fpage>
          -
          <lpage>432</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref47">
        <mixed-citation>
          [46]
          <string-name>
            <given-names>S.</given-names>
            <surname>Reardon</surname>
          </string-name>
          . “
          <article-title>The Alchemical Revolution”</article-title>
          .
          <source>InSc:ience 332</source>
          (
          <year>2011</year>
          ), pp.
          <fpage>914</fpage>
          -
          <lpage>915</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref48">
        <mixed-citation>
          [47]
          <string-name>
            <given-names>O.</given-names>
            <surname>Ronneberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fischer</surname>
          </string-name>
          , and
          <string-name>
            <given-names>T.</given-names>
            <surname>BroxU</surname>
          </string-name>
          . -Net:
          <article-title>Convolutional Networks for Biomedical Image Segmentation</article-title>
          .
          <source>May 18</source>
          ,
          <year>2015</year>
          . url: http://arxiv.org/abs/1505.04597.
        </mixed-citation>
      </ref>
      <ref id="ref49">
        <mixed-citation>
          [48]
          <string-name>
            <given-names>E.</given-names>
            <surname>Rößlin</surname>
          </string-name>
          and
          <string-name>
            <given-names>J. v.</given-names>
            <surname>Cuba</surname>
          </string-name>
          .Kreuterbuch. Franckenfurt,
          <source>am Meyn / Frankfurt am Main: Egenol昀</source>
          , 1536. url: http://data.onb.ac.at/rep/109E61B8.
        </mixed-citation>
      </ref>
      <ref id="ref50">
        <mixed-citation>
          [49]
          <string-name>
            <given-names>E.</given-names>
            <surname>Rublee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Rabaud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Konolige</surname>
          </string-name>
          , and
          <string-name>
            <surname>G. Bradski. “ORB</surname>
          </string-name>
          :
          <article-title>An e昀케cient alternative to SIFT or SURF”</article-title>
          .
          <source>In:2011 International Conference on Computer Vision</source>
          .
          <source>2011 IEEE International Conference on Computer Vision</source>
          (ICCV). Barcelona, Spain: Ieee, Nov.
          <year>2011</year>
          , pp.
          <fpage>2564</fpage>
          -
          <lpage>2571</lpage>
          . doi:
          <volume>10</volume>
          .1109/iccv.
          <year>2011</year>
          .
          <volume>6126544</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref51">
        <mixed-citation>
          [50] [51]
          <string-name>
            <given-names>W. H.</given-names>
            <surname>Ry昀.</surname>
          </string-name>
          <article-title>Das new groß Distillier Buch</article-title>
          .
          <source>Franckfurt / Frankfurt am Main: Christ. Egenol昀</source>
          ,
          <volume>1545</volume>
          . url: http://data.onb.ac.at/rep/107BF132.
        </mixed-citation>
      </ref>
      <ref id="ref52">
        <mixed-citation>
          <string-name>
            <surname>W. H. Ry昀</surname>
          </string-name>
          . New groß distillier-Buch.
          <source>Franckfort / Frankfurt am Main: Christian Egenol昀's Erben</source>
          ,
          <volume>1556</volume>
          . url: http://data.onb.ac.at/rep/105FA8A2.
        </mixed-citation>
      </ref>
      <ref id="ref53">
        <mixed-citation>
          [52]
          <string-name>
            <given-names>P.</given-names>
            <surname>Saranrittichai</surname>
          </string-name>
          et al. “
          <article-title>Zero-Shot Visual Classi昀椀cation with Guided Cropping”</article-title>
          .
          <source>Ianr:Xiv preprint arXiv:2309.06581</source>
          (
          <year>2023</year>
          ). url: http://arxiv.org/abs/2309.0658 1.
        </mixed-citation>
      </ref>
      <ref id="ref54">
        <mixed-citation>
          [53]
          <string-name>
            <given-names>K.</given-names>
            <surname>Sechidis</surname>
          </string-name>
          et al. “
          <article-title>On the Strati昀椀cation of Multi-Label Data”</article-title>
          .
          <source>In:Machine Learning and Knowledge</source>
          Discovery in Databases. Ed. by
          <string-name>
            <given-names>D.</given-names>
            <surname>Gunopulos</surname>
          </string-name>
          et al. Vol.
          <volume>6913</volume>
          . Springer Berlin Heidelberg,
          <year>2011</year>
          , pp.
          <fpage>145</fpage>
          -
          <lpage>158</lpage>
          . doi:https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -23808-6\_1
          <fpage>0</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref55">
        <mixed-citation>
          [54]
          <string-name>
            <given-names>P.</given-names>
            <surname>Smith.</surname>
          </string-name>
          <article-title>The Making and Knowing Project (</article-title>
          <year>2015</year>
          -
          <fpage>2022</fpage>
          ).
          <year>2020</year>
          . url: makingandknowing.o rg.
        </mixed-citation>
      </ref>
      <ref id="ref56">
        <mixed-citation>
          [55]
          <string-name>
            <given-names>P. H.</given-names>
            <surname>Smith</surname>
          </string-name>
          . “Laboratories”.
          <source>InT:he Cambridge History of Science</source>
          <volume>3</volume>
          / Early Modern Science. Ed. by
          <string-name>
            <given-names>K.</given-names>
            <surname>Park</surname>
          </string-name>
          and
          <string-name>
            <given-names>L.</given-names>
            <surname>Daston</surname>
          </string-name>
          . Cambridge: Cambridge University Press,
          <year>2006</year>
          , pp.
          <fpage>290</fpage>
          -
          <lpage>305</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref57">
        <mixed-citation>
          [56]
          <string-name>
            <surname>L.</surname>
          </string-name>
          <article-title>Thurneysser zum ThurnH.istoria sive Descriptio Plantarum Omnium</article-title>
          . Berlini / Berlin: ?, 1578. url: http://data.onb.ac.at/rep/108C96EA.
        </mixed-citation>
      </ref>
      <ref id="ref58">
        <mixed-citation>
          [57]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ulsted</surname>
          </string-name>
          and
          <string-name>
            <surname>J. A.</surname>
          </string-name>
          et al.
          <article-title>Coelum philosophorum</article-title>
          . Lugduni / Lyon: Guilielmus Rovillius,
          <volume>1572</volume>
          . url: http://data.onb.ac.at/rep/1086E9DB.
        </mixed-citation>
      </ref>
      <ref id="ref59">
        <mixed-citation>
          [58]
          <string-name>
            <given-names>M.</given-names>
            <surname>Valleriani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Kräutli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lockhorst</surname>
          </string-name>
          , and
          <string-name>
            <surname>N. Shlomi. “</surname>
          </string-name>
          <article-title>Vision on Vision: De昀椀ning Similarities Among Early Modern Illustrations on Cosmology”S.cIinen:ti昀椀c Visual Representations in History</article-title>
          . Ed. by
          <string-name>
            <given-names>M.</given-names>
            <surname>Valleriani</surname>
          </string-name>
          , G. Giannini, and
          <string-name>
            <given-names>E.</given-names>
            <surname>Giannetto</surname>
          </string-name>
          . Cham: Springer International Publishing,
          <year>2023</year>
          , pp.
          <fpage>99</fpage>
          -
          <lpage>137</lpage>
          . doih: ttps://doi.org/10.1007/978-3-
          <fpage>031</fpage>
          -1131
          <fpage>7</fpage>
          -
          <lpage>8</lpage>
          \_4.
        </mixed-citation>
      </ref>
      <ref id="ref60">
        <mixed-citation>
          [59]
          <string-name>
            <given-names>E.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Anandkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Alvarez</surname>
          </string-name>
          , and P. LuoS.egFormer:
          <article-title>Simple and E昀케cient Design for Semantic Segmentation with Transformers</article-title>
          .
          <source>Oct</source>
          .
          <volume>28</volume>
          ,
          <year>2021</year>
          . url: http://a rxiv.
          <source>org/abs/2105.1520 3.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>