<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Assessing the clinicians' pathway to embed arti cial intelligence for assisted diagnostics of fracture detection</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Carlos Francisco Moreno-Garc a</string-name>
          <email>c.moreno-garcia@rgu.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Truong Dang</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kyle Martin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manish Patel</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrew Thompson</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lesley Leishman</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nirmalie Wiratunga</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Robert Gordon University</institution>
          ,
          <addr-line>Garthdee Road, Aberdeen, Scotland</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>1869</year>
      </pub-date>
      <abstract>
        <p>Fracture detection has been a long-standing paradigm on the medical imaging community. Many algorithms and systems have been presented to accurately detect and classify images in terms of the presence and absence of fractures in di erent parts of the body. While these solutions are capable of obtaining results which even surpass human scores, few e orts have been dedicated to evaluate how these systems can be embedded in the clinicians and radiologists working pipeline. Moreover, the reports that are included with the radiography could also provide key information regarding the nature and the severity of the fracture. In this paper, we present our rst ndings towards assessing how computer vision, natural language processing and other systems could be correctly embedded in the clinicians' pathway to better aid on the fracture detection task. We present some initial experimental results using publicly available fracture datasets along with a handful of data provided by the National Healthcare System from the United Kingdom in a research initiative call. Results show that there is a high likelihood of applying transfer learning from di erent existing and pre-trained models to the new records provided in the challenge, and that there are various ways in which these techniques can be embedded along the clinicians' pathway.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        In recent years, fracture detection has been one of the most
cited challenges in medical imaging analysis, evidenced both
by public competitions [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and clinical trials [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] alike. The
design of a system which aids clinicians in the automatic
detection of fractures is of paramount to reduce the workload of
the front line sta and allow them more time to focus on the
most urgent cases. To address this issue, the Scottish
Government, Opportunity North East (ONE) and the Small Business
Research Initiative (SBRI) announced a challenge to carry out
a project towards looking at this problem in the healthcare
system in the northeast of Scotland3. A team comprised of
members from the industry (Jiva.ai) and academia (Robert
Gordon University) was formed to look at the problem and
design a solution.
      </p>
      <p>In addition, a key contribution of this work is the
modelling of the clinician's pathway, exploring the current
process of radiology imaging for fracture treatment. This was
built through a series of co-creation sessions with a reporting
radiographer, then veri ed by two clinicians and two other
reporting radiographers. By designing this pathway, we
identi ed three key stakeholders (clinician, radiologist, patient)
and four sub-processes: (1) requesting radiology images; (2)
acquiring radiology images; (3) reporting on radiology
images; and (4) decision support. By understanding the current
approach to radiology imaging for fracture treatment, we
consider that we are capable of o ering a valuable contribution
that will allow research groups to pinpoint opportunities for
smart automation of this process in future.</p>
      <p>The contributions of this paper are as follows:
1. Identify the current processes involved with the various
procedures a ected by radiology imaging for fractures.
2. Explore where these processes can be improved through the
implementation of AI.
3. Identify what form of AI would be most applicable in order
to maximise the obtained bene ts by the a ected
stakeholder.
4. Given the limited amount of samples provided by the
challenge proposer, perform initial proof of concept tests using
baseline methods to identify the potential of transfer
learning in this domain.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
    </sec>
    <sec id="sec-3">
      <title>Image recognition</title>
      <p>
        Only a handful of demonstrations of machine learning,
computer vision and natural language processing for bone
fracture detection appear in scienti c literature. Lindsey et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
demonstrated that a deep neural network trained on 256'000
x-rays could detect fractures with a similar diagnostic
accuracy to a sub-specialised orthopaedic surgeon. Also, Olczak et
al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] applied deep learning to analyse 256'458 x-rays, and
concluded that arti cial intelligence methods are applicable
for radiology screening, especially during out-of-hours
healthcare or at remote locations with poor access to radiologists
or orthopedists. Smaller scale studies using tens to low
thousands of images include Lawn et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], Kim and MacKinnon
[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], Tomita et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], Dimililer [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and Bandyopadhyay et al.
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], amongst others.
      </p>
      <p>
        Three technical frames have been described as applicable
for ML in radiology: image segmentation, registration and
computer-aided detection and diagnosis [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Out of these,
graph models, such as Bayesian networks and Markov
random elds, have been identi ed as the two most widely used
techniques for fracture modelling. However, recent advances
in generative deep models (e.g. variational autoencoders) [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
have been applied to annotate both images and text; which
are yet to be exploited in radiology and related applications.
Similarly, multi-modal learners [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] have been used to learn
from image and text to improve recognition of objects; unlike
single modality learners these can combine mixed embedding
spaces that unite di erent modalities. These have been
applied to public text and images but digital health applications
are yet to emerge. Given the relevance of both image and text
to clinical radiology we expect to adapt these algorithms to
create an innovative uni ed embedding suited to automated
annotation by deep generative algorithms. Indeed, we can use
state-of-the-art translation algorithms such as transformers
[
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] which exploit similarity and capitalise on adjacency
information, to generate reports from both radiograms and clinical
text.
3
      </p>
      <p>Modelling a Clinician's Pathway
Any arti cially intelligent solution which is suggested to
resolve some problem within the eld of radiology should be
rooted in deep understanding of the domain and user
requirements. With this in mind, we have received input from
domain-experts to develop a clinician's pathway, detailing
current process of radiology imaging for fracture treatment. In
this paper, we aim to demonstrate the opportunities which
offer potential for smart automation in this eld. In particular,
we aim to highlight the areas where the application of arti
cial intelligence would be impactful for increasing e ciency,
improving patient experience and decreasing cost.
3.1</p>
      <p>Co-Creation of Clinician's Pathway
We performed a multi-stage co-creation process to model the
clinician's pathway which was indicative of real-world
radiology practice. These stages were divided into a design phase,
where we aimed to understand and model the current
processes for the acquisition and reporting of radiology images,
and a validation phase, in which we obtained unbiased
feedback on our model from personnel external to our co-creation.
The result is a clinician's work ow which has been
developed alongside a reporting radiographer, and veri ed by two
clinicians (one consultant radiologist and one senior accident
&amp; emergency doctor) and two reporting radiographers from
within the National Health Service (NHS) Scotland. We are
therefore con dent in its accuracy and its suitability to
describe real radiology processes. Although this pathway has
been built with input from British radiologists, we suggest it
can be generalised to wider radiology practices (within
reason).</p>
      <p>During the design phase, we organised two separate
cocreation sessions. In the rst session, we met with a
reporting radiographer to discuss the complete journey of a patient
who was given an appointment for radiology imaging for a
suspected fracture. This session was useful to establish the
process start-points and end-points. In the second session, we
observed a reporting radiographer reporting on a series of
xray images for suspected fractures. This session was intended
to identify relevant technologies and the role they played in
reporting on radiology images. The outcome of this two stage
design phase was an initial draft of the clinician's pathway
which could be validated by domain experts.</p>
      <p>We then organised three sessions for the validation of the
developed work ow. In the rst session, we obtained
feedback upon the pathway from the reporting radiographer who
was directly involved in its formation. This allowed any
errors which had arisen due to misunderstanding aspects of the
design phase to be corrected. In the second session, a
member of the research team met with a clinician and a reporting
radiographer to explain and discuss the draft pathway and
obtain feedback. This session was designed to ensure that the
pathway could be generalised to more than just the single
radiographer with whom it had been co-designed. In
particular, the session resulted in a number of updates to the role
of the clinician as an actor in the process. Finally, we used
the third session as an opportunity to obtain blind feedback
on the developed pathway as a form of litmus test regarding
its accuracy to radiology practice. We presented the pathway
to a new clinician and reporting radiographer, and requested
feedback on any areas where they felt (a) that the pathway
was not indicative of real-world practice and (b) that there
were opportunities for arti cial intelligence to make the
process more e cient.</p>
      <p>The results of the validation sessions were very valuable
for the design process. As a methodology, by performing our
co-creation of the clinician's pathway in this manner, we are
con dent it is accurate to real-world practice, and
generalisable beyond simply an individual's viewpoint. In the nal
validation session, the clinician did not highlight any areas of
the work ow which were not indicative of real-world practice,
while the reporting radiologist suggested only a minor
amendment to terminology. Furthermore, the areas which both of the
participants suggested were suitable for arti cial intelligence
to make a process improvement very closely overlapped with
our own ndings as researchers. We discuss this in more detail
in Section 3.3. In the following subsection, we will introduce
and discuss the developed clinician's pathway in detail.
3.2</p>
      <p>Resulting Clinician's Pathway
In modelling the process of radiology imaging for fracture
treatment, we identi ed three key stakeholders (clinician,
radiologist, patient) and four sub-processes: (1) requesting
radiology images; (2) acquiring radiology images; (3) reporting on
radiology images; and (4) decision support. The complete
gure can be accessed via this link4 and can be seen in Figure 1.
We summarise our pathway using a work ow diagram which
we will break down into respective processes in Figures 2, 3
and 4.
4 https://www.dropbox.com/s/3gx7bicf43bn0lx/wrk w comp c.
pdf?dl=0
s
e
g
a
m
I
y
g
o
l
o
i
d
a
R
g
n
i
t
s
e
u
q
e
R
s
e
g
a
m
I
y
g
o
l
o
i
d
a
R
g
n
i
r
i
u
q
c
A
s
e
g
a
m
I
y
g
o
l
o
i
d
a
R
n
o
g
n
i
t
r
o
p
e
R
t
r
o
p
p
u
S
n
o
i
s
i
c
e
D</p>
      <sec id="sec-3-1">
        <title>Request</title>
        <p>information
entered and
scheduled on RIS
and ordering info
sent to PACs.</p>
        <p>On day of
appointment</p>
        <p>(or
attendance)</p>
      </sec>
      <sec id="sec-3-2">
        <title>RIS sends information to imaging modality.</title>
      </sec>
      <sec id="sec-3-3">
        <title>Radiographer</title>
        <p>commences
appointment.</p>
        <sec id="sec-3-3-1">
          <title>Start</title>
          <p>1. Requesting Radiology Images: When requesting
radiography images for a patient, a clinician sends an imaging
request to the radiology department. This request contains
information on the patient's demographic (including a medical
history), information about the clinician making the request
and the examinations requested. It allows an appointment to
be scheduled on the Radiology Information System (RIS), and
the ordering information is dispatched to Picture Archiving
and Communications System (PACS). PACS is an end-to-end
system that supports the process of acquiring radiography
images of a patient from referral of the patient until
diagnosis and subsequent treatment are agreed. Contained within
PACS is the RIS, where the textual components of patient
information is stored.</p>
          <p>2. Acquiring Radiology Images: Each day, the RIS
automatically generates a work list for each imaging modality.
This work list gives the radiologist (or the reporting
radiographer) access to the request information created by the
original referring clinician, helping them to understand the
imaging requirements. This enables the radiologist to perform the
requested imaging. After the imaging has been performed,
the medical equipment will generate images in Digital
Imaging and Communication in Medicine (DICOM) standard, and
load them into a graphic user interface where they are made
available for the radiologist to annotate and report. A
graphical representation of this is displayed in Figure 2. DICOM
de nes an image standard and format for medical images.
Images are high resolution and are linked directly with other
patient data (such as name, gender and age). It was
developed by the National Electrical Manufacturers Association
(NEMA) as part of a set of standards that de ne best
practice and inform the international standard for the capture,
retrieval, storage and transmission of medical imaging data.
DICOM is currently the most commonly used standard across
the world for medical imaging, and is implemented in most
radiology, cardiology and radiotherapy devices, as well as
devices in other medical domains such as dentistry.</p>
          <p>3. Reporting on Radiology Images: Radiologists can
access these images by calling up a list of unreported
examinations. For each patient, previous images and reports are also
available to support diagnosis. PACS enables radiologists to
annotate the images to highlight areas of interest or identify
supporting evidence for their diagnosis. These annotations
include the ability to perform simple measurements (length of
objects, angles of intersections, etc) and to mark a Region of
Interest (ROI) on the image. This allows the radiologist to
use tools to capture metadata about the ROI, including its
area, average pixel values, standard deviation, and range of
pixel values.</p>
          <p>The radiologist will then generate a textual report to
summarise and describe their ndings. These reports have no
set template or length, but generally include a statement of
whether a fracture has been detected, what type of fracture
it is, where it is located, and the seriousness of the breakage.
Furthermore, the reports may be appended to existing
documentation on the patient (if previous radiology records exist)
or may be used to begin a radiology record (if no previous
visits have been recorded). Many countries then require the
reports to be authorised by a radiologist before being released
to a clinician. For example, within the United Kingdom the
standards for fail-safe communication of radiology reports are
governed by the Royal College of Radiologists (RCR)5. The
result of these factors is that the reports are a complex textual
data source with limited uniformity and describing a broad
range of diagnosis and observations. We represent this
information as part of the work ow displayed in Figure 3.</p>
          <p>4. Decision Support In the existing pathway, decision
5 http://bit.ly/rcr-standards
support occurs after the radiology reports are generated. This
is non-optimal; often for accident and emergency fracture
cases (which make up the majority of fractures in a
hospital) the clinician will attempt to read and comprehend the
generated radiology images without any input from an expert
radiologist. This can be seen in the current work ow in
Figures 2 and 4. This occurs because experts can be unavailable
- not all radiology sta are su ciently trained to report on
acquired images. As a result the clinician is forced to make
a diagnosis and organise follow up treatment on the basis of
their individual knowledge. This can lead to misdiagnosis, if
the clinician's ndings are not consistent with the
radiologist's, which can have an impact on the patient's health as
well as nancial consequences for the hospital involved.</p>
          <p>Having developed an understanding of the current
procedure of radiology imaging for fracture treatment, we are
motivated to make some recommendations where arti cial
intelligence could make improvements to this process.
3.3</p>
          <p>Opportunities for Arti cial Intelligence
The key outcome of this work is highlighting the
applicability of arti cial intelligence in two places: to reduce burden
on radiologists by (1) autonomously classifying radiology
images and (2) generating understandable and accurate medical
reports to describe the intelligent system's ndings.
Applications of arti cial intelligence to ll these gaps presents an
opportunity to improve decision-support for clinicians by
giving them access to the information immediately. This is a key
factor that is missing from much of the research literature
on this topic; although an arti cial intelligence method for
fracture recognition should enhance the e ciency of
radiologists, it should also improve the decision-making of clinicians.
Therefore, it should be of a suitable form to be absorbed by
that user group.</p>
          <p>This outcome is supported by the veri cation obtained from
our test group of clinicians and radiologists. Based on their
feedback, we have highlighted the most impacted area of the
current clinical pathway in Figure 5.</p>
          <p>
            As seen in our discussion of related work, there has been
much exploration of autonomous classi cation of fracture
images throughout the literature [
            <xref ref-type="bibr" rid="ref1 ref10 ref14 ref3 ref5 ref8 ref9">1, 3, 5, 8, 9, 10, 14</xref>
            ]. However,
few works have considered how this could be integrated with
existing medical processes. It is clear that from a clinician's
perspective, it would be desirable to have the classi cation of
the image and the report in order to support their
decisionmaking. This suggests an ecosystem of arti cial intelligence
processes would be much more suitable than a standalone
method.
4
          </p>
          <p>Experiments on Image Classi cation
The main purpose of the experimental framework was to test
the learning capabilities of di erent baseline algorithms and
settings to classify the images provided in the challenge as
fracture/no fracture. To do so, the rst task consisted of
having a specialist re-annotate the data provided by splitting it
into fracture and no fracture labels, based both on the visual
aspect of the radiography and on the information provided
by the text reports. It was discovered that while most of the</p>
          <p>Start</p>
          <p>Patient</p>
          <p>Report sent to
t clinician.
images corresponrded to "regular" scenarios (where the
o
p No
pose is to assess wphether the patient has su ered a fracture or
u New report
not), some other ScasesRaepRloesrptosertnct otontained fFoolllolwouwp -up reports (ide nap-pended tofracture)
ti ed as POP) wiiscnotrpohere dtmccilhaliainngtineicochoiiaaesnnisis.ssueNo is n oaoprtgpaonintistoemdetnoitdentify
ence/absence of aeu fractur?e, but rathceorrrect treagtmievnet a follow
p to
DiiscSeno YedmcsiRlaiaengtpi?nccohoiarestniss No coraorperFgpocaotllniotnirwsetemadutepmtnoetnt
D
ow Nfoor reporting on radiology images.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>First</title>
          <p>Nerewporertp?ort
appended to
puexris-ting reportas. patient</p>
          <p>which already has had the fracture identi ed in a
previous visit. We labelled 73 images as positive (i.e. with
and 138 negative (i.e. no fracture). Moreover, six
existing reports.
the pres- examples were POP fractures, and thus were not included in
up for our experiments.</p>
          <p>Yes
End</p>
          <p>End
To demonstrate the potential of transfer learning capabilities
of the selected algorithms towards the NHS provided records,
we used the following publicly available dataset.</p>
          <p>MURA: The MURA (MUsculoskeletal RAdiographs)
dataset is a large dataset of bone X-rays. Algorithms are
tasked with determining whether an X-ray study is normal or
abnormal. It consists of X-ray scans on elbow, nger, forearm,
hand, humerus, shoulder and wrist. The training set consists
of 14'873 positive cases and 21'939 negative cases while the
validation set has 1'530 positive examples and 1'667 negative
examples. Among them, forearm and wrist are close to our
problem, which consists of 7'443 negative and 5'094 positive
examples. Images from this dataset can be accessed here6.
6 https://stanfordmlgroup.github.io/competitions/mura/
4.2</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Image Preprocessing</title>
      <p>
        To increase the likelihood of classi cation and the training
sample size, We applied the following preprocessing
techniques, which are the most commonly used in related
literature [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]:
horizontal ip, width shift by 0.1,
height shift by 0:1,
shearing with range 0:1,
zoom with range from 0:9 to 1:25 and
random rotation from 0 to 15 degrees.
4.3
      </p>
    </sec>
    <sec id="sec-5">
      <title>Architecture Details</title>
      <p>
        These followings baseline architectures were used in our
experiments:
A VGG 16 [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] is a Convolutional Neural Network (CNN)
model proposed by Simonyan and Zisserman. The model
achieves 92.7% top-5 test accuracy in ImageNet, which is a
dataset of over 14 million images belonging to 1'000 classes.
It is an improvement over the classical AlexNet [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] by
replacing large kernel-sized lters (11 and 5 in the rst and
second convolutional layer, respectively) with multiple 3 3
kernel-sized lters one after another. The original VGG16
was trained for weeks and was implemented using NVIDIA
Titan Black GPU's.
      </p>
      <p>
        B Resnet 50 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is a CNN architecture of 50 layers deep,
each of which is formulated as learning residual functions
with reference to the layer inputs, instead of learning
unreferenced functions. Because of these residual modules, the
architecture can become very deep. This architecture won
the 1st place on the ILSVRC 2015 classi cation challenge.
C Inception V3 [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ] is a CNN architecture which
achieved improved utilisation of the computing resources
inside the network by carefully crafted design that allows
for increasing the depth and width of the network while
keeping the computational budget constant. To optimize
quality, the architectural decisions were based on the
Hebbian principle and the intuition of multi-scale processing.
The authors also proposed ways to scale up networks in
ways that aim at using the added computation as e
ciently as possible by suitably factorised convolutions and
aggressive regularization. Tests were made on the ILSVRC
2012 dataset, in which with an ensemble of four models
and multi-crop evaluation, authors reported 3.5% top-5
error on the validation set (3.6% error on the test set) and
17.3% top-1 error on the validation set.
4.4
      </p>
    </sec>
    <sec id="sec-6">
      <title>Experiment Details</title>
      <p>
        To have di erent points of comparison, we tested the results
of using the three aforementioned classi ers to classify
images from the MURA dataset. We distinguished between the
following four con gurations:
1. When the networks were pre-trained with ImageNet [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
2. When they were initialised randomly.
3. When the networks were initialised randomly, trained on
all the MURA dataset (except wrist and arm images) and
then retrained on wrist and arm images.
4. When the networks were pre-trained from ImageNet
randomly, trained on all of MURA (except wrist and arm) and
then retrained on wrist and arm images.
      </p>
      <p>The accuracy result can be seen in Table 4.4 and the run
time in Table 4.4:</p>
      <p>VGG16
Resnet50
InceptionV3
regardless of its origin. Moreover, case 2 with these same
architectures also showed good performance, (79% and 78%
respectively), but even with the retraining on wrist and arm
images, results were slightly worse than training only with
ImageNet images. This may be due to the fact that some
wrist/arm images had to be used for such retraining instead
of testing. In terms of run time, we also found out that case
1 overall is faster to train and test.</p>
      <p>After this initial validation, we tested the transfer
learning capability from MURA to the newly acquired images. We
tested the following three cases:
5. When networks were pre-trained on ImageNet, trained on</p>
      <p>MURA and tested on the new dataset.
6. Mixing MURA and new images to generate both training
and test sets (70% train, 30% test).
7. Same as the previous case, however the test set was
composed of 70% of MURA images and 30% from the new
dataset.</p>
      <p>The accuracy results are shown in Table 4.4:</p>
      <p>In contrast to what was expected from the previous test, we
observed that for case 5, all CNNs were unable to learn how
to classify the new images. In contrast, it was more likely to
obtain higher accuracy rates for case 6 and VGG 16 (81%),
although this is a direct result of images from the MURA
dataset being mixed within the test set. Meanwhile, case 7
and Inception V3 obtained 75% accuracy, but keeping in mind
that the test set is only composed of new images, this was a
clear indication that it is possible to transfer a model using a
larger amount of images. In terms of run time, we discovered
that it was faster to train networks through case 6, followed by
case 6 and case 7 respectively. The complete run time results
can be seen in Table 4.4.</p>
      <p>The results showed that case 1 with VGG 16 and ResNet
50 delivered the best accuracy overall (82% and 80%
respectively), implying that it is possible to obtain good accuracy
provided that we can train the systems with su cient data,
5</p>
    </sec>
    <sec id="sec-7">
      <title>Conclusion</title>
      <p>In this paper, we have presented a rst step towards assessing
the most proper way to embed machine learning, computer
vision and natural language processing into the clinicians'
pathway to improve assisted diagnostics of fracture detection. We
have reviewed the most signi cant literature and designed a
pipeline where we have annotated the most relevant action
points where arti cial intelligence can be used to improve the
current practices. In addition, we have carried out some
initial experiments to verify how current methods and transfer
learning perform on identifying fractures in a reduced dataset
provided by the British public health service. Results show
that there is a great likelihood of being able to apply transfer
learning for these purposes, and in the case that more images
are provided by the challenge setter, then the accuracy can
vastly improve.</p>
      <p>We will continue this partnership to explore more ways in
which we can further improve our ndings and including other
technologies to enhance the existing results. Finally, we will
keep working with clinicians and radiographers to correctly
assess their pathways and e ectively applying these
technologies in commercial settings.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgment</title>
      <p>We would like to acknowledge the Scottish Government's
Small Business Research Initiative (SBRI) and Opportunity
North East (ONE) for supporting this work, and to the Data
Safe Haven (DaSH) from the National Health Service (NHS)
for granting access to the test data.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Oishila</given-names>
            <surname>Bandyopadhyay</surname>
          </string-name>
          , Arindam Biswas, and
          <string-name>
            <surname>Bhargab</surname>
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Bhattacharya</surname>
          </string-name>
          .
          <article-title>Long-bone fracture detection in digital X-ray images based on digital-geometric techniques</article-title>
          ,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Socher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          , and
          <string-name>
            <given-names>L.</given-names>
            <surname>Fei-Fei</surname>
          </string-name>
          .
          <article-title>ImageNet: A Large-Scale Hierarchical Image Database</article-title>
          .
          <source>In CVPR09</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Kamil</given-names>
            <surname>Dimililer</surname>
          </string-name>
          . IBFDS:
          <article-title>Intelligent bone fracture detection system</article-title>
          .
          <source>In Procedia Computer Science</source>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Kaiming</given-names>
            <surname>He</surname>
          </string-name>
          , Xiangyu Zhang, Shaoqing Ren, and
          <string-name>
            <given-names>Jian</given-names>
            <surname>Sun</surname>
          </string-name>
          .
          <article-title>Deep residual learning for image recognition</article-title>
          .
          <source>arXiv 1512.03385</source>
          , pages
          <fpage>770</fpage>
          {
          <fpage>778</fpage>
          , 06
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D. H.</given-names>
            <surname>Kim</surname>
          </string-name>
          and
          <string-name>
            <given-names>T.</given-names>
            <surname>MacKinnon</surname>
          </string-name>
          .
          <article-title>Arti cial intelligence in fracture detection: transfer learning from deep convolutional neural networks</article-title>
          .
          <source>Clinical Radiology</source>
          ,
          <volume>73</volume>
          (
          <issue>5</issue>
          ):
          <volume>439</volume>
          {
          <fpage>445</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Diederik</surname>
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>Danilo J.</given-names>
          </string-name>
          <string-name>
            <surname>Rezende</surname>
            , Shakir Mohamed, and
            <given-names>Max</given-names>
          </string-name>
          <string-name>
            <surname>Welling</surname>
          </string-name>
          .
          <article-title>Semi-supervised learning with deep generative models</article-title>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Alex</given-names>
            <surname>Krizhevsky</surname>
          </string-name>
          , Ilya Sutskever, and
          <string-name>
            <surname>Geo</surname>
            rey
            <given-names>E</given-names>
          </string-name>
          <string-name>
            <surname>Hinton</surname>
          </string-name>
          .
          <article-title>Imagenet classi cation with deep convolutional neural networks</article-title>
          . In F. Pereira,
          <string-name>
            <given-names>C. J. C.</given-names>
            <surname>Burges</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bottou</surname>
          </string-name>
          ,
          <article-title>and</article-title>
          K. Q. Weinberger, editors,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>25</volume>
          , pages
          <fpage>1097</fpage>
          {
          <fpage>1105</fpage>
          . Curran Associates, Inc.,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Brian</given-names>
            <surname>Lawn</surname>
          </string-name>
          .
          <article-title>Fracture of brittle solids</article-title>
          . Cambridge university press,
          <year>1993</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Robert</given-names>
            <surname>Lindsey</surname>
          </string-name>
          , Aaron Daluiski, Sumit Chopra, Alexander Lachapelle, Michael Mozer, Serge Sicular, Douglas Hanel, Michael Gardner, Anurag Gupta, Robert Hotchkiss, and
          <string-name>
            <given-names>Hollis</given-names>
            <surname>Potter</surname>
          </string-name>
          .
          <article-title>Deep neural network improves fracture detection by clinicians</article-title>
          .
          <source>Proceedings of the National Academy of Sciences of the United States of America</source>
          ,
          <volume>115</volume>
          (
          <issue>45</issue>
          ):
          <volume>11591</volume>
          {11596, nov
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Jakub</surname>
            <given-names>Olczak</given-names>
          </string-name>
          , Niklas Fahlberg, Atsuto Maki, Ali Sharif Razavian, Anthony Jilert, Andre Stark, Olof Skoldenberg, and Max Gordon.
          <article-title>Arti cial intelligence for analyzing orthopedic trauma radiographs: Deep learning algorithms are they on par with humans for diagnosing fractures?</article-title>
          <source>Acta Orthopaedica</source>
          ,
          <volume>88</volume>
          (
          <issue>6</issue>
          ):
          <volume>581</volume>
          {
          <fpage>586</fpage>
          ,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Karen</given-names>
            <surname>Simonyan</surname>
          </string-name>
          and
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Zisserman</surname>
          </string-name>
          .
          <article-title>Very deep convolutional networks for large-scale image recognition</article-title>
          .
          <source>arXiv 1409.1556</source>
          , 09
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Christian</surname>
            <given-names>Szegedy</given-names>
          </string-name>
          , Wei Liu, Yangqing Jia,
          <string-name>
            <given-names>Pierre</given-names>
            <surname>Sermanet</surname>
          </string-name>
          , Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Rabinovich</surname>
          </string-name>
          .
          <article-title>Going Deeper with Convolutions</article-title>
          .
          <source>arXiv:1409</source>
          .4842 [cs],
          <year>September 2014</year>
          . arXiv:
          <volume>1409</volume>
          .
          <fpage>4842</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Christian</surname>
            <given-names>Szegedy</given-names>
          </string-name>
          , Vincent Vanhoucke, Sergey Io e, Jonathon Shlens, and
          <string-name>
            <given-names>Zbigniew</given-names>
            <surname>Wojna</surname>
          </string-name>
          .
          <article-title>Rethinking the inception architecture for computer vision</article-title>
          . CoRR, abs/1512.00567,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Naofumi</surname>
            <given-names>Tomita</given-names>
          </string-name>
          , Yvonne Y. Cheung, and
          <string-name>
            <given-names>Saeed</given-names>
            <surname>Hassanpour</surname>
          </string-name>
          .
          <article-title>Deep neural networks for automatic detection of osteoporotic vertebral fractures on CT scans</article-title>
          .
          <source>Computers in Biology and Medicine</source>
          ,
          <volume>98</volume>
          (February):
          <volume>8</volume>
          {
          <fpage>15</fpage>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Ashish</surname>
            <given-names>Vaswani</given-names>
          </string-name>
          , Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones,
          <string-name>
            <surname>Aidan N Gomez</surname>
            , L ukasz Kaiser, and
            <given-names>Illia</given-names>
          </string-name>
          <string-name>
            <surname>Polosukhin</surname>
          </string-name>
          .
          <article-title>Attention is all you need</article-title>
          . In I. Guyon,
          <string-name>
            <given-names>U. V.</given-names>
            <surname>Luxburg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wallach</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Fergus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Vishwanathan</surname>
          </string-name>
          , and R. Garnett, editors,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>30</volume>
          , pages
          <fpage>5998</fpage>
          {
          <fpage>6008</fpage>
          . Curran Associates, Inc.,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>O.</given-names>
            <surname>Vinyals</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bohus</surname>
          </string-name>
          , and
          <string-name>
            <given-names>R.</given-names>
            <surname>Caruana</surname>
          </string-name>
          .
          <article-title>Learning Speaker, Addressee and Overlap Detection Models from Multimodal Streams</article-title>
          . In ICMI,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Xingwei</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>Lihua</given-names>
            <surname>Li</surname>
          </string-name>
          , Wei Liu, Weidong Xu,
          <string-name>
            <given-names>Dror</given-names>
            <surname>Lederman</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Bing</given-names>
            <surname>Zheng</surname>
          </string-name>
          .
          <article-title>An interactive system for computeraided diagnosis of breast masses</article-title>
          .
          <source>Journal of Digital Imaging</source>
          ,
          <volume>25</volume>
          :
          <fpage>570</fpage>
          {
          <fpage>579</fpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Xulei</surname>
            <given-names>Yang</given-names>
          </string-name>
          , Zeng Zeng, Sin G. Teo,
          <string-name>
            <surname>Li</surname>
            <given-names>Wang</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vijay Chandrasekhar</surname>
            , and
            <given-names>Steven</given-names>
          </string-name>
          <string-name>
            <surname>Hoi</surname>
          </string-name>
          .
          <article-title>Deep learning for practical image recognition: Case study on kaggle competitions</article-title>
          .
          <source>In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD '18</source>
          , pages
          <fpage>923</fpage>
          {
          <fpage>931</fpage>
          , New York, NY, USA,
          <year>2018</year>
          .
          <article-title>Association for Computing Machinery</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>