<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Modeling of Camera from Image Quality Testing Perspective: Valeo Vision Systems Case</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Muhammad Yahya</string-name>
          <email>muhammad.yahya@valeo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aedán Breathnach</string-name>
          <email>n.breathnach@valeo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Faisal Khan</string-name>
          <email>faisal.khan@valeo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iman Abaspur</string-name>
          <email>iman.abaspur@valeo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rajkumar Ranganathan</string-name>
          <email>rajkumar.ranganathan@valeo.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>SemIIM'23: 2nd International Workshop on Semantic Industrial Information Modelling</institution>
          ,
          <addr-line>7th</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Valeo Vision Systems</institution>
          ,
          <addr-line>Dunmore, Tuam, Galway</addr-line>
          ,
          <country country="IE">Ireland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Advanced Driving Assistance Systems (ADAS) have significantly enhanced the modern driving experience by integrating state-of-the-art technology to bolster vehicle safety and driver comfort. Cameras, serving as the eyes of these systems, are pivotal in capturing real-time visual data. This data is processed and analyzed to make instantaneous decisions, such as object detection and lane departure warnings. The manufacturing of the camera has to pass certain tests to qualify the production. During the production, it generated a huge amount of data which is stored in diferent storage places. This, however take huge eforts and time for image quality team to digest such scattered data. To solve this issue of data integrity, we propose Camera Ontology (CamOnt) with the scope to represent the camera testing domain knowledge. The ontology is built using the knowledge gathered from the domain experts and ISO12233 document, and is evaluated with the catalogue of SPARQL queries provided.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <sec id="sec-2-1">
        <title>CEUR</title>
      </sec>
      <sec id="sec-2-2">
        <title>Workshop</title>
      </sec>
      <sec id="sec-2-3">
        <title>Proceedings</title>
        <p>Systems (ADAS) [4]. Serving as the eyes of the system, cameras capture real-time visual data,
which is then processed and analyzed to make split-second decisions, such as object detection,
lane departure warnings, and pedestrian alerts. The precision and reliability of these cameras
directly influence the system’s ability to prevent potential accidents and ensure driver safety.
However, If an ADAS camera fails, the system may not detect obstacles or misinterpret visual
data, leading to compromised safety features and an increased risk of accidents [5, 6]. Causes of
camera failure can range from physical obstructions like dirt or debris on the lens, software
glitches, and malfunctioning hardware components, to adverse environmental conditions such
as extreme temperatures. Ensuring the consistent performance and reliability of ADAS cameras
is thus paramount for the overall safety and efectiveness of the system.</p>
        <p>Valeo1, an Original Equipment Manufacturer (OEM) and a global automotive supplier is at
the forefront of developing advanced vision systems that play a crucial role in the realm of
ADAS. Their state-of-the-art cameras are designed to enhance vehicle safety by providing a
comprehensive view of the vehicle’s surroundings, enabling features such as lane departure
warnings, trafic sign recognition, and pedestrian detection. Valeo’s vision systems leverage
sophisticated image processing algorithms and high-resolution sensors to ensure accurate
real-time data interpretation, even in challenging lighting or weather conditions. As ADAS
technologies continue to evolve towards fully autonomous driving, Valeo’s commitment to
innovation and quality ensures that its cameras remain an integral component in driving safety
advancements, helping to reduce accidents and save lives on the road.</p>
        <p>Each camera undergoes rigorous testing to meet specified standards in the Valeo production
lines. During these tests a huge amount of data is produced, that is then stored across various
storage locations. The Image Quality (IQ) team is responsible for evaluating this scattered data
to ensure that testing criteria are met. The challenge, however, arises from the dispersion of data
across multiple locations, resulting in the IQ team spending an extended duration on this task.</p>
        <p>To address this issue ontology emerged as a significant tool to integrate the data [ 7].
Ontologies2 appeared as a significant tool to represent the domain knowledge of the manufacturing
domain to support data integration and interoperability [10, 11].</p>
        <p>In this work, we are proposing Camera Ontology (CamOnt) to integrate the data residing in
different places in Valeo vision system. The CamOnt is built with the knowledge gathered from the
Valeo experts with the scope defined to help the image quality team access the data integrated via
a uniform model. At present, the ontology is evaluated with the experts-defined query catalogue.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>2. Related Work</title>
      <p>In the evolving landscape of semantic web technologies, ontologies have emerged as pivotal
techniques for representing domain-specific knowledge [ 12, 13, 14]. The domain of camera
1https://www.valeo.com/en/catalogue/cda/surround-view-camera-solutions/
2Gruber et al. 1993 [8], define an ontology as the formal, explicit specification of a shared conceptualization. The
basic elements of an ontology are a concept and the relation between it, and its axioms. When the instances of
concepts are populated into the developed ontology it becomes a knowledge base also known as knowledge graph
[9]. There are two key elements in Ontology: terminological components (Tbox) and assertional components (Abox)
that specify the concepts and their instances, respectively.
technology and testing is no exception, with several attempts made to formalize and structure
the vast array of concepts and relationships inherent to it.</p>
      <p>Wei et. al explored sentiment analysis within the context of digital camera reviews [15].
They constructed a Sentiment Ontology Tree (SOT) specifically for digital cameras, which
contained nodes representing various attributes of the camera and their associated sentiments.
This work provides a unique perspective on camera ontology, emphasizing the sentiment aspect
of camera attributes rather than just the technical specifications. In the visual domain, the
Spatio-Temporal Visual Ontology was introduced [16]. While not directly related to cameras,
it ofers insights into the semantic representation of visual data, such as images produced by
cameras. A notable contribution in the broader realm of product families comes from Nanda
et al. [17]. They showcased the practicality of the Product Family Ontology Development
Methodology (PFODM) by constructing an ontology for a family of one-time-use cameras.</p>
      <p>There are existing ontologies like the aforementioned that touch on cameras, but they don’t
specifically address the complexity of camera testing for automotive use. However, camera
technology in the automotive sector is crucial for enhancing safety. Our work fills this gap,
focusing on this niche yet vital area. The rapid evolution of camera tech highlights the need for
such a specialized ontology, especially since current literature doesn’t adequately cover this
domain.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Ontology Development Methodology</title>
      <p>We now discuss the ontology development process. We have adopted the ontology development
process from [18]. In the pursuit of a semantic model for camera image quality testing, the Valeo
Vision Systems case study emphasize the significance of a structured ontology development
approach. The initial phase involves defining the functional and non-functional requirements
for an ontology tailored to the complexity of camera image quality. Collaborating with domain
experts in line with ISO12233 document, specific use-cases related to the Image quality is
described. This is strengthen by a thorough examination of ISO12233 standard document, and
datasets, laying the groundwork for the ontology’s purpose and scope.</p>
      <p>The next phase is the formalization of concepts. By scrutinizing existing ontologies, relevant
terminologies are searched, ensuring alignment with the camera domain. This meticulous
formalization is characterized by establishing relationships between concepts. For instance,
a concept like ImageSensor might be linked to ImageProcessingAlgorithm through specific
properties. Leveraging tools such as Protégé, these formalized concepts are seamlessly translated
into RDF/OWL formats. After the development of the ontology, the next step is its evaluation.
In this preliminary work, we have evaluated our ontology based on a competency questions.</p>
    </sec>
    <sec id="sec-5">
      <title>4. Ontology Overview</title>
      <p>The section presents an overview of the CamOnt ontology. The ontology focuses on the domain
of camera testing and visual resolution measurements (see Figure 1 for class hierarchy). The
ontology encapsulates various aspects of camera testing, from the settings and conditions
under which cameras are tested to the specific features of test charts used in visual resolution
measurements.</p>
      <p>The key classes in the ontology include Camera, which represents the primary entity being
tested, and Test, which captures the various tests a camera might undergo. The ontology also
delves into the nuances of test conditions, such as white balancing (AutomaticWhiteBalancing
and ManualWhiteBalancing), gamma correction (GammaCorrection), and camera focusing
(CameraFocusing). Test charts, essential tools in camera testing, are represented with classes
like TestChart, DimensionalSpecification, and HyperbolicWedgeTestPattern. Table 1</p>
      <p>Class Definition
Camera The primary entity being tested, representing any camera device.
Test Captures the various evaluations or procedures a camera may undergo.
AutomaticWhiteBalancing iRmepagreessetnatksetnhebyauthtoemcaamticerma.ethod of adjusting the colour balance in
ManualWhiteBalancing Denotes the manual method of adjusting the colour balance in images.
GammaCorrection Refers to the adjustment of luminance or colour values in the image.
CameraFocusing Represents the techniques or methods used by the camera to focus on subjects.
TestChart A tool or pattern used in camera testing to evaluate various parameters.
DimensionalSpecification Specifies the dimensions or size-related attributes of a test chart.</p>
      <p>HyperbolicWedgeTestPattern A specific pattern on a test chart used for certain evaluations.
shows some of the classes with their definitions.</p>
      <p>The ontology defines several object properties to capture the relationships between these
classes. For instance, hasTestConditions can link a Test to its specific conditions, while
usesWhiteBalancing might specify the type of white balancing a camera employs. Table 2 shows some of
the object properties with their domain and range.</p>
      <p>To exemplify the domain knowledge, consider axiom 1 which states that there exists some
cameras that use the automatic white-balancing method. Axiom 2 represents that there exist
some tests that employ the Hyperbolic Wedge Test Pattern in their test charts. Moreover, axiom
3 states that there exist some cameras that have a specific lens setting of type Camera Framing
and Lens Focal Length Setting and also use a focusing method of type Camera Focusing.
1. Camera ⊓ ∃usesWhiteBalancing.AutomaticWhiteBalancing
2. Test ⊓ ∃employsTestChart.HyperbolicWedgeTestPattern
3.
Camera⊓∃hasLensSetting.CameraFramingandLensFocalLengthSetting⊓∃usesFocusingMethod.CameraFocusing</p>
    </sec>
    <sec id="sec-6">
      <title>5. Evaluation</title>
      <p>According to Gomez-Perez et al. 1995, ontology evaluation is a technical judgment of the
ontology in relation to a frame of reference [19]. This frame of reference can encompass
requirement specifications, competency questions, and its real-world applications. A crucial
aspect of ontology evaluation is the formulation of competency questions from the user’s
perspective to ascertain if the ontology meets its intended purpose. Hammar et. al 2010
proposed the creation of usage examples in natural language to highlight the significance of the
ontology’s concepts [20]. Commonly termed competency questions, these user-centric queries
are instrumental in gauging the ontology’s scope. Essentially, they represent the questions
e
t
i
h</p>
      <p>.
sW g</p>
      <p>n
e i
s c
:u n</p>
      <p>a
m l
a a
c B</p>
      <p>e
ry ra it
e e h
u</p>
      <p>m W
Q a</p>
      <p>c
L c? it
Q { a
R e m
A re t</p>
      <p>o
SP h u
w :A
a m
r
e a</p>
      <p>c
am g</p>
      <p>n
c? i</p>
      <p>c
t
c n
e la
l
e a
s B }
} .</p>
      <p>t
r
g
n .
i
t
t
e
S t
s e
n S
g
n
i
t
a
h e</p>
      <p>h</p>
      <p>L t
tC s g
s a n
e :
t
?
h e</p>
      <p>L
m l
a a
t
ra c c</p>
      <p>o
{ h a F</p>
      <p>r s
re tC e
e s
h e</p>
      <p>n
m e
a L
w s
t y {</p>
      <p>T c d</p>
      <p>? n
ra l</p>
      <p>a
o e g</p>
      <p>r
h p e in
C m h
m ra r</p>
      <p>F
e a
t e
s :
e
t
? a
t c
c
t
s a
e c
e
l
e t
s ? t
w am</p>
      <p>r
m e</p>
      <p>m
? a</p>
      <p>C
c :
e
l</p>
      <p>m
E</p>
      <p>E
]</p>
      <p>]
:
P
D
,
y
t
r
e
p
o
r
p
t
c
e
j
b
o
:
E
P
O
,
n
o
i
s
s
e
r
p
x
E
s
s
a
l
C
:
E
C
w
.
s
n
r
e
t
t
a
p
h
t
i
n
o
i
t
a
t
n
e
s
e
r
p
e
r
L
Q
R
A
P
S
r
i
e
h
t
d
n
a
s
n
o
i
t
:
3
e
l
b
a
T
.
n
o
i
t
c
e
r
r
o
C
a
m
m
a
g
?
n
o
i
{ t</p>
      <p>c
e e
r r
e r
h o
w C</p>
      <p>a
n
o m
i
t m
c a
e
r</p>
      <p>G
r s
o y
C lo
a p
m m
m :e
a
g m
? a
t c
c t
s
]
2
E
C
[
]
C d C
stakeholders aim to answer using the ontology and its linked knowledge base. Hence, designing
a comprehensive set of competency questions that encapsulate most real-world scenarios is
imperative. These questions necessitate thorough scrutiny to eliminate any that are irrelevant.
5.1. Competency Questions: Valeo Use-case
In the context of CamOnt, the competency questions were sourced from Camera IQ domain
experts, eliminating the need for an in-depth analysis of their efectiveness. It is worth noting
that experts who provide domain knowledge are diferent from those gave SPARQL queries.
These questions are tabulated in Table 3. The table’s first column presents the competency
questions in natural language, the second column delineates the triple pattern addressed by
each question, and the final column showcases the corresponding SPARQL query designed
to extract the requisite knowledge. Furthermore, we incorporated instances of camera data
to validate the capability of the CamOnt to represent it. The SPARQL queries are tailored to
examine the camera and their test results. Each test is conducted with specific parameters like
lens setting, white balancing method, gamma correction, and test chart used. These parameters
are interconnected to the test via specific relations. For instance, query 1 reveals that an XYZ 3
camera utilized an automatic white balancing, query 2 returns a test chart of a hyperbolic wedge
test pattern and query 4 with a gamma correction of 2 is used in a particular test.</p>
    </sec>
    <sec id="sec-7">
      <title>6. Conclusion and Future work</title>
      <p>In this paper, we have developed and presented the CamOnt ontology, a semantic model tailored
for the domain of camera testing and visual resolution measurements. Our ontology stands as a
testament to the interplay between camera technology and semantic modeling, ofering a
structured approach to understanding and analyzing camera tests. The ontology is developed with the
knowledge acquired from ISO12233 documents and domain experts. Furthermore, the
competency questions, curated with insights from domain experts, underscore the real-world relevance
and robustness of CamOnt. The catalogue of SPARQL queries, specifically designed for CamOnt,
showcases its ability to extract detailed insights and highlights its potential as a valuable tool for
Valeo end users. In future, we will incorporate more domain knowledge of cameras other than
testing. Its harmonization with DOLCE or BFO will be carried out will be carried out in the future.
[4] J. S. Murthy, G. Siddesh, W.-C. Lai, B. Parameshachari, S. N. Patil, K. Hemalatha,
Objectdetect: A real-time object detection framework for advanced driver assistant systems using
yolov5, Wireless Communications and Mobile Computing 2022 (2022).
[5] A. Ebrahimi, E. Akbari, Design and implementation of an afordable reversing camera
system with object detection and obd-2 integration for commercial vehicles, 2023.
[6] A. Wahid, M. Yahya, J. G. Breslin, M. A. Intizar, Self-attention transformer-based
architecture for remaining useful life estimation of complex machines, Procedia Computer Science
217 (2023) 456–464.
[7] B. Zhou, Z. Tan, Z. Zheng, D. Zhou, Y. He, Y. Zhu, M. Yahya, T.-K. Tran, D. Stepanova, M. H.</p>
      <p>Gad-Elrab, et al., Neuro-Symbolic AI at Bosch: Data Foundation, Insights, and Deployment,
Technical Report, 2022.
[8] T. R. Gruber, A translation approach to portable ontology specifications, Knowledge
acquisition 5 (1993) 199–220.
[9] M. Yahya, J. G. Breslin, M. I. Ali, Semantic web and knowledge graphs for industry 4.0,</p>
      <p>Applied Sciences 11 (2021) 5110.
[10] M. Yahya, B. Zhou, Z. Zheng, D. Zhou, J. G. Breslin, M. I. Ali, E. Kharlamov, Towards
generalized welding ontology in line with iso and knowledge graph construction, in:
European Semantic Web Conference, Springer, 2022, pp. 83–88.
[11] M. Yahya, A. Ali, Q. Mehmood, L. Yang, J. G. Breslin, M. I. Ali, A benchmark dataset with
knowledge graph generation for industry 4.0 production lines, Semantic Web (????) 1–19.
[12] D. Rincon-Yanez, M. H. Gad-Elrab, D. Stepanova, K. T. Tran, C. C. Xuan, B. Zhou, E.
Karlamov, Addressing the scalability bottleneck of semantic technologies at bosch, arXiv
preprint arXiv:2309.10550 (2023).
[13] Z. Zheng, B. Zhou, A. Soylu, E. Kharlamov, Towards a visualisation ontology for data
analysis in industrial applications (2022).
[14] A. Iqbal, A. Shahid, M. Roman, M. T. Afzal, M. Yahya, Exploiting contextual word
embedding for identification of important citations: Incorporating section-wise citation counts
and metadata features, IEEE Access 11 (2023) 114044–114060. doi:10.1109/ACCESS.2023.
3320038.
[15] W. Wei, J. A. Gulla, Sentiment learning on product reviews via sentiment ontology tree, in:
Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics,
2010, pp. 404–413.
[16] J. I. Olszewska, Spatio-temporal visual ontology, in: BMVA/EACL EPSRC Workshop on</p>
      <p>Vision and Language (VL’11), 2011.
[17] J. Nanda, T. W. Simpson, S. R. Kumara, S. B. Shooter, A methodology for product family
ontology development using formal concept analysis and web ontology language (2006).
[18] M. Yahya, B. Zhou, J. G. Breslin, M. I. Ali, E. Kharlamov, Semantic modeling, development
and evaluation for the resistance spot welding industry, IEEE Access (2023).
[19] A. Gómez-Pérez, N. Juristo, J. Pazos, Evaluation and assessment of knowledge sharing
technology, Towards very large knowledge bases (1995) 289–296.
[20] K. Hammar, K. Sandkuhl, The state of ontology pattern research: a systematic review of
iswc, eswc and aswc 2005–2009, in: The Workshop On Ontology Patterns (WOP 2010) At
The 9th International Semantic Web Conference (ISWC 2010), 2010, pp. 5–17.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Farooq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Corcoran</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rotariu</surname>
          </string-name>
          , W. Sharif,
          <article-title>Object detection in thermal spectrum for advanced driver-assistance systems (adas)</article-title>
          ,
          <source>IEEE Access 9</source>
          (
          <year>2021</year>
          )
          <fpage>156465</fpage>
          -
          <lpage>156481</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>W.-Y.</given-names>
            <surname>Chung</surname>
          </string-name>
          , T.-W. Chong,
          <string-name>
            <given-names>B.-G.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Methods to detect and reduce driver stress: a review</article-title>
          ,
          <source>International journal of automotive technology 20</source>
          (
          <year>2019</year>
          )
          <fpage>1051</fpage>
          -
          <lpage>1063</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Barakoti</surname>
          </string-name>
          ,
          <article-title>Enhancing driving safety using artificial intelligence technology (</article-title>
          <year>2023</year>
          ).
          <article-title>3Used dummy data instead of the actual values due to Valeo's data policy. The data can not be shared in any form outside the organization</article-title>
          .
          <article-title>Due to the policy, we are unable to share the figures of the real queries</article-title>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>