<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>MatTag: Practical material tagging using visual fingerprints</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Adam Staš</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniel Pilař</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jiří Filip</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>The Czech Academy of Sciences, Institute of Information Theory and Automation</institution>
          ,
          <addr-line>Prague</addr-line>
          ,
          <country country="CZ">Czech Republic</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Assessment of material properties is essential for tasks such as similar material retrieval or swatch comparison in industrial design, manufacturing, and quality control. While many material similarity measures exist, they often fail to align with human perception. In this paper, we introduce a novel smartphone application using a machine learning model that leverages a perceptual representation known as the visual fingerprint of materials-linking image-based measurements to intuitive, human-understandable attributes. Trained on human ratings collected through psychophysical studies, the model can predict a material's visual fingerprint using just two photographs captured under diferent lighting conditions. The application employ this model to assess any planar material sample using only a printed registration template and a flashlight. The app captures two photographs and predicts the material's perceptual attributes. We demonstrate several practical use cases, including building personal material databases, retrieving visually similar materials, and exploring materials that match user-defined perceptual criteria. By enabling perceptually grounded comparisons and metadata extraction, our application provides a standardized representation of material appearance. This marks a step toward more intuitive and interoperable use of material properties across diverse digital environments.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Material Appearance</kwd>
        <kwd>Perceptual Attributes</kwd>
        <kwd>Visual Fingerprint</kwd>
        <kwd>Smartphone Application</kwd>
        <kwd>Material Retrieval</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>There are currently many systems and data formats for the visualization and digital representation of
real-world materials. However, these systems often lack meta-information about fundamental visual
properties, which is essential for accurate material identification, comparison, and retrieval. Recent
advances in the study of material perception have led to the development of a machine learning model
that predicts how humans evaluate materials using a unified visual identifier a vector of sixteen
perceptual attribute ratings. This perceptually grounded representation enables more intuitive interpretation
of material appearance.</p>
      <p>Despite this progress, there remains a lack of a practical tool that makes these technologies accessible
to a broader audience – including researchers, domain experts, and advanced users in material analysis
and visualization. Such a tool should support:
• Easy and consistent cataloging of personal/company material samples
• Searching for visually similar materials, either from a reference sample or via user-defined
perceptual criteria
• Comparing the visual attributes of two selected materials in an intuitive way
These capabilities have significant potential in industrial applications, such as quality control across
production batches, analyzing appearance diferences between product lines, or selecting products that
visually match customer expectations.</p>
      <p>Mobile platforms are ideally suited for implementing such a tool due to their accessibility, integrated
cameras, and ease of use—without the need for specialized hardware. In this paper, we introduce a
mobile application for Android that builds on the aforementioned machine learning model to enable
intuitive evaluation of real-world materials. The app features a user-friendly interface for capturing
two photographs of a material sample under diferent lighting conditions. The resulting analysis is
visualized as a polar plot, displaying predicted values of the sixteen perceptual attributes. Additional
features include:
• Local storage of analyzed materials for ofline access
• An interactive filtering interface based on the polar plot, allowing users to define desired perceptual
properties by manipulating the attribute axes directly
• Cloud connectivity, enabling storage and search across a shared material database, with all
analysis performed on a remote server
This application bridges the gap between advanced perceptual modeling and practical, accessible tools
for material analysis in both professional and industrial settings. In the following sections we overview
principle of material fingerprinting, its integration to a mobile application, evaluate its performance by
capturing real-materials and outline several practical usecases.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related work</title>
      <p>
        Establishing a connection between perceptual texture spaces and computational representations has
been the focus of numerous psychophysical studies, including work by Tamura et al.[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], Malik and
Perona[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], Ravishankar and Jain [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], Mojsilovic et al.[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and Long et al.[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        More recent approaches combine human perception of materials—often represented through color
and texture—with machine learning methods. Schwartz and Nishino [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] introduced the concept of visual
material traits, which encode characteristic appearance features using convolutional representations of
image patches for use in material recognition tasks. Subsequent work [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ] explored locally recognizable
material attributes, learning perceptual attribute spaces from pairwise perceptual distances and training
classifiers to reproduce them from image features.
      </p>
      <p>
        To reduce reliance on predefined attribute sets, Schwartz and Nishino [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] later proposed a method
for deriving material annotations by probing human visual perception through simple yes/no questions
comparing pairs of small image patches.
      </p>
      <p>
        Filip et al.[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] investigated perceptual dimensions of wood materials using video stimuli and a
combination of similarity and attribute rating studies. In follow-up work[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], they linked these perceptual
ratings with computational statistics and demonstrated the predictive power of statistical features on
an independent test dataset.
      </p>
      <p>
        While recent AI-based methods (e.g., CLIP [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]) provide fast multimodal embeddings for comparing
visual content, they lack direct alignment with human perception of material appearance. For instance,
although CLIP model itself can estimate the likelihood of certain visual attributes in provided image
pairs, these estimates diverge significantly from human ratings. Yellow bars in Fig. 2 show correlations
between model predictions and human ratings of the 16 visual attributes for diferent material categories.
In average the correlation is only 0.321.
      </p>
      <p>
        Deschaintre et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] addressed this gap by linking free-text descriptions to fabric appearance. They
collected a large dataset of human-written descriptions and corresponding fabric images, then derived
a compact lexicon of attributes and structural descriptors used to describe fabric appearance. This
annotated dataset was used to train a language-image model, producing a perceptually meaningful
latent space for fabric retrieval and automatic annotation.
      </p>
      <p>
        This paper is a follows up to a recent work by Filip et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ], which introduced a material
fingerprinting model based on human-perceived visual attributes. The model predicts these attributes from
only two photographs captured under controlled lighting conditions and forms the basis for the MatTag
application introduced in this paper.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Material fingerprinting</title>
        <p>
          This work builds on a previously introduced model for predicting key perceptual features of materials,
as proposed in [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. As outlined in Fig. 1-a, that study involved a psychophysical experiment using
standardized video sequences of 347 diverse real-world materials, including fabrics, wood, and other
surface types. The dataset was designed to cover a broad spectrum of textures, colors, and reflective
properties.
        </p>
        <p>Instead of relying on static imagery, the study used video sequences to capture the authentic
appearance of flat material samples under varying viewing conditions. Sixteen key visual appearance
attributes were identified through psychophysical analysis, supported by over 110,000 human ratings
that mapped perceptual attributes across material categories.</p>
        <p>
          To model this perceptual space, CLIP-derived image features [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] were combined with a multi-layer
perceptron (MLP) trained on the human rating dataset (see Fig.1-a). The resulting model predicts
attribute ratings based on just two input photographs, taken under distinct lighting conditions: one
emphasizing specular reflection, and the other capturing a non-specular view (Fig. 1-b). The predicted
attribute values can be applied to tasks such as perceptually grounded material comparison or
similaritybased retrieval.
        </p>
        <p>The predictive performance of the model is illustrated by the blue bars in Fig. 2, which show the
correlation between model-predicted ratings and human ratings. Each bar represents the performance
for a specific material category, while the overall mean correlation across all 68 test materials is
0.909. These test materials were not included in the model’s training set, ensuring a fair evaluation of
generalization performance. It is worth noting that this level of performance is particularly impressive,
given that the model predicts perceptual attributes from only two input images, whereas human
observers based their ratings on full video sequences of the materials.</p>
        <p>Evaluation of the model is performed server-side and typically completes within a few seconds. The
method is robust to sample rotation, provided the material does not exhibit strong anisotropic efects.
A current limitation lies in the training dataset, which, although carefully curated, is restricted to 347
materials. Nonetheless, this represents one of the largest sets used in perceptual material research to
date. The model is relatively tolerant to minor deviations in capture geometry but may fail to distinguish
between materials with appearance changes in angular configurations outside the sampled conditions.
Furthermore, the trained model assumes a fixed capture scale and resolution, corresponding to a 26 ×
26 mm material area resampled to 512 × 512 pixels. As a result, it may not efectively evaluate materials
with very low spatial frequency patterns or extremely gradual visual transitions.</p>
        <p>
          In this paper, we extend the use of the trained model by introducing a smartphone application,
developed within matester thesis project [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], that enables intuitive capture of the two required images
and provides automatic evaluation of the material’s perceptual fingerprint. We also describe several
real-world use cases of the application.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. MatTag application</title>
      <p>Since the application predicts human judgments for a fixed set of visual attributes, it efectively performs
attribute tagging. For this reason, we call it MatTag (Material Tag). The app consists of three main
functional tabs: Capturing material images, Analyzing and comparing fingerprints, Application settings.</p>
      <sec id="sec-3-1">
        <title>3.1. Capturing workflow</title>
        <p>The capturing process requires a printed paper template with a cut-out window to define the area of
interest (see Fig. 3). This template is placed on the planar surface of the material being analyzed. It
enables automatic detection of the correct camera polar angle, which is set to 45 degrees, matching the
expected lighting geometry.</p>
        <p>Once the camera reaches the correct angle, the frame color changes from red to green, indicating
that the user may proceed with capturing. Two images must be taken: One with the light coming from
the left side, one with the light source placed opposite the camera (to emphasize specular reflection).
These steps are illustrated in Fig.4-a,b. After both images are captured, they are automatically corrected
for perspective distortion and cropped. The user is then prompted to verify the image order (Fig.4-c)
before initiating the analysis (d).</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Analysis and fingerprint comparison</title>
        <p>During analysis, the images are uploaded to a remote server, where the visual fingerprint is computed
in a matter of seconds. The resulting values for the sixteen perceptual attributes are returned and
displayed as a polar plot (Fig. 5-a).</p>
        <p>
          Once the fingerprint is available, users can perform:
• Retrieval of similar materials from either a local or a shared server-side database (b). The retrieval
process uses the original distance function proposed in [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ], which combines correlation and
absolute diferences between attribute values.
        </p>
        <p>• Comparison of two materials, either in a single polar plot (c) or as two side-by-side plots (d)</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Interactive filtering of materials</title>
        <p>Users can also filter materials by manually adjusting individual attribute values on the polar plot. This
creates a custom query fingerprint used for perceptual retrieval. Each attribute can be increased or
decreased by dragging its axis marker outward or inward, respectively, depending on how important
the attribute should be in the search.</p>
        <p>Fig. 6 illustrates retrieval results for two manually defined fingerprints. Examples (a,b) show retrieval
for materials with higher thickness, striped or checkered patterns, and lower shininess. Examples (c,d)
depict a search for materials with high brightness and low shininess. The last edited attribute is shown
beneath the graph, and changes can be easily undone using the Undo button.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Platform and availability</title>
        <p>MatTag is developed for the Android operating system and is compatible with devices running Android
12.0 (API level 31) or higher. The application architecture separates the user interface and capture
logic from the perceptual analysis backend. The mobile front-end, developed in Kotlin and Java for
Android, handles camera control, user interaction, and image preprocessing. The perceptual fingerprint
is computed server-side using a Python-based backend, which runs the trained machine learning model
and returns predicted attribute values to the app. Additional requirements are functional camera and
internet connection for material analysis and remote storing materials (if approved). The app will be
available through the project website, and we plan to release a stable version on the Google Play Store.</p>
        <p>In the Settings tab, users can choose whether to allow their data to be stored on the remote server.
If disabled, only local data is available for retrieval, while the server is used solely for computing the
ifngerprint. Additionally, users can export captured images and corresponding fingerprint values as a
CSV file for further analysis or archiving.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>The performance of the application was evaluated on a set of real-world material samples, as shown
in Fig.7 and Fig.8. For each material, we present: the two input images (captured under specular and
non-specular lighting), the estimated perceptual fingerprint visualized as a polar plot, and the five most
similar materials retrieved from our database of 347 samples. Each displayed similar material includes
both input images and the corresponding ground truth attributes obtained from human ratings. We also
report the typicality of each material, visualized as a gray bar representing the mean absolute similarity
score to the top 10% of most similar materials in the database. Although this typicality measure is not
yet implemented in the current MatTag release, its integration is planned in a future update.</p>
      <p>The results demonstrate the model’s rotation invariance, provided that the material’s main directional
features are aligned with the template edges (see samples 8 and 9, 13 and 15, 14 and 16). The impact of
lighting conditions, such as illumination intensity and direction, is also illustrated in samples 11 and
12, 13 and 14, and 15 and 16. It is important to note that polished wood materials are not retrieved, as
the database does not currently contain such samples. Despite this, the method consistently retrieves
perceptually relevant matches, even for materials that were not present in the training set. Examples
include samples 2 (bubble foil), 5 (rough wood), 10 (cork panel), 21 (carbon fibers), and 22 (embossed
rubber). Additionally, the method supports extended retrieval by allowing users to manually adjust
specific visual attributes—e.g., to retrieve materials that are similar in most respects but with higher
brightness or a stronger stripe pattern.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Practical recommendations and use-cases</title>
      <p>Illumination conditions – The accuracy of the method depends on the consistent use of the same
illumination source and its recommended positioning. Directional lighting is preferred over point-like
sources, which are typical of many LED lights. Improved results can be achieved by placing a difuse
transparent material (e.g., frosted plastic or paper) a short distance in front of the LED to soften and
expand the light.</p>
      <p>For highly specular materials, it is recommended to slightly ofset the light source from the exact
specular direction. This avoids image saturation and improves visibility of surface texture, which is
essential for accurate analysis.</p>
      <p>A dark environment is not essential, but brightly lit rooms should be avoided. The intensity of the
controlled light source should always be significantly higher than the ambient illumination.
Sample placement – To minimize blurring or distortion, the template must be positioned as close to
the sample surface as possible and should remain perfectly flat without bending. While the method can
still operate on images of nearly-flat surfaces, its accuracy may be reduced. In such cases, the approach
is suitable only for relative comparisons between materials, provided they are captured using the same
nearly-flat geometry.</p>
      <p>Accurate alignment of the sample with the template, so that its edges are parallel to the camera’s
optical axis, also improves attribute prediction, especially when comparing visually similar materials.
This alignment is particularly important for anisotropic materials that exhibit directional structures,
such as wood grains or woven fabrics. Note that the model cannot capture the full range of anisotropic
behavior beyond what is represented in the two input images. To extend its capability to handle full
anisotropy, one would need to broaden the angular trajectories of both lighting and viewing directions
in the stimulus videos, repeat the rating experiment, and retrain the MLP model accordingly. Such an
extension would also require a larger number of input images to suficiently represent the material’s
directional appearance.</p>
      <sec id="sec-5-1">
        <title>Below, we outline several real-world use-cases for the MatTag application:</title>
        <p>1. Creating a private material database: When captured materials are stored locally, MatTag
can be used to build a proprietary material database. Since the source code for the server-side
component is publicly available, multiple users within an institution can share a private database
of in-house materials for retrieval and similarity comparison. The reliability of such a database</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>We presented MatTag, a practical mobile application for perceptual analysis and comparison of material
appearance using their visual fingerprints. Built o n a m achine learning m odel trained o n human
perceptual data, MatTag predicts intuitive visual attributes from just two smartphone-captured images
under controlled lighting conditions. The application enables users to capture, analyze, and retrieve
materials based on their perceptual properties—without requiring specialized hardware. Through
features such as interactive filtering, local and remote retrieval, and visual fingerprint comparison
via polar plots, MatTag supports both professional workflows (e.g., material cataloging and quality
control) and exploratory use in design and research. By ofering a perceptually grounded, image-based
representation of materials on a widely accessible platform, MatTag represents a step forward in bridging
material perception research with practical, real-world applications in digital material workflows. In
the future update we plan adding information on fingerprint prediction reliability.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>This research was supported by the Czech Science Foundation under grant GA22-17529S. We thank
Veronika Vilímovská and Jan Kotera for their contributions to the implementation of the material
ifngerprinting model. W e also gratefully acknowledge the volunteers who participated in the user study
of the application.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used GPT-4 in order to grammar and spelling
check. After using this service, the author(s) reviewed and edited the content as needed and take full
responsibility for the publication’s content.</p>
    </sec>
    <sec id="sec-9">
      <title>Online Resources</title>
      <sec id="sec-9-1">
        <title>The source codes for the MatTag application and server are available:</title>
        <p>• GitHub - MatTag application https://github.com/adamstas/material-fingerprint-app
• GitHub - MatTag server https://github.com/adamstas/material-fingerprint-app-server</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>H.</given-names>
            <surname>Tamura</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mori</surname>
          </string-name>
          , T. Yamawaki,
          <article-title>Textural features corresponding to visual perception</article-title>
          ,
          <source>Systems, Man and Cybernetics</source>
          , IEEE Transactions on
          <volume>8</volume>
          (
          <year>1978</year>
          )
          <fpage>460</fpage>
          -
          <lpage>473</lpage>
          . doi:
          <volume>10</volume>
          .1109/TSMC.
          <year>1978</year>
          .
          <volume>4309999</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Malik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Perona</surname>
          </string-name>
          ,
          <article-title>Preattentive texture discrimination with early vision mechanisms</article-title>
          ,
          <source>JOSA A 7</source>
          (
          <year>1990</year>
          )
          <fpage>923</fpage>
          -
          <lpage>932</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ravishankar Rao</surname>
          </string-name>
          , G. Lohse,
          <article-title>Towards a texture naming system: Identifying relevant dimensions of texture</article-title>
          ,
          <source>Vision Research</source>
          <volume>36</volume>
          (
          <year>1996</year>
          )
          <fpage>1649</fpage>
          -
          <lpage>1669</lpage>
          . doi:
          <volume>10</volume>
          .1016/
          <fpage>0042</fpage>
          -
          <lpage>6989</lpage>
          (
          <issue>95</issue>
          )
          <fpage>00202</fpage>
          -
          <lpage>2</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Mojsilovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kovacevic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kall</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Safranek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kicha</surname>
          </string-name>
          <string-name>
            <surname>Ganapathy</surname>
          </string-name>
          ,
          <article-title>The vocabulary and grammar of color patterns, Image Processing</article-title>
          ,
          <source>IEEE Transactions on 9</source>
          (
          <year>2000</year>
          )
          <fpage>417</fpage>
          -
          <lpage>431</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>H.</given-names>
            <surname>Long</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Leow</surname>
          </string-name>
          ,
          <article-title>A hybrid model for invariant and perceptual texture mapping</article-title>
          ,
          <source>in: Pattern Recognition</source>
          ,
          <year>2002</year>
          . Proceedings. 16th International Conference on, volume
          <volume>1</volume>
          , IEEE,
          <year>2002</year>
          , pp.
          <fpage>135</fpage>
          -
          <lpage>138</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>G.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nishino</surname>
          </string-name>
          ,
          <article-title>Visual material traits: Recognizing per-pixel material context</article-title>
          ,
          <source>in: 2013 IEEE International Conference on Computer Vision Workshops</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>883</fpage>
          -
          <lpage>890</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>L.</given-names>
            <surname>Sharan</surname>
          </string-name>
          , C. Liu,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rosenholtz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. H.</given-names>
            <surname>Adelson</surname>
          </string-name>
          ,
          <article-title>Recognizing materials using perceptually inspired features</article-title>
          ,
          <source>International Journal of Computer Vision</source>
          <volume>103</volume>
          (
          <year>2013</year>
          )
          <fpage>348</fpage>
          -
          <lpage>371</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nishino</surname>
          </string-name>
          ,
          <article-title>Automatically discovering local visual material attributes</article-title>
          ,
          <source>in: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR)</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>3565</fpage>
          -
          <lpage>3573</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>G.</given-names>
            <surname>Schwartz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nishino</surname>
          </string-name>
          ,
          <article-title>Recognizing material properties from images</article-title>
          ,
          <source>IEEE transactions on pattern analysis and machine intelligence</source>
          <volume>42</volume>
          (
          <year>2019</year>
          )
          <fpage>1981</fpage>
          -
          <lpage>1995</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Filip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lukavský</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Děchtěrenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Fleming</surname>
          </string-name>
          ,
          <article-title>Perceptual dimensions of wood materials</article-title>
          ,
          <source>Journal of Vision</source>
          <volume>5</volume>
          (
          <year>2024</year>
          )
          <fpage>12</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Filip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vilimovská</surname>
          </string-name>
          ,
          <article-title>Characterization of wood materials using perception-related image statistics</article-title>
          ,
          <source>Journal of Imaging Science and Technology</source>
          <volume>5</volume>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>9</lpage>
          . doi:
          <volume>10</volume>
          .2352/J.ImagingSci.Technol.
          <year>2023</year>
          .
          <volume>67</volume>
          .5.050408.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hallacy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          , G. Goh,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mishkin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          , et al.,
          <article-title>Learning transferable visual models from natural language supervision</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>8748</fpage>
          -
          <lpage>8763</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>V.</given-names>
            <surname>Deschaintre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Guerrero-Viu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gutierrez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Boubekeur</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Masia</surname>
          </string-name>
          ,
          <article-title>The visual language of fabrics</article-title>
          ,
          <source>ACM Trans. Graph</source>
          .
          <volume>42</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .1145/3592391.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Filip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Dechterenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Schmidt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lukavsky</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vilimovska</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kotera</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. W.</given-names>
            <surname>Fleming</surname>
          </string-name>
          ,
          <article-title>Material ifngerprinting: Identifying and predicting perceptual attributes of material appearance</article-title>
          , arXiv, to appear in Royal Society Open Science (
          <year>2024</year>
          ). doi:https://arxiv.org/abs/2410.13615.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Stas</surname>
          </string-name>
          ,
          <article-title>Mobile application for evaluation of material visual properties, Diploma thesis</article-title>
          , Czech Technical University in Prague, Prague, Czech Republic,
          <year>2025</year>
          . URL: https://dspace.cvut.cz/handle/ 10467/122804, accessed:
          <fpage>2025</fpage>
          -07-17.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>