=Paper= {{Paper |id=Vol-3691/paper8 |storemode=property |title=Validation of the Digital Teaching Competence Questionnaire (COMDID-A) in the Mexican context |pdfUrl=https://ceur-ws.org/Vol-3691/paper8.pdf |volume=Vol-3691 |authors=Oscar Daniel Gómez-Cruz,María Concepción Villatoro-Cruz,Ricardo Miguel Maldonado-Domínguez,Eliana Gallardo-Echenique |dblpUrl=https://dblp.org/rec/conf/cisetc/Gomez-CruzVMG23 }} ==Validation of the Digital Teaching Competence Questionnaire (COMDID-A) in the Mexican context== https://ceur-ws.org/Vol-3691/paper8.pdf
                         Validation of the Digital Teaching Competence
                         Questionnaire (COMDID-A) in the Mexican context
                         Oscar Daniel Gómez-Cruz1, María Concepción Villatoro-Cruz2, Ricardo Miguel
                         Maldonado-Domínguez3 and Eliana Gallardo-Echenique4
                         1 Universidad del País Uninnova, Tuxtla Gutiérrez, Chiapas, 29060, México
                         2 Tecnológico Nacional de México/Instituto Tecnológico de Minatitlán, Minatitlán, Veracruz, 96848, México
                         3 Universidad Autónoma de Chiapas, Tuxtla Gutiérrez, Chiapas, 29050, México
                         4 Universidad Peruana de Ciencias Aplicadas, Lima 15023, Perú.



                                                                Abstract
                                                                We present a validation of the Digital Teaching Competence Questionnaire (COMDID-A), used to
                                                                measure the level of digital competencies of teachers at the Autonomous University of Chiapas. A
                                                                documentary review of the existing instruments to measure the digital competencies of teachers was
                                                                conducted. The COMDID-A instrument was selected due to its focus on the evaluation of teachers’ digital
                                                                competences in the university context, as well as its adaptability to different cultural and linguistic
                                                                contexts. Subsequently, a thorough analysis of the instrument was conducted, based on theoretical
                                                                references and the experience of experts in the area of education. Adjustments and modifications were
                                                                made to the original instrument to adapt it to the Mexican context and improve its relevance and
                                                                reliability. The results obtained indicate that the COMDID-A instrument is dependable and relevant for
                                                                its use in the Mexican context. Quantitative and qualitative analyzes show that the instrument is capable
                                                                of effectively measure digital competencies.

                                                                Keywords
                                                                Cross-cultural adaptation, expert judgment, content validity, digital competence1


                         1. Introduction
                         After the outbreak of the COVID-19 pandemic, the prevailing need to train teachers in digital skills
                         became evident [1]. Although the pandemic took many higher education institutions by surprise,
                         the incorporation of digital technologies into classrooms was already underway, although
                         insufficiently [2]. The emerging use of technologies to mitigate the effects of the global shutdown
                         revealed that, in many cases, the assessments conducted were not adequate to determine the
                         level of intervention required and the areas where the need for training is most critical [3]. This
                         scenario has resulted in numerous trainings which not meet the specific demands related to the
                         incorporation of technologies in the education of students [4]. In addition, the lack of a strategic
                         focus on teacher training has produced an unequal adoption of digital tools, which in turn affects
                         the quality of education [5]. Therefore, it is key not only to identify areas for improvement, but
                         also to develop a teacher training model that is comprehensive, flexible and adapted to the
                         specific context of each educational institution [6].
                            In this sense, training in digital skills must go beyond mere technical instruction; it must
                         incorporate pedagogical elements that allow teachers to effectively apply technologies in their
                         educational practice [7]. This holistic approach will not only improve the quality of teaching, but
                         will also contribute to a more inclusive and equitable education, preparing students for the


                         CISETC 2023: International Congress on Education and Technology in Sciences 2023, December 04–06, 2023,
                         Zacatecas, Mexico
                             oscargomez@uninnova.mx (O. D. Gómez-Cruz); maria.vc@minatitlan.tecnm.mx (M. C. Villatoro-Cruz);
                         ricardommd.rmd@gmail.com (R. M. Maldonado-Domínguez); eliana.gallardo@upc.edu.pe (E. Gallardo-Echenique)
                             https://orcid.org/0000-0002-5991-1306 (O. D. Gómez-Cruz); https://orcid.org/0000-0002-8986-6219 (M. C.
                         Villatoro-Cruz); https://orcid.org/0000-0002-8601-8425 (R. M. Maldonado-Domínguez); https://orcid.org/0000-
                         0002-8524-8595 (E. Gallardo-Echenique)
                                                           © 2023 Copyright for this paper by its authors.
                                                           Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
                                                           CEUR Workshop Proceedings (CEUR-WS.org)
                                CEUR
                                            ht
                                             tp:
                                               //
                                                ceur
                                                   -ws
                                                     .or
                                                       g
                                Works
                                    hop     I
                                            SSN1613-
                                                   0073
                                Pr
                                 oceedi
                                      ngs




CEUR
                  ceur-ws.org
Workshop      ISSN 1613-0073
Proceedings
challenges of the 21st century [8]. To realize this comprehensive approach in teacher training, it
is key to have assessment tools that reflect these complexities [9]. In this regard, there are
different proposals to evaluate the level of digital competence, as well as various reference
models or performance standards [10–15]adopted by some Latin American countries.
    In this context, we have chosen to use the Digital Teaching Competence Questionnaire
(COMDID-A) prepared by Lazaro-Cantabrana et al. [16] to evaluate the self-perception of Spanish
teachers [17]. This selection was based on their multidimensional approach, which encompasses
four key dimensions of teaching [18]. It encourages teacher self-reflection and autonomy,
provides instant feedback, contrasted with other instruments [10–15]. In addition, it has been
adapted in other countries, including Chile [16]. This provides an added value for the Latin
American context [19]; however, its specific adaptability to the Mexican environment had not yet
been evaluated. The purpose is to evaluate digital teaching competencies and identify the
condition of the teachers at the Autonomous University of Chiapas (UNACH). Therefore, it/they
was subjected to various tests such as content, concurrent and construct validity, as well as
idiomatic and linguistic validity. By submitting COMDID-A to this process, we intend to provide it
with greater solidity and validity, to be an suitable tool to evaluate the digital competencies of
UNACH teachers [20,21].

    1.1. Digital competences and COMDID-A

    Most authors define digital competencies as an amalgam of skills, knowledge and skills which
enable individuals to employ digital technologies effectively and ethically. In the academic field,
these competencies are key for teachers to efficiently incorporate digital technologies into their
pedagogy, contributing in this way to raising the educational level [22]. To do this, it is necessary
to promote favorable attitudes toward the use of digital technologies and the ability to adapt to
constantly changing technological innovations.
    In order to assess the breadth and depth of digital competencies, multiple approaches to
measurement have emerged[10–15]. Notably, self-perception stands out as an essential
instrument, since it allows teachers to consciously identify their own level of competence in this
field [11,23,24]. For the study case, and responding to the need to evaluate the self-perception of
digital teaching competence (CDD) of teachers of Universidad Autónoma de Chiapas, COMDID-A
becomes a reference [24]. This instrument is organized around four dimensions: D1. Didactic,
curricular, and methodological approach (6 items); D2. Planning, Organization and
Administration of Digital Technology Resources and Spaces (5 items); D3. Relations, Ethics and
Security (5 items); D4. Personal and Professional Development (6 items). Altogether, the
questionnaire includes 22 items that use a five-point Likert scale to determine different degrees
of CDD (non-initiated, beginner, intermediate, expert, and transformer)[16].
    This questionnaire has been applied in multiple contexts and there are several researches and
publications that address its implementation and validation [16,17,25] specifically, to validate
the factorial structure and validity of the construct. Palau et al. [17] Conducted a core component
analysis to simplify the dataset and identify the four dimensions; however, this was intended for
the European context. For the Latin American environment, there is a study in Chile, which
adapted and applied it through focus group [16]. To address the need for a contextualized
evaluation of CDD and to have versions matching the language and sociocultural characteristics
of the Mexican population, this study focuses on the process of cross-cultural adaptation of the
COMDID-A instrument, developed in Spain.

2. Methodology
This study was conducted at Universidad Autónoma de Chiapas (UNACH) as a result of the need
to evaluate the CDD level, to develop a model of professional training in the institution. Initially,
a documentary review was conducted to identify the most appropriate instruments to measure
these competencies. Within this framework, COMDID-A has been chosen as an essential
evaluative instrument; developed by specialists in the field of educational technology at the
Universitat Rovira i Virgili in Tarragona, Spain [16]. This instrument stands out for its ability to
measure teachers’ self-perception in four key areas through 22 descriptors and four levels of
development, and is especially applicable for self-assessments in academic contexts [17]. The
goal is to evaluate digital skills to determine the current level among UNACH teachers.

                                                                         Valida&on Stages

                    Stage 1: Transcultural Adapta%on                                               Stage 2: Content Validity


           Opinion of 3                 Linguis4c and                                                                Digitaliza4on of Results and
                                          Idioma4c.                                   Implementa4on of prac4cal
        Experts in the Area                                                            guide adapted by Escobar-     Website for Distribu4on of
                                         Qualita4ve                                                                           Instrument
                                                                                      Pérez and Cuervo-MarHnez


        Iden4fica4on of                                          Terminology                                            Selec4on of 16 Experts
      confusing ques4ons.               Review of                                                                           and Instrument
                                                                Consistency and             Adding Observa-on
            Propose                    Assessment                                                                          Implementa4on
                                                                   Accuracy                      Sec-on
          adapta4ons

                                                                                                                         Assessment of Judge
                                 Organiza4on and                   Integra4on                                                Consistency
                              Classifica4on according                with cross
                                    to COMDID                         table
                                                                                          Integra4on in Matrix-
                                                                                                                            Data collec4on.
                                                                                            Summary for each
                                                                                                                          Quan4ta4ve Validity
                                                                                        Dimension. Screening for
                                  Integrate results in Visual                                                              using Fleiss’ Kappa
                                                                                             Item Adapta4on
                                       Format (table)



                                                                Iden%fy COMDID Items for Change


Figure 1: Validation phases

    The choice of this instrument was based on its multidimensional approach, which covers four
key aspects of teaching [26]. In addition, its adaptation in other countries, including Chile [16],
adds additional value for its use in Latin American contexts [19]. However, its adaptability to the
Mexican environment has not yet been applied. Since the COMDID-A instrument originated in a
European context, its adaptation to the Mexican environment was imperative, culturally as well
as linguistically. To achieve this, the study was divided into two interconnected phases (see
Figure 1). The first stage focused on the adaptation and validation of the instrument, using the
COMDID-A rubric as proposed by Lázaro-Cantabrana et al. (2018). The adaptation of the
questionnaire to the Mexican environment was key. Two validation phases were conducted: The
first focused on cross-cultural adaptation (linguistic, cultural, and idiomatic) which involved the
participation of three experts in the educational and linguistic field. The second consisted of
content validity through the judgment of 16 experts, using the proposal of Escobar-Perez &
Cuervo-Martinez [27]. This process made it possible to identify the dimensions that required
contextual adaptation, and the results were verified by Fleiss K.

   2.1. Phase 1 Cross-cultural adaptation

    Ensuring the validity of an instrument is a constant concern among researchers. Over time,
validity has been interpreted in various ways and from different fields of study [28,29]. However,
it remains crucial for the choice and use of an instrument. Specifically, the question is whether
the instrument really evaluates what it intends to measure [30]. To achieve an adequate linguistic,
linguistic, and cultural equivalence in an instrument, it needs to be used in different cultural
contexts. There are theoretical proposals that serve as a guide and highlight the importance of
the transition from an instrument from one culture to another [30–32]. This transition is called
transcultural adaptation of the instrument [33]. While in Europe and in English-speaking
countries this process is widely valued, in Latin America it is sometimes not given the necessary
importance[30]. The lack of adequate procedures for translating and adapting the instruments
has led to some research being considered invalid. For COMDID-A to adapt properly to the
Mexican context, it was essential to have the opinion and experience of experts in the linguistic
and educational field [34]. It was decided to consult three recognized UNACH teachers with wide
experience in areas relevant to the study. Table 1 shows a brief description of each expert’s
experience and specialty:

Table 1
Participating experts
 Expert         Experience
 E01            Ph.D. in Philosophy and Educational Sciences
                Universidad Complutense De Madrid
                Professor of the Faculty of Humanities Campus VI, Universidad Autónoma de
                Chiapas. Expert teacher in Psychopedagogy and development of qualitative
                instruments.
 E02            PhD. in Contemporary Philosophy
                Benemérita Universidad Autónoma de Puebla (BUAP), México. Professor of the
                Faculty of Humanities, Campus VI, Universidad Autónoma de Chiapas. 16 years in
                teaching, expert in the phenomenology of education.
 E03            PhD. D. in Education
                Campus Tuxtla, Universidad Autónoma de Chiapas.
                32 years in teaching, expert in Linguistics and Languages. Faculty member of the
                School of Language.

   An e-mail was sent to each of the specialists with a formal invitation to participate. The
COMDID-A instrument was attached to this e-mail requesting its analysis and possible proposals
for modification. The guidelines provided required that they focused on the linguistic and
idiomatic parts based on the following criteria: The first focused on identifying questions that
could generate confusion in the Mexican context, and the second on proposing adjustments for
them. After receiving their assessments, a meticulous review was conducted to discern which
modifications to incorporate and how to do so, ensuring the relevance and clarity of the
instrument in Mexico. The contribution of these experts not only strengthens the linguistic
validity of COMDID-A, but also guarantees its cultural adaptation and specificity in the Mexican
educational context. The reflections of each expert were integrated into a cross table. It is
important to mention that, although observations were scarce, they were essential. For example,
it was perceived that experts agreed to observe the same word twice, which could produce
misinterpretations in the context where the instrument would be applied.
   The proposal of Beaton et al. has been taken as a reference. [35] They established phases to
ensure that the adapted questionnaire is conceptually, idiomatically, semantically, and
operationally equivalent to the original. Having a Spanish version helped to ensure that the
translation process did not have any major problems; however, there were still discrepancies in
the language of the culture itself; therefore, the contribution of the experts helped to confront and
compare the versions to identify and resolve discrepancies. Consistency and precision were
sought in the terminology used, with respect to the original content of COMDID-A [35,36].
   To ensure fidelity in both language and concepts, a stage of adaptation to Mexican Spanish was
implemented through the judgment of experts who did not have prior knowledge of the
instrument in its original version. This made it possible to detect and correct possible deviations
in interpretation or conceptual meaning in the consolidated version of the instrument. The
participation of these experts not only reinforces the linguistic validity of COMDID-A, but also
ensures its adaptation to the culture and particularities of the educational field in Mexico.

    2.2. Phase 2: Content Validity
    Every instrument must go through a validation process to ensure that it is valid, reliable and
that accurately measures what is proposed [37,38]. A validated instrument allows the
generalization of findings [37,38], improves the quality of the study, not only increases the
credibility of the study, but also facilitates efficient data collection and lays the foundation for
future studies [39]. To carry out the second phase of the investigation, we adapted the proposal
of Escobar-Perez and Cuervo-Martinez[27] , who establish a method for the judgment of experts
and that through a practical guide carries out the process of expert judgment that includes:
     1. Prepare instructions and spreadsheets
     2. Select the experts and train them
     3. Explain the context
     4. Enable discussion
     5. Establish agreement between experts by calculating consistency.
    These steps are recommended by several authors [40,41], and are considered essential for
conducting an expert judgment effectively. A space for general observations was added to this
guide; this space ensured that the instrument was applicable to the given Mexican context. This
enriched what was already established in Phase 1. Subsequently, the instrument was digitized
using the LimeSurvey, software which helped its distribution and the data collection. This process
ended with the creation of the www.competenciadigital.com.mx website, which served as a
platform for managing the tool and the database.
    The next step was the thorough selection of expert judges, an essential component to ensure
the validity and reliability of the study. Following the best practices in the selection of judges
methodologically, it was decided to invite specialists with a solid academic background and a vast
experience in the field of educational technology [27,40,42,43]. In total, 16 experts from various
states of the Mexican Republic were added (see Table 2). The inclusion of experts from different
geographical areas and universities allowed for a more complete evaluation, addressing various
aspects of the subject in question. This selection strategy was based on rigorous methodologies
previously established by various authors [27,40,42,43], thus ensuring that the process was
aligned with high quality academic standards. This holistic approach not only strengthened the
validity of the study, but also set a precedent for future research in the field. Each expert was
contacted by email, being mostly members of the Inter-Institutional Committee for the Evaluation
of Higher Education (CIEES) with a specialty in Educational Technology. In the mail, a link to the
digitized instrument was provided, accompanied by a detailed protocol, the theory supporting
the instrument, a specific timeframe to complete the task, as well as clear definitions of the
evaluation criteria.

Table 2
Expert judges according to their university and home state
 State                            University                                             Amount
 Guerrero                         Universidad Autónoma de Gerrero                        1
 Ciudad de México                 UNAM/UNITEC                                            1
 Veracruz                         Universidad Veracruzana                                1
 Tamaulipas                       Regional Center for Teacher Training and Educational   1
                                  Research
                                  Universidad Autónoma de Tamaulipas                     1
 Sonora                           Universidad de Sonora                                  1
 Chihuahua                        Universidad Tecnológica de Ciudad Juárez               1
 Chiapas                          Universidad Autónoma de Chiapas                        6
                                  Universidad Pablo Guardado Chávez                      2
                                  Universidad del País INNOVA                            1

   After the end of the evaluation period, which lasted five months from the first contact, the data
obtained was thoroughly collected. To guarantee the quantitative validity of the COMDID-A
instrument, a mathematical analysis process of the collected data was implemented, and in this
framework, the Fleiss K coefficient emerged as a crucial statistical indicator. Fleiss K is a metric
that evaluates the degree of agreement between multiple judges or evaluators [44–46] . It is used
specifically to measure the consistency of the classifications awarded by different judges to the
same subjects [47,48] . A high value of the Fleiss K coefficient indicates a higher concordance
among the judges, which, in turn, reinforces the reliability of the instrument in question [46].
   To carry out this comprehensive quantitative analysis, a team of mathematics experts was
formed. This team was led by an academic recognized in the National System of Researchers
(SNI), level II and an advanced student of the Bachelor of Mathematics. The analysis focused on
the evaluation of four key dimensions: Didactics, Curriculum and Methodology; Planning,
Organization and Management of Digital Technological Spaces and Resources; Relational, Ethics
and Security; and Personal and Professional. Each dimension was examined under four aspects:
Clarity, sufficiency, coherence, and relevance. Each dimension-aspect pair included between 5
and 7 questions whose answers could be: High level, Moderate level, Low level or Does not meet
the criteria.
   This evaluation process gives rise to 16 matrices, which contain all the evaluations made by
the experts. The results obtained in the first matrix are presented below.

Table 3
Assessments made by experts
 Aspect Competence
 Assessm
 ent        G2Q00001[S    G2Q00001[S     G2Q00001[S     G2Q00001[S      G2Q00001[S     G2Q00001[S
 Questio    Q005]         Q006]          Q007]          Q008]           Q009]          Q010]
 ns
 Judge ID   Level
 4          2. Low        3. Moderate    3. Moderate    3. Moderate     3. Moderate    4. High
 5          3. Moderate   3. Moderate    3. Moderate    3. Moderate     4. High        4. High
 12         4. High       4. High        4. High        4. High         4. High        4. High
 13         3. Moderate   4. High        4. High        4. High         4. High        3. Moderate
 17         3. Moderate   4. High        3. Moderate    4. High         4. High        3. Moderate
 18         4. High       4. High        4. High        4. High         4. High        2. Low
 20         4. High       4. High        4. High        2. Low          3. Moderate    4. High
 22         2. Low        3. Moderate    3. Moderate    3. Moderate     3. Moderate    2. Low
 27         3. Moderate   4. High        4. High        2. Low          4. High        3. Moderate
 31         3. Moderate   3. Moderate    3. Moderate    3. Moderate     3. Moderate    2. Low
 32         3. Moderate   3. Moderate    3. Moderate    3. Moderate     4. High        3. Moderate
 38         4. High       4. High        4. High        4. High         4. High        3. Moderate
 40         4. High       4. High        3. Moderate    3. Moderate     3. Moderate    3. Moderate
 42         4. High       4. High        4. High        4. High         4. High        4. High
 44         4. High       3. Moderate    4. High        4. High         4. High        3. Moderate
 45         2. Low        2. Low         2. Low         4. High         4. High        4. High

   The objective was to assess the level of concordance in the evaluations of the 16 judges. For
this purpose, two statistical coefficients were used: Kendall’s coefficient W of and Fleiss Kappa
coefficient. Considering that Kendall’s W was designed for ordinal trials, its adjustment was used
for repeated trials; however, the results were mostly non-significant, which led to its dismissal.
   The Fleiss Kappa coefficient was identified as the most suitable option for this study, especially
since the collected valuations are presented in nominal form. Two hypotheses were formulated:
The null hypothesis ($H_0$), which holds that there is no significant real agreement beyond
chance, and the alternative hypothesis ($H_1$), which states that the observed agreement is
statistically significant. The p-value was used to evaluate the evidence against the null hypothesis
and to determine the statistical significance of the agreement observed among the judges. In this
way, a summary matrix was constructed using the calculation of the Fleiss Kappa for each
dimension-aspect pair, providing an integral view of the results obtained.

Table 4
Summary matrix of the results of the Fliess Kappa for each dimension-aspect
 Dimension                     Clarity           Competence      Coherence                 Relevance
                                     p valor   K       p valor   K       p valor   K       p valor K
 D1. Didactic, curricular, and       0.000     0.227   0.000     0.305   0.000     0.223   0.000    0.352
 methodological
 D2. Planning, organization and      0.000     0.446   0.000     0.433   0.008     0.229   0.000   0.563
 management        of      digital
 technological    spaces     and
 resources
 D3. Relational, ethics and          0.000     0.425   0.633     0.032   0.000     0.352   0.000   0.304
 security
 D4. Personal and professional       0.000     0.446   0.000     0.289   0.000     0.269   0.000   0.671

    It is important to note that, with one exception, the results obtained are statistically significant,
since the corresponding p-values do not exceed the level of significance established in 0.05. To
better understand the level of agreement between the judges, we resort to the interpretation of
the Fleiss Kappa proposed by Altman (1991). Noting that the median of \( K \) values is 0,352,
Altman suggests that this reflects a generally weak level of agreement among judges. In other
words, there is a certain discrepancy in the evaluations conducted by the different judges. To
illustrate the level of agreement, we created Table 5 which we have called the "Frequency Table
of Agreement Levels according to the Fleiss Kappa":

Table 5
Frequency Table of Agreement Levels according to Fleiss Kappa
                          Agreement Level            Frequency
                          Poor                       0
                          Weak                       9
                          Moderate                   5
                          Good                       1
                          Very Good                  0

   This table allows for a quick and effective visualization of how agreement levels are
distributed among judges. For example, it can be observed that most judges (9 out of 16) have a
"Weak" level of agreement, while only one reaches a "Good" level of agreement. This highlights
the need to review and possibly adjust the evaluation tool to improve consistency among judges.
Finally, after this rigorous process of evaluation and analysis, the final measurements were
conducted, and the corresponding results were obtained. These results will serve as a basis for
future research and methodological adjustments.

3. Results and Discussion
The research was able to adapt and validate the COMDID-A instrument for its application in
Mexico. The results show that the instrument is dependable and relevant to evaluate the digital
competencies of teachers at Universidad Autónoma de Chiapas. The method proposed by
Escobar-Pérez and Cuervo-Martinez was used to validate and adapt COMDID-A, [27],A comment
section was added to collect specific observations from the experts. This approach made it
possible to obtain qualitative data that enriched the qualitative phase of the research. The
comments of the experts were classified according to COMDID-A items and dimensions, which
helped to identify patterns and relevant coincidences in specific items.
Table 6
Table Sample items for change
 New       New             COMDID          Item with observation           Adapted Item
 category subcategory Subcategory
 Meaning N/A               1.3             1.2 Processing of information 1.3 Management, analysis
                                           and creation of knowledge      of information and creation
                                                                          of knowledge
                            1.4 (item 1)   Use digital technologies to Use digital technologies to
                                           increase motivation and increase motivation and
                                           facilitate    learning     for facilitate    learning    for
                                           students with NEE              students with NEI
 Structure Abbreviation 1.1 (item 1)       Design EA activities which Design teaching-learning
                                           involve the use of digital (EA) activities involving the
                                           technologies                   use of digital technologies
            Grammar         3.1 (item 3)   Serve as a model for other Be a reference for other
                                           professionals      on     the professionals       on    the
                                           responsible and safe use of responsible and safe use of
                                           digital technologies           digital technologies

   Categories such as "Meaning" and "Structure" were established to organize the data. Within
"Structure," subcategories such as "Abbreviation" and "Grammar” were created. Items that
required changes were moved to "Change Formats," placing them in the corresponding categories
and subcategories. This meticulous process allowed to have a visual map of the items to be
modified. A relevant change was the adaptation of the term "Special Educational Needs (NEE, for
the Spanish acronyms)" to "Inclusive Education Needs (IND, for the Spanish acronyms)" for the
Mexican context. See table 4. This detailed approach ensured that the COMDID-A instrument was
well-founded and adapted to the Mexican context, ensuring its validity and reliability. A
concordance was observed between the qualitative responses and the dimensions evaluated. In
addition, tables were developed to improve the understanding of the changes made, highlighting
the adaptation of terms and parameters to measure teaching strategies.
   The qualitative results meet the objective of adapting and validating the instrument, since final
actions were determined for item changes based on qualitative and quantitative analysis (Fleiss
Kappa). The next task is to manage updating the instrument for its application in Mexico. See
Figure 2 showing the visual overview of the instrument items that are proposed for changes.
Figure 2: Visual overview of items proposed in return

4. Conclusions
The research focused on evaluating the digital competencies of teachers at Universidad
Autónoma de Chiapas. It began with a documentary review to identify appropriate instruments
to measure these competencies. The COMDID-A instrument developed by Lazaro-Cantabrana et
al. [16] in Spain was considered as the most appropriate; therefore, a cultural and linguistic
adaptation was required to be implemented in Mexico. To validate the instrument in the Mexican
context, two types of validations were conducted: Idiomatic and linguistic, and quantitative
through expert judgment. In the idiomatic validation, three experts were consulted to adapt the
instrument to the linguistic conditions of Mexico. For content validity, 16 experts in educational
technology were resorted to, and statistical methods such as Kendall's W coefficient and Fleiss's
Kappa coefficient were applied.
    It is essential to understand that the validation of an instrument is not an isolated process but
must consider its applicability in a specific context. This thorough approach ensures that the
instrument is both applicable and dependable in the Mexican context. The research not only seeks
to evaluate the digital competencies of teachers, but also to contribute to the body of knowledge
in the field of educational technology and teacher training in Mexico. This Mexican version of
COMDID-A can be considered equivalent to the original; it is linguistically, semantically, and
culturally adapted to the Mexican context. The authors are aware that any validation process is a
process that requires testing other types of validity to ensure the validity of the instrument.
    This research lays the foundations for future studies and the implementation of teacher
training strategies in digital competencies, aligned with the needs and context of the UNACH and
potentially applicable in other educational institutions in the country. That is to say, this method
improves efficiency in data collection and serves as a basis for future research, highlighting the
importance of its application in the various fields of study that should not be ignored.
Acknowledgements
The authors thank the experts who participated voluntarily and anonymously in this study. This
study was partially funded by the Research Direction of the Universidad Peruana de Ciencias
Aplicadas (UPC).

References
[1] Molina Montalvo HI, Macías Villareal JC, CEpeda Hernández AA. Educación en tiempos de
     COVID-19: Una aproximación a la realidad en México: experiencias y aportaciones.
     Comunicaci. Ciudad de México: 2022.
[2] Fernández Escárzaga J, Gabriela J, Varela D, Lorena P, Martínez M. De la educación presencial
     a la educación a distancia en época de pandemia por Covid 19. Experiencias de los docentes.
     Rev Electrónica Sobre Cuerpos Académicos y Grup Investig 2020;7:87–110.
[3] Cruz-Aguayo Y, Hincapé D, Rodríguez C. Profesores a prueba: Claves para una evaluación
     docente exitosa. 2020. https://doi.org/10.18235/0002149.
[4] Guerrero I, Kalman J. La inserción de la tecnología en el aula: Estabilidad y procesos
     instituyentes en la práctica docente. Rev Bras Educ 2010;15:213–29.
     https://doi.org/10.1590/S1413-24782010000200002.
[5] Mendoza R, Bellodas M, Ortiz C, Puelles L, Asnate E, Zambrano J. Desafíos interdisciplinarios
     para los docentes en el aprendizaje virtual. 2023.
[6] Balladares-Burgos J, Valverde-Berrocoso J. El modelo tecnopedagógico TPACK y su
     incidencia en la formación docente: una revisión de la literatura. RECIE Rev Caribeña Investig
     Educ 2022;6:63–72. https://doi.org/10.32541/recie.2022.v6i1.pp63-72.
[7] Díaz Chamorro CM. El Modelo Tpack Como Método Pedagógico Para El Desarrollo De
     Competencias Digitales En Los Docentes De La Unidad Educativa “Víctor Mideros.” 2023.
[8] Esquerre Ramos LA, Pérez Azahuanche MÁ. Retos del desempeño docente en el siglo XXI: una
     visión          del        caso       peruano.          Rev         Educ        2021;45:0–21.
     https://doi.org/10.15517/revedu.v45i1.43846.
[9] Castañeda L, Esteve F, Adell J. Why rethinking teaching competence for the digital world? Rev
     Educ a Distancia 2018:1–20. https://doi.org/10.6018/red/56/6.
[10] Agreda M, Hinojo MA, Sola JM. Diseño y validación de un instrumento de evaluación de
     competencia digital docente. Pixel-Bit, Rev Medios y Educ 2016:39–46.
[11] Ferrari A. DIGCOMP : A Framework for Developing and Understanding Digital Competence
     in Europe. Luxembourg: 2013. https://doi.org/10.2788/52966.
[12] Reixach E, Andrés E, Ribes JS, Gea-Sánchez M, López AÀ, Cruañas B, et al. Measuring the
     Digital Skills of Catalan Health Care Professionals as a Key Step Toward a Strategic Training
     Plan: Digital Competence Test Validation Study. J Med Internet Res 2022;24.
     https://doi.org/10.2196/38347.
[13] Restrepo-Palacio S, de María Segovia Cifuentes Y. Design and validation of an instrument for
     the evaluation of digital competence in Higher Education. Ensaio 2020;28:932–61.
     https://doi.org/10.1590/S0104-40362020002801877.
[14] Zempoalteca Durán B, Barragán López JF, González Martínez J, Guzmán Flores T. Teaching
     training in ICT and digital competences in Higher Education System. Apertura 2017;9:80–96.
     https://doi.org/10.32870/ap.v9n1.922.
[15] Zubieta J, Bautista T, Quijano A. Aceptació n de las TIC en la docencia : una tipologı́a de los
     acadé micos de la UNAM. 2012.
[16] Lázaro-Cantabrana JL, Gisbert-Cervera M, Silva-Quiroz JE. Una rúbrica para evaluar la
     competencia digital del profesor universitario en el contexto latinoamericano. Edutec Rev
     Electrónica Tecnol Educ 2018:1–14. https://doi.org/10.21556/edutec.2018.63.1091.
[17] Palau R, Usart M, Ucar Carnicero MJ. La competencia digital de los docentes de los
     conservatorios. Estudio de autopercepción en España. Rev Electron LEEME 2019:24–41.
     https://doi.org/10.7203/LEEME.44.15709.
[18] Lazáro Cantabrana JL, Gisbert Cervera M. Elaboración de una rúbrica para evaluar la
     competencia digital del docente. Rev Ciéncies LÉducació 2015:19.
[19] Cisneros-Barahona AS, Marqués-Molias L, Samaniego-Erazo N, Mejía-Granizo CM. La
     Competencia Digital Docente. Diseño y validación de una propuesta formativa. Pixel-Bit2
     2023;68:7–41.
[20] Creswell JW. Research design: Qualitative, quantitative, and mixed methods approaches. 4th
     ed. Thousands Oaks, CA: SAGE Publications, Inc.; 2014.
[21] Sireci SG. The construct of content validity. Soc Indic Res 1998;45:83–117.
     https://doi.org/10.1023/a:1006985528729.
[22] Paz Saavedra LE, Gisbert Cervera M, Usart Rodríguez M. Competencia digital docente, actitud
     y uso de tecnologías digitales por parte de profesores universitarios. Pixel-Bit, Rev Medios y
     Educ 2022:91–130. https://doi.org/10.12795/pixelbit.91652.
[23] Janssen J, Stoyanov S. Online Consultation on Experts "Views on Digital Competence 2012.
[24] Lázaro Cantabrana JL, Gisbert Cervera M. Elaboració d’una rúbrica per avaluar la
     competència digital del docent. Univ Tarraconensis Rev Ciències l’Educació 2015;1:48.
     https://doi.org/10.17345/ute.2015.1.648.
[25] Silva J, Usart M, Lázaro-Cantabrana J-L. Competencia digital docente en estudiantes de último
     año de Pedagogía de Chile y Uruguay TT - Teacher’s digital competence among final year
     Pedagogy students in Chile and Uruguay. Comunicar 2019;61:33–43.
[26] Lázaro Cantabrana JL. La competència digital docent com a eina per garantir la qualitat en
     l’ús de les TIC en un centre escolar. vol. 1. 2015. https://doi.org/10.17345/ute.2015.1.667.
[27] Escobar-Pérez J, Cuervo-Martínez Á. Validez de contenido y juicio de expertos: Una
     aproximación a su utilización [Content validity and expert judgement: An approach to their
     use]. Av En Medición 2008;6:27–36.
[28] Aiken LR, Yang W, Soto M, Segovia L, Binomial P, Miller JM, et al. Diseño y validación de un
     cuestionario para analizar la calidad en empleados de servicios deportivos públicos de las
     mancomunidades de municipios extremeñas. Educ Psychol Meas 2011;7:181–92.
     https://doi.org/10.1177/0013164412473825.
[29] Cho J. Validity in qualitative research revisited. Qual Res 2006;6:319–40.
     https://doi.org/10.1177/1468794106065006.
[30] Gallardo-Echenique E, Marqués Molias L, Gomez Cruz OD, De Lira Cruz R. Cross-cultural
     adaptation and validation of the “student communication & study habits” questionnaire to
     the mexican context. Proc - 14th Lat Am Conf Learn Technol LACLO 2019 2019:104–9.
     https://doi.org/10.1109/LACLO49268.2019.00027.
[31] Arribas A. Adaptación Transcultural de Instrumentos. Guía para el Proceso de Validación de
     Instrumentos Tipo Encuestas. Rev Científica La Asoc Médica Bahía Blanca 2006;16:74–82.
[32] Lira MT, Caballero E. Cross-Cultural Adaptation of Evaluation Instruments in Health: History
     and Reflections of Why, How and When. Rev Medica Clin Las Condes 2020;31:85–94.
     https://doi.org/10.1016/j.rmclc.2019.08.003.
[33] International Test Commission (ITC). ITC Guidelines for Translating and Adapting Tests. 2nd
     ed. [Www.InTestCom.Org]: ITC; 2016.
[34] Cardoso Ribeiro C, Gómez-Conesa A, Hidalgo Montesinos MD. Metodología para la
     adaptación      de     instrumentos     de      evaluación.   Fisioterapia   2010;32:264–70.
     https://doi.org/10.1016/j.ft.2010.05.001.
[35] Beaton D, Bombardier C, Guillemin F, Ferraz MB. Recommendations for the Cross-Cultural
     Adaptation of the DASH & QuickDASH Outcome Measures. Toronto: 2007.
[36] Beaton D, Bombardier C, Guillemin F, Ferraz MB. Guidelines for the process of cross-cultural
     adaptation of self-report measures. Spine (Phila Pa 1976) 2000;25:3186–91.
     https://doi.org/10.1097/00007632-200012150-00014.
[37] Sireci SG. The Construct of Content Validity. Soc Indic Res 1998;45:83–117.
     https://doi.org/10.1007/sl.
[38] Almanasreh E, Moles R, Chen TF. Evaluation of methods used for estimating content validity.
     Res Soc Adm Pharm 2019;15:214–21. https://doi.org/10.1016/J.SAPHARM.2018.03.066.
[39] Hambleton RK, Patsula L. Adapting tests for use in multiple languages and cultures. Soc Indic
     Res 1998;45:153–71. https://doi.org/10.1023/A:1006941729637.
[40] Skjong R, Wentworth BH. Expert Judgment and Risk Perception. Proc. Elev. Int. Offshore
     Polar Eng. Conf., vol. IV, Stavanger, Norway: International Society of Offshore and Polar
     Engineers; 2001, p. 537–44.
[41] de Arquer MI. Fiabilidad Humana: métodos de cuantificación, juicio de expertos. 1995.
[42] Cabero J, Llorente M del C. La aplicación del juicio de experto como técnica de evaluación de
     las tecnologías de la información y comunicación (TIC) [The expert’s judgment application
     as a technic evaluate information and communication technology (ICT)]. Eduweb Rev Tecnol
     Inf y Comun En Educ 2013;7:11–22.
[43] Urrutia Egaña M, Barrios Araya S, Gutiérrez Núñez M, Mayorga Camus M. Métodos óptimos
     para determinar validez de contenido. Rev Cuba Educ Medica Super 2015;28:547–58.
[44] Cerda Lorca J, Villarroel Del P. L. Evaluación de la concordancia inter-observador en
     investigación pediátrica: Coeficiente de Kappa. Rev Chil Pediatr 2008;79:54–8.
     https://doi.org/10.4067/s0370-41062008000100008.
[45] Torres J, Perera V. Cálculo de la fiabilidad y concordancia entre codificadores de un sistema
     de categorías para el estudio del foro online en e-learning. Rev Investig Educ 2009;27:89–
     103.
[46] Falotico R, Quatto P. Fleiss’ kappa statistic without paradoxes. Qual Quant 2015;49:463–70.
     https://doi.org/10.1007/s11135-014-0003-1.
[47] López A, Galparsoro DU, Fernández P. Medidas de concordancia : el índice de Kappa. Cad Aten
     Primaria 2001:2–6.
[48] Gwet KL. Large-Sample Variance of Fleiss Generalized Kappa. Educ Psychol Meas
     2021;81:781–90. https://doi.org/10.1177/0013164420973080.
[49] Altman DG. Practical Statistics for Medical Research. 1991.