<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>A. Liu);</journal-title>
      </journal-title-group>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>for Instructional Decision Support⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alex Liu</string-name>
          <email>alexliux@uw.edu</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shawon Sarkar</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lief Esbenshade</string-name>
          <email>lief@uw.edu</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victor Tian</string-name>
          <email>ztian27@uw.edu</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kevin He</string-name>
          <email>kevin@colleague.ai</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zachary Zhang</string-name>
          <email>zac@colleague.ai</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Min Sun</string-name>
          <email>misun@uw.edu</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hensun Innovation</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>This study introduces a practice-informed AI nudging system designed to support responsible, context-aware instructional decision-making during lesson planning. Grounded in a hybrid intelligence framework, the system integrates multiple sources of human insight: individual teacher preferences, collective behavior patterns, and pedagogical research, with computational support from large language models to generate timely and explainable nudges. A co-occurrence network of instructional strategies, derived from annotated teacher-AI interactions, guides contextual nudge generation using behavioral heuristics. Initial analysis shows strong alignment between system recommendations and teacher' pedagogical goals and values, as well as increased diversity in efective instructional practices. By balancing professional autonomy with evidence-based guidance, this work advances responsible AI in education, contributing an explainable, value-sensitive system that fosters sustainable pedagogical improvement in real-world settings.</p>
      </abstract>
      <kwd-group>
        <kwd>design</kwd>
        <kwd>Digital nudging</kwd>
        <kwd>hybrid intelligence</kwd>
        <kwd>context-aware decision support</kwd>
        <kwd>teacher-AI interaction</kwd>
        <kwd>value-sensitive</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Lesson planning is a central activity through which teachers translate pedagogical and content
knowledge into classroom instruction [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. This process involves sequencing and adapting activities to
engage diverse learners, drawing on both expertise and instructional judgment [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. More than routine
preparation, it serves as embedded professional learning, requiring integration of content, pedagogy,
and learner needs [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Without suficient support, this cognitively demanding task may yield
suboptimal instruction, highlighting the need for real-time, context-aware decision support. Traditional
professional development often falls short: it is dificult to access, insuficiently personalized, and
disconnected from day-to-day planning [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Even when training is available, instructional planning
remains time-intensive and complex [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Recent advances in artificial intelligence (AI) present an
opportunity to address these gaps by supporting teachers’ instructional decision-making. However,
without thoughtful design, AI-generated recommendations risk falling short in both efectiveness and
trustworthiness [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Nudging, rooted in behavioral science, refers to subtle prompts that influence decisions without
limiting choice [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. In education, teachers, like all individuals, are subject to cognitive constraints
from limited time, information, and energy [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ]. Nudges can reduce instructional inertia by gently
steering teachers toward better practices while preserving autonomy [11]. When embedded in AI
systems, nudges can be dynamically tailored to a teacher’s context, supporting decisions around content,
pedagogy, and classroom management [12]. Their potential lies in leveraging real-time data to generate
      </p>
      <p>CEUR</p>
      <p>ceur-ws.org
timely, personalized prompts [13], making decision support itself a form of embedded professional
development.</p>
      <p>While research on AI-powered nudging in instructional preparation is limited, studies in classroom
orchestration show that real-time feedback can enhance instructional flow and student engagement
[14, 15]. These findings underscore nudging’s potential in complex, high-stakes environments. However,
teachers vary in experience, pedagogical philosophy, and technological fluency, requiring systems that
not only model behavior but also interpret goals and preferences. Ignoring these diferences risks
eroding trust and reinforcing a one-size-fits-all approach.</p>
      <p>Equally important are the ethical stakes. Without transparency and alignment with teacher values,
nudges risk being perceived as manipulative, raising concerns about agency and power imbalances in
AI-mediated decision-making [16, 17, 13]. If teachers feel their expertise is undermined, they may resist
adoption, especially in autonomy-centered environments [18]. To be trusted, nudging systems must
preserve professional judgment, communicate their rationale clearly, and act as supportive partners
rather than prescriptive agents.</p>
      <p>This underscores the role of human-centered design (HCD) in building AI that complements rather
than constrains teacher decision-making. HCD ensures systems are not only technically efective, but
also aligned with user values, workflows, and autonomy [ 19]. In education, where complex judgment is
central, AI systems must respect expertise and avoid overriding intent [20]. Nudges must be transparent
and explainable [21], and personalization must go beyond surface tailoring to align with evolving
pedagogical needs, classroom contexts, and teacher goals [22].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Design Tensions: Adoption vs. Efectiveness</title>
      <p>A central challenge in designing AI-powered nudges for educators lies in balancing two potentially
competing objectives: alignment with existing practices and values versus introduction of efective but
unfamiliar strategies. On one hand, nudges must resonate with teachers’ current routines and pedagogical
orientations to maintain trust, foster engagement, and uphold professional autonomy. On the other
hand, to meaningfully support instructional improvement, nudges must also surface research-backed
practices that teachers may not typically consider—strategies that go beyond existing habits while
remaining contextually appropriate.</p>
      <p>To address the tension between respecting teacher autonomy and promoting instructional
improvement, we propose an AI-powered nudging system grounded in a hybrid intelligence framework.
This approach integrates four complementary sources of insight: (1) individual teacher preferences
captured through usage patterns, (2) collective behavioral trends derived from educator interactions
on the platform, (3) pedagogical best practices from instructional research, and (4) the reasoning and
generative capabilities of LLMs. By integrating multiple sources of human and LLM contributions,
the system can enhance the contextual relevance and pedagogical validity of its recommendations,
encouraging sustained instructional growth without enforcing prescriptive behavior change.</p>
      <p>To evaluate this system, we investigate the following research questions:
• RQ1: To what extent do the system-generated nudges align with the pedagogical goals, values,
and instructional intent expressed by teachers during interaction with the platform?
• RQ2: Do nudges lead teachers to adopt a more diverse set of high-leverage instructional practices
than they would typically select on their own?
• RQ3: Does the hybrid intelligent framework afect the explainability, contextual alignment, and
adoption of nudges across varied educational contexts?</p>
    </sec>
    <sec id="sec-3">
      <title>3. 3. Hybrid Intelligence Framework</title>
      <sec id="sec-3-1">
        <title>3.1. Data Source</title>
        <p>Our analysis draws on de-identified chat history logs from an AI-powered instructional assistant
used by K-12 educators across grade levels, subject areas, and various instructional contexts. These
interactions reflect authentic teacher goals, instructional challenges, and pedagogical reasoning during
lesson planning.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Identification and Qualitative Annotation</title>
        <p>We conducted an extensive literature review on efective classroom instructional approaches. Drawing
from established research in teacher education and instructional design frameworks, we developed
a codebook of high-leverage instructional strategies, including opportunities for student discourse,
scafolding techniques, the use of meaningful real-world examples, and formative assessment [ 23].</p>
        <p>Using this codebook, we applied automated qualitative coding to teacher-AI dialogues to identify
instances of these instructional strategies, along with contextual information such as subject area,
educational level, and learner needs (e.g., English language learners, students with special education needs,
neurodiverse students). Following automated coding by an LLM, experienced qualitative researchers
manually verified and refined the results to ensure accuracy and interpretive validity. This process
surfaced both explicit instructional moves and implicit pedagogical values, preferences, and intentions
embedded within teacher queries and interactions.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Practice Network Construction</title>
        <p>From the annotated data, we constructed a directed co-occurrence network that models how
instructional practices tend to be sequenced across similar user contexts. The resulting network captures
both collective educator behavior (as it emerges from platform usage) and transitions grounded in
pedagogical research. This structure enables us to generate recommendations that reflect both the
normative behaviors of peers and evidence-based instructional trajectories.</p>
        <p>To enhance personalization and contextual fidelity, we also generate a user-specific preference
network for each teacher with suficient interaction history on the platform. This individualized layer
captures the teacher’s own prior behaviors and choices, demonstrating commonly selected instructional
strategies and frequent content goals. After implementation of this pipeline, we will further incorporate
individuals’ responsiveness to previous nudges to this user preference network. These personal usage
patterns are modeled using a weighted subgraph derived from the global co-occurrence network, giving
additional weight to pathways the user has historically followed. Both networks will be updated
periodically to reflect the recent educators behavior on the platform.</p>
        <p>During nudge generation, the system balances the collective network (peer norms and research-based
best practices) with this personalized subgraph. This compound network architecture allows the
system to recommend actions that are both familiar and growth-oriented: they reflect the teacher’s
preferences while gently nudging toward more efective or underused strategies observed in the broader
community. However, the optimal weight to balance these two networks remains to be investigate.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Nudging Mechanism</title>
        <p>At runtime, the system incorporates LLM tagging the teacher’s current instructional step using
established practice codebook and identifying educational context demonstrated from teachers’ input. The
compound practice network is then queried to predict the most likely and pedagogically meaningful
next actions to nudge. Nudges are selected based on contextual similarity, using the teacher’s subject,
educational levels, and learner needs, and delivered using behavioral heuristics such as status quo,
social proof, and peer efect.</p>
        <p>Nudges are phrased to be suggestive rather than prescriptive, maintaining the teacher’s sense of
agency while introducing efective alternatives grounded in both literature and peer activity. For
example, if a teacher exploring inquiry-based science methods, the system may suggest a transition to
seeking real-world applications, a practice used frequently by other teachers in similar contexts and
recommended by previous studies in science education [24].</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Planned Analysis</title>
      <p>Following implementation of the nudging pipeline within the instructional platform, we will evaluate
its efectiveness through a mixed-methods strategy. To do so, users will be randomly assigned to one
of three groups: a hybrid intelligence group (receiving nudges generated from the practice-informed
network), an LLM-inference group (receiving LLM inferred recommendations), and a control group
(receiving generic or no nudges). Table 1 summarizes the analysis plan for each research question,
integrating controlled experiments like A/B testing [12], platform usage analysis, and qualitative
feedback.</p>
      <sec id="sec-4-1">
        <title>4.1. Expected Results</title>
        <p>Preliminary retrospective analysis using a held-out dataset, which was not used during network
construction, indicates that system-generated recommendations closely align with teachers’ actual
instructional actions, validating the predictive strength of the co-occurrence network. The system also
surfaces a broader range of efective instructional strategies that are pedagogically appropriate for the
given context. This suggests the system is capable of nudging teachers toward efective, research-backed
practices beyond their habitual repertoire.</p>
        <p>We expect this hybrid approach, which combines individual usage patterns, collective educator
behavior, and pedagogical research, to promote both adoption and instructional growth. We expect
that recommendations align with their peers’ behavior are more likely to be perceived as trustworthy
and contextually relevant by teachers, while the integration of evidence-based strategies enhances the
depth and diversity of instructional planning. We anticipate that the practice-informed network will
outperform LLM-only inference in two key areas: (1) contextual precision and teacher acceptance, due
to its grounding in real educator behavior, and (2) pedagogical value, by encouraging the uptake of
strategies aligned with efective teaching frameworks. These findings will be further validated through
controlled experimentation, user behavior analysis, and qualitative teacher feedback. These findings
will be further validated through controlled experimentation, user behavior analysis, and qualitative
teacher feedback.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This study introduces a practice-informed, behaviorally grounded AI nudging pipeline that supports
instructional decision-making without compromising teacher autonomy. By combining a co-occurrence
network of educator behavior with pedagogical research, the system generates nudges that are
explainable, context-aware, and value-sensitive. Our approach exemplifies hybrid intelligence in education,
blending human insight with computational support to guide, not prescribe, teacher action. The findings
ofer practical and theoretical implications for designing responsible AI systems in education and other
professional domains.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work is supported by the Institute of Education Sciences of the U.S. Department of Education,
through Grant R305C240012 and by several awards from the National Science Foundation (NSF #2043613,
#2300291, #2405110) to the University of Washington, and a NSF SBIR/STTR award to Hensun Innovation
LLC (#2423365). The opinions expressed are those of the authors and do not represent views of the
funders.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT-4o in order to: Grammar and spelling
check. After using these tool(s)/service(s), the author(s) reviewed and edited the content as needed and
take(s) full responsibility for the publication’s content.
[11] M. E. Rodriguez, A. E. Guerrero-Roldán, D. Baneres, A. Karadeniz, An intelligent nudging system
to guide online learners, The International Review of Research in Open and Distributed Learning
23 (2022) 41–62. doi:10.19173/irrodl.v22i4.5407.
[12] M. Weinmann, C. Schneider, J. vom Brocke, Digital nudging, Business &amp; Information Systems</p>
      <p>Engineering 58 (2016) 433–436.
[13] R. Rebonato, A critical assessment of libertarian paternalism, Journal of Consumer Policy 37
(2014) 357–396.
[14] M. Tissenbaum, J. Slotta, Supporting classroom orchestration with real-time feedback: A role
for teacher dashboards and real-time agents, International Journal of Computer-Supported
Collaborative Learning 14 (2019) 325–351.
[15] K. Holstein, B. M. McLaren, V. Aleven, Co-Designing a Real-Time Classroom Orchestration Tool to</p>
      <p>Support Teacher-AI Complementarity, Technical Report, Grantee Submission, 2019.
[16] L. Bovens, The ethics of nudge, in: S. O. Hansson, T. Grüne-Yanof (Eds.), Preference Change:
Approaches from Philosophy, Economics and Psychology, Springer Netherlands, Dordrecht, 2009,
pp. 207–219.
[17] H. Farrell, C. R. Shalizi, Pursuing cognitive democracy, in: D. Allen, J. Light (Eds.), From Voice to
Influence: Understanding Citizenship in a Digital Age, University of Chicago Press, Chicago, IL,
2015, pp. 211–231.
[18] F. Furedi, On Tolerance: A Defence of Moral Independence, Bloomsbury Publishing, London, 2011.
[19] Y. Dimitriadis, R. Martínez-Maldonado, K. Wiley, Human-centered design principles for actionable
learning analytics, in: C. Mouza, N. Lavigne (Eds.), Research on E-Learning and ICT in Education:
Technological, Pedagogical and Instructional Perspectives, Springer, Cham, 2021, pp. 277–296.
[20] A. Südkamp, J. Kaiser, J. Möller, Teachers’ judgments of students’ academic achievement: Results
from field and experimental studies, in: S. Krolak-Schwerdt, S. Glock, M. Böhmer (Eds.), Teachers’
Professional Development, Brill, Leiden, 2014, pp. 5–25.
[21] B. Shneiderman, Human-centered artificial intelligence: Three fresh ideas, AIS Transactions on</p>
      <p>Human-Computer Interaction 12 (2020) 109–124.
[22] S. J. B. Shum, R. Luckin, Learning analytics and ai: Politics, pedagogy and practices, British Journal
of Educational Technology 50 (2019) 2785–2793.
[23] N. R. C. U. C. on a Conceptual Framework for New K-12 Science Education Standards, A Framework
for K-12 Science Education: Practices, Crosscutting Concepts, and Core Ideas, National Academies
Press, Washington, DC, 2012.
[24] T. Mirsch, C. Lehrer, R. Jung, Digital Nudging: Altering User Behavior in Digital Environments,
Technical Report, 2017.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. W.</given-names>
            <surname>Raudenbush</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Ball</surname>
          </string-name>
          , Resources, instruction, and research,
          <source>Educational Evaluation and Policy Analysis</source>
          <volume>25</volume>
          (
          <year>2003</year>
          )
          <fpage>119</fpage>
          -
          <lpage>142</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Carlson</surname>
          </string-name>
          ,
          <article-title>Elementary teachers' learning to construct high-quality mathematics lesson plans</article-title>
          ,
          <source>The Elementary School Journal</source>
          <volume>113</volume>
          (
          <year>2013</year>
          )
          <fpage>359</fpage>
          -
          <lpage>385</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Choppin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Amador</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Callard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Carson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gillespie</surname>
          </string-name>
          ,
          <article-title>Synchronous online model for mathematics teachers' professional development</article-title>
          ,
          <source>in: Handbook of Research on Online Pedagogical Models for Mathematics Teacher Education, IGI Global</source>
          , Hershey, PA,
          <year>2020</year>
          , pp.
          <fpage>176</fpage>
          -
          <lpage>202</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Poppink</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Cui</surname>
          </string-name>
          , G. Fan,
          <article-title>Lesson planning: A practice of professional responsibility and development</article-title>
          ,
          <source>Educational Horizons</source>
          <volume>85</volume>
          (
          <year>2007</year>
          )
          <fpage>248</fpage>
          -
          <lpage>258</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Darling-Hammond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. E.</given-names>
            <surname>Hyler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gardner</surname>
          </string-name>
          , Efective Teacher Professional Development, Learning Policy Institute,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>N. D.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Camburn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kelcey</surname>
          </string-name>
          , E. Quintero,
          <article-title>Teachers' time use and afect before and after covid-19 school closures</article-title>
          ,
          <source>AERA Open 8</source>
          (
          <year>2022</year>
          )
          <article-title>23328584211068068</article-title>
          . doi:
          <volume>10</volume>
          .1177/ 23328584211068068.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>R. de Brito Duarte</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Campos</surname>
          </string-name>
          ,
          <article-title>Looking for cognitive bias in ai-assisted decision-making (</article-title>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R. H.</given-names>
            <surname>Thaler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. R.</given-names>
            <surname>Sunstein</surname>
          </string-name>
          , Nudge: Improving Decisions About Health, Wealth, and
          <string-name>
            <surname>Happiness</surname>
          </string-name>
          , Yale University Press, New Haven, CT,
          <year>2008</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>W.</given-names>
            <surname>Samuelson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zeckhauser</surname>
          </string-name>
          ,
          <article-title>Status quo bias in decision making</article-title>
          ,
          <source>Journal of Risk and Uncertainty</source>
          <volume>1</volume>
          (
          <year>1988</year>
          )
          <fpage>7</fpage>
          -
          <lpage>59</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Nickerson</surname>
          </string-name>
          ,
          <article-title>Confirmation bias: A ubiquitous phenomenon in many guises</article-title>
          ,
          <source>Review of General Psychology</source>
          <volume>2</volume>
          (
          <year>1998</year>
          )
          <fpage>175</fpage>
          -
          <lpage>220</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>