<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Drift Dynamics in Denoising Difusion Probabilistic Models for 2D Point Cloud Generation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sanyam Jain</string-name>
          <email>sanyam.jain@dent.au.dk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Khuram Naveed</string-name>
          <email>knaveed@dent.au.dk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Illia Oleksiienko</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Alexandros Iosifidis</string-name>
          <email>Alexandros.Iosifidis@tuni.fi</email>
          <email>io@ece.au.dk</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ruben Pauwels</string-name>
          <email>ruben.pauwels@dent.au.dk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Difusion Probabilistic Models, Interpretability</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Dentistry and Oral Health, Aarhus University</institution>
          ,
          <addr-line>Vennelyst Boulevard 9, 8000 Aarhus C</addr-line>
          ,
          <country country="DK">Denmark</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Electrical and Computer Engineering, Aarhus University</institution>
          ,
          <addr-line>Finlandsgade 22, 8200 Aarhus N</addr-line>
          ,
          <country country="DK">Denmark</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Faculty of Information Technology and Communication Sciences, Tampere University</institution>
          ,
          <addr-line>Korkeakoulunkatu 7, 33720 Tampere</addr-line>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Michela Milano</institution>
          ,
          <addr-line>Alessandro Safiotti, Mauro Vallati</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>This work introduces InJecteD, a framework for interpreting Denoising Difusion Probabilistic Models (DDPMs) by analyzing sample trajectories during the denoising process of 2D point cloud generation. We apply this framework to three datasets from the Datasaurus Dozen - bullseye, dino, and circle - using a simplified DDPM architecture with customizable input and time embeddings. Our approach quantifies trajectory properties, including displacement, velocity, clustering, and drift field dynamics, using statistical metrics such as Wasserstein distance and cosine similarity. By enhancing model transparency, InJecteD supports human-AI collaboration by enabling practitioners to debug and refine generative models. Experiments reveal distinct denoising phases: initial noise exploration, rapid shape formation, and final refinement, with dataset-specific behaviors (e.g., bullseye's concentric convergence vs. dino's complex contour formation). We evaluate four model configurations, varying embeddings and noise schedules, demonstrating that Fourier-based embeddings improve trajectory stability and reconstruction quality. The code and dataset are available at https://github.com/s4nyam/InJecteD.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Denoising Difusion Probabilistic Models (DDPMs) [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ] have become a leading approach in generative
modeling owing to their ability to generate high-quality samples, such as images and point clouds.
Their stability and performance make them a compelling alternative to other generative models, such as
GANs [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and VAEs [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], particularly in applications such as data synthesis and scientific visualization [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
However, the iterative denoising process, often spanning numerous steps, makes DDPMs complex and
dificult to interpret (in terms of how features emerge from pure noise), obscuring how samples evolve
from noise to structured data. This lack of transparency poses challenges for understanding model
behavior, debugging performance issues, and ensuring reliability in applications where interpretability
is crucial, such as scientific data analysis.
      </p>
      <p>
        To address this challenge, we introduce a framework for Interpreting traJectories in Denoising
Difusion ( InJecteD) that is designed to analyze the trajectories of samples during the denoising process
of DDPMs [
        <xref ref-type="bibr" rid="ref1 ref5">1, 5</xref>
        ]. Unlike prior work [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], which focuses on applying DDPMs to point clouds with Fourier
encodings, InJecteD works on a set of quantitative metrics to systematically assess the denoising process.
Specifically, we employ a simplified DDPM architecture with flexible input and time embeddings to
enable explicit tracking of sample evolution. By introducing quantitative metrics, including trajectory
displacement, velocity, clustering, and drift field dynamics, we uncover the underlying patterns of the
      </p>
      <p>CEUR</p>
      <p>
        ceur-ws.org
denoising process. We showcase our approach in 2D point cloud generation process through three
publicly available datasets from the Datasaurus Dozen [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] namely: bullseye, dino, and circle. These
datasets share identical statistical properties but difer in their geometric structures, providing an
opportunity to study how DDPMs capture diverse shapes. Our experiments reveal dataset-specific
behaviors, such as concentric convergence in the bullseye dataset and complex contour formation
in the dino dataset, and identify three distinct phases of denoising: initial noise exploration, rapid
shape formation, and final refinement. We employ four model configurations with varying input and
time embeddings and noise schedules to assess their impact on trajectory stability and reconstruction
quality. While limited to 2D synthetic datasets, the insights gained lay the groundwork for extending
interpretability to more complex data. Our contributions include:
• Customizing an existing lightweight DDPM architecture for interpretability of 2D point cloud
generation.
• Use of relevant statistical and geometric distance metrics for analyzing trajectory properties,
including displacement, velocity, clustering, and drift direction alignment.
• Experimental validation of the proposed InJecteD framework on the bullseye, dino, and circle
datasets, highlighting unique dynamics of the reverse difusion process.
      </p>
      <p>The core contribution of this work involves new insights into the behavior of DDPMs, laying the
groundwork for more interpretable and reliable generative models.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Interpretability in generative models has been explored through various techniques aimed at
understanding model behavior [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Feature importance analysis assesses the relative contribution of input
features to the model’s output, while attention mechanisms highlight regions of focus during data
generation [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. Interactive reconstruction approaches allow users to manipulate latent representations
to reconstruct target instances, providing insights into the model’s internal representations [
        <xref ref-type="bibr" rid="ref10 ref8 ref9">8, 9, 10</xref>
        ].
These methods are particularly relevant to our work, as they ofer ways to probe the evolution of
samples during the denoising process, similar to our focus on analyzing trajectories in DDPMs; however,
they often rely on manual interaction, and by analyzing trajectories more systematically, we aim to
complement these approaches with a more principled understanding of the generative dynamics
      </p>
      <p>
        Trajectory analysis, while well-established in fields such as reinforcement learning [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], biology [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ],
and epidemiology [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], is underexplored in the context of DDPMs. In reinforcement learning, trajectory
analysis predicts agents’ paths [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], while in biology, it tracks cell diferentiation or molecular motion
[
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. In epidemiology, it models health outcome patterns over time [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. These approaches often use
metrics like mean squared displacement or Hidden Markov Models to characterize state transitions
and dynamics. Applying trajectory analysis to DDPMs involves studying how samples evolve through
denoising steps, ofering a novel perspective on the model’s decision-making process and its ability to
capture complex patterns.
      </p>
      <p>
        The Datasaurus Dozen datasets [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] provide a testbed for analyzing properties of complex models
in a simplified and controlled setting [ 14? , 15, 16, 17, 18, 19]. These datasets, consisting of data
points forming shapes such as a bullseye, dinosaur, and circle, are designed to have identical statistical
properties (e.g., mean, variance, correlation) but distinct visual structures when plotted as 2D point
clouds.
      </p>
      <p>
        Chan [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] investigates DDPMs, detailing their denoising mechanics and applying them to 2D point
cloud generation with the Datasaurus dataset, using Fourier encoding to enhance performance. This
work provides a foundational framework for our InJecteD approach, validating our use of similar datasets
and embeddings to analyze trajectory dynamics. Beyond point clouds, interpretability in DDPMs is
advanced through techniques like saliency maps in image generation and latent space analysis in
text-to-image synthesis, revealing feature prioritization and semantic evolution [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ]. These methods
inform our trajectory analysis by ofering complementary perspectives on DDPM dynamics across
contexts.
process.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Dataset Description</title>
        <p>diverse domains. By building on these foundations, our work enhances the transparency of DDPMs,
particularly for 2D point cloud generation, with potential applications in broader generative modeling</p>
        <p>
          Our research builds on these foundations by developing a framework to analyze sample trajectories
in DDPMs, specifically for 2D point cloud generation. By combining insights from interpretability
techniques and trajectory analysis, and using the unique properties of the Datasaurus Dozen, we aim
to enhance the transparency of DDPMs and provide a deeper understanding of their data generation
The bullseye, dino, and circle datasets from the Datasaurus Dozen [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] consist of 2D point clouds, each
with approximately 142 points represented as coordinates (,  ) ∈ ℝ 2. These datasets share identical
statistical properties (mean, variance, correlation) but exhibit distinct geometric structures, making
them ideal for studying structural diversity in DDPMs. Preprocessing normalizes the data to zero mean
and unit variance:
 −
        </p>
        <p>norm =
,  norm =
 −  
 
,
where   ,   are the means and   ,   are the standard deviations of the  and  coordinates. To increase
sample size, each dataset is replicated six times, yielding 852 points per dataset. The data is split into
90% training (766 points) and 10% testing (86 points) sets, with a batch size of 32 for training.</p>
        <p>
          The Datasaurus Dozen is uniquely suited for this study due to its ability to challenge DDPMs to
capture diverse geometric patterns despite statistical uniformity. Alternative datasets, such as MNIST
[
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] or ShapeNet [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ], are less efective: MNIST focuses on class-based patterns (digits), missing
geometric nuances, while ShapeNet for its size (approximately 300 Million). The Datasaurus datasets
provide a controlled, structurally diverse testbed for evaluating trajectory dynamics.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Method</title>
      <p>modeling.</p>
      <sec id="sec-3-1">
        <title>3.1. Difusion Algorithm</title>
        <p>Originally developed to underscore the importance of data visualization over reliance on summary
statistics, Datasaurus datasets are ideal for evaluating how well DDPMs capture diverse geometric
patterns and whether trajectory analysis can diferentiate their denoising behaviors. By applying
DDPMs to these datasets, we dissect the statistical similarity and structural diversity in generative
The DDPM architecture comprises a forward noising process and a learned reverse denoising process,
implemented using a multilayer perceptron (MLP). The algorithm is specifically designed to handle the
2D point clouds of the Datasaurus Dozen datasets (bullseye, dino, and circle), treating each point as an
independent 2D coordinate, since the data lacks the spatial regularity of the structured grid found in
images.</p>
        <p>The standard forward process adds Gaussian noise over  = 50 timesteps using a linear noise schedule.
For a point cloud  0 ∈ ℝ ×2 from a Datasaurus dataset (e.g., circle with  = 852
each point  = (  ,   )represents a 2D coordinate. The state at timestep  is:</p>
        <p>= √ ̄  0 + √1 −  ̄ ,  ∼  (0,  ),

where  ̄ =</p>
        <p>∏=1   ,   = 1 −   , and   ranges linearly from  min = 1 − 0.9999 to  max = 1 − 0.95.
The cumulative product  ̄ transitions from ≈ 1 (original data) at  = 0 to ≈ 0 (near-pure noise)

at  =  . This process is implemented as a function that generates a series {  }=0 for each point
independently, preserving the unstructured nature of the point cloud. Unlike image pixels, which have
(1)
(2)
points after replication),
spatial correlations in a grid, the Datasaurus points are treated as a collection of independent 2D vectors,
with noise applied to each (, )
coordinate pair. This results in a trajectory { 
}
=0 for each point  ,
visualized as scatter plots to capture the evolving geometry (e.g., circle’s uniform ring or dino’s complex
contours).</p>
        <p>The reverse process denoises   ∼  (0,  ) to reconstruct  0 using:
1
where   (  , ) is the noise predicted by the MLP, and   = √
1 −   . The input to the MLP can be a single
point or a batch of points at timestep  ; to optimize this,   ∈ ℝ×2 (where  = 32 is the batch size),
and the output is the predicted noise   ∈ ℝ×2 , representing the 2D noise vector for each point. The
implemented MLP consists of five layers: four hidden layers (64 units, ReLU activations) and one output
layer predicting a 2D noise vector. This allows the model to learn the distribution of points that form
shapes such as the dino’s arms or bullseye’s concentric rings. Input embeddings are:
• Identity: Uses   ∈ ℝ2 directly (2D), feeding the raw (, ) coordinates of each point.
• Fourier: Projects   to a 64-dimensional space:</p>
        <p>emb = [sin( ⊤),cos( ⊤)],  ∈ ℝ 32×2,  ∼  (0,  ).</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. InJecteD Framework</title>
        <p>denoising process.</p>
        <p>We present set of metrics to quantify trajectory and drift dynamics and critical for understanding the
1. Trajectory Displacement: Measures the total Euclidean distance traveled by each point:</p>
        <p>emb = [sin(   ⊤),cos(   ⊤)],   ∈ ℝ16×1,   ∼  (0,  ).</p>
        <p>The working procedure involves: (1) sampling a batch of points  0 from the dataset, (2) applying the
forward process to generate noisy points   at a random timestep  , (3) predicting the noise   (  , ),
and (4) computing the MSE loss to optimize the MLP. The sampling process iteratively denoises  
to  0, generating new point clouds that match the target distribution (e.g., reconstructing the circle’s
ring structure). Four model configurations are trained as follows (read as input-time embedding): (1)
identity-zero,  min = 0.95, (2) Fourier-linear,  min = 0.95, (3) Fourier-Fourier,  min = 0.95, and (4)
Fourier-Fourier,  min = 0.98. Training uses Adam optimization (learning rate 4 × 10−4, gradient clipping
norm 1.0) for 2000 epochs, minimizing the mean squared error (MSE):</p>
        <p>=   0,, [‖ −   (  , )‖22] .</p>
        <p>The MLP-based DDPM is chosen for its simplicity and eficiency in 2D spaces. For the Datasaurus
datasets, the algorithm’s ability to treat points as independent 2D coordinates allows it to capture diverse
geometric patterns (e.g., dino’s complex contours) without relying on spatial correlations, ensuring
lfexibility and robustness in modeling unstructured point clouds.
(3)
(4)
(5)
(6)
(7)
where</p>
        <p>() is the position of the  -th point at timestep  . The distribution of   is visualized as a
histogram, revealing the extent of movement. High displacement, as expected in complex datasets
like dino, indicates intricate trajectory patterns.
2. Trajectory Velocity: Computes the average displacement per timestep:</p>
        <p>1
 =1
 () =</p>
        <p>()
∑ ‖ +1 −  
()
‖
2
Plotted over timesteps,  () identifies denoising phases: high values indicate rapid shape
formation, while low values indicate refinement. This metric is essential for detecting transitions in the
generative process.
3. Trajectory Clustering: Applies K-means clustering (with  = 5 ) to flattened trajectories
{ 
=0 ∈ ℝ ×2 , reshaped to ℝ2 . The resulting labels are visualized on the final point cloud (  0),
highlighting spatial patterns in trajectory behavior. This reveals whether points in similar regions
follow consistent paths, critical for datasets like bullseye with radial structures. The choice of
 = 5</p>
        <p>was made after testing for  = 2, 3, 4, 5, 6 , as higher  values yielded diminishing returns.
4. Wasserstein Distance: Quantifies similarity between original and generated point clouds
(collection of all points):
 =</p>
        <p>( 1( orig,  gen) +  1( orig,  gen)),
where  1 is the 1D Wasserstein distance for  and  coordinates. Lower values indicate higher
ifdelity, essential for evaluating generative performance.
5. Drift Magnitude : For the forward process, the drift at timestep  on a grid point   is:
with magnitude ‖  −   ‖2, weighted by:
For the backward process, the drift is:
1
2
 =
  = √  (1 − ̄−1 )  + √ ̄−1 (1 −   ) 0</p>
        <p>,
1 −  ̄
exp (− ‖  −√ ̄  0‖2 /2)</p>
        <p>1− ̄
∑ 0 exp (− ‖  −√ ̄  0‖2 /2)
1− ̄</p>
        <p>.
1
with magnitude ‖ ̂ −   ‖2. Magnitudes are visualized as heatmaps, showing the strength of
movement across a grid, crucial for understanding denoising dynamics.
6. Drift Direction : Measures alignment between backward drift vectors and the direction to the
ifnal point cloud using cosine similarity:</p>
        <p>CS() =
1

∑
(  ̂
() −  
() ) ⋅ (0
() −</p>
        <p>() )
 =1 ‖ ̂
() −  
()
‖2‖ 0
() −  
()
‖
2
toward final positions, critical for assessing model accuracy.
where  ̂
() is interpolated at</p>
        <p>() using linear interpolation. High CS()indicates efective guidance</p>
        <p>As will be shown in the following, these metrics collectively provide a detailed understanding of the
denoising process. Displacement and velocity quantify movement scale and speed, clustering reveals
spatial patterns, Wasserstein distance evaluates generative fidelity, and drift metrics analyze movement
direction and strength, essential for interpreting DDPM behavior.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Experimental Setup</title>
        <p>The MLP is trained on a CPU with  = 50 timesteps, a batch size of 32, and 2000 epochs. Visualizations
(scatter plots, quiver plots, heatmaps) are saved as SVG files in dataset-specific directories. The sampling
generates 1000 samples per configuration, tracking trajectories for analysis. Noise prediction error is
computed as the MSE per timestep, visualized to assess model performance.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>Our analysis reveals distinct denoising behaviors across the three datasets, with each figure
comprehensively presenting the results for one dataset. We examine the original structure, generation quality,
trajectory dynamics, and denoising metrics for each case. We report evaluation metrics for three datasets
- Bullseye, Circle and Dinosaurs. Figure 1 presents the complete analysis for the bullseye dataset. Key
ifndings include (1) Drift Alignment: Near-perfect radial alignment (cosine similarity &gt;0.9) in later
steps (a). Curve (b) helps to assess whether the learned drift aligns with intended denoising behavior.
The diagnosis in (b) can be divided into three phases: early, middle, and late timesteps. In the early
timesteps (t=0-15) (high noise), the mean cosine similarity is around 0.2, indicating weak alignment
between drift and the final direction. The system is still noisy, but there is a faint guiding signal. During
the middle timesteps (t=15-30), the similarity peaks around 0.4, showing strong alignment. This is
likely the most efective phase, where the drift actively pulls samples toward the final state of denoising.
In the late timesteps (t=30-50) (low noise), similarity drops to 0. The sample is already close to the
target, so the drift mostly fine-tunes details and is no longer directionally aligned, (2) Trajectories: Clear
concentric patterns in clustering (d-e) with 82% of points following class-specific paths, (3) Velocity:
Two-phase velocity curve (f) - sharp drop (t=0-30), final refinement (t=31-50), and (4) Model Comparison:
Fourier-Fourier ( min = 0.98) shows the best-fit displacement distribution (c). We discuss the Drift
Alignment metric more in Appendix 5.</p>
      <p>Similar observations done for Circle in Figure 2 and Dinosaurs in Figure 3. We also illustrate the
formation of Dinosaurs point cloud from pure noise in Figure 4.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>InJecteD addresses the critical need for interpretability in DDPMs, which are increasingly vital for
data synthesis in scientific visualization and computational biology. By quantifying trajectory and
drift dynamics, InJecteD reveals how DDPMs capture geometric structures, enabling improved model
design, debugging, and application in domains requiring transparency. We showcased its applicability
by conducting experiments for 2D point cloud generation on the Datasaurus Dozen datasets, which
revealed three consistent denoising phases and showed that Fourier embeddings significantly improve
trajectory stability. The metrics provide insights into model behavior with minimal computational
overhead, with the Fourier-Fourier configuration emerging as the most efective approach. Future work
could extend to higher dimensions and explore trajectory steering methods. Another future work can
be designing processes analyzing the trajectory dynamics of the data generation learning processes in
other families of generative models.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT from OpenAI in order to: grammar
and spelling check, paraphrase, and reword. After using this tool/service, the author(s) reviewed and
edited the content as needed and take(s) full responsibility for the publication’s content.
(a) Distribution Comparison
(a) Distribution Comparison
(b) Drift Direction</p>
      <p>Clustered Trajectories in Final Point Cloud
0
10
20
30
40</p>
      <p>50
Step</p>
      <p>8
Total Displacement</p>
      <p>9
5
6
7
10
11
12
0.4
2
1
0
1
2
70
60
50
cy40
eeunq
rF30
20
10
50 0 4</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgement</title>
      <p>This work was funded by the Independent Research Fund Denmark, project “Synthetic Dental
Radiography using Generative Artificial Intelligence”, grant ID 10.46540/3165-00237B.</p>
    </sec>
    <sec id="sec-8">
      <title>A. Drift Direction Explained</title>
      <p>The evaluation of drift direction in Figure 1 (b) (for example) is divided into three phases based on the
mean cosine similarity between the drift vector and the direction to the final point: in early timesteps
with high noise, the similarity is around 0.2, indicating a weak but noticeable alignment and faint pull
toward the final state amid chaos; in middle timesteps, it peaks at about 0.4, showing the strongest
alignment and efectiveness in guiding samples along the true trajectory as the ”sweet spot” of denoising;
and in late timesteps with minimal noise, it drops to near 0, becoming nearly orthogonal as the drift
shifts to fine-tuning local details rather than directional guidance.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Abbeel</surname>
          </string-name>
          ,
          <article-title>Denoising difusion probabilistic models</article-title>
          ,
          <source>Advances in neural information processing systems</source>
          <volume>33</volume>
          (
          <year>2020</year>
          )
          <fpage>6840</fpage>
          -
          <lpage>6851</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Weng</surname>
          </string-name>
          ,
          <article-title>What are difusion models?, lilianweng</article-title>
          . github. io (
          <year>2021</year>
          )
          <fpage>21</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Goodfellow</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pouget-Abadie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Mirza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Warde-Farley</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ozair</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Courville</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          , Generative adversarial nets,
          <source>Advances in neural information processing systems</source>
          <volume>27</volume>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Welling</surname>
          </string-name>
          , et al.,
          <source>Auto-encoding variational bayes</source>
          ,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Chan</surname>
          </string-name>
          , Difusion models,
          <year>2024</year>
          . URL: https://andrewkchan.dev/posts/diffusion.html, accessed:
          <fpage>2025</fpage>
          -06-09.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Gillespie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Locke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Davies</surname>
          </string-name>
          , L.
          <string-name>
            <surname>D'Agostino McGowan</surname>
          </string-name>
          ,
          <source>datasauRus: Datasets from the Datasaurus Dozen</source>
          ,
          <year>2025</year>
          . URL: https://github.com/jumpingrivers/datasauRus,
          <source>r package version 0.1</source>
          .9, https://jumpingrivers.github.io/datasauRus/.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-J.</given-names>
            <surname>Ju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-W.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Explaining generative difusion models via visual analysis for interpretable decision-making process</article-title>
          ,
          <source>Expert Systems with Applications</source>
          <volume>248</volume>
          (
          <year>2024</year>
          )
          <fpage>123231</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sohl-Dickstein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. P.</given-names>
            <surname>Kingma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ermon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Poole</surname>
          </string-name>
          ,
          <article-title>Score-based generative modeling through stochastic diferential equations</article-title>
          , arXiv preprint arXiv:
          <year>2011</year>
          .
          <volume>13456</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>C.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-Y.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ermon</surname>
          </string-name>
          ,
          <article-title>Sdedit: Guided image synthesis and editing with stochastic diferential equations</article-title>
          ,
          <source>arXiv preprint arXiv:2108.01073</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Liew</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , V. Y.
          <string-name>
            <surname>Tan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bai</surname>
          </string-name>
          , Dragdifusion:
          <article-title>Harnessing difusion models for interactive point-based image editing</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>8839</fpage>
          -
          <lpage>8849</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>K. A.</given-names>
            <surname>Sadek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nulli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Velja</surname>
          </string-name>
          , J. Vincenti, '
          <article-title>explaining rl decisions with trajectories': A reproducibility study</article-title>
          ,
          <source>arXiv preprint arXiv:2411.07200</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S. V.</given-names>
            <surname>Stassen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. G.</given-names>
            <surname>Yip</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Wong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Ho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Tsia</surname>
          </string-name>
          ,
          <article-title>Generalized and scalable trajectory inference in single-cell omics data with via</article-title>
          ,
          <source>Nature communications 12</source>
          (
          <year>2021</year>
          )
          <fpage>5528</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>H. L. Nguena</given-names>
            <surname>Nguefack</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Pagé</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Katz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Choinière</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vanasse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Dorais</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O. M.</given-names>
            <surname>Samb</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lacasse</surname>
          </string-name>
          ,
          <article-title>Trajectory modelling techniques useful to epidemiological research: a comparative narrative review of approaches, Clinical epidemiology (</article-title>
          <year>2020</year>
          )
          <fpage>1205</fpage>
          -
          <lpage>1222</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Luo</surname>
          </string-name>
          , W. Hu,
          <article-title>Difusion probabilistic models for 3d point cloud generation</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>2837</fpage>
          -
          <lpage>2845</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fu</surname>
          </string-name>
          , X. Han,
          <string-name>
            <given-names>H.</given-names>
            <surname>Liang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chang</surname>
          </string-name>
          , Difusionpointlabel:
          <article-title>Annotated point cloud generation with difusion model</article-title>
          , in: Computer Graphics Forum, volume
          <volume>41</volume>
          , Wiley Online Library,
          <year>2022</year>
          , pp.
          <fpage>131</fpage>
          -
          <lpage>139</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vahdat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gojcic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Litany</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Fidler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kreis</surname>
          </string-name>
          , et al.,
          <article-title>Lion: Latent point difusion models for 3d shape generation</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>35</volume>
          (
          <year>2022</year>
          )
          <fpage>10021</fpage>
          -
          <lpage>10039</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>H.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Huai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Meng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Pei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <article-title>Dif-tree: A difusion model for diversified tree point cloud generation with high realism</article-title>
          ,
          <source>Remote Sensing</source>
          <volume>17</volume>
          (
          <year>2025</year>
          )
          <fpage>923</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Lyu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Change Loy, Gaussiananything: Interactive point cloud latent difusion for 3d generation</article-title>
          , arXiv e-prints (
          <year>2024</year>
          ) arXiv-
          <fpage>2411</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>P.</given-names>
            <surname>Schröppel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wewer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. E.</given-names>
            <surname>Lenssen</surname>
          </string-name>
          , E. Ilg, T. Brox,
          <article-title>Neural point cloud difusion for disentangled 3d shape and appearance generation</article-title>
          ,
          <source>in: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>8785</fpage>
          -
          <lpage>8794</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>L.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <article-title>The mnist database of handwritten digit images for machine learning research [best of the web]</article-title>
          ,
          <source>IEEE signal processing magazine</source>
          <volume>29</volume>
          (
          <year>2012</year>
          )
          <fpage>141</fpage>
          -
          <lpage>142</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A. X.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Funkhouser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Guibas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Hanrahan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Savarese</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Savva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <source>ShapeNet: An Information-Rich 3D Model Repository</source>
          ,
          <source>Technical Report arXiv:1512.03012 [cs.GR]</source>
          , Stanford University - Princeton University - Toyota Technological Institute at Chicago,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>