<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>V. Pavlyshyn);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>AI‑Driven Traffic Signal Control System to Reduce CO2 Emissions*</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Oleksander Ryzhanskyi</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vitaliy Pavlyshyn</string-name>
          <email>vitaliy@ualeaders.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavlo Radiuk</string-name>
          <email>radiukp@khmnu.edu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eduard Manziuk</string-name>
          <email>manziuk.e@khmnu.edu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Oleksander</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Barmak</string-name>
          <email>barmako@khmnu.edu.ua</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Iurii Krak</string-name>
          <email>iurii.krak@knu.ua</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Glushkov Cybernetics Institute</institution>
          ,
          <addr-line>40, Glushkov Ave., Kyiv, 03187</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Khmelnytskyi National University</institution>
          ,
          <addr-line>11, Institutes str., Khmelnytskyi, 29016</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>64/13, Volodymyrska str., Kyiv, 01601</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Rapid urbanization and the resulting traffic congestion have made the reduction of transport‑related carbon‑dioxide (CO2) emissions a pivotal goal for sustainable‑city initiatives. This paper introduces an artificial‑intelligence (AI) traffic‑signal control system that exploits deep reinforcement learning (DRL) to achieve dynamic, emission‑aware phase scheduling at signalized intersections. Unlike conventional fixed‑time or actuated controllers, the proposed framework continuously perceives multi‑modal state information, queue length, average speed, arrival rates and projected emergency‑vehicle trajectories, through a digital twin fed by microscopic traffic simulation. A convolutional DRL agent, trained with proximal‑policy optimization and shaped by a composite reward that penalizes total CO₂, queue growth and delay while rewarding throughput, learns an adaptive policy that balances environmental and mobility objectives. In controlled experiments mirroring a mid‑size European arterial, the AI agent lowers cumulative CO2 emissions by 18% without sacrificing pedestrian service levels. The learned policy exhibits robust generalization under varying traffic demand patterns and mild sensor noise, suggesting strong transferability.</p>
      </abstract>
      <kwd-group>
        <kwd>Traffic light control</kwd>
        <kwd>Deep reinforcement learning</kwd>
        <kwd>SUMO</kwd>
        <kwd>CO2 emissions</kwd>
        <kwd>Hilly topography of the road 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Optimizing traffic light systems to curb carbon dioxide (CO2) emissions is a pressing challenge,
essential not only for current urban sustainability but also for the long-term preservation of
ecosystems and the global biosphere [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Achieving this objective is not straightforward due to several
competing constraints, such as road gradient variations, prioritization for emergency vehicles, and
the need for fair treatment of drivers, pedestrians, and other road users [
        <xref ref-type="bibr" rid="ref2 ref3">2, 3</xref>
        ].
      </p>
      <p>
        One of the most promising techniques for intelligent traffic signal management is reinforcement
learning (RL), particularly its deep learning-enhanced variant, deep reinforcement learning (DRL)
[
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. DRL algorithms are capable of dynamically adapting to fluctuating traffic patterns by
continuously interacting with the environment to learn optimal decision-making policies that
maximize cumulative rewards [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ]. These learned strategies aim to enhance traffic flow and reduce
vehicular emissions.
      </p>
      <p>The primary contribution of this study is the development of a DRL-based simulation framework
for determining traffic light control policies at a single intersection, with a targeted goal of minimizing
CO2 emissions.</p>
      <p>The remainder of the article is organized as follows: the “Related Works” section surveys recent
efforts in traffic signal control aimed at emission reduction. The “Materials and Methods” section
introduces the proposed model tailored for local road intersections. Finally, “Results and Discussion”
presents the simulation outcomes and their interpretation.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        This section synthesizes prior investigations into traffic light signal control focused on reducing
CO2 emissions. A common thread among these studies is the widespread adoption of the SUMO
simulation model [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Researchers have employed a variety of optimization strategies, such as fuzzy
logic [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], game theory (utilizing Nash equilibrium) [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], extreme learning machines [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], genetic
algorithms [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], and an array of modeling techniques [
        <xref ref-type="bibr" rid="ref12 ref9">9, 12, 14</xref>
        ]. Nevertheless, RL and its derivatives
dominate the field [
        <xref ref-type="bibr" rid="ref13">13–15</xref>
        ].
      </p>
      <p>
        For example, reference [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] developed a three-phase fuzzy traffic control framework utilizing the
SUMO model, refining fuzzy rules to achieve superior CO2 emission reductions. Meanwhile, study
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] distinguished itself by factoring pedestrians into traffic light management, applying a
gametheoretic method to synchronize and reduce wait times for vehicles and pedestrians alike, with SUMO
simulations confirming decreases in delays and CO2 emissions.
      </p>
      <p>
        In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], a dual-algorithm approach was proposed, integrating a passive extreme learning machine
with periodic mini-batch learning (PB-ELM) for traffic forecasting and a maximum pressure control
(MPA) algorithm for traffic light regulation, validated through SUMO for enhanced congestion relief.
Reference [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] devised an intersection management framework that optimizes both traffic flow and
energy use within a vehicle-infrastructure setting, employing a genetic algorithm. Simulation
outcomes revealed reductions in total queued vehicles, maximum queue sizes, and average travel
times by 77.81%, 33.33%, and 10.95%, respectively.
      </p>
      <p>Using SUMO, study [16] illustrated that adaptive traffic lights markedly lower fuel use and
emissions compared to fixed systems, achieving reductions in travel time by 82.84% and CO2
emissions by 51.2%. In [17], an intelligent traffic system prioritizing emergency vehicles was
introduced, capable of adjusting signal timings based on vehicle classification (e.g., “emergency”
versus “passenger”).</p>
      <p>
        Work [18] tackled traffic light arrangements to ease congestion, proposing models to optimize
signal timing graphs for specific goals [19], with SUMO assessing their efficacy. Study [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] utilized
RL algorithms and SUMO to refine urban traffic management in Taipei, yielding enhancements in
traffic capacity and travel duration. In [14], the D3QN (Dual Deep Q Network) algorithm effectively
managed traffic light timings across varying traffic densities, outperforming conventional methods
like Webster’s fixed-time control in SUMO trials.
      </p>
      <p>Focusing on isolated intersections, [20] optimized traffic light signals using a double deep
Qnetwork (DDQN), an RL variant. Notably, [15] stressed the importance of tailoring reward functions
in RL to alig n agent actions with CO2 reduction objectives, testing various reward structures with
SUMO and its emission model to gauge performance sensitivity. Reference [21] offered a novel traffic
light control alternative by incorporating predictive noise modeling with RL and a SeqtoSeq-LSTM
framework [22], significantly improving noise levels, CO2 emissions, and fuel efficiency over
traditional systems.</p>
      <p>Further insights come from works [23–25], which showcase successful DQN applications to
analogous challenges. Study [23] introduced a “pressure” concept for regional signal coordination,
supported by extensive SUMO experiments, including a real-world Manhattan scenario with 2510
traffic lights. Reference [24] provided a toolkit for modern DQN implementations, while [25]
leveraged value-based RL for online learning with interpretable policy functions [26]. These findings
highlight the demand for intelligent traffic light systems that reduce CO2 emissions while addressing
vehicle and pedestrian queues, emergency vehicle access [27], and other factors, necessitating further
exploration. The prevalent use of RL and SUMO underscores their promise.</p>
      <p>Thus, this work is aimed at advancing sustainable development by curbing CO2 emissions through
an intelligent traffic signal control model tailored for complex terrains, balancing the needs of drivers,
pedestrians, and emergency services. This will be accomplished by devising a simulation and DQN
model for local intersection traffic light control to minimize CO2 emissions.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Materials and Methods</title>
      <p>In this study, we delineate the elements of a holistic approach crafted by the authors to address traffic
light control at local intersections, targeting CO2 emission reductions (see Figure 1). A key attribute
of this method is the creation of a unified metric and solution that, beyond lowering CO2 emissions,
accounts for complex road topographies (e.g., slopes); drivers’ desires for swift destination arrivals;
pedestrians’ needs, especially near crowded locales like schools and malls; rapid transit for emergency
services; the presence of heavy vehicles; and road emergencies. Moreover, this approach is engineered
for adaptability to encompass additional priorities as they arise.</p>
      <p>
        Note that for traffic signal control (TSC), one of the most widely used solution tools is the
reinforcement learning algorithm, namely its modern implementation Deep Reinforcement Learning
(DRL), since DRL can learn to adapt to changing traffic demands [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6, 28</xref>
        ]. Another tool used in this
work is the traffic modeling package Simulation of Urban Mobility (Eclipse SUMO or simply SUMO
v1.21.0 [29]) with built-in  2 emission models (e.g., Handbook Emission Factors for Road
Transportation – HBEFA 3.1 [30]) and complex road terrain models.
      </p>
      <p>Figure 1 shows the decomposition of the proposed approach into stages. Note that this article
presents research results for stages 3 and 4. Other stages will be considered in future studies.</p>
      <p>To test the approach’s feasibility, it is proposed to simulate traffic through an intersection using
Algorithm 1.</p>
      <sec id="sec-3-1">
        <title>Algorithm 1.</title>
        <p>Simulating the movement of cars through intersections.</p>
        <p>1: Input: intersection properties and traffic scenarios
2: Initialize event: Evnt about the appearance of a new car, relative to the start of the simulation with
the following properties: lane from which the car is declared; lane into which the car is leaving; speed
of the car
3: Set time counter at 0
4: For simulation
5: // Check if there is an event that corresponds to the current simulation time. If there is such an
event, add the corresponding car to the simulation on the specified lane with the specified
parameters:
6: If  .Time = current simulation time Then Add the car to the lane
7: // Update for each vehicle inside the simulation:
8: If the vehicle is near the intersection Then:
9: // Checking the condition of the traffic light:
10: If the green signal allows movement in its direction Then:
11: A car crosses an intersection into a lane
12: If the signal is red then:
13: The car stops
14: If the car is not near the intersection then:
15: If the path is free then:
16: The car continues to move at the current speed
17: If Obstacle Ahead Then:
18: The car stops
19: // Updating the state of the traffic light (at each moment of time we check if it is time to change the
phase of the traffic light)
20: If it’s time to change the phase of the traffic light Then:
21: Moving the traffic light to the next phase
  ℎ,  
 ,       )
channels, 
for the lanes,</p>
        <p>2
emissions, considering the properties of the observation environment.</p>
        <p>In Stage 3, it is proposed to synthesize a reward function for the intersection to reduce 
2</p>
        <p>Note that the motion simulation iteration is performed every ∆ t, but the reward is calculated not
for each iteration, but after N∆t, i.e. the system is controlled for a predetermined number of iterations
and the state of the system changes during this period at certain points in time.
an integral indicator consisting of the following intermediate components:</p>
        <sec id="sec-3-1-1">
          <title>Therefore, the value of</title>
          <p />
          <p>(reward at certain steps of the simulation) is proposed to be defined as
where   is a weighting factor determined empirically,   is a fine or reward.</p>
          <p>In Stage 4, it is proposed to synthesize a DQN model to achieve the goal of reducing
taking into account the properties of the observation environment and the integral loss function. It
should be noted that the study hypothesized that optimal solutions for traffic light control should be
sought not for a specific intersection, but for a sequence of intersections.</p>
          <p>The study proposes to use the independent DQN (IDQN) model for a sequence of intersections
and vehicle traffic scenarios. The essence of IDQN is that we will use several networks – each
intersection’s own DQN. An approach is proposed where each agent learns independently and
simultaneously its own policy, considering other agents as part of the environment.</p>
          <p>In IDQN, each agent observes the observation space of its intersection and chooses a separate
action and receives a reward. Since the agent “sees” the observation space only from its own
perspective, IDQN can be implemented by assigning each agent its own observation history of
actions. In deep reinforcement learning, this can be implemented by giving each agent “its own” DQN
to perform the observations and produce the actions.</p>
          <p>So, we will form IDQN models for intersection sequences and traffic scenarios:

 = ∑     ,
(1)
2
emissions,
  =
 
 ,  
3,    
= { 1,  2, … ,   },
, 
1, 
 , 
2, 
 ,</p>
          <p>(2)
space (at each observation time) in the form of a matrix (
where  – is the number of intersections in the sequence,  
× 
– is the number of lanes that the traffic lights regulate, 
– is the agent’s observation
×  ), where</p>
          <p>– is the number of
– is the number of properties
– is a convolutional layer that detects spatial dependencies between lanes,
takes into account the relationships between the properties of the lanes and prepares data for fully
connected layers that make decisions about changing the phases of the traffic light; convolution with
a kernel (2×2), stride (stride) 1 and no padding (padding) is used,
(ReLU(x) = max(0, x) i.e. this function replaces all values less than zero with zero and leaves values
greater than zero unchanged),  
 – is a layer that converts the output of the  
2 layer into
a one-dimensional vector of size 64 ∗ ( − 1) ∗ (
− 1), which is then fed to the input of the fully
connected layers, 
vector of dimension 64 ∗ (
1, 
− 1) ∗ (
2,</p>
          <p>3are fully connected layers: the first layer converts a
− 1) into a vector of 64 elements, the second layer supports
dimension 64, the third linear layer outputs a vector whose size is equal to the number of actions,

– is a transfer function
   
that the final 
Next, we present an Algorithm 2 for training IDQN models   (2).</p>
          <p>– is a layer that processes the output of the last linear layer, ensuring
values for discrete actions (traffic light phase switching) are obtained.</p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>Algorithm 2.</title>
        <p>Training Reinforcement Deep Learning Models   .</p>
        <p>1: Input: properties of the intersection and traffic light and properties of the traffic scenario at each
moment in time of the simulation according to Algorithm 1
2: For i=1… 
3: Forming the input matrix    for  
4: The input matrix InputLayer passes through the convolutional layer Conv2d to extract features and
learn patterns between traffic lanes, and then through the Flatten layer and fully connected layers
(Linear1, Linear2, Linear3), which form a vector of Q-values.
5: Each element of the vector Q corresponds to the expected reward for a particular action, so its size is
equal to the number of possible actions.
6: Based on the obtained Q-values, an action is selected – that is, the action where Q is maximum. The
observation space determines the number of possible actions. It is used to form the last linear layer of
the network. Each element of this layer corresponds to a Q-value for a specific action.
7: End For
8: Apply the Reward function – the reward is received from the environment after performing the action
and is used to correct Q-values during training.
9: Imprint: Trained Models</p>
        <p>
          Thus, M-networks learn to predict the future reward for each action, allowing for optimal
decisions. Note that during training M uses: reward (from the environment), next state (calculated
after performing the action) obtained from motion simulation, discount factor γ∈[
          <xref ref-type="bibr" rid="ref1">0,1</xref>
          ] (if γ~1, then
the agent focuses on long-term benefits if γ~0 the agent prefers immediate rewards), which shows
the difference in the importance of future and current rewards, and target Q-values. We also note
that the agent’s observation space for an intersection will be a matrix of all lanes and their porperties.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results and Discussion</title>
      <p>To validate the proposed approach, a simple intersection with one lane in each direction was used
(Figure 2).</p>
      <p>Through this intersection, the flow of vehicles with different  2 emissions were simulated using
Algorithm 1. 12 routes were defined (from each direction to each possible target direction), and
traffic flows were generated in four-time intervals (0–25000, 25000–50000, 50000–75000, 75000–
100000 seconds). For each flow, the following were specified: route (in direction of travel), simulation
start and end times, intensity, initial speed, lane position, and lane selection.</p>
      <p>Thus, vehicles were generated at different times and with different intensities, following specific
routes through the intersection.</p>
      <p>Then, the IDQN model trained by Algorithm 2 on the generated traffic was compared with known
similar algorithms [23–25], according to the criteria: (a)  2 emissions, (2) the total waiting time of
all cars, and (3) the total time required for all cars to pass through the intersection.</p>
      <p>For the experimental study presented below, the following component reward functions were
used: (1):    
intersection,   
2 – total emissions</p>
      <p>ℎ – total length of the queue at the
 – total waiting time at the intersection,   
 
– average speed at
− 3 3 = −0.1 ∙ 
 

the intersection. Accordingly: − 1 1 = −1 ∙ 
/10 2, − 4 4 = 0.1 ∙</p>
      <p>2</p>
      <p>/10 3, − 2 2 = −0.1 ∙
/10 2.</p>
      <p>Next, we present the obtained results. Figure 3 shows a comparison of 
emissions for four
approaches (approaches [23–25] and ours) after 100 training cycles. The vertical axis shows the 
level in mg, and the horizontal axis shows the training cycles of the neural network. As can be seen
2

  ℎ
from the graph, for our proposed approach,</p>
      <p>2
approaches (ours and approaches [23–25] after 100 training cycles.</p>
      <p>Next, we present the results obtained. Figure 3 shows a comparison of 
emissions are the lowest.</p>
      <p>2
2
 

emissions for four</p>
      <sec id="sec-4-1">
        <title>The vertical axis shows the</title>
        <p>2 level in mg, and the horizontal axis shows the neural network
training cycles. From Figure 3, for our proposed approach, 
2
emissions are the lowest.
cycles. The vertical axis shows the total waiting time of the full simulation in seconds, and the
horizontal axis shows the neural network training cycles. As can be seen from the graph, the proposed
approach performs at the same level as other approaches.</p>
        <p>(a)
(b)
for all cars to travel through the intersection
emissions.
axis plots time in seconds, and the horizontal axis plots neural network training cycles. As can be
seen from the graph, the proposed approach performs slightly better than the others.</p>
        <p>The results obtained confirm the ability of the proposed intelligent model to obtain optimal traffic
light signal control for a local intersection using simulation modeling and DQN to reduce</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>The results in Table 2 show the ability of our approach to reduce  2 emissions by 18%. The
limitations might include the individual DQN for each intersection, the lack of synchronization
between intersections, and the non-transparency of decision-making in the DQN model.
This study presents an intelligent traffic light control approach based on DRL, specifically employing
a DQN, to reduce CO2 emissions while considering the interests of all road users. The primary
contribution lies in the development and simulation-based validation of a DQN model and a tailored
loss function that enables adaptive traffic signal control at a local intersection. Through extensive
simulation experiments using realistic traffic patterns, the proposed model achieved a measurable
reduction in CO2 emissions by approximately 18%, highlighting the potential of DRL in addre ssing
environmental and operational challenges in urban traffic management. The limitations include the
need to train a separate DQN model for each intersection, the absence of coordination between
adjacent intersections affecting overall traffic flow, and the lack of interpretability in the model’s
decision-making process, which hinders transparency, trust, and practical deployment.</p>
      <p>Future work will focus on addressing these challenges by incorporating real-time input data from
surveillance cameras and other sensors to enable online learning and context-aware signal control.
 2, mg</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This research was supported by the U_CAN –“Towards carbon neutrality of Ukrainian cities” project,
under Grant Agreement No. 101148374, funded by the Horizon Europe program.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
[14] Z. Wang, L. Xu, J. Ma, Carbon dioxide emission reduction-oriented optimal control of traffic
signals in mixed traffic flow based on deep reinforcement learning, Sustainability 15.24 (2023)
16564. doi:10.3390/su152416564.
[15] M. Schumacher, C. M. Adriano, H. Giese, Challenges in reward design for reinforcement
learning-based traffic signal control: An investigation using a CO2 emission objective, SUMO
Conf. Proc. 4 (2023) 131–151. doi:10.52825/scp.v4i.222.
[16] P. C. Sherimon, V. Sherimon, J. Joy, A. M. Kuruvilla, G. Arundas, Efficient deep learning
methods for detecting road accidents by analyzing traffic accident images, Int. J. Comput. 23.3
(2024) 440–449. doi:10.47839/ijc.23.3.3664.
[17] A. Sachenko, V. Kochan, V. Turchenko, Intelligent distributed sensor network, in: IMTC/98
Conference Proceedings. IEEE Instrumentation and Measurement Technology Conference.
Where Instrumentation is Going (Cat. No.98CH36222), IEEE, New York, NY, USA, 1998, pp. 60–
66. doi:10.1109/IMTC.1998.679663.
[18] H. Abohashima, A. Eltawil, M. Gheith, A mathematical programming model and a firefly-based
heuristic for real-time traffic signal scheduling with physical constraints, IEEE Access 9 (2021)
128314–128327. doi:10.1109/access.2021.3112600.
[19] E. Manziuk, O. Barmak, I. Krak, O. Mazurets, T. Skrypnyk, Formal model of trustworthy artificial
intelligence based on standardization, in: Proc. of the 2nd International Workshop on Intelligent
Information Technologies &amp; Systems of Information Security with CEUR-WS, CEUR-WS.org,
Aachen, 2021, pp. 190–197. URL: https://ceur-ws.org/Vol-2853/short18.pdf.
[20] T. A. Haddad, D. Hedjazi, S. Aouag, A new deep reinforcement learning-based adaptive traffic
light control approach for isolated intersection, in: 2022 5th International Symposium on
Informatics and its Applications (ISIA), IEEE, New York, NY, USA, 2022, pp. 1–6.
doi:10.1109/isia55826.2022.9993598.
[21] C. Ounoughi, G. Touibi, S. B. Yahia, EcoLight: Сco-friendly traffic signal control driven by urban
noise prediction, in: Database and Expert Systems Applications. DEXA 2022, Vol. 13426: Lecture
Notes in Computer Science, Springer International Publishing, Cham, 2022, pp. 205–219.
doi:10.1007/978-3-031-12423-5_16.
[22] I. Sutskever, O. Vinyals, Q. V. Le, Sequence to sequence learning with neural networks, in:
Proceedings of the 28th international conference on neural information processing systems
volume 2, MIT Press, Cambridge, MA, USA, 2014, pp. 3104–3112. doi:10.5555/2969033.2969173.
[23] C. Chen, H. Wei, N. Xu, G. Zheng, M. Yang, Y. Xiong, K. Xu, Z. Li, Toward a thousand lights:
Decentralized deep reinforcement learning for large-scale traffic signal control, Proc. AAAI Conf.</p>
      <p>Artif. Intell. 34.04 (2020) 3414–3421. doi:10.1609/aaai.v34i04.5744.
[24] J. Ault, J. Hanna, G. Sharon, Learning an interpretable traffic signal control policy, in: Proc. of
the 19th International Conference on Autonomous Agents and MultiAgent Systems (AAMAS
‘20), International Foundation for Autonomous Agents and Multiagent Systems, Richland, SC,
USA, 2020, pp. 88–96. URL: https://ifaamas.org/Proc./aamas2020/pdfs/p88.pdf.
[25] S. Vladov, L. Scislo, V. Sokurenko, O. Muzychuk, V. Vysotska, S. Osadchy, A. Sachenko, Neural
network signal integration from thermogas-dynamic parameter sensors for helicopters
turboshaft engines at flight operation conditions, Sensors 24.13 (2024) 4246.
doi:10.3390/s24134246.
[26] P. Radiuk, O. Barmak, E. Manziuk, I. Krak, Explainable deep learning: A visual analytics
approach with transition matrices, Mathematics 12.7 (2024) 1024. doi:10.3390/math12071024.
[27] J. Ault, G. Sharon, Reinforcement learning benchmarks for traffic signal control, in: Proc. of the
Thirty-fifth Conference on Neural Information Processing Systems (NeurIPS 2021) Datasets and
Benchmarks Track, 2021, pp. 1–11. URL:
https://people.engr.tamu.edu/guni/papers/NeurIPSsignals.pdf.
[28] V. Mnih, K. Kavukcuoglu, D. Silver, E. Al, Human-level control through deep reinforcement
learning, Nature 518.7540 (2015) 529–533. doi:10.1038/nature14236.
[29] P. Alvarez Lopez, A. Banse, M. Barthauer, M. Behrisch, B. Couéraud, J. Erdmann, Y.-P. Flötteröd,
R. Hilbrich, R. Nippold, P. Wagner, Simulation of urban mobility (SUMO), Software, v. 1.21.0,
Zenodo, 2024. doi:10.5281/zenodo.13907886.
[30] D. Krajzewicz, M. Behrisch, P. Wagner, R. Luz, M. Krumnow, Second generation of pollutant
emission models for SUMO, in: Modeling Mobility with Open Data, Lecture Notes in Mobility,
Springer International Publishing, Cham, 2015, pp. 203–221. doi:10.1007/978-3-319-15024-6_12.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>General</given-names>
            <surname>Assembly United Nations</surname>
          </string-name>
          ,
          <article-title>Progress towards the Sustainable Development Goals</article-title>
          .
          <article-title>Report of the Secretary-</article-title>
          <string-name>
            <surname>General</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          /79/79-E/
          <year>2024</year>
          /54,
          <string-name>
            <given-names>General</given-names>
            <surname>Assembly</surname>
          </string-name>
          . Economic and
          <string-name>
            <given-names>Social</given-names>
            <surname>Council</surname>
          </string-name>
          , New York City, NY, USA,
          <year>2024</year>
          . URL: https://unstats.un.org/sdgs/files/report/2024/SG-SDGProgress
          <string-name>
            <surname>-Report-</surname>
          </string-name>
          2024
          <string-name>
            <surname>-</surname>
          </string-name>
          advanced-unedited-version.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>O.</given-names>
            <surname>Ryzhanskyi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Manziuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          , I. Krak,
          <string-name>
            <given-names>N.</given-names>
            <surname>Bacanin</surname>
          </string-name>
          ,
          <article-title>An approach to optimizing CO2 emissions in traffic control via reinforcement learning</article-title>
          ,
          <source>in: Proc. of the 5th International Workshop on Intelligent Information Technologies &amp; Systems of Information Security with CEUR-WS, CEUR-WS.org, Aachen</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>137</fpage>
          -
          <lpage>155</lpage>
          . URL: https://ceur-ws.org/Vol3675/paper10.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>V.</given-names>
            <surname>Pavlyshyn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Manziuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Damasevicius</surname>
          </string-name>
          ,
          <article-title>Modeling environment intelligent transport system for eco-friendly urban mobility</article-title>
          ,
          <source>in: Intelligent Information Technologies \&amp; Systems of Information Security</source>
          <year>2024</year>
          ,
          <article-title>CEUR-WS</article-title>
          .org, Khmelnytskyi,
          <year>2024</year>
          , pp.
          <fpage>118</fpage>
          -
          <lpage>136</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3675</volume>
          /paper9.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>I.</given-names>
            <surname>Krak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Kuznetsov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kondratiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Azarova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Barmak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Padiuk</surname>
          </string-name>
          ,
          <article-title>Analysis of deep learning methods in adaptation to the small data problem solving</article-title>
          , in: S.
          <string-name>
            <surname>Babichev</surname>
          </string-name>
          , V. Lytvynenko (Eds.),
          <source>Lecture Notes in Data Engineering, Computational Intelligence, and Decision Making</source>
          , Vol.
          <volume>149</volume>
          : LNDECT, Springer International Publishing, Cham,
          <year>2023</year>
          , pp.
          <fpage>333</fpage>
          -
          <lpage>352</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          - 16203-9_
          <fpage>20</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Haydari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yilmaz</surname>
          </string-name>
          ,
          <article-title>Deep reinforcement learning for intelligent transportation systems: A survey</article-title>
          ,
          <source>IEEE Trans. Intell. Transp. Syst. 23.1</source>
          (
          <year>2020</year>
          )
          <fpage>11</fpage>
          -
          <lpage>32</lpage>
          . doi:
          <volume>10</volume>
          .1109/tits.
          <year>2020</year>
          .
          <volume>3008612</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>C.</given-names>
            <surname>Louw</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Labuschagne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Woodley</surname>
          </string-name>
          ,
          <article-title>A comparison of reinforcement learning agents applied to traffic signal optimisation</article-title>
          ,
          <source>SUMO Conf. Proc. 3</source>
          (
          <year>2022</year>
          )
          <fpage>15</fpage>
          -
          <lpage>43</lpage>
          . doi:
          <volume>10</volume>
          .52825/scp.v3i.
          <fpage>116</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B.</given-names>
            <surname>Setiyono</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Sulistyaningrum</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Soetrisno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. W.</given-names>
            <surname>Wicaksono</surname>
          </string-name>
          <article-title>, Multi vehicle speed detection using euclidean distance based on video processing</article-title>
          ,
          <source>Int. J. Comput. 18.4</source>
          (
          <year>2019</year>
          )
          <fpage>431</fpage>
          -
          <lpage>442</lpage>
          . doi:
          <volume>10</volume>
          .47839/ijc.18.4.1613.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P. A.</given-names>
            <surname>Lopez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Behrisch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bieker-Walz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Erdmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.-P.</given-names>
            <surname>Flötteröd</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Hilbrich</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Lücken</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rummel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wagner</surname>
          </string-name>
          , E. Wiessner,
          <article-title>Microscopic traffic simulation using SUMO</article-title>
          ,
          <source>in: 2018 21st International Conference on Intelligent Transportation Systems (ITSC)</source>
          , IEEE, New York, NY, USA,
          <year>2025</year>
          , pp.
          <fpage>2575</fpage>
          -
          <lpage>2582</lpage>
          . doi:
          <volume>10</volume>
          .1109/ITSC.
          <year>2018</year>
          .
          <volume>8569938</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Chala</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. T.</given-names>
            <surname>Kóczy</surname>
          </string-name>
          ,
          <article-title>Intelligent fuzzy traffic signal control system for complex intersections using fuzzy rule base reduction</article-title>
          ,
          <source>Symmetry</source>
          <volume>16</volume>
          .9 (
          <year>2024</year>
          )
          <article-title>1177</article-title>
          . doi:
          <volume>10</volume>
          .3390/sym16091177.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Game theory-based signal control considering both pedestrians and vehicles in connected environment</article-title>
          ,
          <source>Sensors</source>
          <volume>23</volume>
          .23 (
          <year>2023</year>
          )
          <article-title>9438</article-title>
          . doi:
          <volume>10</volume>
          .3390/s23239438.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Faqir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Loqman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Boumhidi</surname>
          </string-name>
          ,
          <article-title>Combined extreme learning machine and max pressure algorithms for traffic signal control</article-title>
          ,
          <source>Intell. Syst. With Appl</source>
          .
          <volume>19</volume>
          (
          <year>2023</year>
          )
          <article-title>200255</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.iswa.
          <year>2023</year>
          .
          <volume>200255</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>B.</given-names>
            <surname>Großwindhager</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rupp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tappler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tranninger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Weiser</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. K.</given-names>
            <surname>Aichernig</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Boano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Horn</surname>
          </string-name>
          , G. Kubin,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mangard</surname>
          </string-name>
          , et al.,
          <article-title>Dependable internet of things for networked cars</article-title>
          ,
          <source>Int. J. Comput. 16.4</source>
          (
          <year>2017</year>
          )
          <fpage>226</fpage>
          -
          <lpage>237</lpage>
          . doi:
          <volume>10</volume>
          .47839/ijc.16.4.911.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>I.</given-names>
            <surname>William</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kozhevnikov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sontheimer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chou</surname>
          </string-name>
          ,
          <article-title>Enhancing urban traffic management in Taipei: A reinforcement learning approach</article-title>
          , in: 2024
          <source>Smart City Symposium Prague (SCSP)</source>
          , IEEE, New York, NY, USA,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/scsp61506.
          <year>2024</year>
          .
          <volume>10552687</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>