<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Predicting Pseudo-Random and Quantum Random Number Sequences using Hybrid Deep Learning Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Dmytro</forename><surname>Proskurin</surname></persName>
							<email>proskurin.d@stud.nau.edu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">National Aviation University</orgName>
								<address>
									<addrLine>1 Liubomyra Huzara Ave</addrLine>
									<postCode>03058</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sergiy</forename><surname>Gnatyuk</surname></persName>
							<email>s.gnatyuk@nau.edu.ua</email>
							<affiliation key="aff0">
								<orgName type="institution">National Aviation University</orgName>
								<address>
									<addrLine>1 Liubomyra Huzara Ave</addrLine>
									<postCode>03058</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tetiana</forename><surname>Okhrimenko</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">National Aviation University</orgName>
								<address>
									<addrLine>1 Liubomyra Huzara Ave</addrLine>
									<postCode>03058</postCode>
									<settlement>Kyiv</settlement>
									<country key="UA">Ukraine</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Predicting Pseudo-Random and Quantum Random Number Sequences using Hybrid Deep Learning Models</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">9923A238083C2976DBACCDFF290F4746</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-12-29T06:05+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Random numbers</term>
					<term>RNN</term>
					<term>CNN</term>
					<term>LSTM</term>
					<term>GRU</term>
					<term>Hybrid model</term>
					<term>Secure communication</term>
					<term>PRNG</term>
					<term>QRNG</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Predicting random number sequences has significant implications for cryptography and secure communication systems. In this paper, a hybrid deep learning model was proposed, it combines Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, and RNNs to predict pseudo-random number generator (PRNG) and quantum random number generator (QRNG) sequences. Proposed model was compared with traditional CNNs, LSTMs, and RNNs models. Given results showed that the hybrid model outperforms the other models, providing better prediction accuracy for PRNG and QRNG sequences.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Random number generation is a crucial component of many applications, including cryptography, secure communication systems, simulations, and probabilistic algorithms. Pseudo-random number generators (PRNGs) and quantum random number generators (QRNGs) are two main types of random number generators, with QRNGs providing better security due to their inherent unpredictability <ref type="bibr" target="#b0">[1]</ref>. However, predicting PRNG and QRNG sequences remains an essential task to assess their security and reliability. Deep learning techniques, such as Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, and RNNs, have been extensively used in various time series prediction tasks <ref type="bibr" target="#b1">[2]</ref>. In this paper, was proposed a hybrid deep learning model that combines CNNs, LSTMs, and RNNs to predict PRNG and QRNG sequences. The model is trained and evaluated on a dataset containing both PRNG and QRNG sequences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>Several studies have explored the use of deep learning techniques for predicting random number sequences. For instance, the use of CNNs and LSTMs has been reported in predicting PRNG sequences <ref type="bibr" target="#b2">[3]</ref>. In another study, RNNs have been employed to predict QRNG sequences <ref type="bibr" target="#b3">[4]</ref>. However, there is limited research on hybrid deep learning models that combine multiple neural network architectures to predict PRNG and QRNG sequences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Goal of the Research</head><p>The primary goal of this research is to investigate the effectiveness of various deep learning architectures, including MLP, CNN, LSTM, and RNNs models, for the task of predicting the next value in a sequence of random numbers generated by a combination of PRNG and QRNG sources. By exploring different neural network architectures, was aimed to identify the most suitable model for this problem, considering aspects such as predictive accuracy, model complexity, and training time. Another objective is to assess whether the trained models can achieve better prediction results than a random baseline, indicating that they have learned meaningful patterns in the data. To ensure a fair comparison, will be used appropriate evaluation metrics, such as Mean Squared Error (MSE) and Mean Absolute Error (MAE), to quantify the performance of each model and compare it against a random prediction benchmark. Finally, was aimed to provide insights into the practical implications of using deep learning models to predict random number sequences generated from quantum sources, as well as discussing potential future research directions in this field. By understanding the strengths and limitations of various models for this task, authors hope to contribute to the development of more advanced techniques for analysing and predicting random number sequences in different contexts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology</head><p>The dataset used in this study consists of PRNG and QRNG sequences generated using various algorithms, such as the Mersenne Twister, Linear Congruential Generator, and a commercial QRNG device <ref type="bibr" target="#b4">[5]</ref>. The dataset is divided into training, validation, and test sets, ensuring a balanced representation of PRNG and QRNG sequences in each set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Model Architecture</head><p>The proposed hybrid deep learning model combines the strengths of CNNs, LSTMs, and RNNs to predict PRNG and QRNG sequences. The model consists of a CNN layer for feature extraction, followed by an LSTM layer to capture temporal dependencies, and a RNNs layer for capturing long-range dependencies. The final output is a single linear activation unit that produces the predicted value. The model is trained using the Adam optimizer and mean squared error (MSE) as the loss function <ref type="bibr" target="#b5">[6]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Results Analysis</head><p>The first step in analysing the model's performance was to visually inspect the predicted values against the true values. This was achieved by plotting the first 100 true values and the corresponding predicted values on the same graph. This visualization allows us to assess the overall fit of the model to the data and identify any noticeable discrepancies between the predicted and true values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Similarity Assessment</head><p>To quantify the similarity between the true values and the predicted values, was calculated the Pearson correlation coefficient. This metric measures the linear relationship between two datasets, with a value close to 1 indicating a strong positive relationship. A pre-defined threshold of 0.9 was used to determine whether the predicted values were considered close to the true values. Based on the computed correlation coefficient, was concluded whether the model's predictions were close to the true values or not.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Model Performance Comparison</head><p>To assess the effectiveness of the model, its performance was compared against a random prediction baseline. This was done by generating random predictions within the same range as the true values and calculating the Mean Squared Error (MSE) for both the model's predictions and the random predictions. By comparing these MSE values, we were able to determine whether the GRU model's predictions were better than random ones. The results from the visual inspection, similarity assessment, and model performance comparison provide a comprehensive analysis of the model's performance in predicting the next value in a sequence of random numbers. These findings contribute to our understanding of the model's effectiveness for this specific task and offer insights into potential improvements or alternative approaches.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Experiments and Results</head><p>The hybrid model is trained on the dataset and its performance is compared with traditional RNNs, CNNs, and LSTMs. The models are evaluated using the Pearson correlation coefficient and mean squared error (MSE) to assess the similarity between the true and predicted values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.1">Simple RNN</head><p>The simple RNN is the most basic form of a recurrent neural network, characterized by its single hidden layer that takes input from the previous time step and feeds it back into the network for the next time step (Fig. <ref type="figure">1</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 1: Simple RNN architecture</head><p>Despite its simplicity, the performance of simple RNNs in predicting PRNG and QRNG sequences is limited due to their inability to capture long-range dependencies as a result of the vanishing gradient problem (Fig. <ref type="figure" target="#fig_0">2</ref>). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.2">Gated Recurrent Unit (GRU)</head><p>The GRU is an advanced RNN architecture that addresses the vanishing gradient problem observed in simple RNNs. With the introduction of gating mechanisms, GRUs can learn when to update the hidden state and when to maintain the existing state, allowing them to capture longer-range dependencies more effectively (Fig. <ref type="figure" target="#fig_1">3</ref>).  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.3">Bidirectional RNN</head><p>Bidirectional RNNs process the input sequence in both forward and backward directions, enabling the network to capture information from both past and future time steps. This capability proves useful for tasks where context from both directions is important, such as natural language processing and speech recognition (Fig. <ref type="figure">5</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 5: Bidirectional RNN architecture</head><p>In the context of PRNG and QRNG sequence prediction, bidirectional RNNs exhibit enhanced performance due to their ability to incorporate information from the entire sequence (Fig. <ref type="figure" target="#fig_3">6</ref>). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.4">Stacked RNN</head><p>A stacked RNN architecture consists of multiple layers of RNNs stacked on top of each other, allowing the network to learn more complex features and representations of the input sequence. This increased complexity can lead to improved prediction performance for PRNG and QRNG sequences (Fig. <ref type="figure" target="#fig_4">7</ref>).  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.5">Convolutional Neural Networks</head><p>CNNs have shown success in time series prediction tasks due to their ability to capture local patterns and dependencies <ref type="bibr" target="#b6">[7]</ref>. In our experiments, a CNN model is trained on the PRNG and QRNG sequences dataset (Fig. <ref type="figure">9</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 9: CNN architecture</head><p>The results indicate that the CNN model can capture some local patterns in the sequences, but struggles to predict long-range dependencies, leading to suboptimal prediction accuracy (Fig. <ref type="figure" target="#fig_6">10</ref>). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.6">Long Short-Term Memory Networks</head><p>LSTMs are designed to capture long-term dependencies in time series data <ref type="bibr" target="#b7">[8]</ref><ref type="bibr" target="#b8">[9]</ref><ref type="bibr" target="#b9">[10]</ref>. Was trained an LSTM model on the PRNG and QRNG sequences dataset and evaluate its performance (Fig. <ref type="figure">11</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 11: Long Short-Term Memory architecture</head><p>The results show that the LSTM model can capture temporal dependencies in the sequences, but its performance is limited by the absence of feature extraction capabilities (Fig. <ref type="figure" target="#fig_7">12</ref>) <ref type="bibr" target="#b10">[11]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.7">Hybrid Deep Learning Model</head><p>The proposed hybrid model combines the strengths of CNNs, LSTMs, and RNNs to predict PRNG and QRNG sequences (Fig. <ref type="figure" target="#fig_8">13</ref>) <ref type="bibr" target="#b11">[12,</ref><ref type="bibr" target="#b12">13]</ref>. The model's performance is compared with the other models, and the results show that the hybrid model outperforms the traditional CNNs, LSTMs, and RNNs models, providing better prediction accuracy for PRNG and QRNG sequences (Fig. <ref type="figure" target="#fig_2">14</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Figure 14: Hybrid Deep Learning performance</head><p>It can be observed numerous instances where the models were able to predict the exact value or a very close trend in PRNG and QRNG sequences (Fig. <ref type="figure" target="#fig_10">15, 16</ref>). These instances demonstrate the effectiveness of the models in understanding the underlying patterns and dependencies within the data, as well as their capability to generalize and make accurate predictions on unseen data.  Furthermore, it was observed that the models were often able to predict a close trend in the sequences, even if the exact value was not pinpointed (Fig. <ref type="figure" target="#fig_12">17</ref>  This indicates that the models have a strong grasp of the overall dynamics and structure of the data, enabling them to generate predictions that closely follow the actual trajectory of the PRNG and QRNG sequences <ref type="bibr" target="#b13">[14]</ref><ref type="bibr" target="#b14">[15]</ref><ref type="bibr" target="#b15">[16]</ref>. This level of trend identification can prove beneficial in scenarios where understanding the general direction or pattern of the data is more critical than pinpointing individual values <ref type="bibr" target="#b16">[17]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conclusions and Future Work</head><p>In this paper, was presented a novel hybrid deep learning model that combines the strengths of Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) networks, and RNNs to predict Pseudo-Random Number Generator (PRNG) and Quantum Random Number Generator (QRNG) sequences. Our results demonstrate that the hybrid model outperforms traditional CNNs, LSTMs, and RNNs models in terms of prediction accuracy for both PRNG and QRNG sequences.</p><p>As part of the future work, it is planned to explore other hybrid model architectures that could further enhance the performance of our current model. Also was aimed to investigate the use of additional features, such as information from the frequency domain, to improve the prediction capabilities of our model. Furthermore, authors intend to study the generalizability of our hybrid model to other sequence prediction tasks. These tasks may include predicting cryptographic keys, secure communication protocols, and other securityrelated applications. Additionally, will be considered the development of more robust and efficient training strategies to ensure that proposed model remains effective even in the face of rapidly evolving security threats. By continuing to enhance and refine our hybrid deep learning model, authors hope to contribute to the advancement of secure communications and data protection in the digital age.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Simple RNN performance</figDesc><graphic coords="3,99.20,492.64,456.10,124.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Gated Recurrent Unit architecture</figDesc><graphic coords="4,230.52,56.70,193.14,175.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Gated Recurrent Unit performance</figDesc><graphic coords="4,97.03,297.75,459.92,124.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Bidirectional RNN performance</figDesc><graphic coords="5,98.93,94.65,456.62,124.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: Stacked RNN architecture</figDesc><graphic coords="5,214.02,336.46,226.23,345.80" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 8 :</head><label>8</label><figDesc>Figure 8: Stacked RNN performance</figDesc><graphic coords="6,97.78,94.65,458.70,124.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 10 :</head><label>10</label><figDesc>Figure 10: CNN performance</figDesc><graphic coords="7,100.02,451.37,454.22,124.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figure 12 :</head><label>12</label><figDesc>Figure 12: Long Short-Term Memory performance</figDesc><graphic coords="8,101.52,568.00,450.83,124.70" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 13 :</head><label>13</label><figDesc>Figure 13: Hybrid Deep Learning architecture</figDesc><graphic coords="9,215.32,125.74,223.90,528.21" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Figure 15 :</head><label>15</label><figDesc>Figure 15: Exact match 1</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Figure 16 :</head><label>16</label><figDesc>Figure 16: Exact match 2</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_11"><head></head><label></label><figDesc>).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_12"><head>Figure 17 :</head><label>17</label><figDesc>Figure 17: Close trend example</figDesc></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>This work is carried out within the framework of research grant №0122U002361"Intelligent system of secure packet data transmission based on reconnaissance UAV", funding by the Ministry of Education and Science of Ukraine during 2022-2023.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Quantum random number generators</title>
		<author>
			<persName><forename type="first">M</forename><surname>Herrero-Collantes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>Garcia-Escartin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Reviews of Modern Physics</title>
		<imprint>
			<biblScope unit="volume">89</biblScope>
			<biblScope unit="page">15004</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Deep Learning</title>
		<author>
			<persName><forename type="first">I</forename><surname>Goodfellow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Courville</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>MIT Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Forecasting Sunspot Numbers with LSTM</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Pérez-Jiménez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Membrane Computing</title>
				<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="153" to="166" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="5998" to="6008" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Random Number Generators</title>
		<author>
			<persName><forename type="first">G</forename><surname>Marsaglia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Modern Applied Statistical Methods</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="2" to="13" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">A method for stochastic optimization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adam</forename></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6980</idno>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Deep learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="volume">521</biblScope>
			<biblScope unit="page" from="436" to="444" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Research of Machine Learning Algorithms for the Development of Intrusion Detection Systems in 5G Mobile Networks and Beyond</title>
		<author>
			<persName><forename type="first">A</forename><surname>Imanbayev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tynymbayev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Odarchenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Sensors</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">24</biblScope>
			<date type="published" when="2022">2022. 9957</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Sequence to sequence learning with neural networks, Advances in neural information processing systems</title>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Vinyals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">27</biblScope>
			<biblScope unit="page" from="3104" to="3112" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Empirical evaluation of gated recurrent neural networks on sequence modelling</title>
		<author>
			<persName><forename type="first">J</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gulcehre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.3555</idno>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Real-time ML Algorithms for the Detection of Dangerous Objects in Critical Infrastructures</title>
		<author>
			<persName><forename type="first">I</forename><surname>Azarov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gnatyuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Aleksander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Azarov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukasheva</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3373</biblScope>
			<biblScope unit="page" from="217" to="226" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Content-Based Machine Learning Approach for Hardware Vulnerabilities Identification System</title>
		<author>
			<persName><forename type="first">G</forename><surname>Iashvili</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Avkurova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Iavich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bauyrzhan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gagnidze</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gnatyuk</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lecture Notes on Data Engineering and Communications Technologies</title>
		<imprint>
			<biblScope unit="volume">83</biblScope>
			<biblScope unit="page" from="117" to="126" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Integrated QKD and QRNG Photonic Technologies</title>
		<author>
			<persName><forename type="first">J</forename><surname>Aldama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sarmiento</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>López Grande</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Signorini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">T</forename><surname>Vidarte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Pruneri</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Lightwave Technology</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="issue">23</biblScope>
			<biblScope unit="page" from="7498" to="7517" />
			<date type="published" when="2022">1 Dec.1, 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Cryptographic key exchange method for data factorial coding</title>
		<author>
			<persName><forename type="first">E</forename><surname>Faure</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shcherba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Vasiliu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fesenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2654</biblScope>
			<biblScope unit="page" from="643" to="653" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Pseudo Quantum Random Number Generator with Quantum Permutation Pad</title>
		<author>
			<persName><forename type="first">R</forename><surname>Kuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mckenzie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Redding</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Quantum Computing and Engineering (QCE)</title>
				<meeting><address><addrLine>Broomfield, CO, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="359" to="364" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Random Sequence Generation using Superconducting Qubits</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Nene</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Third International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV)</title>
				<meeting><address><addrLine>Tirunelveli, India</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="640" to="645" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
