<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Redshift estimation on astronomy spectral data using CNN-based architecture with correction⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Iryna Yurchuk</string-name>
          <email>i.a.yurchuk@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vladyslav Skyruta</string-name>
          <email>vlad.skyruta@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>Volodymyrska str. 64/13, Kyiv, 01601</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <fpage>0000</fpage>
      <lpage>0001</lpage>
      <abstract>
        <p>Redshift is a fundamental characteristic of cosmic objects, particularly galaxies, stars, and quasars. By measuring the redshift of a galaxy, it is possible to determine its distance, velocity, and the expansion of the universe, as well as study the relative motion of cosmic structures. For investigating objects with high redshifts - where visible light is shifted into the infrared spectrum - scientists rely on advanced orbital telescopes, such as the James Webb Space Telescope. In this study, an efficient algorithm for estimating the redshift of galaxies based on their spectra has been developed. The algorithm has an architecture built upon CNNs, and incorporates a correction mechanism for the final results using laboratory-calibrated positions of key emission lines in galaxy spectra. Trained on a dataset of 10,000 samples consisting of starburst galaxies with redshifts ranging from 0 to 0.6, the model achieves a mean absolute error of 0.0169 without and 0.0139 with correction. This corresponds to an average relative accuracy of approximately 97.1% without and 97.7% with correction. The proposed method demonstrates significant potential for improving the accuracy and efficiency of redshift estimation in large-scale astronomical surveys.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;redshift estimation</kwd>
        <kwd>spectral analysis</kwd>
        <kwd>emission lines</kwd>
        <kwd>machine learning 1</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The study of redshift is a cornerstone of modern astrophysics and cosmology. It provides critical
insights into the structure, dynamics, and evolution of the universe. High-redshift objects, such as
distant galaxies and quasars, are particularly valuable for understanding the early stages of cosmic
evolution [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, accurately estimating redshift from galaxy spectra remains a challenging
task, especially for large datasets generated by modern astronomical surveys, such as SDSS, Euclid,
NASA, etc. In this context, redshift estimation stays as the central research object of the study.
      </p>
      <p>
        And this is where machine learning algorithms are jumping in. Machine learning is good at
handling large and complex datasets. It is an ideal tool for modern astronomical research, where huge
amounts of spectral data are generated by large-scale surveys and advanced telescopes ([
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]).
These algorithms are particularly good at identifying complicated patterns and correlations within
data that may be difficult or impossible to distinguish using traditional analytical methods. As such,
the application of machine learning techniques to galaxy spectra serves as the research subject
explored in this work.
      </p>
      <p>
        There have already been numerous studies that apply machine learning techniques to the problem
of redshift estimation. Many of these approaches utilize traditional machine learning models, such as
Random Forests or Support Vector Machines, as well as more recent deep learning architectures like
standard CNNs ([
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and [5]). These methods have demonstrated significant success in automating
redshift estimation and improving its accuracy compared to classical techniques. However, they
often face limitations in terms of scalability, computational efficiency and their ability to generalize
features across diverse datasets.
      </p>
      <p>In contrast, the approach presented in this work make use of advanced Inception network
architecture, which is specifically designed to capture multi-scale features within data. By
incorporating a correction mechanism based on laboratory-calibrated positions of key emission lines,
our method achieves higher precision in redshift estimation. This combination of advanced
architecture and post-processing refinement allows the algorithm to outperform existing methods,
particularly when applied to large datasets of galaxies with varying spectral characteristics. The
purpose of this study is to refine redshift estimation techniques by enhancing accuracy and scalability
for large astronomy datasets.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Theoretical Background and Related Work</title>
      <sec id="sec-2-1">
        <title>2.1. Theory and concepts</title>
        <p>When a beam of white light passes through a triangular prism, it disperses into its fundamental
components, known as the spectrum. Spectra play crucial role in modern astronomical research,
since they allow scientists to determine the qualitative and quantitative chemical composition of
celestial objects, their temperature, the presence of magnetic fields, their velocity and much more.</p>
        <p>As light leaves the surface of a cosmic object, it passes through its atmosphere. During this
process, certain wavelengths are absorbed by the chemical elements present in the atmosphere. This
absorption results in gaps or dark lines in the spectrum, which are known as Fraunhofer absorption
lines.</p>
        <p>In contrast to absorption lines, emission lines are bright lines that appear in the spectrum when
atoms or molecules emit light at specific wavelengths. These lines are produced when electrons in
atom transition from a higher energy level to a lower one, releasing energy in the form of photons.
Emission lines are typically observed in regions of hot, ionized gas, such as nebulae or the outer
layers of certain stars.</p>
        <p>Each chemical element has a unique spectral signature, consisting of specific absorption or
emission lines. In contrast, astronomical spectra, for example of galaxies, are complex combinations
of the spectra from numerous sources, including stars, gas clouds, and dust. These spectra include a
continuous background, absorption and emission lines.</p>
        <p>By analyzing both absorption and emission lines of astronomical spectra, astronomers can gain a
comprehensive understanding of the physical and chemical properties of celestial objects. For
instance, they can easily identify the chemical elements that absorbed or emitted the light, thereby
determining the chemical composition of the atmosphere or surrounding gas. Since the laws of
physics are assumed to operate similarly throughout the Universe, it is expected that the absorption
and emission lines of common chemical elements should appear the same everywhere.</p>
        <p>However, when astronomers observe the spectra of celestial objects, they almost always find that
their spectral lines are “shifted” toward the red (longer wavelength) end of the spectrum. This
phenomenon is known as the redshift.</p>
        <p>The redshift of light can occur due to several well-known phenomena. The most common cause is
the Doppler effect, where light from an object moving away from the observer is stretched to longer
wavelengths. Another significant cause is the cosmological redshift, which arises from the expansion
of the Universe, stretching the wavelengths of light as it travels through space. Additionally,
gravitational redshift occurs when light escapes a strong gravitational field, losing energy and
shifting to longer wavelengths.</p>
        <p>Redshift is a fundamental concept in astronomy. By observing how the color of a celestial object
changes over time, or how it differs from what was expected, researchers can discover a lot of
interesting facts about the object. Redshift quantifies how much the wavelength of electromagnetic
radiation detected on Earth has changed compared to the wavelength emitted by the source.</p>
        <p>It is, of course, impossible to physically capture the light beam of a distant star before it has been
redshifted during its journey to Earth. Therefore, astronomers use alternative methods to measure
redshift, the primary ones being the spectrographic and photometric methods.</p>
        <p>Spectrographs are highly powerful tools for studying light from cosmic objects. They split light
into its constituent colors, allowing scientists to analyze each color range in detail. By comparing the
absorption lines of the most common chemical elements in the Universe with their standard values,
astronomers can determine how much the entire spectrum has been “shifted” in comparison to its
original state when the light left the celestial body. This approach to measure redshift is called
spectrographic method.</p>
        <p>Photometers, on the other hand, measure the total intensity of light within a specific range of
frequencies. While this method is less detailed and affects the accuracy of redshift calculations – since
it considers only averaged values of the spectrum rather than its full details – it is faster and simpler.</p>
        <p>The photometric method of estimating redshift is particularly useful when analyzing large
clusters of cosmic objects, where precise measurements are not mandatory. However, when
astronomers conduct detailed studies of individual objects, they rely on spectral data and the
spectrographic method for determining redshift.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Review of Existing Approaches</title>
        <p>First of all, there is a significant number of studies that have been conducted to estimate photometric
redshift. Usually they take advantage of using traditional machine learning methods, as in [6] or [7].
Data Release 17 of SDSS also use kd-tree nearest neighbor fit, described in [8], to estimate
photometric redshifts efficiently. However, there are also plenty of examples of applying deep
learning techniques [9].</p>
        <p>Despite these features, photometric redshift estimation remains challenging due to the limitations
of photometric data. The lack of detailed spectral information often introduces uncertainties and
biases, which can impact the accuracy of the calculated redshifts.</p>
        <p>On the other hand, some studies, focus on calculating redshift using spectral data. This approach
is often considered more accurate than photometric methods, because spectral data provides detailed
information about the light emitted by celestial objects across a wide range of wavelengths, unlike
photometric methods, which rely on broad-band magnitudes.</p>
        <p>Since CNNs are good at extracting features from 1D data, there have been already conducted some
researches using them. For instance, in [10] besides the fact that authors transform the redshift
estimation problem from regression to classification by dividing the redshift range into discrete
intervals based on Euclid spectroscopic instrument, they use CNNs to classify galaxies into these
intervals as well.</p>
        <p>In addition to that, there is a study that utilize Bayesian CNNs inspired by VGG architectures to
estimate spectroscopic redshift [11]. It provides both predictions and uncertainty estimates, enabling
the identification of problematic spectra and balancing prediction accuracy with coverage. The
method outperforms traditional template-fitting techniques and helps discover misclassified or
unrecognized quasars.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Method</title>
      <p>The proposed approach aims to extract as much useful information as possible from spectral data to
achieve high-precision redshift estimation. Since spectra are essentially 1D data, CNNs are
wellsuited for analyzing them. In this work, we propose not only using a standard one- or multi-layer
CNN for feature extraction but also incorporating the Inception network [12].</p>
      <p>The Inception network is designed to capture features at multiple scales by applying
convolutional filters of different sizes in parallel. This architecture potentially is advantageous for
spectral data, as it allows the model to simultaneously detect both narrow features, such as emission
or absorption lines, and broader patterns, such as continuum variations.</p>
      <p>In addition to the primary model, an additional correction method is introduced to further refine
the redshift estimation. After the initial redshift is predicted by the main model, the spectrum is
“shifted” back to its born state based on the calculated redshift. In this initial state, emission lines are
basically aligned with their expected laboratory-calibrated positions. Using these corrected spectra,
the model adjusts its predictions by recalculating the redshift based on the ideal positions of the
emission lines and considering spectra characteristics, such as peaks and flat regions. This additional
correction step ensures more precise redshift value.</p>
      <sec id="sec-3-1">
        <title>3.1. Data preparation</title>
      </sec>
      <sec id="sec-3-2">
        <title>3.1.1. Key notes about the dataset</title>
        <p>The model will be evaluated on Data Release 18 of SDSS [13], the latest and most comprehensive
dataset from the Sloan Digital Sky Survey, offering high-quality spectral data across a wide range of
celestial objects.</p>
        <p>Since stars, QSOs, and galaxies exhibit fundamentally different features in their spectra, this study
focuses only on galaxy spectra with redshifts up to 1. It is worth mentioning that the proposed
approach can be adapted for stars and quasars as well, making it flexible for broader applications.</p>
        <p>To ensure a homogeneous dataset, only starburst galaxies are to be considered in this work.
Starburst galaxies are characterized by a recent and transient increase in their SFR, often by a factor
of up to 50 compared to regular galaxies. The strong radio emissions of starburst galaxies make it
easier for machine learning models to extract meaningful features from their spectra, thus enhancing
the reliability and precision of redshift estimation. It is expected that the proposed approach will also
work for regular galaxies, though this would require reconsidering the data preprocessing methods
described in the next subsubsection.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.1.2. Models inputs</title>
        <p>First of all, let’s define the structure of the input data and highlight its key features relevant to this
work. Since there are two models involved – the main model and the correction model – each
requires a distinct set of inputs. However, both models share a common input – the spectrum, which
represents the flux values (in 10-17 erg/s/cm2/Å) corresponding to specific wavelengths (in angstroms
Å).</p>
        <p>Typically, galaxy spectra provide flux information within the range of approximately 4000 to 9000
angstroms, as this range covers key spectral features such as the Balmer series, prominent emission
lines and other important indicators of galaxy properties. This range is also well-suited for optical
observations, as it falls within the sensitivity of most ground-based telescopes and spectrographs.</p>
        <p>In addition to the spectrum, the main model requires the redshift value as an input. Redshift,
denoted as z, is defined as the fractional change in the wavelength of light due to various physical
phenomena. It can be calculated using the formula
z= λobs− λrest ,
λrest
(1)
where λobs is a wavelength measured by the observer, and λrest is the wavelength at the source.</p>
        <p>The correction model, on the other hand, takes as input the spectra shifted back to their emitted
states, along with the template positions of several well-known and widely distributed emission lines
(e.g. Hα, Hβ, [OIII], [NII]). After shifting the spectra to their source states, any empty values that arise
due to the shift are filled with zeros, while values that fall outside the defined above wavelength range
are removed.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.1.3. Data preprocessing</title>
        <p>Since spectra are 1D signals often containing noise from various sources, wavelet filtering is an
effective solution for denoising while preserving important features. Wavelet filters are a
wellestablished tool for noise reduction and are widely used in various applications. For instance, in [15],
wavelet filtering is applied to detect anomalies and emergency states in signals by isolating key
characteristics after removing noise.</p>
        <p>For noise reduction, the Daubechies wavelet is commonly used due to its ability to capture smooth
and localized features in 1D signals. The filtering process involves decomposing the signal into
wavelet coefficients across multiple scales, where noise is typically concentrated in the
highfrequency components. A soft thresholding technique is applied to suppress these noisy coefficients,
ensuring that the essential features of the signal are preserved. After thresholding, the signal is
reconstructed by combining the modified coefficients, resulting in a cleaner version of the original
data.</p>
        <p>It’s worth mentioning that there is no need to completely remove all features from the spectra, as
some degree of noise is acceptable if it preserves the essential characteristics of the signal. The chosen
threshold, set to 10% of the maximum coefficient at the finest scale, keeps a balance between effective
noise reduction and retaining meaningful spectral features, as shown on Figure 3. This value ensures
that weaker but important components, such as emission lines, are not completely suppressed, while
still keeping the majority of high-frequency noise.</p>
        <p>The next step in the data preprocessing pipeline involves applying spline approximation to the
spectrum. This technique is very useful for starburst galaxies, whose spectra are characterized by a
relatively flat continuum with distinct emission lines standing out prominently.</p>
        <p>Then, by calculating the difference between the wavelet-filtered spectrum and the spline
approximation line, we align the spectrum with the zero axis, removing base flux variations and
focusing on relative flux changes (Figure 4). While this process sacrifices some physical context, it is
highly beneficial for further training the convolution-based networks, as they are better suited to
analyze relative variations rather than absolute flux values. Finally, the resulting difference is
normalized by dividing it by its maximum value, ensuring consistent scaling across all spectra and
preparing the data for efficient feature extraction and model training.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.2. Building the machine learning model</title>
        <p>As mentioned earlier, 1D CNNs are widely used for processing sequential data, such as time series or
audio signals, to detect important patterns or identify unwanted anomalies. Given the sequential
nature of astronomical spectra, they are excellent candidates for extending the scope of 1D CNN
applications to the field of astronomy.</p>
        <p>In general, CNNs (Convolutional Neural Networks) are a class of deep learning models designed
to automatically and adaptively learn spatial hierarchies of features from data. The core operation in
CNNs is the convolution, which involves applying a filter to the input data to extract features.
Mathematically, a convolution operation for a 1D input can be expressed as
y ( i )=∑ x ( j )∗w ( i− j ) , (2)</p>
        <p>j
where x(j) is the input signal, w(i-j) is a filter, and y(i) is the resulting feature map.</p>
        <p>A 1D CNN is a specialized type of CNN where the convolution operation is applied along a single
dimension.</p>
        <p>One of the enhancements of CNN demonstrated in this work is the Inception network
architecture [12]. The main idea behind it is to process input data at multiple scales simultaneously by
using parallel convolutional layers with different kernel sizes.</p>
        <p>The Inception architecture was chosen for redshift prediction due to its ability to detect in
multiple scales both narrow spectral features, such as emission lines, and broader patterns, such as
continuum variations. Additionally, its modular design allows for efficient computation.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.3. Correction method</title>
        <p>The correction method is designed to further enhance the accuracy of the overall algorithm. In
regression tasks, convolutional networks typically return a single redshift value, which is not directly
tied to specific spectral features, such as emission or absorption lines. To address this limitation, we
propose a correction mechanism that aligns the predicted redshift value with the spectral
characteristics.</p>
        <p>But, in the first place it is necessary to calculate the positions of emission lines based on the
estimated redshift value using Formula 1. These calculated positions allow us to adjust the emission
lines, aligning them more closely with the observed spectrum peaks, thereby refining the redshift
estimation.</p>
        <p>To begin with, the most naïve method to adjust emission line positions is a greedy algorithm. The
idea behind this approach is to iteratively refine the predicted positions of emission lines by
searching for the best match within a local region of the spectrum.</p>
        <p>For each predicted position, the algorithm evaluates nearby points within a defined search radius,
considering three main factors:



proximity to the initial prediction;
intensity of the spectrum at the point;
relative distance between neighboring emission lines.</p>
        <p>The algorithm assigns a cost to each candidate position based on these criteria and selects the one
with the lowest cost. This process ensures that the corrected positions align more closely with the
actual peaks in the spectrum while maintaining consistent spacing between emission lines.</p>
        <p>Another approach to refine emission line positions involves training a machine learning model
using local regions of the spectrum. For that, gradient boosting method is used. It is an ensemble
learning algorithm that builds a series of decision trees. They can be trained to solve either
classification or regression tasks. In our case, the regressor is built to predict positions of emissions
lines.</p>
        <p>Precise estimation of all mentioned algorithms will be presented in the next section.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experimental Evaluation</title>
      <p>Important, that the input spectra do not share a common wavelength grid, which complicates direct
comparisons and processing. Each spectrum is sampled at different wavelengths, making it necessary
to align all spectra to a unified grid before further analysis. To fix this issue, a common wavelength
grid was defined in the range from 4000 Å to 9000 Å with step equal to 1 Å. The flux values are
interpolated onto the common grid using a linear interpolation method. Chosen grid has direct
impact on the model performance and accuracy. A finer grid with smaller step sizes keeps more
precise representation of spectral features, improving the model performance. At the same time,
increased accuracy comes at the cost of higher computational complexity and memory usage, as the
model must process a larger number of data points.</p>
      <sec id="sec-4-1">
        <title>4.1. CNN model evaluation</title>
        <p>The evaluation of the algorithm begins with testing the main model, which is based on a CNN. For
this purpose, a single-layer CNN was implemented. The performance of the model varies depending
on the number of the convolutional filters and their sizes, as shown in Table 1.</p>
        <p>Trainings were conducted on the CNN using the ReLU activation function and the mean absolute
error loss function over 20 epochs. In addition to convolution filter, average pooling is applied with
filter size of 2.</p>
        <p>As it is shown in the Table 1, smaller convolution filters result in higher MAE values since they are
less effective at capturing broader spectral features. On the other hand, larger filters demonstrate
improved accuracy, with the lowest MAE achieved using 64 filters and a filter size of 150. Increasing
the number of filters to 128 does not consistently improve performance. While increasing the number
of filters can enhance feature extraction, it may also introduce redundancy or overfitting, depending
on the filter size.</p>
        <p>To achieve a more precise estimation of redshift, the Inception network was used. Since the
model's complexity and computational efficiency are critical for handling large datasets, the
architecture with dimensionality reduction was used (Figure 5). It consists of three to four parallel
towers: two or three convolutional filters of different sizes applied after a 1x1 convolution, and a max
pooling operation followed by another 1x1 convolution.</p>
        <p>Trainings was conducted on the Inception network consisting of three towers using the ReLU
activation function again as well as the mean absolute error loss function over 20 epochs. Results can
be found in Table 2.</p>
        <p>Considering the results, the Inception network demonstrates significantly higher MAE values
compared to the single-layer CNN model. The lowest MAE achieved by the Inception network is
0.0504, which is worse than the best result of 0.0169 obtained by the CNN with 64 filters and a filter
size of 150. This imbalance can be explained by the simplicity and efficiency of the CNN architecture,
which is better suited for the straightforward task of redshift estimation. In contrast, the Inception
network’s complex architecture may introduce unnecessary computational overhead and fail to
capture the specific features required for accurate redshift prediction.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Correction method evaluation</title>
        <p>Since there is no difference which CNN architecture choose to test correction approach, single layer
CNN with 64 filters and a filter size of 150 will be used further.</p>
        <p>The CNN model outputs the estimated redshift value for a given spectrum. It is then used to
calculate the positions of emission and absorption lines. These calculated positions are not aligned
with specific spectral characteristics, such as peaks, as shown on Figure 6.</p>
        <p>As described earlier, two approaches are going to be compared – the gradient boosting model and
the greedy algorithm.</p>
        <p>The gradient boosting is trained using 100 estimator trees and a 0.1 shrinkage parameter. Mean
squared error is a loss function. The greedy algorithm assigns a cost to each candidate position based
on its proximity to the initial prediction, the intensity of the wavelength at the point, and how well
the new position maintains the spacing between neighboring emission lines. The positions with the
lowest cost are selected as the new candidates for the emission lines.</p>
        <p>Eight emission lines are chosen for performing correction: Hβ, [O III] 4959, [O III] 5007, [N II]
6548, Hα, [N II] 6583, [S II] 6716, [S II] 6730.</p>
        <p>Table 3 contains the results of the correction model using different approaches. The region size
indicates the size of the local window around each emission line where adjustments can be made. It
can be observed from the table that both the greedy method and gradient boosting show
improvements in redshift prediction accuracy as the region size increases. These results show that
both methods are effective, however still gradient boosting provides more consistent corrections,
especially for larger region sizes.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>This paper describes the effective approach of redshift estimation using spectral data and machine
learning techniques. The proposed algorithm combines convolutional neural networks (CNNs) for
initial redshift prediction with correction methods designed to refine the estimated values based on
spectral characteristics.</p>
      <p>The final results of the model achieve up to 97.7% accuracy in redshift estimation when utilizing
gradient boosting correction.</p>
      <p>Future work could explore hybrid approaches or further optimization of the Inception network to
better align its capabilities with the requirements of redshift estimation. Additionally, the correction
methods could be enhanced by incorporating more sophisticated spectral features or involving
advanced machine learning techniques, such as attention mechanisms, to better capture the
relationships between emission lines and spectral peaks.</p>
      <p>Overall, the combination of CNN-based redshift estimation and correction methods provides a
promising framework for accurate and efficient analysis of spectral data.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The authors have not employed any Generative AI tools.
Communications and Image Processing, Macau, China, 2020, pp. 294-297,
doi:10.1109/VCIP49819.2020.9301849.
[5] A. D’Isanto and K. L. Polsterer, Photometric redshift estimation via deep learning, A&amp;A 609,</p>
      <p>A111 (2018). doi:10.1051/0004-6361/201731326.
[6] Hu Zou et al, Photometric Redshifts and Galaxy Clusters for DES DR2, DESI DR9, and HSC-SSP</p>
      <p>PDR3 Data, Res. Astron. Astrophys. 22 065001 (2022). doi:10.1088/1674-4527/ac6416.
[7] Vikhyat Agarwal et al, SpyderZ: An Efficient Support Vector Machine Library for Photometric
Redshift Estimation and Redshift Probability Information, in Res. Notes AAS 8 126 (2024).
doi:10.3847/2515-5172/ad47fb.
[8] Csabai, I., Multidimensional indexing tools for the virtual observatory, Astronomische</p>
      <p>Nachrichten, vol. 328, no. 8, p. 852, 2007. doi:10.1002/asna.200710817.
[9] Q. Lin, D. Fouchez, J. Pasquet, M. Treyer, R. Ait Ouahmed, S. Arnouts, O. Ilbert, Photometric
Redshift Estimation with Convolutional Neural Networks and Galaxy Images: A Case Study of
Resolving Biases in Data-Driven Methods. A&amp;A 662, A36 (2022).
doi:10.1051/00046361/202142751.
[10] R. Stivaktakis, G. Tsagkatakis, B. Moraes, F. Abdalla, J. -L. Starck and P. Tsakalides,
Convolutional Neural Networks for Spectroscopic Redshift Estimation on Euclid Data, in IEEE
Transactions on Big Data, vol. 6, no. 3, pp. 460-476, 2020, doi:10.1109/TBDATA.2019.2934475.
[11] O. Podsztavek, P. Škoda, P. Tvrdík, Spectroscopic redshift determination with Bayesian
convolutional networks, Astronomy and Computing 40 100615 (2022).
doi:10.1016/j.ascom.2022.100615.
[12] Szegedy, Christian: Going deeper with convolutions, in The IEEE Conference on Computer</p>
      <p>Vision and Pattern Recognition (CVPR). 1-9, 2015. doi:10.1109/CVPR.2015.7298594.
[13] Andrés Almeida et al, The Eighteenth Data Release of the Sloan Digital Sky Surveys: Targeting
and First Spectra from SDSS-V, ApJS 267 44 (2023). doi:10.3847/1538-4365/acda98.
[14] Ivana Orlitova, Starburst galaxies, in Reviews in Frontiers of Modern Astrophysics: From Space</p>
      <p>Debris to Cosmology; Springer 2020. doi:10.1007/978-3-030-38509-5.
[15] Olha Oliinyk, Yuri Taranenko, Valerii Lopatin, Analysis of Discrete Wavelet Spectra of Broand</p>
      <p>Signals, Computer Modeling and Intelligent Systems, 3392, 188-198. doi:10.32782/cmis/3392-16.
[16] Carter L Rhea et al, A machine learning approach to galactic emission-line region classification,
RAS Techniques and Instruments, vol. 2, issue 1, 345–359. doi.org:10.1093/rasti/rzad023.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Haojing</given-names>
            <surname>Yan</surname>
          </string-name>
          et al,
          <source>First Batch of z ≈ 11-20 Candidate Objects Revealed by the James Webb Space Telescope Early Release Observations on SMACS 0723-73, ApJL 942 L9</source>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .3847/2041- 8213/aca80c.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Jiménez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. Torres</given-names>
            <surname>Torres</surname>
          </string-name>
          ,
          <string-name>
            <surname>R. John and I. Triguero</surname>
          </string-name>
          ,
          <article-title>Galaxy Image Classification Based on Citizen Science Data: A Comparative Study</article-title>
          ,
          <source>IEEE Access 8</source>
          (
          <year>2020</year>
          ),
          <fpage>47232</fpage>
          -
          <lpage>47246</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2020</year>
          .
          <volume>2978804</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Abel</given-names>
            <surname>Dana</surname>
          </string-name>
          <article-title>Ruiz: Detection of Gravitational Wave signals using Machine Learning methods</article-title>
          and Generative Pre-trained
          <string-name>
            <surname>Transformers</surname>
          </string-name>
          , Facultat de Fisica, Universitat de Barcelona,
          <source>Diagonal</source>
          <volume>645</volume>
          , 08028 Barcelona, Spain.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Jin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <article-title>Machine Learning for Photometric Redshift Estimation of Quasars with Different Samples</article-title>
          , in: 2020 IEEE International Conference on Visual
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>