=Paper= {{Paper |id=Vol-3282/icaiw_waai_3 |storemode=property |title=Satellite Orbit Prediction Using a Machine Learning Approach |pdfUrl=https://ceur-ws.org/Vol-3282/icaiw_waai_3.pdf |volume=Vol-3282 |authors=Giridhar Jadala,Gowri Namratha Meedinti,Radhakrishnan Delhibabu |dblpUrl=https://dblp.org/rec/conf/icai2/JadalaMD22 }} ==Satellite Orbit Prediction Using a Machine Learning Approach== https://ceur-ws.org/Vol-3282/icaiw_waai_3.pdf
Satellite Orbit Prediction using a Machine Learning
Approach
Giridhar Jadala* , Gowri Namratha Meedinti and Radhakrishnan Delhibabu
Vellore Institute Of Technology, Vellore, India


                                      Abstract
                                      The most cutting-edge approaches for orbit prediction right now use physics-based models, which call for
                                      exceedingly precise information about the object’s trajectory, its operating environment, and its intended
                                      course of action. We don’t actually have access to this information. Our knowledge of space weather and
                                      air density is limited, trajectories are only seldom obtained from noisy ground-based radar systems, and
                                      satellite operators are reluctant to disclose their maneuvering intentions. Current orbit predictions that
                                      are only based on physics-based models may not achieve the required accuracy for collision avoidance
                                      and have already resulted in satellite collisions due to the lack of information regarding the state of the
                                      space environment and resident space objects’ (RSOs’) body characteristics. Two line element sets(TLE)
                                      made accessible to the public lack any related error or correctness information. The majority of TLE
                                      prediction techniques used today fit polynomials, which cannot capture periodic properties. This paper
                                      has presented a methodology for orbital prediction using curve fitting and LSTM on historical orbital
                                      data. The proposed machine learning approaches are used with various TLE parameters where the LSTM
                                      model is trained to learn through large amounts of historical TLE data. The fitted data is synthesized
                                      and then compared with the SGP4 predictions. The two proposed methods focus on reducing prediction
                                      errors. The results of the study demonstrate that the proposed machine learning approaches can improve
                                      orbital prediction accuracy with good performance in most cases. We go into further optimization and
                                      the computing needs for using all-on-all conjunction analysis on the whole TLE collection and visualize
                                      when and where conjunction may occur, both currently and in the near future.

                                      Keywords
                                      TLE, SGP4, Orbit, Prediction, Satellite Collisions, LSTM, Curve Fitting




1. Introduction
Spacecraft, astronauts, and the global commons of near-earth space are all at risk from collisions
in orbit. Several collisions between spacecraft and debris have already occurred, with the Iridium
33/Cosmos 2251 collision in January 2009 being the first known satellite-satellite collision. The
total number of collisions so far is consistent with the foresight of the runaway chain reaction
[1].
   The number of resident space objects (RSOs) is fast increasing, as is the frequency of collision
warnings between RSOs. Improved orbit prediction accuracy is a major challenge for Space
Situational Awareness (SSA). Iridium 33 and Cosmos 2251 collided in 2009, revealing the lack
ICAIW 2022: Workshops at the 5th International Conference on Applied Informatics 2022, October 27–29, 2022, Arequipa,
Peru
*
  Corresponding author
$ jadala.giridhar2019@vitstudent.ac.in (G. Jadala); gowri.namratha2019@vitstudent.ac.in (G. N. Meedinti);
r.delhibabu@vit.ac.in (R. Delhibabu)
                                    © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0).
 CEUR
 Workshop
 Proceedings
               http://ceur-ws.org
               ISSN 1613-0073
                                    CEUR Workshop Proceedings (CEUR-WS.org)



                                                                                                       28
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46


of high-accuracy orbit prediction capability [2]. The majority of existing orbit prediction
methods are physics-based. Fundamentally, the failure is due to a lack of required data, such as
the state of the space object at the start of trajectory computation, environmental conditions
such as atmospheric drag and solar radiation pressure, RSO body characteristics such as mass,
geometry, and material, and maneuvering object intent information. Their success is contingent
on a thorough understanding of forces including Earth’s gravity, atmospheric drag, and solar
radiation pressure. For high-accuracy forecasts, the RSO’s maneuvering information, geometric
information, and other aspects are also required. The analytical model is well-developed, but the
RSO’s data is limited or difficult to obtain. Satellite collision would produce orbiting fragments,
every of which might increase the chance of more collision, resulting in the expansion of belt
trash around the earth. Increased in-order collisions easily can reach the collisional cascading
point called “Kessler’s syndrome”, because the random collision frequency is non-linear with
debris accumulation rates, the development can eventually become the foremost necessary long
supply of debris accumulation rates unless the buildup rate of larger, non-operational objects in
earth orbit was considerably reduced.
   In comparison to the physics-based method, machine learning (ML) offers a new modeling
and prediction path. Without specifically describing the geometry of RSOs, movements, or the
space environment, the forecast can be made. Instead, the models are created using a significant
amount of historical data, much like how humans learn from their past experiences to predict
future events. Machine learning approaches have proven their worth in a variety of applications
[3], with many in the aerospace field [4, 5].
   The orbits of RSOs are continually developing and are impacted by both conservative and non-
conservative processes. As a result, learning orbit prediction errors is distinct from, and perhaps
more difficult than, many other machine learning problems. Nonetheless, we believe that part
of the information can be learned because the prediction mistakes are implicitly included in
RSOs’ historical data, such as measurements, estimations, and predictions errors. The orbits
of tracked RSOs around Earth are specified and updated in the US space catalog as TLEs. TLE
data is readily available for satellite owners and operators at Space-Track.Org published by US
strategic command (USSTRATCOM). Numerous space companies (private and government) are
launching their satellites into orbit for advancement in space exploration and communication.
These satellite companies have inadequate technology and procedures to avoid close encounters
between objects. So the only source of information at their fingertips from the spec of paint to
the international space station is the publicly available TLE data. NORAD two-element sets
are used for increasing the number of satellite operations and analyses. General perturbation
element sets generated by NORAD can be used to predict the position and velocity of earth-
orbiting objects. Around five mathematical models for the prediction of satellite position and
velocity are available. Among these SGP4 model is widely used to generate NORAD element
sets. It was developed by Kencrankford in 1970 and is used for near-earth satellites [6]. The
analytical SGP4 model is computationally fast and efficient however its advantage within the
computation potency is shaded by a lack of high accuracy in propagated orbits [7]. So operators
can use the SGP4 algorithm to address this issue interestingly we demonstrated two machine
learning approaches for predicting and improving RSOs orbit information.
   The machine learning methodology provides unique modeling and prediction capability
compared to physics-based approaches. One of the approaches used is based on various TLE



                                                   29
Giridhar Jadala et al. CEUR Workshop Proceedings                                             28–46


parameters such as Inclination (i) is an independent mean element of a TLE that defines the
orientation of the orbital plane with respect to the Earth [8]. The right ascension of the ascending
node (𝜔) is a mean element of a TLE that defines the orientation of an orbit with respect to
the ^z axis, which is parallel to the rotation axis of the Earth [8], Argument of perigee (𝜔) is a
mean element of a TLE that defines the orientation of the orbital plane with respect to the Earth
around the axis parallel to the angular momentum vector of the orbit, Mean anomaly (m) is a
mean element of a TLE that defines the position of a space object in its orbit [8], eccentricity
(e) is a mean element of a TLE that defines the shape of the orbit while semi-major axis (a)
defines the size of the orbit, mean motion (n) is one of the 6 independent mean elements, which
include the inclination, right ascension of the ascending node, the argument of perigee, mean
anomaly and eccentricity, that is required for computations of TLEs [9]. Curve fitting is used
for some parameters such as right ascension of the node, argument of pedigree, mean anomaly,
and eccentricity, to look at the standardized data and extract physical parameters. Curve fitting
tries to adjust a smooth and balancing model function to provide the best-optimized fit for the
data. A non-linear least square minimization process is used to find a plausible minimization
quantity. LSTM layers are used to fit the data of some TLE parameters as these are more efficient
in the case of time series data. First, the model is trained with historical TLE data, and then
validation is done which outputs loss value and metric values for the model. Finally, the actual
predictions for the future epochs which output the fitted data with higher accuracy, and then
new synthesized TLE can be computed. This work considers each orbital parameter individually.
LSTM networks are a sort of recurrent neural network that can learn order dependence in
sequence prediction issues. This is an essential behavior in difficult problem areas like machine
translation, speech recognition, and more complex problems such as orbital prediction.
   The prime contributions of this study are summarised as follows:
    • Using LSTM and curve fitting for making predictions of satellite’s TLE (The proposed ML
      approach orbit is closer to the true orbit. This concept is illustrated in Figure 1).
    • Obtaining a comparable accuracy to the SGP4 model.


2. Background
2.1. SGP4
Simplified perturbation models (SGP, SGP4, SDP4, SGP8, and SDP8) are a series of five mathe-
matical models for calculating orbital state vectors of satellites and space debris relative to the
Earth-centred inertial coordinate system. Because of its widespread use, especially with two-line
element sets provided by NORAD and NASA, this group of models is commonly referred to as
SGP4.
   The influence of perturbations induced by the Earth’s form, drag, radiation, and gravitational
forces from other bodies such as the sun and moon is predicted by these models. Simplified
General Perturbations (SGP) models are used to calculate the orbital period of near-earth
objects having orbital periods of less than 225 minutes. Simplified Deep Space Perturbations
(SDP) models are applicable to objects with an orbital period of more than 225 minutes, which
corresponds to an altitude of 5,877.5 km in a circular orbit.



                                                   30
Giridhar Jadala et al. CEUR Workshop Proceedings                                           28–46




Figure 1: Predicted ML approach orbit vs Actual Orbit


   The TLE sets include mean orbital elements that were derived by specifically eliminating
periodic fluctuations. These periodic changes must be recreated by the prediction model
precisely in the same manner as they were eliminated in order to produce accurate forecasts.
One of these orbit models for satellites in near-Earth space (period 225 minutes) is the SGP4
model.
   Using ’ideal’ TLE data sets that were created based on the precise orbit information available
for the operational satellites, it was first determined whether the SGP4 analytical orbit model’s
intrinsic modeling accuracy could be improved.
   Current SGP4 implementations are based on Brouwer’s gravity solution [1] and Lane’s
atmospheric model [10], but with Lyddane’s adjustments to avoid loss of precision in the
assessment of periodic correlations [3], which are also significantly simplified to improve
evaluation efficiency. The NORAD mean element sets can be used for prediction with SGP.
   The unit orientation vectors are calculated by:

                                    𝑈 = 𝑀 𝑠𝑖𝑛𝑈𝑘 𝑧 + 𝑁𝑐𝑜𝑠𝑈𝑘                                    (1)


                                     𝑣 = 𝑀 𝑐𝑜𝑠𝑈𝑘 − 𝑁𝑠𝑖𝑛𝑈𝑘                                     (2)
  Where
                                     ⎧
                                     ⎨𝑀𝑥 = −𝑠𝑖𝑛Ω𝑘 𝑐𝑜𝑠𝑖𝑘
                                     ⎪
                                  𝑀 = 𝑀𝑦 = 𝑐𝑜𝑠Ω𝑘 𝑐𝑜𝑠𝑖𝑘                                        (3)
                                     ⎪
                                      𝑀𝑧 = 𝑠𝑖𝑛𝑖𝑘
                                     ⎩




                                                   31
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46


                                         ⎧
                                         ⎨𝑁𝑥 = 𝑐𝑜𝑠Ω𝑘
                                         ⎪
                                      𝑁 = 𝑁𝑦 = 𝑠𝑖𝑛Ω𝑘                                           (4)
                                         ⎪
                                          𝑁𝑧 = 0
                                         ⎩

  Then the position and velocity are calculated by

                                             𝑟 = 𝑟𝑘 𝑈                                          (5)

                                                       ̇) 𝑣
                                       Γ̇ = 𝑟̇ 𝑅 𝑈 + (𝑟𝑓                                       (6)
                                                         𝑘


2.2. LSTM
Long short-term memory, LSTM algorithm, which is a type of Recurrent Neural Network (RNN).
An input layer, a hidden layer, and an output layer make up the three layers. By constructing
weight coefficients between hidden layers, LSTM solves the problem of long-distance dependence
that RNNs cannot manage. It indicates that while forecasting satellite telemetry parameter time
series data, LSTM can develop long-term connections between distant nodes in the time series,
improving the accuracy of time series data prediction [11].
  Our intuition to use LSTM was because of its effectiveness on the time series. We have only
used linear regression in some cases where data looked linear. Looking closely at the linear
data points, one would notice they usually return back to a value after some time abruptly.
Like a modulo operation so delinearize helps with taking a linear result and applying a modulo
operation to the result.

2.3. Curve Fitting
The process of creating a curve, or mathematical function, that best fits a set of data points,
sometimes subject to constraints, is termed curve fitting. Either interpolation, where an exact
fit to the data is required or smoothing, where a "smooth" function is generated that roughly
fits the data, can be used to fit a curve. Fitted curves can be used to explain the relationships
between two or more variables, to infer values of a function when no data are available, and as
a tool for data visualization. Extrapolation is the use of a fitted curve outside the range of the
observed data, and it carries some risk because it could represent the construction process of
the curve just as much as it does the actual data.

2.4. Dataset
The TLE dataset was employed in this study. A two-line element set (TLE) is a standardized
format for describing a satellite’s orbit and trajectory. Below is an example of the International
Space Station. There are 14 fields in a TLE; however, our method only needs 9 of them. The
9 fields are Epoch Year - The year the TLE was calculated, Epoch Day - The day and fraction
of the day the TLE was calculated, B- star - The drag term or radiation pressure coefficient,
Inclination - Satellite orbit tilt between 0 and 180 degrees, Right Ascension of the Ascending
Node - Ranging from 0 to 360 degrees, Eccentricity - A measure of how circular the orbit is



                                                   32
Giridhar Jadala et al. CEUR Workshop Proceedings                                           28–46




Figure 2: An orbit in 3D defined by TLE parameters


ranging from 0 to 0.25 for LEO satellites, Argument of Perigee - The angle from the ascending
node ranging from 0 to 360 degrees, Mean Anomaly - The angular position measured from
pericenter if the orbit was circular ranging from 0 to 360 degrees, Mean Motion - The angular
speed necessary to complete one orbit measured in revolutions per day with a minimum of
11.25 for LEO satellites.


3. Methodology
3.1. Methods and Procedure
There are a few studies that investigate different approaches to estimating TLEs. USSTRATCOM
publishes TLE data on the spacetrack.org website. We retrieve TLE data of RSOs by their
SATNAME within a specific period of time. After getting the data, we then used it to visualize
over time. We observed various dependencies inTLE data propagation. TLEs, by their nature,
can only be converted to position data using SGP4. TLEs include the mean states estimated by
fitting observations to the dynamics provided by SGP4, and they can only be used with SGP4.
As a result, we attempted to use curve fitting to ensure that our models were invariant with
respect to the dependencies. To construct a linear or non-linear function based on the data, we
tried different fitting methods. We tried polynomial fitting and sinusoidal fitting on the data.
   The linear or non-linear curve fitting method is a mere global minimization of the weighted
sum of squares. So in our study, for some parameters such as the right ascension of the node, the
argument of perigree, mean anomaly, eccentricity, the weighted sum of squares, and root mean
squared error is used to assess the goodness of fit. TLE data is input to either adjust a smooth
and balancing model function that describes the data adequately or train a special kind of
recurrent neural network that is capable of learning long-term dependencies in data. Applying
this method to all parameters produced disappointing results. So curve fitting worked with



                                                   33
Giridhar Jadala et al. CEUR Workshop Proceedings                                             28–46




Figure 3: LSTM Model


some parameters, but for others, such as inclination, mean motion, etc, it gave some reasonable
deviations. So we tried LSTM (Figure 3) to fit the data. LSTM is a special kind of recurrent neural
network and it is good at predicting sequential and temporal data. A fully connected neural
network architecture is trained with 9 hidden layers using tanh as an activation function. To
modify each epoch weight and the loss function efficiently, we have used the Adamax optimizer,
a variant of Adam based on the infinity room [12]. As this problem is large in terms of data
parameters, the Adam optimizer is a primary algorithm for optimization to estimate lower-order



                                                   34
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46


moments.
   Training the model would continue by comparing the output with the target, calculating
the error, and optimizing the weights, reiterating the process again and again. We believe that
we did a thorough hyper-parameter search and optimization in order to seek the best overall
parameters in order to minimize our loss on the validation set.
   We have also chosen an ideal number of epochs after which the accuracy stopped increasing.
When determining optimized weights for the trained model, the loss function is the most
important factor to consider. The MAE loss function is used as it is more robust to outliers. This
fully-connected neural network learns the mapping between the input and output by connecting
all neurons available, and it has 201,052,641 trainable parameters (total number of weights and
biases). The LSTM architecture is shown in (Figure 3). To reduce the over-fitting of the model, I
used a dropout of 0.2, which significantly reduced the complex co-adaptions of training data.
Neural networks are sensitive to the scaling of the input data; therefore, the input data is either
standardized or normalized based on the resultant performance of the model. For this neural
network model, normalization is done through Z-score normalization. The arithmetic mean
and standard deviation of the given input data are calculated first. The standard Z-score is
calculated as follows: The Z-score indicates how many standard deviations the data element is
from the mean. As follows, min-max normalization is done for each data element. It is observed
that input data for neural network standardization outperforms normalization:

                                             𝑥 − 𝑚𝑖𝑛(𝑥)
                                    𝑧𝑛 =                                                        (7)
                                           𝑚𝑎𝑥(𝑥) − 𝑚𝑖𝑛(𝑥)
                                               𝑥 − 𝑢(𝑥)
                                           𝑧𝑠 =                                                (8)
                                                   𝜎
   Table 1 shows the parameters which are used to standardize and normalize the input data.
A ratio of 80/20 is used for data splitting, such that 80% goes to the training subset and 20%
goes to the testing subset. The input to LSTM is a 3-dimensional array to train and predict the
output. The model is tested with n data points from the TLE dataset so that the rest of the TLE
data can be used for validation and testing. The fitted data is returned for every parameter
of the input TLE data for n data points. The reliability of conjunction messages and collision
risk estimation is largely dependent on the accuracy of the fitted data to estimate the orbit and
associated covariance information.

Table 1
Parameters to normalize and standardize the input data for LSTM model
                 S.No                Parameter                  Mean        Std
                   1     First time derivative of mean motion   -5.8e-13   2.8e-12
                   2            Ballistic drag coefficient      1.2e-05    5.9e-05
                   3                  Mean motion                0.062     1.5e-05
                   4                    Inclination              0.0639    2.07e-5




                                                   35
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46


3.2. Proposed model
Figure 4 depicts the overall framework of the model proposed in this paper. The framework of
the orbit prediction through machine learning is shown in Figure 4. The components and the
process is explained in their respective section.




Figure 4: Framework



3.3. Task Flow
Figure 5 depicts the task flow diagram of the proposed method.


4. Results
4.1. Residual plots and Scatter Plots
The absolute residual plots and the scatter plots with regression of the TLE parameters prediction
made using machine learning models are presented below Figures (6, 7, 8, 9, 10, 11, 12, 13). The



                                                   36
Giridhar Jadala et al. CEUR Workshop Proceedings                                           28–46




Figure 5: Process Flow


absolute residual plot detects incorrect behavior of residuals, the graph depicts the density of
residuals predicted by ML models and their values are displayed along the x-axis for every
TLE parameter. The residuals are concentrated close to 0.00001(rad/min) and their mean is
scattered approximately between 0.0000 and 0.0001, hence the predictions are not biased and
uncorrelated. The percentage of residuals smaller than 0.0001 is 31.9% and residuals smaller than
0.0007 is 89.2% for B* model. The peak outside the confidence interval for particular lag k means
the output y(t) that originates from the input u(t-k) is not properly described by the LSTM
model. The residuals smaller than 0.0001 are 90.1% for mean motion and the RMSE observed is
2.2e-07. Table 3 shows the mean and standard deviation of the residuals of the predictions for
the machine learning model.
   The patterns of both predicted output and actual output (1/Earth Radius) are close to the
regressed diagonal line. The first-time derivative of motion model including other TLE parameter
models has high goodness of fit so that the predicted values are not heavily dispersed away from



                                                   37
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46


Table 2
RMSE of TLE parameters
                                           RMSE of the proposed         RMSE of Gradient boost-
 S.No      Training Approach
                                           model                        ing tree model
           First time derivative of mean
 1                                         1.7e-12                      -
           motion
 2         Ballistic drag coefficient      2.3e-05                      2.4e-05
           Right ascension of ascend-
 3                                         0.004                        0.01
           ing node
 4         Argument of perigee             0.998                        0.22
 5         Mean anomaly                    1.2741                       0.3
 6         Mean motion                     2.2e-07                      5.3e-9
 7         Inclination                     2.06e-6                      9.8e-5
 8         Eccentricity                    7.5e-6                       2e-7


Table 3
Mean and standard deviation of residuals for the given parameters
                                                               Mean of gradi-     Variance
                           Mean of pro-        Std of pro-
 S.No    Parameter                                             ent   boosting     of   gradient
                           posed model         posed model
                                                               trees              boosting trees
         First     time
 1       derivative of     4.766425e-13        2.943803e-12    -                  -
         mean motion
         Ballistic drag
 2                         8.627849e-06        5.971809e-05    0.0002             0.0009
         coefficient
         Right       as-
         cension      of
 3                         -0.009              0.011680        0.0045             0.0029
         ascending
         node
         Argument of
 4                         0.320074            101.875286      0.24               1.898
         perigee
         Mean
 5                         1.049211            8.9800481       0.19               0.23
         anomaly
 6       Mean motion       -7.972256e-08      1.668083e-05     6.5e-5             1.8e-8
 7       Inclination       4.01e-6            5.14e-6          0.002              1.6e-5
                           7.82378080292368e-
 8       Eccentricity                         0.000078         0.008              1.23e-5
                           05


the line, this indicates the values are non-random around the line. The machine learning models
output the values close to the target values. The points are located around the diagonal line
which means that the proposed model predicted well. The regression plot of actual vs predicted
output shows that the RMSE predictions of the LSTM model are quite smaller compared to other
naive and Bayesian neural networks and Gradient Boosting Tress as shown in Table 2. The
density plot of residuals and scatter plot with regression of Right ascension of ascending node



                                                     38
Giridhar Jadala et al. CEUR Workshop Proceedings                                          28–46




Figure 6: Performance of first time derivative of the mean motion




Figure 7: Performance of Ballistic drag coefficient B* in inverse earth radii




Figure 8: Performance of Right ascension of ascending node


is shown in Figure 8. The absolute mean of residuals is 0.009 radians and the standard deviation
is 0.0116. The larger variance is due to the ineffectiveness of the machine learning model to
recognize the cyclic behavior when the mean element switches values between 2𝜋 and 0 as
the orbit propagate in time. Due to similar performances in residuals of their predictions both
the argument of perigee and the mean anomaly machine learning models have relatively poor



                                                   39
Giridhar Jadala et al. CEUR Workshop Proceedings                                             28–46




Figure 9: Performance of Argument of perigee




Figure 10: Performance of Mean Anomaly




Figure 11: Performance of Mean Motion


performance in determining the decision boundary that can map the orbital evolution to the
mean elements accurately. As discussed above for the right ascension of ascending node, due to
difficulty in capturing the cyclic behavior of the angle between 0 and 2𝜋 due to 𝐿𝑣 (Ω, 𝑤, 𝑣)
feature i.e. equinoctial element, all machine learning models learn a representation that has
large fluctuations in the residuals of its predictions and this may very well signify that argument




                                                   40
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46




Figure 12: Performance of Inclination




Figure 13: Performance of Eccentricity


of perigee and mean anomaly is modified to fit the perturbed orbit. The RMSE values of the
proposed machine learning model output values close to the average of all target value There
are outliers observed in the data which are kept between a comparable range of - 0.1000 to 0.
1000. For these parameters which behave randomly with LSTM it is observed that linear fitting
works well. If one looks at the linear data points, one would notice that they usually return to a
value at some time abruptly.
   Large fluctuations in the residuals of predictions are a feature of every machine learning
model, and this may signify that, like in the case of B*, the mean anomaly and the argument of
perigee have been altered to fit the perturbed orbit. Unlike the other parameters, eccentricity
has no significant relationship with the machine learning models so the polynomial sin fitting
function can approximate 𝑠𝑖𝑛(𝑥) between 0 and 2𝜋. The development of value in the above-
mentioned mean elements between 0 and 2𝜋 occur for about 0.73% of the test data. Future
research will solve this problem by incorporating other characteristics into the features such as
(Ω, 𝑤, 𝑣) by utilizing data transformations.




                                                   41
Giridhar Jadala et al. CEUR Workshop Proceedings                                            28–46


4.2. RMSE results of Proposed Model vs Gradient Boosting trees
Table 2 summarises the RMSE results of the proposed model with gradient boosting trees, the
values tell you how concentrated the data is around the line of best fit. As you can observe
the RMSE values of the argument of perigee have a high error which gives relatively high
weight. Depending on the features chosen, different machine learning techniques perform
differently. The proposed machine learning model outperforms the gradient-boosting trees
regarding prediction accuracy unless the absolute difference between the prediction of the
proposed model and the prediction of gradient-boosting trees is larger than the threshold value.
The LSTM machine learning model is trained to predict each of the respective mean elements
of a TLE and the model is chosen to be validated by the errors in the orbital evolution. The
error in the orbital evolution is computed by propagating TLEs estimated by the selected best
machine learning models and the associated Standard TLEs backward in time using SGP4 for
150 orbital periods. Figure 14 shows the comparison of orbital propagated using SGP4 model
satellite TLE data with the proposed machine learning model predictions of TLE data. The
below propagations of TLE data compare the actual RASAT RSO’s TLE data with the predicted
TLE data of our proposed model.

4.3. Evaluation of Prediction Error and Error Handling of the proposed model
From Table 4 we can see that the prediction accuracy using the proposed model significantly
outperforms that using TLE/SGP4 [Space-track.org]. For the RASAT satellite, the absolute
maximum error for 50 days in the along-track direction is 35.6 km using TLE/SGP4, while it
is 3.8km using the proposed machine learning model. A reduction of 89.325% for the absolute
maximum along track error is achieved using the proposed machine learning model. For
Starlette, the absolute maximum prediction error for 50 days in the along-track direction is 13.9
km using TLE [spacetrack.org] and 2.9 km using the proposed model. The absolute maximum
along track error is decreased by 79.1367% using the proposed model.

Table 4
Averages of the maximum 50-day position prediction error from 150 computations (in Kilometres)
                                  TLE          (space                          Average Improve-
 S.No      Satellite                                    Proposed Model
                                  track.org)                                   ment Percentage
 1         RASAT                  35.6                  3.8                    89.3258%
 2         STARLETTE              13.9                  2.9                    79.1367%
 3         STELLA                 2.7                   1.1                    59.2593%
 4         GOKTURK-2              11.3                  3.9                    65.4867%


   Table 4 gives the averages of maximum 50-day position prediction errors using the proposed
machine learning models and TLE [spacetrack.org] from 50 computations. Compared with
spacetrack.org TLE, the averages of the maximum 50-day position prediction errors of the
proposed model propagated orbits are 3.8 km, 2.9 km, 1.1 km, and 3.9 km for Rasat, Starlette,
Stella, Gokturk-2, which have been improved by 89.32%, 79.1367%, 59.2593%, 64.4867%, respec-
tively. These findings suggest that the proposed model’s TLE generated from the accurate orbit



                                                   42
Giridhar Jadala et al. CEUR Workshop Proceedings                                               28–46


predictions may be able to provide even more precise orbit predictions. Further examination
shows that prediction errors in the along-track, cross-track, and radial directions have periodic
properties. This indicates that the predicted errors could be modeled by proper correction
functions and the absolute relative errors can be improved by non-linear least squares and
differential correction methods. For example, it is determined that n = 10 is suitable for the
error fitting for the 50-day forecast period. The unknown coefficients ai, bi, and ci I = (1, 10) are
estimated using the nonlinear least square technique using the prediction errors at each epoch
as observations. The prediction errors at any epoch are estimated using a correction function
after calculating the coefficients in the correction functions and are then used to correct the
predicted orbits to provide more accurate orbit predictions. The amount and variance of the
prediction errors are drastically decreased when the adjustments from the correction function
are used. In conclusion, the performance of our model for different parameters of TLE is very
precise and accurate this allows our proposed model to provide an estimate of the TLE data
at various epochs. Our proposed LSTM has significant regression capabilities due to its more
flexible structure so when dealing with different aspects of error handling with very large
measurement noise it shows stable performance. So this ML approach can be applied without
knowing the accurate noise in the dataset.

4.4. Comparison and Evaluation of proposed Machine Learning model with
     other models
This work is the first effort to use a unique machine learning methodology to predict satellite
TLE data to improve the orbit prediction accuracy of the RSO at future epochs and focuses on
reducing prediction errors. Depending on the TLE parameters, machine learning techniques
function differently. The mean components of TLE are predicted by two different machine
learning techniques and the performance of the models is compared with other models based
on errors in orbital evolution. In Table 5 we have presented the comparison of the percentage
of improvement for position prediction errors of the satellites with generalized polynomial
algorithm [6] and data-driven model.

Table 5
Comparison of Improvements of Prediction Errors
                                                                             Other      Model
                                  Our        Proposed     Other Model - Gen-
 S.No      Satellite                                                         -     Data-Driven
                                  Model                   TLE [13]
                                                                             Model [14]
 1         STARLETTE              79.1367%                74.9%                  61.87%
 2         STELLA                 59.2593%                NA                     51.267%
 3         AJISA                  71.275%                 74.285%                41.98%
 4         LARETS                 77.110%                 70.611%                NA


  After performing predictions on different LEO satellites, the summary statistics say that the
typical prediction error of our method is 1.1 km at 30 days out corresponding to ∼28 m/day
prediction error growth. The worst case is observed in STELLA where the median error grows
<90 m/day and the best case is observed in RASAT at 9m/day. These are compared to typical



                                                   43
Giridhar Jadala et al. CEUR Workshop Proceedings                                          28–46




Figure 14: Comparison of fitted data with actual TLE Datapoints


growth of 1500m/day for TLEs propagated with SGP4 for these satellites and 200m/day when
compared to other ML and data-driven approaches as shown in Table 6. This table comparatively
indicates that there is no secular error growth(10-180m/day) in the orbital evolution of TLEs
predicted by our machine learning model. The proposed method exhibits approximately three
orders of magnitude improvement in prediction accuracy over TLEs propagated with SGP4 and
one order of magnitude improvement in prediction accuracy over data-driven approaches. The
trained model can reduce more than half of the prediction error as shown in Table 4. With this
prediction, we have estimated the error and limited the maximum error we can have in the
prediction. It allows us to know the current position of the satellite, and to predict the final
position with a bounded error. The intersection of both curves in Figure 14 shows that for
around 80% of the testing data, the residual error percentage is less than 100%, indicating that
our trained model has been effective in lowering their prediction errors. In general, the orbit
prediction accuracy for the majority of future epochs can be improved by the suggested ML
technique.

Table 6
Comparison of accuracies (error growth in meters/day) with other Prediction Methods
          S.No           Prediction Methodology            Accuracy [in meters/day]
            1              Our proposed model                        10-180
            2            TLE + SGP4 Model [15]                      100-3000
            3            TLE + Fitted Model [14]                     50-200
            4            Least Square Model [16]                     10-300
            5      Bayesian Neural Network Model [17]                10-200
            6     High Accuracy Catalog + SP Model [15]              50-200




                                                   44
Giridhar Jadala et al. CEUR Workshop Proceedings                                               28–46


5. Conclusion
In this paper, a new and unique approach for estimating the TLE parameters of an RSO has been
developed. We have presented a prediction technique that allows us to more accurately estimate
the satellite TLE parameters in the future. As a general evaluation, the pros of this paper are
the distinctive study that we have performed to predict new data parameters in base on the
data we had with comparable accuracy and limiting error in such prediction. Using RASAT
and STELLA as an example, the results show that the performance is stable on testing data
and this prediction technique is effective for RSOs for most position and velocity components.
RSOs in low earth orbit are simulated to examine the performance of the machine learning
models. After the corrections to the predicted orbit, the final prediction accuracy can also
be improved significantly With the obtained results, it is possible to conclude that Curve
fitting and Long short-term memory are promising machine learning approaches for estimating
TLE parameters. The machine learning model’s projected TLEs can accurately determine the
reference orbit. Different machine learning architectures, non-discontinuous transformations
of cyclic parameters, and more input parameters might all improve model performance.
   The efficiency of the proposed ML technique varies with the orbit’s RSO and further research
is required to shed light on their relationship among various TLE parameters.
   Further research is suggested to combine physics and machine learning models to enhance
the capabilities of the model. To improve the orbit propagation accuracy it is suggested to
model the orbit propagation errors for the biased results. Future research will also examine the
viability of utilizing physics-integrated machine learning models to improve predictions of the
right ascension of the ascending node and the inclination, as well as the conjunction search
component for collision avoidance.


References
 [1] D. Brouwer, Solution of the problem of artificial satellite theory without drag, Technical
     Report, YALE UNIV NEW HAVEN CT NEW HAVEN United States, 1959.
 [2] D. J. Kessler, B. G. Cour-Palais, Collision frequency of artificial satellites: The creation of a
     debris belt, Journal of Geophysical Research: Space Physics 83 (1978) 2637–2646.
 [3] R. Lyddane, Small eccentricities or inclinations in the brouwer theory of the artificial
     satellite, The Astronomical Journal 68 (1963) 555.
 [4] C. Ampatzis, D. Izzo, Machine learning techniques for approximation of objective functions
     in trajectory optimisation, in: Proceedings of the ijcai-09 workshop on artificial intelligence
     in space, Citeseer, 2009, pp. 1–6.
 [5] J. Hartikainen, M. Seppanen, S. Sarkka, State-space inference for non-linear latent force
     models with application to satellite orbit prediction, arXiv preprint arXiv:1206.4670 (2012).
 [6] B. Coffee, R. Bishop, K. Cahoy, Propagation of cubesats in leo using norad two line element
     sets: Accuracy and update frequency, in: AIAA Guidance, Navigation, and Control (GNC)
     Conference, 2013, p. 4944.
 [7] R. Wang, J. Liu, Q. Zhang, Propagation errors analysis of tle data, Advances in Space
     Research 43 (2009) 1065–1069.




                                                   45
Giridhar Jadala et al. CEUR Workshop Proceedings                                          28–46


 [8] R. Abay, S. Balage, M. Brown, R. Boyce, Two-line element estimation using machine
     learning, The Journal of the Astronautical Sciences 68 (2021) 273–299.
 [9] D. A. Vallado, Fundamentals of astrodynamics and applications, volume 12, Springer
     Science & Business Media, 2001.
[10] M. LANE, The development of an artificial satellite theory using a power-law atmospheric
     density representation, in: 2nd Aerospace Sciences Meeting, 1965, p. 35.
[11] F. R. Hoots, R. L. Roehrich, Models for propagation of NORAD element sets, Technical
     Report, Aerospace Defense Command Peterson AFB CO Office of Astrodynamics, 1980.
[12] H. Peng, X. Bai, Improving orbit prediction accuracy through supervised machine learning,
     Advances in Space Research 61 (2018) 2628–2646.
[13] B. Li, J. Sang, J. Ning, Tle generation from sparse tracking data and its performance,
     Advances in the Astronautical Sciences 158 (2016) 4003–4014.
[14] C. Levit, W. Marshall, Improved orbit predictions using two-line elements, Advances in
     Space Research 47 (2011) 1107–1115.
[15] J. C. Chan, D. Navarro, Comparisons of norad two-line elements with intelsat orbital
     elements, in: Space Debris, volume 473, 2001, pp. 771–779.
[16] C. Puente, M. A. Sáenz-Nuño, A. Villa-Monte, J. A. Olivas, Satellite orbit prediction using
     big data and soft computing techniques to avoid space collisions, Mathematics 9 (2021)
     2040.
[17] H. Peng, X. Bai, Comparative evaluation of three machine learning algorithms on improving
     orbit prediction accuracy, Astrodynamics 3 (2019) 325–343.




                                                   46