<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Research on the application of Machine Learning in predicting diabetes</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Wenfeng</forename><surname>Ye</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">School of Computer Science</orgName>
								<orgName type="institution">University of Nottingham Malaysia</orgName>
								<address>
									<settlement>Kuala Lumpur</settlement>
									<country key="MY">Malaysia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Hui</forename><surname>Zeng</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">School of Midwifery</orgName>
								<orgName type="institution">Ganan Medical University</orgName>
								<address>
									<settlement>Ganzhou</settlement>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Research on the application of Machine Learning in predicting diabetes</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">742E61550B89729FDEE86F8A4D38F2CE</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:45+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Machine learning</term>
					<term>Linear regression</term>
					<term>Polynomial regression</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This article aims to establish a predictive model for diabetes occurrence using machine learning methods. We utilized a clinical dataset from patients at Sylhet Diabetes Hospital in Sylhet, Bangladesh, including various clinical features related to diabetes such as Age, Gender, Polyuria, Polydipsia, sudden weight loss, weakness, Polyphagia, Genital thrush, visual blurring, Itching, Irritability, delayed healing, partial paresis, muscle stiffness, Alopecia, and Obesity. We performed data preprocessing and feature engineering, converting categorical variables into numerical form and standardizing the data. Subsequently, we experimented with various machine learning algorithms including logistic regression, decision trees, and support vector machines. Through cross-validation and grid search for parameter optimization, we selected multiple linear regression as the final predictive model. We evaluated the model's performance on the test set using metrics such as mean squared error (MSE), mean absolute error (MAE), R-squared (R^2), and root mean squared error (RMSE). Experimental results indicate that our model demonstrates high accuracy and reliability in predicting diabetes occurrence. Through this model, we can promptly identify individuals at risk of diabetes, providing doctors with more accurate diagnostic and treatment recommendations, and potentially offering crucial decision support for diabetes prevention and management.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Diabetes is a common and serious chronic disease that significantly impacts the quality of life and health of millions of people worldwide <ref type="bibr" target="#b0">[1]</ref>. According to data from the World Health Organization (WHO), diabetes has become a global epidemic, with an estimated 320 million people expected to be affected globally by 2030. Diabetes not only imposes physical health burdens on patients but also leads to various complications such as cardiovascular diseases, kidney diseases, and retinopathy, posing a serious threat to patients' lives <ref type="bibr" target="#b1">[2]</ref>.</p><p>Early prevention and diagnosis are particularly crucial for the prevention and control of diabetes. However, due to the complex pathogenesis of diabetes and the lack of obvious early symptoms, many patients are not aware of their health <ref type="bibr" target="#b2">[3]</ref>.</p><p>problems in the early stages of the disease, leading to missing the optimal treatment window. Therefore, establishing effective diabetes prediction models to timely identify individuals at higher risk of the disease is of great significance for reducing the incidence of diabetes and improving patients' quality of life.</p><p>Traditional diabetes prediction methods mainly rely on doctors' experience and clinical indicators such as blood glucose levels and insulin sensitivity. However, these methods have drawbacks including long diagnosis times, high costs, and reliance on medical resources. With the continuous development of machine learning and artificial intelligence technologies, it has become possible to construct diabetes prediction models using big data and deep learning methods. Machine learning algorithms can learn from massive clinical data to discover potential patterns and features of diabetes occurrence, providing doctors with more accurate and rapid diabetes diagnosis and prediction services, thereby contributing to improving diabetes prevention and control efforts <ref type="bibr" target="#b3">[4]</ref>.</p><p>Therefore, this study aims to utilize machine learning methods, based on clinical data, to construct a diabetes prediction model, achieving fast and accurate prediction of diabetes occurrence. This will provide medical institutions and patients with more effective prevention and management strategies, ultimately reducing the incidence of diabetes and improving public health. Through the conduct of this research, it is hoped to provide important theoretical and practical support for further research and application in the field of diabetes prediction <ref type="bibr" target="#b4">[5]</ref>.</p><p>The dataset utilized in this study comprises 520 samples. After data cleansing, anomalies and missing values were addressed, and features in the dataset were processed, including transformation, encoding, and standardization. Additionally, we explored the distribution of age data by plotting the histogram of the target variable. The histogram revealed a normal distribution of age data. For categorical variables, we employed one-hot encoding for transformation. For numerical variables, standardization was conducted to ensure uniform scales across different features <ref type="bibr" target="#b5">[6]</ref>.</p><p>In response to the issue of excessive or redundant features, feature selection was performed to identify features with significant impact on the target variable. As a result, 16 features were retained, as depicted in the following figure. These features appear to exhibit strong multicollinearity among them.</p><p>For categorical variables, we employed one-hot encoding for transformation. For numerical variables, standardization was conducted to ensure uniform scales across different features.</p><p>In response to the issue of excessive or redundant features, feature selection was performed to identify features with significant impact on the target variable. As a result, 16 features were retained, as depicted in the following figure. These features appear to exhibit strong multicollinearity among them. In this study, we adopted the mean square error (MSE) as the index to evaluate the prediction performance of the model. The mean squared error is a measure of the squared difference between the predicted value of the model and the actual observed values. It is calculated by summing the squared differences between the predicted value and the actual observed value for each sample and then dividing by the number of samples. A smaller MSE indicates a more accurate model prediction <ref type="bibr" target="#b6">[7]</ref>.</p><formula xml:id="formula_0">𝑀𝑀𝑀𝑀𝑀𝑀 = 1 𝑚𝑚 ��𝑦𝑦 𝑖𝑖 − 𝑦𝑦 �� 2 𝑚𝑚 𝑖𝑖=1<label>(1)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Model Establishment</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Normal equations in linear regression</head><p>Linear regression is a statistical method used to establish a relationship between an independent variable (or feature) and a dependent variable. It assumes that there is a linear relationship between the independent variable and the dependent variable, that is, it can be described by a straight line. The goal of linear regression is to find the best-fit line that minimizes the error between the predicted value and the observed value. Normal equations are a method used to solve the parameters of linear regression models. It finds the best fitting line by minimizing the sum of squared residuals.</p><p>The linear regression model is of the form.</p><formula xml:id="formula_1">𝑦𝑦 = β 0 + β 1 𝑥𝑥 1 + β 2 𝑥𝑥 2 + ⋯ + β 𝑛𝑛 𝑥𝑥 𝑛𝑛 +<label>(2</label></formula><p>) By taking the partial derivatives of the model parameters and setting them equal to zero, we can obtain the normal equation.</p><formula xml:id="formula_2">β = (𝑋𝑋 𝑇𝑇 𝑋𝑋) −1 𝑋𝑋 𝑇𝑇 𝑦𝑦 (3)</formula><p>By taking the partial derivative of the loss function with respect to the parameter vector 𝛽𝛽, setting it equal to zero, and then solving for the parameter vector 𝛽𝛽, we can obtain the normal form of the above equation.</p><p>We fit a linear regression model to the training data in our study and evaluate the model performance on the test set. The accuracy and generalization ability of the model were evaluated by calculating the coefficients and intercepts of the model, as well as using the mean squared error <ref type="bibr" target="#b7">[8]</ref>.</p><p>Below is the confusion matrix for the mean variance of the linear regression equation model </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Linear Regression: Gradient Descent</head><p>Stochastic Gradient Descent (SGD) is an optimization algorithm used to train machine learning models, especially on large-scale datasets. It is an iterative optimization algorithm in which the objective function of each round calculation is no longer the whole sample error, but only a single sample error, that is, only one sample at a time is substituted to calculate the gradient of the objective function to update the weight, and then the next sample is repeated until the loss function value stops decreasing or the loss function value is less than a tolerable threshold. Its equation can be described as Stochastic Gradient Descent (SGD) is an optimization algorithm used to train machine learning models, especially on large-scale datasets. It is an iterative optimization algorithm in which the objective function of each round calculation is no longer the whole sample error, but only a single sample error, that is, only one sample at a time is substituted to calculate the gradient of the objective function to update the weight, and then the next sample is repeated until the loss function value stops decreasing or the loss function value is less than a tolerable threshold. Its equation can be described as</p><formula xml:id="formula_3">θ = θ − α∇𝐽𝐽�θ; 𝑥𝑥 (𝑖𝑖) , 𝑦𝑦 (𝑖𝑖) � (4)</formula><p>After the model is trained, the coefficients and bias terms of the model are obtained, and the test set is used to make predictions. Then, the mean square error (MSE) was used as the evaluation index to calculate the difference between the model prediction results and the true value. The final evaluation results show that the model has a small mean square error, indicating that it performs well in predicting the target variable. Compared to the previous model using the normal equation method, this model based on stochastic gradient descent is more accurate and generalizable, better able to adapt to changes in the data and provide more reliable predictions</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3">Ridge Regression Models in L2 Regularization</head><p>Ridge regression is an extended form of linear regression. It introduces an L2 norm penalty term to constrain the complexity of the model, so as to avoid the overfitting problem. In ridge regression, the goal is to minimize the sum of a loss function and an L2-norm penalty term, where the loss function is usually the sum of squared residuals (RSS) <ref type="bibr" target="#b8">[9]</ref>.</p><formula xml:id="formula_4">min β �|𝑦𝑦 − 𝑋𝑋β|� 2 2 + α�|β|� 2 2 (5)</formula><p>Our goal is to minimize a loss function that consists of a squared loss term and an L2 regularization term of the following form.</p><formula xml:id="formula_5">𝐽𝐽(𝜽𝜽) = 1 2𝑚𝑚 �(ℎ 𝜽𝜽 (𝐱𝐱 𝐢𝐢 ) − 𝑦𝑦 𝑖𝑖 ) 2 𝑚𝑚 𝑖𝑖=1 + 𝜆𝜆 � 𝜃𝜃 𝑗𝑗 2 𝑛𝑛 𝑗𝑗=1<label>(6)</label></formula><p>The analytical solution of ridge regression can be obtained by least squares method. By taking the derivative of the objective function and setting the derivative to zero, an analytical expression for the regression coefficients can be obtained. 𝜽𝜽 = (𝑋𝑋 𝑇𝑇 𝑋𝑋 + 𝜆𝜆𝜆𝜆) −1 𝑋𝑋 𝑇𝑇 𝑦𝑦 <ref type="bibr" target="#b6">(7)</ref> Ridge regression can effectively reduce the amplitude of regression coefficients by introducing L2 regularization term, thereby reducing the complexity of the model and avoiding overfitting. In addition, ridge regression can also deal with multicollinearity problems and improve the stability and generalization ability of the model <ref type="bibr" target="#b9">[10]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4">Lasso (Least Absolute Shrinkage and Selection Operator) mode.</head><p>Lasso model is a regularization method based on linear regression. Its core idea is to minimize the loss function while adding the L1 norm penalty term, so that the model parameters tend to be sparse, and the coefficients of some features are compressed to zero to realize feature selection and model simplification.</p><p>The objective is to minimize a loss function, which consists of a squared loss term and an L1 regularization term of the following form.</p><formula xml:id="formula_6">J(𝛉𝛉) = 1 2m �(h 𝛉𝛉 (𝐱𝐱 𝐢𝐢 ) − y i ) 2 m i=1 + λ ��θ j � n j=1<label>(8)</label></formula><p>An efficient solution to Lasso is coordinate descent. The method updates only one coefficient at each step and treats the other coefficients as constants. Specifically, for each coefficient we fix it in turn, then minimize the objective function, and update this process iteratively until convergence.</p><formula xml:id="formula_7">θ j = argmin θ j � 1 2m �(h θ (x i ) − y i ) 2 m i=1 + λ ��θ j � n j=1 �<label>(9)</label></formula><p>An important property of Lasso Regression is that it tends to eliminate unimportant weights <ref type="bibr" target="#b10">[11]</ref>.</p><p>For example: for relatively large values of α, higher-order polynomials degenerate to quadratic or even linear: higher-order polynomial features. The weight of is set to 0.</p><p>That is, Lasso Regression can automatically perform feature selection and output a sparse model (only A few features have nonzero weights) Sub gradient vectors for Lasso Regression</p><formula xml:id="formula_8">𝐽𝐽(β) = 1 2𝑛𝑛 �|𝑦𝑦 − 𝑋𝑋β|� 2 2 + α�|β|� 1 (<label>10</label></formula><formula xml:id="formula_9">)</formula><p>ROC curve is a common tool used to evaluate the performance of classification models. It shows the relationship between True Positive Rate and False Positive Rate. Based on these results, we can conclude that ridge regression model and Lasso regression model are better choices for predicting the occurrence of diabetes, they can explain the data more accurately and have smaller prediction error</p><p>The Ridge Regression and Lasso Regression models perform well on this dataset, with low mean square error, mean absolute error, and root mean square error, and r-squared values close to 1, indicating a good fit to the data. In contrast, the Normal Equation and Stochastic Gradient Descent models perform poorly, with negative values of R² indicating that the model fails to fit the data well <ref type="bibr" target="#b11">[12]</ref>.</p><p>Here is an overview of the training of the four models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Solutions And Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">The Solution of Trending model</head><p>1. Comparison between normal equation model and Stochastic Gradient Descent (SGD) model:</p><p>The normal equation model and the SGD model have the same performance metrics, which indicates that they produce similar prediction results <ref type="bibr" target="#b12">[13]</ref>.</p><p>However, these models have negative R-squared values (-0.479319), indicating that their performance is below the horizontal line that passes through the mean of the data.</p><p>The mean square error (MSE) and root mean square error (RMSE) are relatively high, indicating a large difference between the predicted and actual values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Ridge regression model and Lasso regression model</head><p>The ridge regression and Lasso regression models significantly outperform the normal equation and SGD models in terms of performance.</p><p>The R-squared values for these two models (0.651793) indicate that they explain about 65%, indicating that the models fit the data very well.</p><p>The mean square error (MSE) and root mean square error (RMSE) are relatively low, indicating that the difference between the predicted and actual values is small compared to the normal equation and the SGD model.</p><p>Furthermore, the two models have relatively low mean absolute error (MAE), indicating their high accuracy in predicting the target variable <ref type="bibr" target="#b13">[14]</ref>. In general, indicators of the models in the experiment, ridge regression and Lasso regression models are better than the normal equation and SGD models in terms of prediction accuracy and fitting </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Model tuning as well as supplemental experiments</head><p>(1) In this experiment, we explored one of the most common regularization techniques used in linear regression models, Lasso regression introduces the L1 term into the loss function to make the model parameters sparse, so as to achieve the effect of feature selection and dimensionality reduction. However, in practice, the performance of Lasso regression is affected by the regularization parameter alpha. Therefore, this experiment aims to further optimize the performance of the Lasso regression model by tuning the regularization parameter alpha <ref type="bibr" target="#b14">[15]</ref>.</p><p>Firstly, we selected a representative dataset and performed data preprocessing and preparation. We then built a basic Lasso regression model and used grid search with k-fold cross-validation to find the best regularization parameter over a range of predefined alpha values. This process aims to minimize the fitting error of the model on the training data while maintaining the ability to generalize to new data.</p><p>After the parameter tuning was completed, we retrained the Lasso regression model and evaluated the performance on an independent test set. We evaluated the prediction accuracy, generalization ability, and robustness of the model and quantified the performance of the model using metrics such as mean square error, mean absolute error, R-squared score, and root mean square error.</p><p>Through the analysis of the Test results, we conclude that the model is better than that of the basic model. It shows that by adjusting the regularization parameter appropriately, the fitting effect and generalization ability of Lasso regression model can be improved, and its application value in practical problems can be improved. In addition, we discuss the limitations of the experimental results and future research directions in order to further refine and advance the research in this area. mean square error, mean absolute error, Rsquared score, and root mean square error <ref type="bibr" target="#b15">[16]</ref>.</p><p>(2) Feature selection is a key step in machine learning and statistical modeling. It aims to pick out the most predictive features from the original feature set to improve the performance and generalization ability of the model. Feature selection can help reduce the dimensionality of the data, reduce the complexity of the model, improve the interpretability of the model, and speed up the model training process. In practice, feature selection is one of the key steps in building efficient and reliable machine learning models.</p><p>Lasso regression is a commonly used linear regression method, which has the ability of automatic feature selection. By adding an L1 regularization term to the objective function, Lasso regression can compress the coefficients of some features to zero, thus achieving feature sparsity. This allows Lasso regression to perform well on datasets with a large number of features and to discover the most relevant features.</p><p>The purpose of this experiment is to explore the effect of different feature selection methods on the performance of Lasso regression models. We will compare three commonly used feature selection methods: Feature selection is a key step in machine learning and statistical modeling. Its purpose is to select the most predictive features from the original feature set to improve the performance and generalization of the model. Feature selection helps to reduce the dimensionality of the data, reduce the complexity of the model, improve the interpretability of the model, and speed up the training process of the model. In practice, feature selection is one of the key steps in building efficient and reliable machine learning models.</p><p>Lasso regression is a commonly used linear regression method, which has the ability of automatic feature selection. By adding an L1 regularization term to the objective function, Lasso regression can compress the coefficients of some features to zero, thus achieving feature sparsity. This allows Lasso regression to perform well on datasets with a large number of features and to discover the most relevant features.</p><p>The purpose of this experiment is to explore the effect of different feature selection methods on the performance of Lasso regression model. We will compare three commonly used feature selection methods: Wrapper method, Filter method and Embedded method, and analyze their effect in Lasso regression model Before performing feature selection, we first trained a Lasso regression model on the original feature set as a baseline model. We recorded the performance metrics of the baseline model, including mean squared error (MSE), Mean absolute error (MAE), Rsquared (R-²), and others <ref type="bibr" target="#b16">[17]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Comparison of feature selection methods</head><p>Next, we used three different feature selection methods and compared their impact on the performance of the Lasso regression model:</p><p>Wrapper method (RFE) : Recursive Feature Elimination (RFE) method is used for feature selection. RFE works by repeatedly training the model and gradually eliminating the least important features until the desired number of features is reached. We chose an appropriate number of features and trained the Lasso regression model on these features.</p><p>Filter method (correlation coefficient) : Use correlation coefficient for feature selection. The correlation coefficient measures how linearly correlated the features are with the target variable. We selected the features with the highest correlation with the target variable and trained a Lasso regression model on these features.</p><p>Embedded method (L1 regularization) : A feature selection mechanism that uses L1 regularization itself. L1 regularization can compress the coefficients of some features to zero to achieve feature sparsity. We trained a Lasso regression model and recorded the features corresponding to nonzero coefficients.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4">Performance comparison</head><p>Finally, we compare the Lasso regression model performance under the three feature selection methods. We analyze the differences in model performance metrics among the various methods and explore the reasons behind these differences. Through the performance comparison, we draw conclusions and recommendations regarding the selection and application of feature selection methods <ref type="bibr" target="#b17">[18]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion and Outlook</head><p>The method of feature selection using Lasso regression model in our study has shown good results on this problem. By regularizing the penalty, the model is able to automatically select features that have a significant impact on the prediction of diabetes, thus improving the generalization ability and interpretability of the model <ref type="bibr" target="#b18">[19]</ref>.</p><p>Combined with the feature selection mechanism of Lasso regression, the model has strong interpretability <ref type="bibr" target="#b19">[20]</ref>. We can get a clear picture of which features play a key role in predicting diabetes, which can help medical researchers to understand the pathogenesis of the disease, and evaluate the results based on our model, We can see that the model performs well on metrics such as mean squared Error (MSE) Mean absolute error (MAE Rsquared (R^2) root mean squared error (RMSE) <ref type="bibr" target="#b20">[21]</ref>. This indicates that our model can predict the occurrence of diabetes relatively accurately while maintaining high precision and recall</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head> Further research</head><p>Based on the performance of the current model, we can further explore how to improve the model and improve its predictive performance <ref type="bibr" target="#b21">[22]</ref>. For example, we can try other regularization methods or use more complex model architectures to improve the prediction accuracy of our model <ref type="bibr" target="#b22">[23]</ref>.</p><p>In summary, our developed model shows good performance and high explanatory power in predicting the occurrence of diabetes, which provides an important reference for further medical research. We can use machine learning methods for analysis and prediction. Through experiments, it is not difficult to find that the second-order polynomial regression <ref type="bibr" target="#b23">[24]</ref>.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Distribution histogram of age.</figDesc><graphic coords="2,185.83,497.25,223.66,168.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Heavy correlations between features.</figDesc><graphic coords="3,164.18,298.65,266.87,198.75" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Confusion matrix for the mean variance of the positive regression equation model.</figDesc><graphic coords="4,178.20,475.64,238.95,199.60" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Mean error comparison between the SGD model and the general linear mode.</figDesc><graphic coords="5,176.03,409.61,243.26,192.85" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: About the roc curve of this model.</figDesc><graphic coords="7,176.03,321.95,243.30,162.05" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: Four evaluation metrics for the four regression models.</figDesc><graphic coords="8,95.93,446.66,402.94,89.00" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 7 :</head><label>7</label><figDesc>Figure 7: training of the four models.</figDesc><graphic coords="9,101.80,85.05,391.73,198.20" type="bitmap" /></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A New Hybrid Feature Selection Method Based on Lasso Regression and Rough Set Theory for Microarray Data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="17006" to="17016" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Hybrid Forecasting of Ozone Concentration Using Extreme Learning Machine and Lasso Regression Based on Deep Feature Selection</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="14142" to="14154" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Comparative Study for the LASSO Regression Method for the Number of Bedrooms Determination in Real Estate Market in Egypt</title>
		<author>
			<persName><forename type="first">H</forename><surname>Abdelaal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nassar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Elshoush</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hamed</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Science, Technology and Innovation (IEREK Interdisciplinary Series for Sustainable Development)</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="109" to="116" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">An Ensemble Learning Based Feature Selection Method for Internet of Things Data</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Internet of Things Journal</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="3715" to="3723" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Acceleration of Lasso-Based Feature Selection on Cloud</title>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Parallel and Distributed Systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="62" to="74" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A Novel Feature Selection Method Based on Elastic Net for Enhancing Brain-Computer Interface Performance</title>
		<author>
			<persName><forename type="first">L</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Neural Systems and Rehabilitation Engineering</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="page" from="1727" to="1735" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Hyperspectral Image Classification with Deep Convolutional Neural Networks and Lasso Regression Based on Feature Selection</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Lin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page">679</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Efficiently Feature Selection for Lasso-Based Machine Learning on Internet of Things Data</title>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Industrial Informatics</title>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">An Improved Feature Selection Method for Magnetic Resonance Brain Image Classification</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Liu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Medical Imaging and Health Informatics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="641" to="650" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Multi-Source Heterogeneous Data Feature Fusion Based on Elastic Net Regularization and Cross-Validation Algorithm</title>
		<author>
			<persName><forename type="first">L</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zheng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="3895" to="3905" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Predicting the Organic Solar Cell Efficiency via Machine Learning and Feature Selection</title>
		<author>
			<persName><forename type="first">W</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advanced Materials Technologies</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="issue">12</biblScope>
			<biblScope unit="page">2100737</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Feature Selection Based on Deep Learning and Elastic Net for Breast Cancer Classification</title>
		<author>
			<persName><forename type="first">X</forename><surname>Tian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Healthcare Engineering</title>
		<imprint>
			<biblScope unit="volume">2021</biblScope>
			<biblScope unit="page">5566511</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Lasso Regression Feature Selection Based on Optimal Weight of DCT Transform and Extreme Learning Machine</title>
		<author>
			<persName><forename type="first">X</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="144501" to="144510" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A Machine Learning Method Based on Lasso Regression and Ridge Regression for the Prediction of Groundwater Level</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Water</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="issue">21</biblScope>
			<biblScope unit="page">3032</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">An Improved Elastic Net Method with Features Selection for Power System Stability Margin Prediction</title>
		<author>
			<persName><forename type="first">X</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="134380" to="134388" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">A Hybrid Feature Selection Method Based on Lasso and Genetic Algorithm for Text Classification</title>
		<author>
			<persName><forename type="first">H</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Information Sciences</title>
		<imprint>
			<biblScope unit="volume">546</biblScope>
			<biblScope unit="page" from="160" to="179" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A Hybrid Machine Learning Method Based on Feature Selection and Lasso Regression for Classifying Brain Disorder</title>
		<author>
			<persName><forename type="first">W</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dong</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Biomedical Engineering</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="page" from="3719" to="3727" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">A New Hybrid Feature Selection Method Based on Lasso and FCBF for Bioinformatics Data Classification</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
			<publisher>BioMed Research International</publisher>
			<biblScope unit="page">5535983</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">A Novel Hybrid Feature Selection Method Based on Lasso and Rough Sets Theory for Fault Diagnosis of Rotating Machinery</title>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Complexity</title>
		<imprint>
			<biblScope unit="page" from="1" to="14" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">A New Hybrid Feature Selection Method Based on Lasso and Fast Relief for Mechanical Fault Diagnosis</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Complexity</title>
		<imprint>
			<biblScope unit="page" from="1" to="11" />
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">A New Hybrid Feature Selection Method Based on Lasso Regression and Rough Set Theory for Microarray Data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="17006" to="17016" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Feature Selection for Intrusion Detection Based on Lasso Regression and Ant Colony Optimization</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Shang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="105499" to="105509" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">A New Feature Selection Method Based on Lasso and Principal Component Analysis for Stock Market Prediction</title>
		<author>
			<persName><forename type="first">L</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="130051" to="130061" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">A Novel Hybrid Feature Selection Method Based on Lasso and Recursive Feature Elimination for Tumor Classification</title>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Complexity</title>
		<imprint>
			<biblScope unit="page" from="1" to="10" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
