<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Deep Spatio-Temporal Encoding: Achieving Higher Accuracy by Aligning with External Real-World Data</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Chen</forename><surname>Jiang</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Auburn University</orgName>
								<address>
									<settlement>Auburn</settlement>
									<region>AL</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Wenlu</forename><surname>Wang</surname></persName>
							<email>wenlu.wang@tamucc.edu</email>
							<affiliation key="aff1">
								<orgName type="institution">Texas A&amp;M University-Corpus Christi</orgName>
								<address>
									<addrLine>Corpus Christi</addrLine>
									<region>TX</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jingjing</forename><surname>Li</surname></persName>
							<email>jingjingli@meta.com</email>
							<affiliation key="aff2">
								<address>
									<addrLine>Meta</addrLine>
									<settlement>Menlo Park</settlement>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Naiqing</forename><surname>Pan</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Auburn University</orgName>
								<address>
									<settlement>Auburn</settlement>
									<region>AL</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Wei-Shinn</forename><surname>Ku</surname></persName>
							<email>weishinn@auburn.edu</email>
							<affiliation key="aff0">
								<orgName type="institution">Auburn University</orgName>
								<address>
									<settlement>Auburn</settlement>
									<region>AL</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Deep Spatio-Temporal Encoding: Achieving Higher Accuracy by Aligning with External Real-World Data</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">1EFE386AD5DB4EF4E57B8CB63CCF3079</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:17+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Spatial-temporal Analysis</term>
					<term>Deep Learning</term>
					<term>Transformer</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Spatio-temporal deep learning has drawn a lot of attention since many downstream real-world applications can benefit from accurate predictions. For example, accurate prediction of heavy rainfall events is essential for effective urban water usage, flooding warning, and mitigation. In this paper, we propose a strategy to leverage spatially connected real-world features to enhance prediction accuracy. Specifically, we leverage spatially connected real-world climate data to predict heavy rainfall risks in a broad range in our case study. We experimentally ascertain that our Trans-Graph Convolutional Network (TGCN) accurately predicts heavy rainfall risks and real estate trends, demonstrating the advantage of incorporating external spatially-connected real-world data to improve model performance, and it shows that this proposed study has a significant potential to enhance spatio-temporal prediction accuracy, aiding in efficient urban water usage, flooding risk warning, and fair housing in real estate.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Spatio-temporal predictions have been extensively studied due to their impact on real-world applications <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5]</ref>. For example, heavy rainfall events can cause significant damage to infrastructure and pose serious threats to human safety. Predicting these events with greater accuracy allows better preparation and response <ref type="bibr" target="#b5">[6]</ref>, ultimately saving lives and reducing the economic impact of such events.</p><p>Deep learning methods, such as deep spatio-temporal prediction models <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref>, have improved the performance of rainfall forecasting over the years. However, the role of external data in enhancing the prediction accuracy is still controversial. Some argue that external data can provide more useful information for the prediction model, while others claim that external data can introduce more noise and complexity to the learning process. In this study, we propose to improve spatio-temporal predictions by combin-ing spatially-linked external real-world data along with a TGCN to learn the spatio-temporal dependencies from the combined data. As it has been proven that utilizing more multi-source real-world data is more likely to lead to higher accuracy <ref type="bibr" target="#b8">[9]</ref>, our study aims to introduce a fresh perspective on integrating external real-world data into the proposed framework. We use heavy rainfall prediction as a case study for our proposed method, and overall we aim to provide accurate spatio-temporal predictions by leveraging as much information as possible, enabling better decision-making for a broad range of spatio-temporal applications and at the same time offering a novel angle and a comprehensive evaluation to demonstrate the feasibility of integrating additional external real-world data without the necessity of customizing transformer attention mechanisms. Our approach is experimentally validated by predicting heavy rainfall events and real estate hotspots.</p><p>The traditional method for predicting heavy rainfall involves manually engineering features from weather data, including temperature, pressure, humidity, etc. Meteorologists rely on their expertise to interpret this data and forecast future weather patterns. This process entails observing and analyzing atmospheric factors to predict weather patterns. However, this traditional approach is time-consuming, labor-intensive, and susceptible to human error, especially when dealing with large datasets. As data grows, it becomes increasingly challenging to analyze large amounts of information by hand.</p><p>Previous research has investigated using deep learning for precipitation prediction <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref> with promising results. However, some limitations can be significantly improved to enhance deep model performance. One area with room for enhancement is leveraging spatial dependencies. To tackle this challenge, we propose a model that integrates both Graph Convolution Networks (GCNs) and a Transformer. This model enables combining external spatially-linked data for spatio-temporal predictions.</p><p>Specifically, we employ a GCN to analyze the adjacency matrix on a grid level and generate correlations between each grid element. The GCN captures the spatial relationships and dependencies among neighboring grid points, allowing for a comprehensive understanding of the data's spatial dynamics. We then utilize a Transformer model to encode the temporal precipitation data and combine it with the spatial correlations obtained from the GCNs. By combining the GCNs and the Transformer within the proposed TGCN model, we create a framework that harnesses both the spatial and temporal dimensions of the data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Graph Neural Networks</head><p>Graph Convolutional Networks (GCNs) are a type of deep learning model designed to process data represented in a graph structure, such as social or sensor networks <ref type="bibr" target="#b11">[12]</ref>. GCNs have demonstrated their effectiveness in various applications, including node classification, link prediction, and recommendation systems <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15,</ref><ref type="bibr" target="#b15">16]</ref>. The concept of Graph Neural Networks (GNNs) was initially introduced in <ref type="bibr" target="#b16">[17]</ref> and further expanded upon in subsequent research by <ref type="bibr" target="#b17">[18]</ref>. GNNs, a type of recurrent neural network (RNN), iteratively propagate information from neighboring nodes until reaching a stable fixed point. This iterative process has traditionally been computationally expensive, but recent studies, such as <ref type="bibr" target="#b18">[19]</ref>, have made significant improvements in this area. Inspired by the success of Convolutional Neural Networks (CNNs) in computer vision, which extract highlevel features from images using convolution and pooling layers, current models aim to adapt these layers to directly process graph inputs. GCNs can be categorized into two types of graph convolution layers: spectral graph convolution and localized graph convolution, as discussed in <ref type="bibr" target="#b19">[20]</ref>. Early research primarily focused on spectral graph convolutions, pioneered by <ref type="bibr" target="#b20">[21]</ref>. The current state-of-the-art model, GCN, further simplified the graph convolution operation by employing a localized first-order approximation. However, spectral methods require operations on the entire graph Laplacian during training, which can be computationally expensive. Several subsequent works, such as FastGCN <ref type="bibr" target="#b21">[22]</ref> have aimed to alleviate this issue. Recently, researchers have explored the application of GCNs in time series prediction. For example, spatio-temporal GCNbased approaches have been proposed for traffic flow prediction <ref type="bibr" target="#b22">[23]</ref>, and the integration of time-aware topological information into GCNs using the mathematical framework of zigzag persistence <ref type="bibr" target="#b23">[24]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Spatial Temporal Prediction</head><p>In this section, we discuss various existing temporal and spatial-temporal forecasting methods. For example, Recurrent Neural Networks (RNNs), especially long-short-term memory (LSTM) <ref type="bibr" target="#b24">[25]</ref>, have gained popularity in time series forecasting <ref type="bibr" target="#b25">[26]</ref>. Convolutional Neural Networks (CNN) and its variant Temporal Convolutional Neural Networks (TCN) are another option for sequence prediction <ref type="bibr" target="#b26">[27]</ref>, offering parallel computations compared to RNNs <ref type="bibr" target="#b27">[28]</ref>. In recent years, researchers have explored Transformers and its variants in time series forecasting, achieving state-ofthe-art performance in tasks like energy consumption and stock market <ref type="bibr" target="#b28">[29,</ref><ref type="bibr" target="#b29">30,</ref><ref type="bibr" target="#b30">31]</ref>. Designing a model capable of comprehensively capturing both spatial and temporal patterns represents another emerging trend in spatial-temporal prediction tasks <ref type="bibr" target="#b31">[32,</ref><ref type="bibr" target="#b32">33]</ref>. For example, <ref type="bibr" target="#b32">[33]</ref> introduced a spatial-temporal graph neural network for predicting traffic flow.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>In this section, we detail our model architecture and the benefits of our design.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Overview</head><p>The architecture we propose, illustrated in Figure <ref type="figure" target="#fig_0">2</ref>, incorporates a combination of techniques to enhance the prediction model. We begin by utilizing a transformer encoder to effectively encode the time series precipitation data, and then integrate local climate features into the model, enabling a comprehensive understanding of the factors influencing heavy rainfall.</p><p>To address spatial dependencies and relationships among grid points, a GCN is introduced. This GCN learns the spatial dependencies within the dataset, considering the interconnectedness of grids based on their spatial locations. By leveraging the GCN, the model becomes capable of capturing and integrating spatial information, thereby enhancing prediction accuracy.</p><p>The latent code, which combines the encoded time series precipitation data and the spatially connected local climate features learned through the GCN, is fed into a multi-layer perceptron (MLP) for prediction. This integrated architecture allows the MLP model to leverage the fused information, including temporal precipitation data, other climate features, and spatial factors, to effectively learn and infer future heavy rainfall areas.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Model Architecture</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Preliminaries</head><p>Our proposed TGCN model consists of Encoder, GCNs and Multi-layer Perceptron (MLP) layers. The major component in the transformer is the Multi-head self-attention.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>𝐴𝑡𝑡𝑒𝑛𝑡𝑖𝑜𝑛(𝑄, 𝐾, 𝑉</head><formula xml:id="formula_0">) = 𝑠𝑜𝑓 𝑡𝑚𝑎𝑥( 𝑄𝐾 𝑇 √ 𝑑 𝑘 )𝑉<label>(1)</label></formula><p>Where the K and V are matrices that store the keys and values. Q is the query that will map against a set of keys.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2.">Transformer-based Encoder</head><p>We have developed a predictive model using the Transformer architecture, tailored for heavy rainfall forecasts. Unlike traditional methods that only use past rainfall data, our model factors in numerous external variables to boost accuracy. We examine local features, including geography, atmospheric conditions (pressure, temperature, wind), humidity, and topography, all of which influence heavy rainfall likelihood in a specific area. Therefore, we have developed a transformer-based prediction model <ref type="bibr" target="#b33">[34]</ref> that incorporates GCNs to process the spatial features. By doing so, our model can capture the spatial relationships among various features in a graph structure, such as the dependencies between grid point locations and their corresponding climate data. The integration of the GCNs enhances our model's ability to capture both temporal and spatial information. Our model design starts with a transformer encoder capturing temporal precipitation patterns, followed by embedding this data and merging it with local climate data like moisture and humidity. We enhance prediction accuracy with this added context. As illustrated in Figure <ref type="figure" target="#fig_1">3</ref>, GCNs involve learning a linear transformation of the feature vectors of each node in a graph, which is then used to update the node features by aggregating information from the node's neighbors. Mathematically, this can be expressed as:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3.">Graph Convolutional Networks</head><formula xml:id="formula_1">ℎ (𝑙+1) 𝑣 𝑖 = 𝜎 ⎛ ⎝ ∑︁ 𝑣 𝑗 ∈𝒩 (𝑣 𝑖 ) 1 𝑐𝑖𝑗 𝑊 (𝑙+1) ℎ (𝑙+1) 𝑣 𝑗 ⎞ ⎠<label>(2)</label></formula><p>In the equation, ℎ</p><formula xml:id="formula_2">(𝑙+1) 𝑣 𝑖</formula><p>represents the feature vector of node 𝑣𝑖 at layer 𝑙 + 1, 𝑊 (𝑙+1) denotes the learnable weight matrix for layer 𝑙 + 1, 𝒩 (𝑣𝑖) represents the set of neighbors of node 𝑣𝑖, and 𝑐𝑖𝑗 is a normalization constant that ensures proper scaling of the aggregated information. The function 𝜎 denotes a non-linear activation function, which introduces non-linearity into the model. In our specific case, we utilize the ReLU activation function. This equation can be interpreted as calculating a weighted sum of the feature vectors of the neighbors of node 𝑣𝑖 at layer 𝑙 + 1, where the weights are determined by the learned weight matrix 𝑊 (𝑙+1) . Then, a non-linear activation function is applied to obtain the updated feature vector ℎ</p><formula xml:id="formula_3">(𝑙+1) 𝑣 𝑖</formula><p>for node 𝑖 at layer 𝑙 + 1. This process is repeated across multiple layers to learn expressive representations of the graph data.</p><p>For the final prediction, we utilize a four-layer MLP model that combines time series data with other features, effectively leveraging both temporal and spatial information captured by our model for more accurate predictions.</p><p>By leveraging the transformer architecture, incorporating GCNs, and utilizing a four-layer MLP model, our approach enables the effective integration of temporal and spatial information for improved prediction accuracy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.4.">Jointly Learning</head><p>As illustrated in Figure <ref type="figure" target="#fig_0">2</ref>, we propose to map temporal data and non-temporal data into the same latent space and merge the latent vectors for the subsequent prediction task.</p><p>To encode the local climate features and capture the spatial dependencies among the grid points for data 𝑥𝑐, we employ a GCN to learn the relationships and dependencies within the spatial domain. The output hidden features at a specific layer 𝐿 can be denoted as ℎ (𝐿) 𝑐 . Equation 2 is applied in this context. Assuming we use 𝐿 𝑐 layers in total, and we use the final layer to summarize climate information, which is defined as:</p><formula xml:id="formula_4">hc = ℎ (𝐿 𝑐 ) 𝑣 (3)</formula><p>where ℎ  <ref type="formula" target="#formula_1">2</ref>. We encode temporal precipitation data using a transformer encoder <ref type="bibr" target="#b33">[34]</ref>,</p><formula xml:id="formula_5">ht = 𝑇 𝑟𝑎𝑛𝑠𝑓 𝑜𝑟𝑚𝑒𝑟𝐸𝑛𝑐𝑜𝑑𝑒𝑟(𝑥𝑡) (4) ht ∈ R 𝑑 𝑡<label>(5)</label></formula><p>. Since 𝑥𝑡 and 𝑥𝑐 are encoded as ht and hc, we define the merged hidden state as hm hm = 𝐶𝑂𝑁 𝐶𝐴𝑇 (ht, hc)</p><p>To further process the merged information, we use another multi-layer perceptron specifically trained for the prediction task. Similarly, we define the 𝑙-th layer network as (assuming 𝐿 𝑛 layers in total)</p><formula xml:id="formula_7">ℎ (𝑙) 𝑛 = 𝑅𝑒𝐿𝑢(𝑊 (𝑙) 𝑛 ℎ (𝑙−1) 𝑛 + 𝑏 (𝑙) 𝑛 )<label>(7)</label></formula><p>where ℎ We use the output from the last layer for prediction</p><formula xml:id="formula_8">𝑦 ¯= 𝑠𝑖𝑔𝑚𝑜𝑖𝑑(ℎ (𝐿 𝑛 ) 𝑛 )<label>(8)</label></formula><p>Loss is measured with the Binary Cross-Entropy loss (BCE)</p><formula xml:id="formula_9">𝑙𝑜𝑠𝑠 = 𝐵𝐶𝐸𝑙𝑜𝑠𝑠(𝑦 ¯, 𝑦)<label>(9)</label></formula><p>The binary cross entropy (BCE) loss can be formulated as follows:</p><formula xml:id="formula_10">𝐵𝐶𝐸𝑙𝑜𝑠𝑠 = − 1 𝑁 𝑁 ∑︁ 𝑖=1 [𝑦𝑖 log(𝑝𝑖) + (1 − 𝑦𝑖) log(1 − 𝑝𝑖)]<label>(10</label></formula><p>) where: 𝑁 is the total number of samples, 𝑦𝑖 is the true label for sample 𝑖, 𝑝𝑖 is the predicted probability 𝑖, log denotes the natural logarithm.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experimental Validation</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Datasets</head><p>Our data and code are publicly available 1 . In our dataset, the train and test split ratio is 7:3.</p><p>1 https://github.com/jiang28/Deep-Spatio-Temporal-Encoding</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.1.">Precipitation Dataset</head><p>Our precipitation dataset is sourced from the NOAA HRRR dataset<ref type="foot" target="#foot_0">2</ref> , offering real-time climate data at a 3 km spatial resolution and 1-hour temporal resolution. This dataset <ref type="bibr" target="#b34">[35]</ref> encompasses total precipitation, precipitation rate, and nine additional climate variables, including humidity (%), moisture availability (%), pressure (Pa), wind speed (m/s), and total cloud cover (%). Simulated brightness temperature data is acquired from the GOES 11 satellite<ref type="foot" target="#foot_1">3</ref> . The precipitation data consist of the following three types:</p><p>• Temporal precipitation data, denoted as 𝑥𝑡, as shown in Table <ref type="table" target="#tab_1">1</ref> and Figure <ref type="figure" target="#fig_6">5</ref>. It captures the historical patterns and fluctuations in precipitation over time. Specifically, we define the temporal precipitation rate and total accumulated precipitation over the past 6 hours as 𝑥 ℎ , which consists of 𝑁 timestamps: dataset represents a specific location within the study area, such as a region or a cell. To represent the relationships between these grid points, we used an adjacency matrix. In the adjacency matrix, a value of 0 indicates that two grid points are not neighbors, while a value of 1 denotes their neighboring relationship.</p><formula xml:id="formula_11">𝑥𝑡 = {𝑥 1 𝑡 , 𝑥</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.2.">Real-estate Dataset</head><p>The real estate dataset captures the dynamics of the U.S. real estate market by collecting spatially correlated data from multiple sources. It consists of 7,436 neighborhoods, 567 cities, 304 counties, 225 metros, and 50 states across the U.S. The data are connected through spatial locations, forming a multi-level spatial hierarchy. The dataset consists of three main components: census data, pricing history, and school district information. Here are some statistics about the real estate dataset:</p><p>• Spatial Hierarchy Levels: The dataset includes a multi-level spatial hierarchy, including information at the state, metro, county, city, and neighborhood levels. • Census Data: The census data consists of 16 variables related to various aspects of housing prices, personal income, demographics, and spatial information. • Pricing History: The dataset includes temporal housing price history for each neighborhood, spanning from 1996 to 2019. • School District Information: The dataset incorporates school district information. It provides details on the number of school districts present in each county within the studied area. Additionally, the dataset includes information on the top school district(s) within the region. Temporal data format. It has data on the grid id, longitude, latitude, grid points, grid spacing, vertical level, timestamps, total precipitation, and precipitation rate.</p><p>To facilitate the task of predicting real estate hotspots, the dataset is classified into two classes based on the house price increase rate for each neighorhood: 1 for hotspots and 0 for non-hotspots. The detailed settings of the Real-estate Dataset can be found in <ref type="bibr" target="#b35">[36]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Evaluation Metrics</head><p>We evaluate the performance of a classification system using various metrics, including Accuracy, Recall, Precision, F1-score, and ROC. These metrics are calculated based on the number of true positives (𝑡𝑝), false positives (𝑓𝑝), false negatives (𝑓𝑛), and true negatives (𝑡𝑛). Accuracy measures the proportion of observations, both positive and negative, that were correctly classified by the system, and can be computed using the formula:</p><formula xml:id="formula_12">𝑎𝑐𝑐 = 𝑡𝑝 + 𝑡𝑛 𝑡𝑝 + 𝑓𝑝 + 𝑡𝑛 + 𝑓𝑛</formula><p>Recall measures the proportion of true positives that were correctly identified by the system, and can be computed using the formula:</p><formula xml:id="formula_13">𝑟𝑒𝑐𝑎𝑙𝑙 = 𝑡𝑝 𝑡𝑝 + 𝑓𝑛</formula><p>Precision measures the proportion of identified positives that were actually true positives, and can be computed using the formula:</p><formula xml:id="formula_14">𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 = 𝑡𝑝 𝑡𝑝 + 𝑓𝑝</formula><p>F1-score is a weighted average of precision and recall, and provides a single measure of the system's accuracy on the dataset, and can be computed using the formula:</p><formula xml:id="formula_15">𝐹 1 = 2 * 𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 * 𝑟𝑒𝑐𝑎𝑙𝑙 𝑝𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 + 𝑟𝑒𝑐𝑎𝑙𝑙</formula><p>ROC (Receiver Operating Characteristic) curve is a graphical plot that illustrates the performance of a binary classifier system. It is created by plotting the True Positive Rate (TPR) against the False Positive Rate (FPR), which can be computed using the formulas:</p><formula xml:id="formula_16">𝑇 𝑃 𝑅 = 𝑡𝑝 𝑡𝑝 + 𝑓𝑛 𝐹 𝑃 𝑅 = 𝑓𝑝 𝑓𝑝 + 𝑡𝑛</formula><p>Overall, these metrics provide a comprehensive evaluation of a classification system's performance and can help identify areas for improvement.   Our study identifies heavy rainfall risk areas based on precipitation rate. Following the United States Geological Survey (USGS) standard <ref type="foot" target="#foot_2">4</ref> , we define the heavy rainfall risk as follows:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Heavy Rainfall Prediction</head><formula xml:id="formula_17">Class = {︃ 0, if 𝑅 &lt; 4 mm/hr 1, if 𝑅 ≥ 4 mm/hr</formula><p>Recognizing the significance of precipitation rate as a critical factor, our objective is to pinpoint areas that are susceptible to encountering heavy rainfall within the next hour. The classification into two classes simplifies the problem and provides a clear distinction between areas with different levels of heavy rainfall risk. Using a 4 mm/hour</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>When comparing model performance on the Precipitation dataset, the proposed model has achieved an accuracy of 86.6%.</p><p>threshold, we classify areas as either low-risk (labeled as 0) or high-risk (labeled as 1). For example, out of 10,000 grid points in the study area, 4,798 have a potential for heavy rain risk, while 5,202 do not. This classification simplifies decision-making and resource allocation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Baselines</head><p>We use the following baseline methods:</p><p>• Random Forest (RF) <ref type="bibr" target="#b36">[37]</ref> • Support Vector Machine (SVM) <ref type="bibr" target="#b37">[38]</ref> • Decision Tree (DT) <ref type="bibr" target="#b38">[39]</ref> • Linear Regression (LR) <ref type="bibr" target="#b39">[40]</ref> • Multilayer Perceptron (MLP) <ref type="bibr" target="#b40">[41]</ref> • Long Short Term Memory (LSTM) <ref type="bibr" target="#b24">[25]</ref> • Transformer <ref type="bibr" target="#b33">[34]</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Performance Analysis</head><p>Based on the results presented in Table <ref type="table">2</ref> and Table <ref type="table">3</ref>, we can analyze the performance of different models on the Real Estate dataset and the Precipitation dataset, respectively. In Table <ref type="table">2</ref>, the proposed model outperforms all the baseline models with an accuracy of 95.6%. The proposed model also exhibits the highest precision for both classes (0 and 1), achieving 0.93 and 0.97, respectively. It demonstrates high recall values for both classes as well. The F1 scores are also higher for the proposed model compared to the baseline models, indicating a better balance between precision and recall. The TGCN model's performance is further reflected in the ROC score of 0.954, which indicates its ability to discriminate between the two classes effectively.</p><p>Table <ref type="table">3</ref> shows that the proposed model again achieves the highest accuracy of 86.6%. Similar to the Real Estate dataset, the TGCN model demonstrates superior precision and recall values for both classes compared to the baseline models. It achieves precision scores of 0.9 and 0.83 for classes 0 and 1, respectively, along with recall scores of 0.82 for class 0 and 0.85 for class 1. The F1 scores also indicate the TGCN model's overall better performance. The ROC score for the TGCN model is 0.867.</p><p>These results demonstrate that the proposed TGCN model consistently outperforms the other models on both datasets in terms of accuracy, precision, recall, F1 score, and ROC score. The TGCN model's ability to capture temporal, nontemporal, and spatial information through its integration of the transformer layer and the graph convolutional network contributes to its good performance in identifying and predicting hotspots and heavy rainfall areas.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In conclusion, the accurate prediction of heavy rainfall events is crucial for effective urban water usage, disaster response, and mitigation efforts. This paper proposed a prediction model that leverages spatially connected features and real-world climate data to predict heavy rainfall risks across a broad range. Through extensive experimentation, it was observed that the TGCN model outperformed the other machine learning methods in forecasting both heavy rainfall events and real estate trends.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Future Work and Limitations</head><p>While this study successfully demonstrated the effectiveness of the proposed TGCN model in predicting heavy rainfall risks, there are several avenues for future research and improvement.</p><p>We plan to incorporate more diverse and comprehensive datasets, including additional meteorological and geographical features. This expansion has the potential to enhance the accuracy and generalizability of the TGCN model. Furthermore, we are considering the integration of real-time data streams and the utilization of advanced data fusion techniques to further enhance the model's forecasting capabilities.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Design Flow of the Trans-Graph Convolutional Prediction Model: The Trans-Graph Convolutional Prediction Model incorporates a transformer layer for time-series precipitation data, a GCN for local climate features and spatial relationships among grid points, and a four-layer MLP model for the final prediction.</figDesc><graphic coords="3,83.50,65.61,428.28,186.98" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Graph Convolutional Network Architecture: The input data consists of the spatial relation matrix and spatially connected climate data. The nodes in the figure are for illustrative purposes.</figDesc><graphic coords="3,101.02,374.42,155.64,269.40" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>In this equation, ℎ𝑐 represents the hidden features at layer 𝐿, which are obtained by applying the ReLU activation function to the sum of the weighted input features 𝑊</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head></head><label></label><figDesc>𝑛is the input of the (l-1)-th layer in the i-th position. 𝑊 (𝑙) 𝑛 and 𝑏 (𝑙) 𝑛 are model parameters.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Study Area:</head><label></label><figDesc>Figure 4 presents the location of the study area in this study. It consists of 10,000 grids across the state of Florida in the U.S.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The study area consists of 10,000 grids across South Florida in the United States. The figure shows the observed precipitation values in each county within this area.</figDesc><graphic coords="5,337.67,231.53,157.52,157.22" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: Study Area Precipitation Rate Heatmap: 100x100 grid region on September 28, 2022, at 13:00 (mm/s).</figDesc><graphic coords="5,315.32,449.75,202.23,111.24" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>2 𝑡 , ..., 𝑥 𝑁 𝑡 } 𝑥 𝑖 𝑡 𝑖∈{1..𝑁 } represents the average price for the 𝑖-th timestamp. • Local climate data 𝑥𝑐: The dataset comprises twelve local climate variables, including temperature, humidity, wind speed, atmospheric pressure, and various other meteorological factors. • Spatial location data 𝑥𝑠: Each grid point in the</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc></figDesc><table><row><cell>GridID</cell><cell>Longitude</cell><cell>Latitude</cell><cell>Grid Points</cell><cell>Grid Spacing</cell><cell>Vertical Level</cell></row><row><cell>1</cell><cell>122.71</cell><cell>21.13</cell><cell>1799 × 1059</cell><cell>3 km</cell><cell>50</cell></row><row><cell>Time Stamps</cell><cell>2022/09/23 00:00</cell><cell>2022/09/23 01:00</cell><cell>2022/09/23 02:00</cell><cell>...</cell><cell>2022/10/02 00:00</cell></row><row><cell>Precipitation rate (mm/hour)</cell><cell>0.0</cell><cell>0.72</cell><cell>0.94</cell><cell>...</cell><cell>0</cell></row><row><cell>Total Precipitation (mm)</cell><cell>0.01</cell><cell>1.88</cell><cell>4.3</cell><cell>...</cell><cell>31.61</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">https://rapidrefresh.noaa.gov/hrrr/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://www.goes.noaa.gov/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://www.usgs.gov/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgement</head><p>This work was partially supported by the National Science Foundation (NSF) under Grant No. 2318641. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not reflect the views of the National Science Foundation.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0" />			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Bert-trip: Effective and scalable trip representation using attentive contrast learning</title>
		<author>
			<persName><forename type="first">A.-T</forename><surname>Kuo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE 39th International Conference on Data Engineering (ICDE)</title>
				<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="2023">2023. 2023</date>
			<biblScope unit="page" from="612" to="623" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Freeway travel time prediction using deep hybrid model-taking sun yat-sen freeway as an example</title>
		<author>
			<persName><forename type="first">P.-Y</forename><surname>Ting</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y.-L</forename><surname>Chiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-T</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sakai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename></persName>
		</author>
		<author>
			<persName><forename type="first">.-K</forename><surname>Jeng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-S</forename><surname>Hwu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Vehicular Technology</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="page" from="8257" to="8266" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Hierarchical nearest-neighbor gaussian process models for large geostatistical datasets</title>
		<author>
			<persName><forename type="first">A</forename><surname>Datta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Banerjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">O</forename><surname>Finley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Gelfand</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American Statistical Association</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<biblScope unit="page" from="800" to="812" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Spatiotemporal interpolation using gstat</title>
		<author>
			<persName><forename type="first">B</forename><surname>Gräler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Pebesma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">B</forename><surname>Heuvelink</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">R J</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page">204</biblScope>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Dynamic spatial-temporal graph convolutional neural networks for traffic forecasting</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Diao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>He</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI conference on artificial intelligence</title>
				<meeting>the AAAI conference on artificial intelligence</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="890" to="897" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A deep reinforcement learning system for the allocation of epidemic prevention materials based on ddpg</title>
		<author>
			<persName><forename type="first">K</forename><surname>Kitchat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-H</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-T</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sakai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Surasak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">242</biblScope>
			<biblScope unit="page">122763</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">A novel framework for spatio-temporal prediction of environmental data using deep learning</title>
		<author>
			<persName><forename type="first">F</forename><surname>Amato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guignard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Robert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kanevski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific reports</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">22243</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Smart deep learning based wind speed prediction model using wavelet packet decomposition, convolutional neural network and convolutional long short term memory network</title>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Mi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Energy Conversion and Management</title>
		<imprint>
			<biblScope unit="volume">166</biblScope>
			<biblScope unit="page" from="120" to="131" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Accurate medium-range global weather forecasting with 3d neural networks</title>
		<author>
			<persName><forename type="first">K</forename><surname>Bi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Tian</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nature</title>
		<imprint>
			<biblScope unit="page" from="1" to="6" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A deep learning multimodal method for precipitation estimation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Moraux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dewitte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Cornelis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Munteanu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page">3278</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Deep learning for precipitation nowcasting: A benchmark and a new model</title>
		<author>
			<persName><forename type="first">X</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lausen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D.-Y</forename><surname>Yeung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-K</forename><surname>Wong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-C</forename><surname>Woo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">N</forename><surname>Kipf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Welling</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1609.02907</idno>
		<title level="m">Semi-supervised classification with graph convolutional networks</title>
				<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Transgcn: Coupling transformation assumptions with graph convolutional networks for link prediction</title>
		<author>
			<persName><forename type="first">L</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Janowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 10th international conference on knowledge capture</title>
				<meeting>the 10th international conference on knowledge capture</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="131" to="138" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Bayesian joint estimation of multiple graphical models</title>
		<author>
			<persName><forename type="first">L</forename><surname>Gan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Narisetty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Liang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Large-scale learnable graph convolutional networks</title>
		<author>
			<persName><forename type="first">H</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ji</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 24th ACM SIGKDD international conference on knowledge discovery &amp; data mining</title>
				<meeting>the 24th ACM SIGKDD international conference on knowledge discovery &amp; data mining</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1416" to="1424" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Knowledge graph convolutional networks for recommender systems</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Guo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The world wide web conference</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3307" to="3313" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">A new model for learning in graph domains</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Monfardini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Scarselli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IEEE International Joint Conference on Neural Networks</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="729" to="734" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">The graph neural network model</title>
		<author>
			<persName><forename type="first">F</forename><surname>Scarselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">C</forename><surname>Tsoi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagenbuchner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Monfardini</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Trans. Neural Networks</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="61" to="80" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Tarlow</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brockschmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">S</forename><surname>Zemel</surname></persName>
		</author>
		<title level="m">Gated graph sequence neural networks</title>
				<imprint>
			<publisher>ICLR</publisher>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Graph convolutional neural networks for web-scale recommender systems</title>
		<author>
			<persName><forename type="first">R</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Eksombatchai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">L</forename><surname>Hamilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">SIGKDD, ACM</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="974" to="983" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Bruna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaremba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Szlam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<title level="m">Spectral networks and locally connected networks on graphs</title>
				<imprint>
			<publisher>ICLR</publisher>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Fastgcn: Fast learning with graph convolutional networks via importance sampling</title>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xiao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ICLR, OpenReview.net</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhu</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1709.04875</idno>
		<title level="m">Spatio-temporal graph convolutional networks: A deep learning framework for traffic forecasting</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Attention based spatial-temporal graph convolutional networks for traffic flow forecasting</title>
		<author>
			<persName><forename type="first">S</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI conference on artificial intelligence</title>
				<meeting>the AAAI conference on artificial intelligence</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="922" to="929" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Long short-term memory</title>
		<author>
			<persName><forename type="first">S</forename><surname>Hochreiter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural computation</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="1735" to="1780" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Predicting the price of bitcoin using machine learning</title>
		<author>
			<persName><forename type="first">S</forename><surname>Mcnally</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Roche</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Caton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2018 26th Euromicro International Conference on Parallel, Distributed and Network-based Processing (PDP), IEEE</title>
				<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="339" to="343" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Traffic accident hotspot prediction using temporal convolutional networks: A spatio-temporal approach</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">D</forename><surname>Yeddula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Hui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st ACM International Conference on Advances in Geographic Information Systems</title>
				<meeting>the 31st ACM International Conference on Advances in Geographic Information Systems</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Borovykh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bohte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">W</forename><surname>Oosterlee</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1703.04691</idno>
		<title level="m">Conditional time series forecasting with convolutional neural networks</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Adversarial sparse transformer for time series forecasting</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Ding</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="17105" to="17115" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Accurate multivariate stock movement prediction via data-axis transformer with multi-level contexts</title>
		<author>
			<persName><forename type="first">J</forename><surname>Yoo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Soun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>-C. Park</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Kang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining</title>
				<meeting>the 27th ACM SIGKDD Conference on Knowledge Discovery &amp; Data Mining</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2037" to="2045" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Rice yield prediction and model interpretation based on satellite and climatic indicators using a transformer method</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Hao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sun</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Remote Sensing</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page">5045</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">Selfattention convlstm for spatiotemporal prediction</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Cheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Yuan</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the AAAI conference on artificial intelligence</title>
				<meeting>the AAAI conference on artificial intelligence</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="11531" to="11538" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Traffic flow prediction via spatial temporal graph neural network</title>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the web conference 2020</title>
				<meeting>the web conference 2020</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1082" to="1092" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1706.03762</idno>
		<title level="m">Attention is all you need</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">A multimodal geo dataset for high-resolution precipitation forecasting</title>
		<author>
			<persName><forename type="first">C</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st ACM International Conference on Advances in Geographic Information Systems</title>
				<meeting>the 31st ACM International Conference on Advances in Geographic Information Systems</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1" to="4" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Modeling real estate dynamics using temporal encoding</title>
		<author>
			<persName><forename type="first">C</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W.-S</forename><surname>Ku</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th International Conference on Advances in Geographic Information Systems</title>
				<meeting>the 29th International Conference on Advances in Geographic Information Systems</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="516" to="525" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Random decision forests</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">K</forename><surname>Ho</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 3rd international conference on document analysis and recognition</title>
				<meeting>3rd international conference on document analysis and recognition</meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="1995">1995</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="278" to="282" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">A training algorithm for optimal margin classifiers</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">E</forename><surname>Boser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">M</forename><surname>Guyon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">N</forename><surname>Vapnik</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the fifth annual workshop on Computational learning theory</title>
				<meeting>the fifth annual workshop on Computational learning theory</meeting>
		<imprint>
			<date type="published" when="1992">1992</date>
			<biblScope unit="page" from="144" to="152" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Classification and regression trees</title>
		<author>
			<persName><forename type="first">W.-Y</forename><surname>Loh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Wiley interdisciplinary reviews: data mining and knowledge discovery</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="14" to="23" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Generalized linear models</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Nelder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">W</forename><surname>Wedderburn</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the Royal Statistical Society: Series A (General)</title>
		<imprint>
			<biblScope unit="volume">135</biblScope>
			<biblScope unit="page" from="370" to="384" />
			<date type="published" when="1972">1972</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<analytic>
		<title level="a" type="main">C1. 2 multilayer perceptrons</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Almeida</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Handbook of Neural Computation C</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
