=Paper=
{{Paper
|id=Vol-3248/paper22
|storemode=property
|title=Wi-Fi, CCTV and PDR Integrated Pedestrian Positioning System
|pdfUrl=https://ceur-ws.org/Vol-3248/paper22.pdf
|volume=Vol-3248
|authors=Max J. L. Lee,Meiling Su,Li-Ta Hsu
|dblpUrl=https://dblp.org/rec/conf/ipin/LeeSH22
}}
==Wi-Fi, CCTV and PDR Integrated Pedestrian Positioning System==
Wi-Fi, CCTV and PDR Integrated Pedestrian Positioning System Max J. L. Lee 1, Meiling Su 1 and Li-Ta Hsu 1 1 Department of Aeronautical and Aviation Engineering, The Hong Kong Polytechnic University, Hung Hom, Kowloon, Hong Kong, China Abstract Navigation inside a closed area with no GNSS-signal accessibility is a highly challenging task. In order to tackle this problem, the Wi-Fi beacon-based methods have grabbed the attention of many researchers. However, it suffers from lack of accuracy and robustness against environmental changes and dynamic object movements. With the advancement of inference speed in real-time deep learning applications, in this paper, a multi-modal end- to-end system for large-scale indoor positioning has been proposed, namely Wi-Fi, CCTV and PDR integrated pedestrian positioning system, which increases positioning accuracy by overcoming the difficulties of environment changes. Firstly, a user request positioning and sends the detect access points (APs) and associated signal strength to a server. A Wi- Fi one-shot model is used to estimate an initial position for the user, then the nearby CCTV cameras are activated to detect and position pedestrians. A CCTV and PDR integrated particle filter is used to determine a valid correspondence to the user, simultaneously collecting Wi-Fi fingerprints from the user to passively update the Wi-Fi one-shot support set in real-time. By implementing the proposed system, we could achieve a highly accurate integrated indoor positioning with precision level of 0.3m. Keywords 1 Indoor Positioning, Visual Positioning, Particle Filter, Wi-Fi Positioning 1. Introduction Indoor positioning systems have attracted a great interest from the researchers over the past decade. These systems can provide positioning, navigation, tracking services where global navigation satellite systems (GNSSs) could not reach [1]. To overcome the limitation of GNSS in indoor environments, various indoor positioning systems have been developed using Wi-Fi, Bluetooth (BLE), ultra-wideband (UWB), and radio-frequency identification (RFID). Typically, beacon nodes are a prerequisite to localize in the indoor environment [2]. Out of the above mentioned radio-frequency based systems, Wi- Fi is the most popular one due to its scalability [3]. The Wi-Fi localization is usually performed by four main techniques: triangulation, received signal strength (RSS), scene analysis (fingerprinting) and proximity based. However, the radio ranging measurements contain noise on the order of several meters [4]. This noise occurs because radio propagation tends to be highly non-uniform. Physical obstacles such as walls, furniture, etc. reflect and absorb radio waves [5]. As a result, the distance estimation is deteriorated by the diffractions and reflections. The fingerprinting method attempts to overcome this problem by making use of diffractions and reflections as additional features for positioning, however, it suffers when there are changes in the environment, as the features also change. Hence, an alternative positioning method is suggested in this paper to overcome the difficulties of environment changes. IPIN 2022 WiP Proceedings, September 5 - 7, 2022, Beijing, China EMAIL: maxjl.lee@connect.polyu.hk (M. J. L. Lee); meiling.su@connect.polyu.hk (M. Su); lt.hsu@polyu.edu.hk (L.T. Hsu) ORCID: 0000-0002-5524-6724 (M. J. L. Lee); 0000-0002-9095-5681 (M. Su); 0000-0002-0352-741X (L.T. Hsu) ©️ 2020 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings (CEUR-WS.org) Figure 1: Overview of The Proposed Wi-Fi and CCTV Integrated Pedestrian Positioning System. In the recent years, CCTV cameras have matured to a stage where it is playing an important role in monitoring and security [6]. These CCTV cameras can capture real-time information of the environment, which provides a reliable basis for setting up a digital twin (DT) because they can integrate information ranging from geometric changes in the building layout to the occupancy and use of rooms and spaces. DT for buildings can be seen as an extension to capture real-world data and feed it back into a 3D model, thus neatly closing the information loop [7-9]. In addition, CCTV cameras also capture the position of people and assets, hence, a position of the objects relative to the camera (and therefore the global position) can be obtained. If the correspondence to the user is found, it can directly transmit the person’s position to the person’s smartphone. This position can also be used to collect real- time data for Wi-Fi fingerprinting. Therefore, the main objective of the study presented in this paper is to develop a Wi-Fi, CCTV and PDR integrated pedestrian positioning system that enables navigation of the indoor environment using in-built Wi-Fi access points (APs) and CCTV cameras. The proposed Wi-Fi, CCTV and PDR integrated pedestrian positioning system attempts to make full use of the existing infrastructure of the indoor environment for positioning. The proposed method offers several major advantages over the existing Wi-Fi standalone positioning methods. • Firstly, the integration of CCTV and PDR can provide 0.3m positioning accuracy, enough for pedestrian positioning purposes. • Secondly, the positioning can be used to update the Wi-Fi one-shot support set, eliminating the need for continuous manual Wi-Fi fingerprinting collection. • Thirdly, as the CCTV is integrated for positioning, it is less susceptible to environment changes and dynamic object movements, overcoming the poor positioning due to the change in Wi-Fi features. 2. The Proposed Indoor Positioning System The flowchart of the proposed Wi-Fi and CCTV integrated pedestrian positioning system is shown in Fig. 1. The algorithm can be divided into the offline and online processes. The offline stage includes estimating the CCTV cameras’ intrinsic parameter by a checkboard, and extrinsic parameter using perspective and point from the known pose of visible landmarks in the indoor environment and the BIM (Building Information Modelling model) in Sec. 4. In the online stage, the smartphone requests a position by performing a Wi-Fi scan, then sending the detected APs and signal strength to the server. The server calculates an initial position based on a trained one-shot learning model (Sec. 3). The CCTV cameras near the initial position is activated to detect people (Sec. 5). The query pixel position of each detected person classified as “holding a phone” are corrected (Sec. 4), then their 3D positions are calculated based on the inverse perspective projection (Sec. 5). Assuming there are multiple detected persons (candidate positions), the problem is reformulated to estimate the correct correspondence from a list of candidates. We used a particle filter (PF) to estimate the correct correspondence based on the smartphone heading and velocity (Sec. 6). The PF positioning solution will then be used to estimate the state, and if given a small state covariance, the position will be used to update the Wi-Fi few-shot support set (Sec. 6). Several assumptions were made in the research: • Firstly, the Wi-Fi one-shot model was pre-trained in another environment as describe in [10]. The research only updates the support set. • Secondly, the research requires initial Wi-Fi fingerprints for the Wi-Fi on-shot model support set. However, it can eliminate the need for continuous Wi-Fi fingerprinting collection. • Thirdly, the world coordinate system was assumed to be the BIM 3D cartesian coordinate system for ease of calculation. The calculated cartesian positions can be converted to WGS84. • Fourthly, the BIM/3D model used to estimate the extrinsic parameter of the CCTV was assumed to be accurate and up to date. • Fifthly, the query pixel position correction assumes the query image shares the approximate same scene with the reference image at different viewpoints. • Sixthly, the pedestrian that request positioning are holding a smartphone for the deep learning model to classify pedestrians “holding a smartphone”. • Seventhly, the calculated position of pedestrian is assumed to be the bottom centre pixel of the detected bounding box. • Eighthly, the smartphone heading estimation is assumed to provide within ±90° uncertainty. The paper is organized as follow: Section 3 introduces the Wi-Fi one-shot positioning training and inference. Section 4 introduces the BIM and CCTV cameras’ intrinsic and extrinsic parameters estimation. Section 5 presents the person detection and position estimation. Section 6 presents the PF. Section 7 tests the proposed method with data obtained in a laboratory. Finally, conclusions and future perspectives are presented in Sections 8 and 9, respectively. 3. Wi-Fi One-shot Positioning Training and Inference A key challenge to machine learning-based fingerprinting approaches for wireless indoor localization is data labeling [10]. This includes the need to collect timely labeled data in one environment as the environment dynamics may change over time (e.g., for real-time localization), and the need to collect a new set of labeled data for localization in each new environment (e.g., for multi-environment localization). This time-sensitive and environment-dependent nature of wireless data incurs significant data collection and maintenance cost for supervised learning which requires a large amount of labeled data. The recently introduced few-shot transfer learning overcomes this problem by reformulating classification as a similarity [11]. More specifically, a graph neural network (GNN) model can be trained to measure the similarity of a query data to each of the 𝑛 classes in the support set. Inspired by the recent advance in few-shot learning, the proposed Wi-Fi, CCTV and PDR integrated pedestrian positioning system attempts to eliminate the need for data collection by passively collecting Wi-Fi fingerprints from users, while obtaining the approximate ground truth location estimated from the PF (detailed in Section 6). The estimated position and uncertainty using Wi-Fi signals are detailed in [10] and described as Eq. (1) in this paper. 𝑤 𝐱𝑤𝑖𝑓𝑖 , 𝜀𝑤𝑖𝑓𝑖 = 𝑂𝑛𝑒𝑆ℎ𝑜𝑡(𝐀, 𝐁) (1) 𝑤 Where 𝐱𝑤𝑖𝑓𝑖 is the estimated Wi-Fi 2D state (𝑥, 𝑦) in the BIM 3D cartesian coordinate system, and 𝜀𝑤𝑖𝑓𝑖 is the estimated uncertainty. 𝐀 refers to the query data, a list of detected access points and signal strength from an unknown location. 𝐁 refers to the support set, where each class corresponds to one known location, and each class has a list of detected access points and signal strength at that known location. 𝑂𝑛𝑒𝑆ℎ𝑜𝑡 is the function that accepts a query data and calculates the similarity to each class in the support set, the class with the highest similarity is 𝑤𝐱𝑤𝑖𝑓𝑖 . Once an accurate position is estimated from the PF (Sec. 6), the collected Wi-Fi fingerprints will be used to update the support set (𝐁) of the Wi-Fi one-shot model. 4. CCTV Intrinsic and Extrinsic Estimation To calculate the position of pedestrians based on the cameras, the intrinsic and extrinsic parameters of the cameras must be known. To estimate the intrinsic parameter (𝐊), a checkerboard can be used to calibrate each CCTV camera. To estimate the extrinsic parameter (𝐓), we made use of visible landmarks in the indoor environment to infer the pose of the camera. A BIM model stores information of objects in the indoor environment, including the objects’ class, shape, appearance, size and pose in the world coordinate frame [12]. Therefore, we can use perspective and point from the BIM to estimate the pose of the CCTV camera. Once this is complete, we can obtain an intrinsic and extrinsic associated reference image for each CCTV camera as described in Eq. 2. c𝑘 −1 𝐱 = ( w𝐓 ) ∙ w𝐱 (2) 𝑗 c𝑘 𝑗 i𝑘 𝐩𝑗 = c𝑘𝐊 ∙ c𝑘𝐱𝑗 Where w and c is the world and camera coordinate frame where each coordinate is expressed as a 3D cartesian vector c𝑘𝐱, w𝐱 ∈ (𝑥, 𝑦, 𝑧). i is the image coordinate frame, where each coordinate is expressed a pixel vector 𝐩 ∈ (𝑢, 𝑣). c𝑘𝐊 is the 3x3 intrinsic matrix of camera 𝑘. w𝐓c𝑘 is the 4x4 extrinsic matrix of camera 𝑘 in the world coordinate frame. We can therefore take a point 𝑗 in the world frame w𝐱𝑗 and express it as a pixel vector i𝑘𝐩𝑗 in the image frame (i𝑘 ) for camera 𝑘. The expanded equation is expressed in Eq. (3). c𝑘 −1 w (3) 𝑥𝑗 w 𝑟11 𝑟12 𝑟13 𝑡1 𝑥𝑗 c𝑘 𝑟 𝑟 𝑟 𝑡 w 𝑦𝑗 𝑦𝑗 = | 21 22 23 2 | c𝑘 𝑧𝑗 𝑟31 𝑟32 𝑟33 𝑡3 w 𝑧𝑗 [ 1 ] 0 0 0 1 c𝑘 [ 1 ] ( ) c𝑘 c𝑘 i𝑘 𝑢𝑗 𝑓𝑥 0 𝑢0 𝑥𝑗 c𝑘 c𝑘 𝑧𝑗 [ 𝑘𝑣𝑗 ] = [ 0 𝑓𝑦 𝑣0 ] [ 𝑗 ] i 𝑦 1 0 0 1 c𝑘𝑧𝑗 It is assumed that each camera’s intrinsic and extrinsic parameter may change overtime. The former can be caused by optical zoom, whereas the latter from camera tilt and panning. We denote c𝑘 as the reference camera, and the query (changed) camera as c𝑘 ′. The query camera’s intrinsic and extrinsic parameters are difficult to estimate simultaneously from the reference camera due to many unknowns, leading to local minima [13]. However, in this paper, the position of the pedestrian is assumed and calculated based on the 3D position of a pixel, as described in Sec. 5. Therefore, rather than estimating the query camera parameters to calculate the 3D position of a pixel, we estimate the pixel-to-pixel correspondence between the new and reference image. Since the 3D position is known for each pixel in the reference image, once a pixel is detected in the query image, it can correspond to a pixel in the reference image to estimate its 3D position. State-of-the-art dense matching was used to estimate the pixel-to-pixel correspondence as described in [14] and shown in Fig. 2. We assume the Percentage of Correct Key-points (PCK) is above 70% as evaluated in [14]. Figure 2: Example of Using a Dense Matching to Match Pixels of Query Image with Reference Image. The detailed equation can be found in [14] and described as Eq. (4) in this paper. i𝑘 𝐂i𝑘 ′ = 𝐷𝑒𝑛𝑠𝑒𝑀𝑎𝑡𝑐ℎ𝑖𝑛𝑔(i𝑘 , i′𝑘 ) (4) i𝑘 i𝑘 ′ 𝑢𝑗 𝑢𝑗 i𝑘 [ 𝑘𝑣𝑗 ] = 𝐂i𝑘 ′ ∙ [ 𝑘 𝑣𝑗 ] i i ′ 1 1 Where i𝑘𝐂i𝑘 ′ is the correspondence matrix to transform pixel coordinate in new image coordinate frame i𝑘 ′ to reference image coordinate frame i𝑘 . 5. Person Detection and Position Estimation In the online process, pedestrians are first detected using a primary PeopleNet network [15, 16]. To reduce the number of candidate positions (hypothesizes), we also assumed pedestrians that request positioning are holding a smartphone. Therefore, we employed a secondary classification model, that classifies a person either “holding a phone” or “not holding a phone” after being detected. Given a person is detected and is “holding a phone”, we calculate the person’s position using the center bottom pixel of the bounding box. It is assumed that the detection accuracy is 80% or above as stated in [15]. Figure 3: Classification and Pedestrian Detection. The 3D position of the pixel is the intersection between the ray emanating from the normalized pixel and the 3D ground plane. We define a plane in Eq. (5). 𝑎𝑥 + 𝑏𝑦 + 𝑐𝑧 + 𝑑 = 0 (5) Since the world coordinate ground plane is known from the BIM model, we can transform it into the camera coordinate frame in Eq. 6. c𝑘 w 𝑎𝑔 𝑎𝑔 (6) c𝑘 −1 w 𝑏𝑔 −1 T 𝑏𝑔 c𝑘 = ((( w𝐓c𝑘 ) ) ) ∙ w 𝑐𝑔 𝑐𝑔 c𝑘 w [ 𝑑𝑔 ] [ 𝑑𝑔 ] Where 𝑔 denotes the ground plane. The candidate pixel is normalized by multiplying the inverse intrinsic parameters. We can express the normalized pixel in the camera coordinate frame in Eq. (7). c𝑘 𝑥̂𝑗 i𝑘 𝑢𝑗 (7) c𝑘 c𝑘 )−1 𝑦̂ [ 𝑗] = ( 𝐊 ∙ [ 𝑘𝑣𝑗 ] i c𝑘 𝑧̂𝑗 1 c The camera coordinate of the normalized pixel is 𝑘𝐱̂𝑗 ∈ ( c𝑘𝑥̂𝑗 , c𝑘𝑦̂𝑗 , c𝑘𝑧̂𝑗 ). The camera coordinate of the intersection between the normalized pixel and ground plane is calculated in Eq. (8). c𝑘 𝑐𝑘 − c𝑘 𝑑𝑔 (8) 𝐱𝑗 = 𝑡 ∙ 𝐱̂𝑗 ; 𝑡 = c c c c c 𝑘𝑎 𝑔 ̂𝑗 + 𝑏𝑔 𝑦 𝑘𝑥 𝑘 𝑘 ̂𝑗 + 𝑐𝑔 𝑘 Then the camera coordinate of the intersection can be expressed in the world coordinate as Eq. (9). c w 𝐱𝑘,𝑗 = w𝐓c𝑘 ∙ 𝑘𝐱𝑗 (9) 𝑤 The calculated positions are then filtered based on the Wi-Fi initial guess search area 𝐱𝑤𝑖𝑓𝑖 , 𝜀𝑤𝑖𝑓𝑖 . In the remainder, as the positions are located on the ground plane, they will be expressed as a sequence of measurements in 2D Cartesian world coordinate frame as 𝐳𝑘,𝑛 for ease of notation in Eq. (10). 𝐳𝑘,𝑛 ∈ (𝑥𝑘,𝑛 , 𝑦𝑘,𝑛 ) (10) 𝐙𝑘 ∈ (𝐳𝑘,1 , 𝐳𝑘,2 , … 𝐳𝑛 ) Where 𝑘 is the index of the camera. 𝑛 is the index of the filtered positions from 𝑗 that are within the Wi-Fi search area, and 𝐙𝑘 are the measurement inputs of camera 𝑘. 𝐙1…𝑘 is the measurement inputs of all cameras. 6. Particle Filter We formulated a recursive state estimation problem by using a PF. The goal is to estimate a 2D state vector 𝐱. More specifically, the goal is to track the hidden state sequence {𝐱 𝑡 } of a dynamical system, where 𝑡 is a discrete time step. The process model that encodes prior knowledge on how the state 𝐱 is expected to evolve over time can be written as Eq. (11). 𝑥𝑡 = 𝑥𝑡−1 + 𝑣𝑡−1 ∙ cos(𝜃𝑡−1 ) ∙ 𝑑𝑡 + 𝜖𝑥 (11) 𝑦𝑡 = 𝑦𝑡−1 + 𝑣𝑡−1 ∙ sin(𝜃𝑡−1 ) ∙ 𝑑𝑡 + 𝜖𝑦 𝐮 ∈ (𝑣, 𝜃) Where 𝑣 is the pedometer velocity and 𝜃 is the magnetometer heading from the client smartphone [17]. The cumulative uncertainty is assumed to be within ±0.2m every second. The measurements 𝐙1...𝑘,𝑡 is defined in Sec. 5. The particle filter is described in Alg. 1. Algorithm 1 Particle Filter Input: Prior particles set {𝑆𝑡−1 =< 𝐱𝑖,𝑡−1 , 𝓌𝑖,𝑡−1 >, 𝐮𝑡 , 𝐙1...𝑘,𝑡 } Output: 𝑆𝑡 1. 𝑆𝑡 = ∅, 𝜂 = 0 2. For 𝑗 = 1 … 𝑛 3. Sample particles 𝑖(𝑗) from the discrete distribution given by 𝓌𝑡−1 4. Sample 𝐱𝑗,𝑡 from 𝑝(𝐱𝑡 |𝐱𝑖(𝑗),𝑡−1 , 𝐮𝑡 ) 5. 𝓌𝑗,𝑡 = 𝑝(𝐙1...𝑘,𝑡 |𝐱𝑗,𝑡 ) 6. 𝜂 = 𝜂 + 𝓌𝑗,𝑡 7. 𝑆𝑡 = 𝑆𝑡 ∪ {< 𝐱𝑗,𝑡 , 𝓌𝑗,𝑡 >} 8. For 𝑗 = 1 … 𝑛 𝓌 9. 𝓌𝑗,𝑡 = 𝑛𝑗,𝑡 Where 𝑆 is a set of prior particles (𝐱𝑖,𝑡−1 ) with a corresponding weight (𝓌𝑖,𝑡−1 ), input (𝐮𝑡 ) and measurements (𝐙1...𝑘,𝑡 ). 𝜂 is the normalization constant. Then, hidden state 𝐱𝑡 is calculated based on the particle with the highest weight (𝓌𝑗,𝑡 ). To calculate the weight (𝓌𝑗,𝑡 ) of each particle, we need to understand the observation uncertainty (𝜎𝑘,𝑛,𝑡 ) using the known height of the camera ( c𝑘 ′𝑧𝑡 ) and measured 3D distance from the camera to the detected person 𝑑( c𝑘′𝐱𝑡 , 𝐳𝑘,𝑛,𝑡 ) . A Monte-Carlo simulation was used to sample the correlation. Figure 4: Observation Uncertainty of CCTV Positioning System and Fit for Eq. 12. A lowess (linear) smoothing regression was used to fit the data samples. It can be seen that when the height of camera is low and measured distance increases, the positioning error also increases. This is likely due to the pixels representing more distance when the height of camera is low. The fit has the following properties: Table 1 Lowess Fit Properties of Observation Uncertainty of CCTV Positioning System Fit Type SSE (m2 ) R-square DFE Adj R-sq RMSE Coeff (m) Lowess 78.77 0.3481 694 0.3397 0.3369 10 The fit was then used to estimate the positioning error in real-time for each detected person in Eq. 12. 𝜎𝑘,𝑛,𝑡 = 𝑓( c𝑘𝑧𝑡 , 𝑑( ck𝐱𝑡 , 𝐳𝑘,𝑛,𝑡 )) (12) We assumed a multivariable gaussian distribution of the positioning errors, where the covariance matrix variance is equal to 𝜎𝑘,𝑛,𝑡 to generate a likelihood for a given position in Eq. 13. 𝜎𝑘,𝑛,𝑡 0 (13) 𝜇𝑘,𝑛,𝑡 = [𝑥𝑘,𝑛,𝑡 , 𝑦𝑘,𝑛,𝑡 ]; 𝚺𝑘,𝑛,𝑡 = [ ] 0 𝜎𝑘,𝑛,𝑡 1 1 ℓ𝐱,𝑘,𝑛,𝑡 = 𝑒𝑥𝑝 (− (𝐱 − 𝜇𝑘,𝑛,𝑡 )𝚺𝑘,𝑛,𝑡 −1 (𝐱 − 𝜇𝑘,𝑛,𝑡 )T ) 2 √|𝚺𝑘,𝑛,𝑡 |(2𝜋)2 Where 𝜇𝑘,𝑛,𝑡 is the mean and 𝚺𝑘,𝑛,𝑡 is the covariance matrix of the multivariable gaussian distribution. ℓ𝐱,𝑘,𝑛,𝑡 is the calculated likelihood of a given position 𝐱 from camera (𝑘) and person (𝑛). All distribution were compared such that the highest likelihood at each position was used to combine into a global distribution, then normalized such that the total area under the distribution is equal to 1. It is important to note that there is no minimum distance required between two pedestrians as each pedestrian is represented as a distribution. The final weighting (𝓌𝑗,𝑡 ) in Alg. 1 of a given position can be calculated based on Alg. 2. Algorithm 2 Calculation of 𝓌𝑗,𝑡 Input: 𝐙1...𝑘,𝑡 , 𝐱𝑗,𝑡 Output: 𝓌𝑗,𝑡 1. Calculate ℓ𝐱,1…𝑘,1…𝑛,𝑡 for every position 2. Find the maximum ℓ𝐱,1…𝑘,1…𝑛,𝑡 for every position and create new distribution such that ℓ𝑚𝑎𝑥 𝐱,𝑡 for every position 3. Normalize distribution such that ℓ̂𝑚𝑎𝑥 𝐱,𝑡 for every position 4. ̂ 𝑚𝑎𝑥 𝓌𝑗,𝑡 = ℓ𝐱𝑗,𝑡 for specific position index (𝑗) The weighting calculation is visualized in Fig. 5. Where the likelihood distribution of each person is combined into a normalized global distribution. Figure 5: Example of A Likelihood Function Heatmap of 3 Pedestrians Captured by Two CCTV Cameras. The particle with the highest weighting as described in algorithm 1 will be the PF estimated state 𝒙𝑡 . Given that the particle estimated state covariance is less than 1m, it will be used to update the support set described in Sec. 3. 7. Experimental Results In this study, the experimental trajectory was conducted in a laboratory as shown in Fig. 8. The laboratory is a small indoor environment which contains numerous dynamic objects. Three commercial CCTV cameras are placed across the indoor environment to test the feasibility of the proposed positioning system. The ground truth positions of the trajectory were recorded, and the positioning quality of the proposed method was analyzed with other positioning methods, including: 1. Ground Truth, provided by markers labelled on the floor and timestamp from the CCTV camera. 2. Pedestrian Dead Reckoning (PDR), provided by Samsung Galaxy Note 20 smartphone and PDR app in [17]. 3. CCTV, provided by commercial Tapo smart cameras [18]. 4. Proposed Particle Filter (CCTV & PDR). 5. Wi-Fi Extended Naïve Bayes Positioning (ML) [19]. 6. Wi-Fi One-Shot Positioning (One-Shot) [10]. 7. Particle Filter (Wi-Fi ML & PDR). 8. Particle Filter (Wi-Fi One-Shot & PDR). Methods 5, 6 are obtained through a succession of fixed points, whereas the other methods integrated with PDR are obtained during a movement. To compare with traditional Wi-Fi ML positioning, Wi-Fi data were collected one month in advance at each location with 2-meter separation as shown in Fig. 6. 200 data were collected at each location to train an Extended Naive Bayes positioning model to classify new Wi-Fi fingerprints to a location. To compare with the Wi-Fi One-Shot positioning, two Samsung Galaxy smartphones were used to collect Wi-Fi data simultaneously during the experiment, where one smartphone collects data for the support set input, and the other phone will act as query input for the one-shot model. Figure 6: Positions That Collected Wi-Fi Signals to Train A Wi-Fi Extended Naïve Bayes Model. In addition, to mimic a real indoor environment with multiple pedestrians, in this study we have a pedestrian walk around the client to create a false-positive dynamic likelihood to the CCTV measurements as shown in Fig. 7. A false-positive static likelihood at a single location was also added post-experiment to the CCTV measurement. The former represents a pedestrian walking, and the latter represents a pedestrian standing, commonly seen in all indoor environments. This was performed to test whether the correct correspondence to the client can be found using the proposed method. Figure 7: Measurement Likelihood with Dynamic and A Mimic Static Interference. The positioning results are plotted onto a bird’s-eye view of the laboratory in Fig. 8. There are two performance metrics used: mean and standard deviation (SD) of the 2D positioning error. Figure 8: Positioning Results of PDR and Wi-Fi Based Positioning Methods. The PDR requires an initial position and was set to the ground truth initial position. The results show that the PDR (yellow marker) begins with great positioning accuracy but starts to accumulate drift error as it progresses as shown after time epoch 20. This has led to a significant cumulative positioning mean error of 2.4m. Therefore, it needs to be integrated with an absolute positioning measurement to correct its cumulative errors. The Wi-Fi Extended Naïve Bayes point positioning (cyan marker) is on average 3.81m to the ground truth (black marker) as analyzed in Figs. 8, 9 and Table II. This is mainly due to the change in the physical environment from dynamic objects, which are common in many indoor environments, thus creating NLOS and multipath signals different from the signals captured previously at the location. The integration of the Wi-Fi Extended Naïve Bayes and PDR (blue marker) is relatively accurate from time 0 to 14 seconds was perhaps due to the more unique Wi-Fi signatures then compared to the open space counterpart at time 20 seconds and after, leading to 3.64m positioning inaccuracy. The collected Wi-Fi signals from the CCTV & PDR PF, were then used as the support set of the Wi-Fi One-Shot positioning, improving the accuracy to 1.89m mean error and 1.56 standard deviation. It was then integrated with PDR to provide 1.98m positioning accuracy with a smaller 1.11m standard deviation. Figure 9: Positioning Error of The Proposed Positioning System and Other Positioning Systems. Figure 10: Positioning Results of The Proposed Positioning System and Other Positioning Systems. The CCTV positioning solution is shown in the green marker, where a correct correspondence is required for positioning. The proposed method (red marker) integrates the PDR with CCTV positioning solution via PF. Consequently, the PDR can estimate the correspondence for the CCTV, whereas the CCTV can correct the positioning error of the PDR. The results of the proposed method provide an accurate and continuous positioning of 0.29m mean error. The collected Wi-Fi fingerprints were then used as the support set of the Wi-Fi One-Shot positioning, improving the accuracy to 1.89m. Table 2 Accuracy of The Compared Positioning Systems Method 2D position error Mean (m) SD (m) PDR 2.40 2.50 PF (CCTV & PDR) 0.29 0.27 PF (Wi-Fi ML & PDR) 3.64 1.62 PF (Wi-Fi One-Shot & PDR) 1.98 1.11 Wi-Fi ML 3.81 2.01 Wi-Fi One-Shot 1.89 1.56 8. Conclusions This paper proposes a Wi-Fi and CCTV integrated indoor positioning solution for a 2D position estimation. In short, pedestrians were detected, and a particle filter was used to estimate the correct correspondence to the user. The estimated state can then be used to update a Wi-Fi one-shot model. The potential advantages of the proposed method are: • The integration of Wi-Fi, CCTV and PDR improves positioning accuracy compared to Wi-Fi standalone positioning. • The Wi-Fi one-shot positioning can improve upon the traditional Wi-Fi positioning due to the use of real-time Wi-Fi data. In this paper, a laboratory indoor environment was used to test the proposed method, however, in the real world there would be additional challenges that needs to be addressed including: • Up-to-date BIM model that reflects the building to calculate the accurate extrinsic parameters of the CCTV cameras. • Areas with no Wi-Fi or CCTV coverage. • True negatives detection of pedestrians, in other words failure to detect pedestrians. Considering the preliminary results presented in this paper, we believe the proposed method can provide accurate positioning and to support various indoor applications, that can be extended to digital twin. 9. Future Works We will conduct a real-time experiment in a large-scale mall in Hong Kong to validate the proposed method. Several potential future developments on the proposed method are suggested. • To make use of state-of-the-art visual trackers that can provide higher weighting to correspondences. • To make use of 3D object detection algorithms that are more robust to positioning estimation compared to 2D object detection [20]. References [1] A. Hameed and H. A. Ahmed, "Survey on indoor positioning applications based on different technologies," in 2018 12th International Conference on Mathematics, Actuarial Science, Computer Science and Statistics (MACS), 24-25 Nov. 2018 2018, pp. 1-5, doi: 10.1109/MACS.2018.8628462. [2] V. Renaudin et al., "Evaluating Indoor Positioning Systems in a Shopping Mall: The Lessons Learned From the IPIN 2018 Competition," IEEE Access, vol. 7, pp. 148594-148628, 2019, doi: 10.1109/ACCESS.2019.2944389. [3] M. Xue et al., "Locate the Mobile Device by Enhancing the WiFi-Based Indoor Localization Model," IEEE Internet of Things Journal, vol. 6, no. 5, pp. 8792-8803, 2019, doi: 10.1109/JIOT.2019.2923433. [4] Q. Yang, S. Zheng, M. Liu, and Y. Zhang, "Research on Wi-Fi indoor positioning in a smart exhibition hall based on received signal strength indication," EURASIP Journal on Wireless Communications and Networking, vol. 2019, no. 1, p. 275, 2019/12/17 2019, doi: 10.1186/s13638-019-1601-3. [5] E. R. Magsino, I. W. H. Ho, and Z. Situ, "The effects of dynamic environment on channel frequency response-based indoor positioning," in 2017 IEEE 28th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC), 8-13 Oct. 2017 2017, pp. 1-6, doi: 10.1109/PIMRC.2017.8292442. [6] D. Acharya, K. Khoshelham, and S. Winter, Real-time detection and tracking of pedestrians in CCTV images using a deep convolutional neural network. 2017. [7] D. Jones, C. Snider, A. Nassehi, J. Yon, and B. Hicks, "Characterising the Digital Twin: A systematic literature review," CIRP Journal of Manufacturing Science and Technology, vol. 29, pp. 36-52, 2020/05/01/ 2020, doi: https://doi.org/10.1016/j.cirpj.2020.02.002. [8] M. J. L. H. H. Y. H. L.-T. A. S. L. M. Lee, "BIPS: Building Information Positioning System," presented at the In Proceedings of the 2021 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Online, 2021. [9] M. J. L. Lee, L.-T. Hsu, H.-F. Ng, and S. Lee, "Semantic-Based VPS for Smartphone Localization in Challenging Urban Environments," 2020. [Online]. Available: http://arXiv.org/abs/ https://arxiv.org/ct?url=https%3A%2F%2Fdx.doi.org%2F10.1109%2FTRO.2021.3075644&v=a8cc9 408. [10] B.-J. Chen and R. Y. Chang, "Few-Shot Transfer Learning for Device-Free Fingerprinting Indoor Localization," 2022. [Online]. Available: http://arXiv.org/abs/. [11] O. Vinyals, C. Blundell, T. Lillicrap, K. Kavukcuoglu, and D. Wierstra, "Matching Networks for One Shot Learning," 2016. [Online]. Available: http://arXiv.org/abs/. [12] S. Alizadehsalehi, A. Hadavi, and J. C. Huang, "From BIM to extended reality in AEC industry," Automation in Construction, vol. 116, p. 103254, 2020/08/01/ 2020, doi: https://doi.org/10.1016/j.autcon.2020.103254. [13] S. Tang, C. Tang, R. Huang, S. Zhu, and P. Tan, "Learning Camera Localization via Dense Scene Matching," 2021. [Online]. Available: http://arXiv.org/abs/. [14] P. Truong, M. Danelljan, L. V. Gool, and R. Timofte, "Learning Accurate Dense Correspondences and When to Trust Them," 2021. [Online]. Available: http://arXiv.org/abs/. [15] NVIDIA. "PeopleNet." https://catalog.ngc.nvidia.com/orgs/nvidia/models/tlt_peoplenet (accessed 2022). [16] A. Bochkovskiy, C.-Y. Wang, and H.-Y. M. Liao, "YOLOv4: Optimal Speed and Accuracy of Object Detection," 2020. [Online]. Available: http://arXiv.org/abs/. [17] N. Patel. "Dead Reckoning, a location tracking app for Android smartphones." GitHub. https://github.com/nisargnp/DeadReckoning (accessed 2022). [18] Tapo, "Pan/Tilt Home Security Wi-Fi Camera," 2022. [Online]. Available: https://www.tapo.com/en/product/smart-camera/tapo-c210/. [19] Find3. "Framework for Internal Navigation and Discovery." https://github.com/schollz/find3 (accessed 2022). [20] J. Ku, A. Pon, and S. Waslander, Monocular 3D Object Detection Leveraging Accurate Proposals and Shape Reconstruction. 2019.