<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Identification of NLOS Acoustic Signal Using CNN and Bi-LSTM</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Hucheng Wang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Suo Qiu</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Haoran Shu</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lisa ( Jingjing) Wang</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xiaonan Luo</string-name>
          <email>luoxn@guet.edu.cn</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhi Wang</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lei Zhang</string-name>
          <email>zhlei0202@163.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Chang'an University</institution>
          ,
          <addr-line>Xi'an, 710061</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Guilin University of Electronic Technology</institution>
          ,
          <addr-line>Guilin, 541004</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Research China</institution>
          ,
          <addr-line>Signify holding, Shanghai, 200233</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Zhejiang University</institution>
          ,
          <addr-line>Hangzhou, 310027</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Compared with other indoor positioning techniques, acoustic signal is an ideal medium for indoor position systems due to its high compatibility and low deployment cost. The most vital reason for the degradation of performance in the process of acoustic signal propagation is non-line-of-sight (NLOS). The traditional signal filtering process is tedious and time-consuming. At the same time, deep learning has shown excellent performance in acoustic signal processing and classification tasks. In this letter, an acoustic signal line-of-sight (LOS)/NLOS identification method based on a convolutional neural network (CNN) and bi-directional long short-term memory (Bi-LSTM) models is proposed. Instead of the spectrogram, the acoustic signal spectrum matrix was fed into the network. The CNN was employed to extract the features from the two-dimensional image-like spectrum matrix automatically, and Bi-LSTM was utilized for classification. We evaluated the classification accuracy of the CNN and Bi-LSTM with diferent architectures, and found that the best one achieved 97.34% in classification performance.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Acoustic</kwd>
        <kwd>NLOS</kwd>
        <kwd>CNN</kwd>
        <kwd>Bi-LSTM</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        In indoor Internet of Things (IoT) technology, the location of the user is crucial privacy data for
humanized services [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Owing to the inability of GNSS signals to penetrate walls and urban
shielding efects, indoor positioning often requires additional signal media. An acoustic signal
has a natural low synchronization cost compared with other indoor positioning techniques. Part
of the electromagnetic positioning methods that are queried through the fingerprint database has
poor accuracy based on the size of the grid. In contrast, Time of Arrival (ToA)/Time Diferential
of Arrival (TDoA) -based acoustic positioning usually only has a positioning accuracy of
decimeters to centimeters [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. More significantly, the acoustic signal is fully compatible with
smart terminals, such as smartphones currently on the market. Users do not need to install
other hardware devices, which is more conducive to promotion and dissemination.
      </p>
      <p>Acoustic signals usually have a frequency of approximately 20 Hz–20 kHz and a wavelength
of approximately 17 mm–17 m. The attenuation is evident after being blocked by obstacles.
The accuracy of raw acoustic positioning is usually not ideal in the absence of any correction
treatment. When the transmitter source, like a speaker, and the receiving source, like a
smartphone microphone, are not directly reachable, the acoustics may be reflected on the surrounding
walls multiple times to reach it, which causes signal delay, and may also exhibit signal loss and
degeneration.</p>
      <p>In the acoustics of indoor positioning systems (IPSs), the NLOS is one of the most formidable
factors that decrease positioning accuracy. NLOS communication is a situation where wireless
signals cannot reach the receiver directly due to the presence of obstacles. Detecting, filtering,
or correcting NLOS signals has become a crucial part of IPSs. The accuracy of the detection
algorithm will directly afect every link. In academia, the following judgment schemes often
exist.</p>
      <p>
        ∙ Most methods will record the ranging values of the previous data and compare them
with the value of the current data. The record passes through the moving or static state
of the measurement target, and refers to the speed or step length of users, the moving
direction, and so on. The data can be obtained themselves [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], or through other external
sensors, such as installing an inertial measurement unit to obtain inertial data [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ], and
correcting NLOS or missing data according to the coarse-grained coupling algorithm [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>
        This approach is not suitable for flexible maneuvering targets.
∙ Another research hotspot is signal features extraction, such as Channel State Information
(CSI) [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], propagation delay [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], channel quality [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], energy intensity [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], and
statistical data, such as machine-learning training samples [
        <xref ref-type="bibr" rid="ref12">12, 13</xref>
        ]. The signal propagation
distance will be estimated through huge data analysis and calculation of generalized
crosscorrelation (GCC), finding the correlation peak, and calculating the time-delay interval of
the direct signal from the messy, raw signal. Support Vector Machines (SVMs), Variational
Autoencoders (VAEs), decision trees are often employed. [14] collected the time-delay
characteristics, waveform characteristics, Rician K-factors, and frequency characteristics
of relative channel gain and summarized them into the Radial-based Function (RBF) core.
[15] proposed a structured Bi-LSTM to train a three-dimensional (3D) terahertz signal.
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] improved [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] and obtained better results after denoising.
∙ The building structure of the room and the indoor map are also ways to distinguish NLOS
information [16].
      </p>
      <p>An acoustic signal has obvious time relevance, and the spectrum of acoustics has solid
characteristic information, which reminds us to use deep learning to distinguish NLOS data.
In this letter, we propose a novel sound NLOS signal recognition method that combines a
convolutional neural network (CNN) and Bi-LSTM. The CNN extracts the spectral features of
the acoustics, and Bi-LSTM classifies the NLOS recognition with strong time relevance. The
Figure 1 shows the main structure.</p>
      <p>The remaining parts of this letter are organized as follows: Section II describes the acoustic
signal and its spectral characteristics; Section III shows how to choose and use the CNN and</p>
      <p>Bi-LSTM classification; the training results, data analysis, and planned future work are presented
in Section IV; and Section V concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Characterization of acoustic signal</title>
      <p>The acoustics in this work is an autonomously modulated chirp signal from [17]. A single chirp
signal can be modulated as
1
() = (2 (0 + 2 02)),
where 0 and 0 are the initial frequency and modulation rate, respectively.</p>
      <p>To facilitate the analysis of the waveform, we inserted a silent interval instead of Frequency
Modulated Continuous Wave (FMCW) to form a complete period  . Then, the transmitted
signal can be described as</p>
      <p>∞
( ) = ∑︁ ( −  +  )( −  ),
=0
(1)
(2)
where (· ) is a step function and  denotes th chirps.</p>
      <p>Considering a complex indoor environment, there are  reflected rays and  difracted
rays received by a microphone, and then the channel response of a microphone can be described
where the subscripts , , and  denote the parameters related to LOS ray, reflection rays, and
difusion rays, respectively, as shown iLnOSFigurNeLO3S, and  denotes the attenuation of diferent
rays. The black-man window (· ) is employed to erase the slight multi-way fluctuation. The
residual noise, such as electromagnetic vibration noise, is represented by (· ).  ,  , and  
others (· )
refer to the propagation delays from diferent paths, and are calculated by  ((· )) = (·) , where
the Bsiu-LpSeTrMscript is the FtChlaoyrerth paCtlhas,sainficdattihoen rseusbusltcript is the way of the arri·val path, and  is
the velocity of sound.
(3)
Microphone</p>
      <p>LOS
NLOS</p>
      <p>Shelter or Screen</p>
      <p>Wall
Speaker</p>
      <p>Microphone
0.0
0.2
0.8</p>
      <p>1.0
0.4 0.6
time (seconds)
(a) LOS
0.4 0.6
time (seconds)</p>
      <p>The separation of coupled signals has always been a complicated issue. Typical filters cannot
separate the aliased signal to an accuracy of more than 90%, while neural networks give us new
ideas.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed neural network method</title>
      <p>(a) LOS
(b) NLOS</p>
      <p>It can be seen that the 2D image-like spectrograms in Figure 5 have prominent block area
characteristics. However, not every pixel and every color in the spectrogram has significance.
Training the spectrogram images will produce many redundant and useless features and decrease
the training accuracy. We further purified the spectrum information to obtain the spectrum
matrix with STFT in Figure 6. This refines the training data and incidentally filters out part of
the background noise.</p>
      <sec id="sec-3-1">
        <title>LSTM unit L(i-1)</title>
      </sec>
      <sec id="sec-3-2">
        <title>LSTM unit L(i) LSTM unit L(i+1)</title>
      </sec>
      <sec id="sec-3-3">
        <title>CNN block(i-1)</title>
        <p>512 channels</p>
        <sec id="sec-3-3-1">
          <title>3.1. Feature extraction of spectrum matrix based on CNN</title>
          <p>A CNN is a multi-layer deep-learning neural network, which implies multiple convolutional
layers and multiple pooling layers. A CNN uses gradient descent to minimize the layer-by-layer
reverse adjustment of the weight parameters in the network by the loss function and improves
Authdieonceatwptourrkedaccuracy throughInfpruetquent iterative trCaiNnNing. Feature map
from Tmhicerocopnhovnoelutional lasypeercitsrudmesmiganterdixto exwtritahcotufteaFtCurlaesyeorf the sp5e1c2truchmanmnaetlrsix sequenceB.i-LSTM
Multiple kernels are employed in the convolutional layers. The Rectified Linear Units (ReLU)
activation function converts the spectrum matrix into characteristics. The MaxPooling layers
down-sample the size of the characteristics.</p>
          <p>The fully connected (FC) layer is usually the last layer of the CNN with the SoftMax function.
However, the traditional FC layer in a CNN ignores contextual relevance. NLOS rarely appears
by itself, but does so continuously. Based on the above characteristics, in the present work we
canceled the FC layer of the CNN and replaced it with Bi-LSTM for the NLOS classification task.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.2. NLOS classification based on Bi-LSTM</title>
          <p>In the virtual environment, occlusion is not independent. For a long-term sound wave sequence,
LSTM is one of the best solutions. In fact, we found that both forward and backward information
can help determine occlusion, so Bi-LSTM is adopted herein.</p>
          <p>
            A com
s(t)
where   ∈ Rℎ is the output vector in the forget gate,  ∈ Rℎ× ℎ and  ∈ Rℎ×  are the
updating weight vectors, and  ∈ Rℎ denotes the bias vector. The activation function   is
selected as the Sigmoid function. The forget gate contains the values of   ⊂ [
            <xref ref-type="bibr" rid="ref1">0, 1</xref>
            ], which
decides the keeping degree of the memory cell through the next operation.
          </p>
          <p>The input gate regulates the input data  and the processed state vector   from the forget
gate, which is described as
 =  ( + ℎ− 1 + ),
˜ =  ℎ(  +  ℎ− 1 +  ),
 =  − 1 + ˜,
where ,  ∈ Rℎ× ℎ, ,  ∈ Rℎ×  denote the updating weight vector that iterates through
training in the input gate, and  ,  ∈ Rℎ denotes the bias. The activation function  ℎ is
selected as the tanh(· ) function. When the candidate cell state vector ˜ is computed, the real
cell state vector  can be updated with last cell state − 1.</p>
          <p>The third gate in a single LSTM unit is the output gate. The output  will be based on the
above cell state, but it is also a filtered version. The output gates are written as</p>
          <p>The bi-directional RNN can use the information before and after the current moment, which
further promotes the accuracy of information judgment. The traditional LSTM will pass ( − 1)
state information before the th moment. If the current ( − 1) states are all LOS, and the user
is switched from LOS to NLOS, the previous information is not suficient to obtain a correct
judgment. Introducing the backward LSTM layer, ( + 1) and the future state can amend the
current judgment. The output of Bi-LSTM can be described as
 =  ( + ℎ− 1 + ),
ℎ =  ·
  ℎ().
(7)
The  ∈ Rℎ× ℎ,  ∈ Rℎ× , and  ∈ Rℎ also represent the weight and bias, correspondingly.
The hidden state ℎ will be updated in this gate from the output  and the new cell state 
with the activation function  ℎ. Then, we insert  into Eq. 4 and obtain the total output O in
Bi-LSTM.
(4)
(5)
(6)</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Dataset and experimental results</title>
      <p>To assess the proposed method, we designed an experiment and collected audio data. The
LOS and NLOS data were collected from four microphones in diferent locations and four
diferent indoor rooms: Laboratory 1, Laboratory 2, Ofice 1, and Ofice 2. For each microphone,
more than 400 pieces of LOS and 400 pieces of NLOS data were collected. Diferent rooms
were selected to extend room impulse response (RIR) information and prevent over-fitting.
Microphones in diferent locations mean that the training is generalized to every part of the
entire room. Each piece of data was washed and sliced to a length of 1s and labeled. A total of
12,800 raw audio samples were composed. We shufled all the data for each epoch and selected
8,960 samples (70% of 12,800) as a training set and the rest samples comprised a testing set. All
the sampling processes were entirely random to prevent over-fitting the model. Based on the
above data, the model training takes 1 hour and 22 minutes. After putting a single data in the
test dataset into the model, it took 0.98 seconds to get the classification result. The dataset is
available on IEEE Dataport; more detailed descriptions of the experiments and collections can
be found by contacting the authors.</p>
      <sec id="sec-4-1">
        <title>4.1. From raw audio wave to spectrum matrix</title>
        <p>Generally, audio data will be analyzed with the help of spectrograms of PSD. We trained the
data based on the spectrogram and obtained the result shown at the top of Figure 8. We assumed
that this occurred for two reasons. First, most of the data captured in the spectrogram are
useless information, and background noise accounts for the majority of it, such as human voices,
mechanical vibrations. Second, the acoustic positioning system is arranged to operate as far as
several dozens of meters usually. The target signal captured by the microphone may be weaker
than other background noises, leading to the testing loss decreasing to zero. Changing the input
data from the signal processing level can significantly improve the classification efect. The
bottom panel of Figure 8 shows that the spectrum matrix results in an apparent gain in the
classification training result.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Network design and configuration</title>
        <p>For the CNN segment, which was sequential at first, we set the size of the kernel to 7 × 7 in the
convolution layer to quickly reduce the input dimension. We then set the batch normalization
and ReLU layers to accelerate training and prevent the internal covariate shift phenomenon.
The only maxpooling layer was used for down-sampling and highlighting essential information.
Next, we re-used the convolution, batch normalization, and ReLU layers to enhance the feature
matrix from 64, 128, 256, and 512. Finally, in the CNN, adaptive average pooling was configured
to extract the feature block of 512 × 1 × 1 and put it into the Bi-LSTM network.</p>
        <p>The design of Bi-LSTM is detailed in Section III.B. Bi-LSTM compressed the CNN block into
32 states and classified it by FC layer. Interestingly, we found that double Bi-LSTM gives better
training results than single or double LSTM. Therefore, finally, we adopted the proposed model
as double Bi-LSTM underlying the CNN.</p>
        <p>Regarding the hyper-parameters, the batch size was set to 256 and the hidden layers of LSTM
100
80
y
c
a
r
c60
u
c
A
40</p>
        <p>Training and testing accuracy for spectrogram
5
10
15
20
25
30
35</p>
        <p>40
Number of Epochs</p>
        <p>(a)
Training and testing accuracy for spectrum matrix
20
40
60
80</p>
        <p>100
Number of Epochs
(b)
testing set
training set
testing set
training set
Comparison of different structures
single Bi-LSTM
double Bi-LSTM
double LSTM
10
8
6
4
2
s
s
o
l
g
n
i
t
s
e
T
to 256. We used stochastic gradient descent to optimize the network weights, with an initial
learning rate of 0.01, momentum of 0.9, and weight decay of 3e-4.</p>
        <p>The value of the spectrum matrix was normalized to 0–255 so that the images could be stored
on disk as grayscale images. Moreover, as the sizes of the images were diferent, these images
were padded to 256 × 256 before feeding to the network model.</p>
        <p>All the experiments and models were implemented in PyTorch on two NVIDIA RTX 3090
graphical processing units. It is worth mentioning that, in order to accelerate the training
process, we selected Apex [18] to implement mixed precision and distributed training.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Discussion</title>
        <p>It can be seen from Figure 9 that the testing set experienced massive jitters before stabilization.
This sharp deterioration did not accompany the training set, nor did it fail to converge or
explode in gradients. Instead, it returned to normal spontaneously and this process repeated
continuously. As the number of epochs increases, the testing loss and learning rate gradually
stabilize, the jitters disappear, and the testing set tends to be stable. We suspect that this may be
due to the low efective signal-to-noise (SNR) ratio. Since the experimental scene is close to the
genuine circumstance, we did not clear the sound of wind and people talking, but deliberately
mixed it as noise, and arising the dificulty of identification.</p>
        <p>Fortunately, the proposed model is robust enough to eliminate this phenomenon within
about 80 epochs (at least 46 epochs in the VGG 19 model) and guarantee an accuracy rate of
approximately 94% (up to 97.34% in the ResNet 18 model).</p>
        <p>Although the Doppler efect is widely used in the vehicle FWCM radar, the static state cannot
produce the Doppler efect. The method proposed in this paper does not require Doppler radar,
and can efectively determine whether NLOS for stationary and dynamic targets.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>In this letter, a novel method of identifying LOS/NLOS acoustics is proposed. We use a CNN
and Bi-LSTM in the deep neural network and the adapted training model to increase the signal
recognition accuracy to at least 97.34%. As the input of the neural network, instead of the
raw audio signal and spectrogram, a 2D image-like spectrum matrix is proposed to obtain
the classification precisely. In the field of acoustics in an IPS, this letter reports, to our best
knowledge, the first use of a focusing neural network to classify NLOS signals.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This work was supported in part by the Fundamental Research Funds for the Central Universities
(Zhejiang University NGICS Platform) and by the National Natural Science Foundation of China
under Grant Nos. 61773344, 61273079, 61772149, 61936002, and 6202780103.
[13] V. Barral, C. J. Escudero, J. A. García-Naya, R. Maneiro-Catoira, Nlos identification and
mitigation using low-cost uwb devices, Sensors 19 (2019) 3464.
[14] L. Zhang, D. Huang, X. Wang, C. Schindelhauer, Z. Wang, Acoustic nlos identification
using acoustic channel characteristics for smartphone indoor localization, Sensors 17
(2017) 727.
[15] S. Fan, Y. Wu, C. Han, X. Wang, A structured bidirectional lstm deep learning method for
3d terahertz indoor localization, in: IEEE INFOCOM 2020-IEEE Conference on Computer
Communications, IEEE, 2020, pp. 2381–2390.
[16] N. Rajagopal, P. Lazik, N. Pereira, S. Chayapathy, B. Sinopoli, A. Rowe, Enhancing indoor
smartphone location acquisition using floor plans, in: 2018 17th ACM/IEEE International
Conference on Information Processing in Sensor Networks (IPSN), IEEE, 2018, pp. 278–289.
[17] L. Zhang, M. Chen, X. Wang, Z. Wang, Toa estimation of chirp signal in dense multipath
environment for low-cost acoustic ranging, IEEE Transactions on Instrumentation and
Measurement 68 (2018) 355–367.
[18] N. Apex, Nvidia apex: Tools for easy mixed-precision training in pytorch, https://developer.
nvidia.com/blog/apex-pytorch-easy-mixed-precision-training/, 2022.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>X.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ansari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. R.</given-names>
            <surname>Elikplim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>A survey on fusion-based indoor positioning</article-title>
          ,
          <source>IEEE Communications Surveys &amp; Tutorials</source>
          <volume>22</volume>
          (
          <year>2019</year>
          )
          <fpage>566</fpage>
          -
          <lpage>594</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Cao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>Combined weighted method for tdoa-based localization</article-title>
          ,
          <source>IEEE Transactions on Instrumentation and Measurement</source>
          <volume>69</volume>
          (
          <year>2019</year>
          )
          <fpage>1962</fpage>
          -
          <lpage>1971</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Kui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Zomaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Nothing blocks me: Precise and real-time los/nlos path recognition in rfid systems</article-title>
          ,
          <source>IEEE Internet of Things Journal</source>
          <volume>6</volume>
          (
          <year>2019</year>
          )
          <fpage>5814</fpage>
          -
          <lpage>5824</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <article-title>Pals: high-accuracy pedestrian localization with fusion of smartphone acoustics and pdr</article-title>
          .,
          <source>in: IPIN (Short Papers/Work-in-Progress Papers)</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>291</fpage>
          -
          <lpage>298</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zheng</surname>
          </string-name>
          , S. Hranilovic, Idyll:
          <article-title>Indoor localization using inertial and light sensors on smartphones</article-title>
          ,
          <source>in: Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>307</fpage>
          -
          <lpage>318</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , G. Wang,
          <article-title>Nlos error mitigation for toa-based source localization with unknown transmission time</article-title>
          ,
          <source>IEEE Sensors Journal</source>
          <volume>17</volume>
          (
          <year>2017</year>
          )
          <fpage>3605</fpage>
          -
          <lpage>3606</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.-S.</given-names>
            <surname>Choi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          , S.-C. Kim,
          <article-title>Deep learning based nlos identification with commodity wlan devices</article-title>
          ,
          <source>IEEE Transactions on Vehicular Technology</source>
          <volume>67</volume>
          (
          <year>2017</year>
          )
          <fpage>3295</fpage>
          -
          <lpage>3303</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>F.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wang</surname>
          </string-name>
          , Ampn:
          <article-title>Real-time los/nlos identification with wifi</article-title>
          ,
          <source>in: 2017 IEEE International Conference on Communications (ICC)</source>
          , IEEE,
          <year>2017</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Wen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>Nlos identification and compensation for uwb ranging based on obstruction classification</article-title>
          ,
          <source>in: 2017 25th European signal processing conference (EUSIPCO)</source>
          , IEEE,
          <year>2017</year>
          , pp.
          <fpage>2704</fpage>
          -
          <lpage>2708</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bo</surname>
          </string-name>
          ,
          <article-title>Uwb nlos/los classification using deep learning method</article-title>
          ,
          <source>IEEE Communications Letters</source>
          <volume>24</volume>
          (
          <year>2020</year>
          )
          <fpage>2226</fpage>
          -
          <lpage>2230</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C.</given-names>
            <surname>Jiang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bo</surname>
          </string-name>
          ,
          <article-title>An uwb channel impulse response de-noising method for nlos/los classification boosting</article-title>
          ,
          <source>IEEE Communications Letters</source>
          <volume>24</volume>
          (
          <year>2020</year>
          )
          <fpage>2513</fpage>
          -
          <lpage>2517</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12] V.
          <string-name>
            <surname>-H. Nguyen</surname>
            , M.-
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Choi</surname>
            ,
            <given-names>Y.-H.</given-names>
          </string-name>
          <string-name>
            <surname>Kim</surname>
          </string-name>
          ,
          <article-title>Nlos identification in wlans using deep lstm with cnn features</article-title>
          ,
          <source>Sensors</source>
          <volume>18</volume>
          (
          <year>2018</year>
          )
          <fpage>4057</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>