<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Speech Enhancement Based on Minimum Variance Distortionless Response Beamformer in Adverse Recording Scenario</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>QuanTrong The</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Post and Telecommunication Institute of Technology</institution>
          ,
          <addr-line>Hanoi</addr-line>
          ,
          <country country="VN">Vietnam</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>In many hands-free speech communication systems such as surveillance device, voice controlled equipment, teleconference system, mobile phones, smart - home, hearing aid, the captured speech signal often degraded due to the existence of third - party talker, unwanted interference, noise, such that the speech enhancement technique are required to enhance the speech quality, speech intelligibility and perceptual listener. The microphone array (MA) technology has been commonly applied to various types of acoustic applications, because of the exploiting the priori information of MA distribution, the designed configuration of geometry, the characteristics of surrounding environment to obtain better noise reduction and speech enhancement at the same time. By using MA, the problem of sound source localization, estimation of direction of arrival of interest useful signal, the steered beampattern toward the desired target speaker, the suppressing of total all background noise are easy resolved. In MA beamforming technique, Minimum Variance Distortionless Response (MVDR) beamformer is one of the most useful methods for extracting the desired talker at certain direction while eliminating all background noise with speech distortion. However, the ideal performance of MVDR beamformer is usually corrupted in realistic recording situations due to the microphone mismatches, the different microphone sensitivities, the displacement of MA configuration. In this article, the author proposed an accurate calculation of steering vector to improve MVDR beamformer's evaluation in complex and annoying environment. The numerical result has confirmed the effectiveness of the author's suggested method in increasing the speech quality from 10.8 (dB) to 12.2 (dB) and reducing the speech distortion to 5.2 (dB). The superiority of this method can be integrated into a multi-channel system to achieve sustainable signal processing.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Microphone array</kwd>
        <kwd>minimum variance distortionless response</kwd>
        <kwd>beamforming</kwd>
        <kwd>speech enhancement</kwd>
        <kwd>steering vector</kwd>
        <kwd>speech quality</kwd>
        <kwd>the signal-to-noise ratio</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Nowadays, the using of hearing aids, smartphones, voice – controlled devices, cellular
communication, teleconferencing equipment, smart-vehicle present new complex challenges
for speech enhancement algorithms in recovering the original clean speech component while
suppressing background noise without speech distortion. The single channel, which is often based
on spectral subtraction, owns the simplicity of performance and easily installed in almost acoustic
instruments. However, this approach can perfectly work in stationary environments. In the
nonstationary situations and rapidly changing characteristics, this method causes speech distortion or
corrupted signal. Therefore, MA technology has been installed for overcoming this drawback. MA
beamforming uses the priori information to achieve better noise reduction and speech
enhancement at the same time.</p>
      <p>
        The beamforming exploits the designed MA distribution, the direction of arrival of speech,
the properties of surrounding noise environments, the constrained formulation of signal
processing, the additive spectral mask, post – filtering, preprocessing to process the captured
signals for obtaining the highest quality of speech enhancement. The MA beamformer can be
categorized into two groups: the fixed beamformer with delay and sum DAS [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1-4</xref>
        ], the adaptive
beamformer with differential microphone array DIF [
        <xref ref-type="bibr" rid="ref5 ref6 ref7 ref8">5-8</xref>
        ], generalized sidelobe canceller [
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref9">9-12</xref>
        ],
minimum variance distortionless response [
        <xref ref-type="bibr" rid="ref13 ref14 ref15 ref16">13-16</xref>
        ], linear constrained minimum variance LCMV
[
        <xref ref-type="bibr" rid="ref16 ref17 ref18 ref19 ref20">16-20</xref>
        ].
      </p>
      <p>The use of adaptive beamfoming technique depends on the particular configuration and the
purpose of the signal processing system on saving the target talker while minimizing the effect
of background noise. MVDR may be in the commonly utilized beamforming designs for
numerous speech applications, which minimize the noise power at the output while maintaining
the original speech component. In more complicated situations, where several sources,
nondirectional noise, competing talkers, it was applied to separate sound sources without speech
distortion.</p>
      <p>Although theoretical the MVDR provides optimal spatial diversity, when the interfering
source locations are constantly changing, the problem of speech enhancement is more
challenging due to the computation and source localization. Assuming the direction of arrival
(DoA) is known, the MVDR beamformer estimates the desired signal while minimizing the
variance of noise component at the output of beamformer. In practice, the DoA of the target
desired signal is not calculated exactly, which significantly degrades the performance of. A lot of
devoted research has been done to improve the robustness of MVDR beamformer by extending
the region where the source can be detected, determined according to the characteristics of
environment. Nevertheless, even assuming perfect source localization (SSL), the fact that
microphone sensors may have distinct, impulsive response, different directional gain and
another level of uncertainty that the MVDR beamformer is not able handle all scenarios well.</p>
      <p>Because of the microphone mismatches, the difference between microphone gain –
sensitivities, the inaccurate MA distribution, the error of computing about the DoA and the
undetermined characteristics of background environments significantly decrease the MVDR
beamformer’s performance. In this paper, the author proposed a new approach for adaptively
estimating the steering vector to enhance speech enhancement.</p>
      <p>The rest of this paper is organized as follows. The first section introduces the problem of
speech enhancement by using MA technology. The second section describes the principal
evaluation of MVDR beamforming in the frequency domain. The author’s suggested method
was presented in section III. The experiments were conducted in section IV with perspective
numerical results. Section V concludes and the author’s work in the future.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Minimum Variance Distortionless Response</title>
    </sec>
    <sec id="sec-3">
      <title>Beamformer</title>
      <p>In this section, the author describes the principle of working of MA in the frequency domain. In
general case, the dual - microphone system (DMA2) was illustrated to understand the problem
of signal processing by MVDR beamformer.</p>
      <p>At the current frequency – frame (, ), the observed MA signals on two microphones: 1(,
),2(, ) can be presented as:
(1)
(2)
(3)</p>
      <p>Where (, ) is the original clean speech signal, 1(, ), 2(, ) is the additive noise, which
significantly degrades on the speech quality. = 0(),  is the direction of arrival of
interest useful signal relative to the axis of DMA2, 0 = ⁄ is the time delay,  is the range
between two mounted microphones,  = 343 (⁄) is sound speed propagation in the air.</p>
      <p>With the definition:(, ) = [1(, ) 2(, )], (, ) = [1(, ) 2(, )], (, ) =
[ −], these equations (1) – (2) can be rewritten as:
(, ) = (, )(, ) + (, )
1(, ) = (, ) + 1(, )
2(, ) = (, )− + 2(, )</p>
      <p>The requirement of speech enhancement is finding an optimum filter’s coefficients(, ),
which ensures the estimated signal (M, ) = (, )(, ) approximately the original clean
speech.</p>
      <p>MVDR beamformer based on the constrained criteria of minimum the total out noise power
while preserving the speech component without speech distortion. The formulation of MVDR
beamformer can be expressed as the following way:</p>
      <p>min
W ( f , k )</p>
      <p>W H ( f , k )ΦNN ( f , k )W ( f , k ) st W H ( f , k ) Ds ( f , θs )=1
(4)</p>
      <sec id="sec-3-1">
        <title>With (, ) is the spectral covariance matrix of noise.</title>
        <p>From the constrained problem of MVDR beamformer, the optimum coefficients were defined as:</p>
        <p>Unfortunately, the priori information about noisy environment is not always available, so
the captured MA signals were used instead of. Finally, the optimum MVDR beamformer’s
coefficients were derived as:</p>
        <p>W MVDR ( f , k )=</p>
        <p>Φ−NN1 ( f , k ) DS ( f , θS )</p>
        <p>DSH ( f , θS )Φ−NN1 ( f , k ) DS ( f , θS )
W MVDR ( f , k )=</p>
        <p>Φ−NN1 ( f , k ) DS ( f , θS )
DSH ( f , θS )Φ−XX1 ( f , k ) DS ( f , θS )
(5)
(6)
Where Φ ❑XX ( f , k )= E {X H ( f , k ) X ( f , k )}=</p>
        <p>E {|X 1( f , k )|2}
E {X ¿2( f , k ) X 1( f , k )}</p>
        <p>E {X 1¿( f , k ) X 2( f , k )}</p>
        <p>E {|X 2( f , k )|2}
(⬚) is conjugate operator.</p>
        <p>The auto – cross power spectral densities can be calculated as the recursive formulation:
P Xi Xi ( f , k )=α P Xi Xi ( f , k −1)+(1−α ) X i¿ ( f , k ) X ❑i( f , k )
(7)
P Xi X j ( f , k )=α P Xi X j ( f , k )+(1−α ) X i¿ ( f , k ) X ❑j( f , k )
(8)</p>
      </sec>
      <sec id="sec-3-2">
        <title>Where  is the smoothing parameter, which in the range {0… 1}.</title>
        <p>In realistic recording situations, due to the microphone mismatches, the differences of
microphone sensitivities, the error of estimation of preferred steering vector, the displacement
of MA distribution, the imprecise of sampling frequency, the moving head of speaker, the
overall MVDR beamformer’s performance usually corrupted. In the next section, the author
proposed a new method for determining accurate steering vectors, which seriously affects signal
processing by beamforming technique.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>The proposed method of estimating the steering vector</title>
      <p>Steering vector (, ) play an important role in MVDR beamformer. However, the preferred
steering vector often changed in the frame with the presence of speech component. Therefore,
the author proposed a recursive formulation for estimating steering vectors as:
(, , ) = (,  − 1, ) + (1 − )(, ) (9)
With smoothing parameter  and the initial steering vector (, 0, ) = (, ).</p>
      <p>The steering vector in frame – speech  (, ) is computed by generalized eigenvalue
decomposition as the following way:
 (, ) = (, ){−1 (, )(, )}
(10)</p>
      <p>Where {−1 } extracts the principal eigenvector by applying generalized eigenvalues
decomposition to the spectral matrix of noise and speech.</p>
      <p>
        In practical applications, the priori information of noise and clean speech is not always
available. Hence, the author proposed using EM algorithm [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] for calculating the spectral
mask, which according to the presence/absence of speech enhancement. Let (, ) denote the
time – frequency mask for speech and (, ) represents the background noise probability.
Then, the matrix (, ),(, ) yields as the following equations:
      </p>
      <p>RSS ( f , k )=
1</p>
      <p>∑ M s ( f , k ) X ( f , k ) , X H ( f , k )
∑ M S ( f , k ) k
k
(11)
1
RNN ( f , k )= ∑ M n ( f , k ) X ( f , k ) , X H ( f , k )
∑ M n ( f , k ) k (12)
k</p>
      <p>The appealing properties of the author’s post – Filtering is tracking, updating the steering
vector according to the speech presence probability in the frame – speech, which leads to
enhance MVDR beamformer’s performance. In the next section, the author demonstrates
experiments to confirm the advantages of suggested technique.</p>
    </sec>
    <sec id="sec-5">
      <title>Experiment results</title>
      <p>
        The purpose of the conducted experiment is to compare the effectiveness of the traditional
MVDR beamformer (tMVbe) and the author suggested method (ausv) in increasing the speech
quality and energy. An objective measurement of SNR [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] was used for computing the obtained
processed signals. The experiment was performed in room dimensions about 9.5 x 8.0 x 3.6 m3.
The dual – microphone array (DMA2) was used for capturing the original clean speech
component, which degraded by noisy environment. The desired speaker stands at distance L =
3.5 (m) to the axis of DMA2 and the direction of arrival of interest signal is  = 600 in the
condition of existence of noise, interference. The range between two mounted microphones is 
= 5(). For recording the mixture of speech and noise, the sampling rate is set  = 16  ,
overlap 50%. The observed MA signals were depicted in Figure 6.
      </p>
      <p>= 512, smoothing parameter  = 0.1 for calculating the auto – cross power spectral
densities and  = 0.92 to track the changing steering vector. After using the MVDR beamformer,
the output signal was derived as Figure 7.</p>
      <p>As a result, in the recording scenarios of presence of complex environments, the moving
head of talker, the coherent difference sensitivities of microphone arrays, the error of estimation
of preferred steering vector significantly affected on the MVDR beamformer’s evaluation.
Although the noise level was suppressed, the original speech component was not recovered.
A comparison of energy between these signals was illustrated in Figure 9.</p>
      <p>By using the author’s method, the steering vector was recursively updated frame - by - frame,
which provides sustainable robustness beamforming for MVDR beamformer. The obtained
result was shown in Figure 8.</p>
      <p>Table 1 presents the receives SNR between the MA signals, the processed signal by tMVbe
and ausv.</p>
      <p>As a result, the obtained SNR increased from 10.8 (dB) to 12.2 (dB) in comparison with tMVbe.
The speech distortion decreased to 5.2 (dB). The updated steering vector allowed the high
directional beampattern adaptively steered toward the sound source according to the rapidly
changing of recording environment and ensured alleviation of the surroundings environment.
The appealing property of the suggested method is tracking, updating and changing
immediately filter’s coefficients for extracting the desired target speaker. The experiment has
presented the ability of the author’s method in reducing speech distortion, improving the MVDR
beamformer’s performance in adverse environments.</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>In this contribution, the author proposed using a new method for computing exactly steering
vector, which plays an important role in MVDR beamformer for forming a high directional
beampattern towards the sound source while removing the background noise. The appealing
property of the suggested method is overcome the heuristic drawback of MVDR beamformer is
very sensitive with the direction of arrival of useful speech component, which often decreases
the output signal’s quality. The numerical result has verified the advantages of the suggested
technique for accurately calculating MVDR beamformer’s coefficients in complex environments
to obtain the original speech component, reduce speech distortion to 5.2 (dB), increase the SNR
from 10.8 (dB) to 12.2 (dB). In the future, the author will investigate the characteristics of diffuse
noise field and the effects of reverberation in living room to further enhance the above method.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Chodingala</surname>
            <given-names>P. K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaturvedi</surname>
            <given-names>S. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patil</surname>
            <given-names>A. T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Patil H</surname>
          </string-name>
          .
          <source>A. Robustness of DAS Beamformer Over MVDR for Replay Attack Detection On Voice Assistants//Proc 2022 IEEE International Conference on Signal Processing and Communications (SPCOM)</source>
          , Bangalore, India,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          , doi: 10.1109/SPCOM55316.
          <year>2022</year>
          .
          <volume>9840757</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Zeng</surname>
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hendriks R. C.</surname>
          </string-name>
          <article-title>Distributed Delay and Sum Beamformer for Speech Enhancement via Randomized Gossip</article-title>
          .
          <source>IEEE/ACM Transactions on Audio, Speech, and Language Processing</source>
          , vol.
          <volume>22</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>260</fpage>
          -
          <lpage>273</lpage>
          , Jan.
          <year>2014</year>
          , doi: 10.1109/TASLP.
          <year>2013</year>
          .
          <volume>2290861</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Rakotoarisoa</surname>
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fischer</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Valeau</surname>
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Marx</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prax</surname>
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brizzi</surname>
            <given-names>L.E.</given-names>
          </string-name>
          <article-title>Time-domain delayandsum beamforming for time-reversal detection of intermittent acoustic sources in flows</article-title>
          .
          <source>J. Acoust. Soc. Am</source>
          .
          <volume>136</volume>
          ,
          <fpage>2675</fpage>
          -
          <lpage>2686</lpage>
          (
          <year>2014</year>
          ). https://doi.org/10.1121/1.4897402.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>António</surname>
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramos</surname>
            <given-names>L</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Holm</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gudvangen</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Otterlei</surname>
            <given-names>R</given-names>
          </string-name>
          .
          <article-title>Delay-and-sum beamforming for direction of arrival estimation applied to gunshot acoustics //</article-title>
          <source>Proc Proceedings Volume</source>
          <volume>8019</volume>
          ,
          <string-name>
            <surname>Sensors</surname>
          </string-name>
          , and
          <string-name>
            <surname>Command</surname>
          </string-name>
          , Control, Communications, and
          <string-name>
            <surname>Intelligence</surname>
          </string-name>
          (C3I)
          <article-title>Technologies for Homeland Security and Homeland Defense X; 80190U (</article-title>
          <year>2011</year>
          ) https://doi.org/10.1117/12.886833.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Huang</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cohen</surname>
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benesty</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <source>Chen J. Kronecker Product Beamforming with Multiple Differential Microphone Arrays // Proc 2020 IEEE 11th Sensor Array and Multichannel Signal Processing Workshop</source>
          (SAM), Hangzhou, China,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          , doi: 10.1109/SAM48682.
          <year>2020</year>
          .
          <volume>9104333</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Zhao</surname>
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luo</surname>
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benesty J. Differential</surname>
          </string-name>
          <article-title>Beamforming with Null Constraints for Spherical Microphone Arrays //</article-title>
          <source>Proc ICASSP 2024 - 2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          , Seoul, Korea, Republic of,
          <year>2024</year>
          , pp.
          <fpage>776</fpage>
          -
          <lpage>780</lpage>
          , doi: 10.1109/ICASSP48485.
          <year>2024</year>
          .
          <volume>10446768</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Luo</surname>
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jin</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benesty</surname>
            <given-names>J</given-names>
          </string-name>
          .
          <article-title>Design of Steerable Linear Differential Microphone Arrays With Omnidirectional and Bidirectional Sensors</article-title>
          .
          <source>IEEE Signal Processing Letters</source>
          , vol.
          <volume>30</volume>
          , pp.
          <fpage>463</fpage>
          -
          <lpage>467</lpage>
          ,
          <year>2023</year>
          , doi: 10.1109/LSP.
          <year>2023</year>
          .
          <volume>3267969</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Wang</surname>
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cohen</surname>
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benesty</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen J. Robust Steerable</surname>
          </string-name>
          <article-title>Differential Beamformers with Null Constraints for Concentric Circular Microphone Arrays</article-title>
          .
          <source>Proc ICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          , Toronto, ON, Canada,
          <year>2021</year>
          , pp.
          <fpage>4465</fpage>
          -
          <lpage>4469</lpage>
          , doi: 10.1109/ICASSP39728.
          <year>2021</year>
          .
          <volume>9414119</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>The</surname>
            <given-names>Q. T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huy</surname>
            <given-names>N. B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anh P. T. Spectral</surname>
          </string-name>
          <article-title>Mask - Based Technique for Improving Generalized Sidelobe Canceller Beamformer's Evaluation. 2023 Seminar on Signal Processing</article-title>
          , Saint Petersburg, Russian Federation,
          <year>2023</year>
          , pp.
          <fpage>106</fpage>
          -
          <lpage>110</lpage>
          , doi: 10.1109/IEEECONF60473.
          <year>2023</year>
          .
          <volume>10366094</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Wang</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guo</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            <given-names>J</given-names>
          </string-name>
          .
          <article-title>Robust Adaptation Control for Generalized Sidelobe Canceller with Time-Varying Gaussian Source Model //</article-title>
          <source>Proc 2023 31st European Signal Processing Conference (EUSIPCO)</source>
          , Helsinki, Finland,
          <year>2023</year>
          , pp.
          <fpage>16</fpage>
          -
          <lpage>20</lpage>
          , doi: 10.23919/EUSIPCO58844.
          <year>2023</year>
          .
          <volume>10289801</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Dai</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Li</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abbasi</surname>
            <given-names>Q. H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Imran M. A. A Fast Blocking</surname>
          </string-name>
          <article-title>Matrix Generating Algorithm for Generalized Sidelobe Canceller Beamformer in High Speed Rail Like Scenario</article-title>
          .
          <source>IEEE Sensors Journal</source>
          , vol.
          <volume>21</volume>
          , no.
          <issue>14</issue>
          , pp.
          <fpage>15775</fpage>
          -
          <lpage>15783</lpage>
          ,
          <fpage>15</fpage>
          <lpage>July15</lpage>
          ,
          <year>2021</year>
          , doi: 10.1109/JSEN.
          <year>2020</year>
          .
          <volume>3002699</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Middelberg</surname>
            <given-names>W.</given-names>
          </string-name>
          , Doclo S.
          <article-title>Comparison of Generalized Sidelobe Canceller Structures Incorporating External Microphones for Joint Noise</article-title>
          and
          <string-name>
            <given-names>Interferer</given-names>
            <surname>Reduction</surname>
          </string-name>
          .
          <source>Speech Communication; 14th ITG Conference</source>
          , online,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Shankar</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Küçük</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Reddy</surname>
            <given-names>C. K. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhat</surname>
            <given-names>G. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Panahi</surname>
            <given-names>I. M. S.</given-names>
          </string-name>
          <article-title>Influence of MVDR beamformer on a Speech Enhancement based Smartphone application for</article-title>
          <source>Hearing Aids// Proc 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC)</source>
          , Honolulu,
          <string-name>
            <surname>HI</surname>
          </string-name>
          , USA,
          <year>2018</year>
          , pp.
          <fpage>417</fpage>
          -
          <lpage>420</lpage>
          , doi: 10.1109/EMBC.
          <year>2018</year>
          .
          <volume>8512369</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Shankar</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhat</surname>
            <given-names>G. S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Panahi</surname>
            <given-names>I. M. S.</given-names>
          </string-name>
          <article-title>Real-time dual-channel speech enhancement by VAD assisted MVDR beamformer for hearing aid applications using smartphone</article-title>
          .
          <source>2020 42nd Annual International Conference of the IEEE Engineering in Medicine &amp; Biology Society (EMBC)</source>
          , Montreal, QC, Canada,
          <year>2020</year>
          , pp.
          <fpage>952</fpage>
          -
          <lpage>955</lpage>
          , doi: 10.1109/EMBC44109.
          <year>2020</year>
          .
          <volume>9175212</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Wang</surname>
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bao</surname>
            <given-names>C.</given-names>
          </string-name>
          <article-title>Multi-channel Speech Enhancement Based on the MVDR Beamformer</article-title>
          and Postfilter //
          <source>Proc 2020 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC)</source>
          , Macau, China,
          <year>2020</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          , doi: 10.1109/ICSPCC50002.
          <year>2020</year>
          .
          <volume>9259489</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Araki</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ono</surname>
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kinoshita</surname>
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Delcroix</surname>
            <given-names>M</given-names>
          </string-name>
          .
          <article-title>Meeting Recognition with Asynchronous Distributed Microphone Array Using Block-Wise Refinement of Mask-Based MVDR Beamformer /</article-title>
          / Proc 2018 IEEE International Conference on Acoustics,
          <source>Speech and Signal Processing (ICASSP)</source>
          , Calgary, AB, Canada,
          <year>2018</year>
          , pp.
          <fpage>5694</fpage>
          -
          <lpage>5698</lpage>
          , doi: 10.1109/ICASSP.
          <year>2018</year>
          .
          <volume>8462458</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Hassani</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bertrand</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moonen</surname>
            <given-names>M.</given-names>
          </string-name>
          <article-title>LCMV beamforming with subspace projection for multi-speaker speech enhancement /</article-title>
          <source>/ Proc 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          , Shanghai, China,
          <year>2016</year>
          , pp.
          <fpage>91</fpage>
          -
          <lpage>95</lpage>
          , doi: 10.1109/ICASSP.
          <year>2016</year>
          .
          <volume>7471643</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Xiao</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pu</surname>
            <given-names>W.</given-names>
          </string-name>
          , Luo
          <string-name>
            <given-names>Z. Q.</given-names>
            ,
            <surname>Zhang</surname>
          </string-name>
          <string-name>
            <surname>T</surname>
          </string-name>
          .
          <article-title>Evaluation of the Penalized Inequality Constrained Minimum Variance Beamformer for Hearing Aids /</article-title>
          /Proc 2018 IEEE International Conference on Acoustics,
          <source>Speech and Signal Processing (ICASSP)</source>
          , Calgary, AB, Canada,
          <year>2018</year>
          , pp.
          <fpage>3344</fpage>
          -
          <lpage>3348</lpage>
          , doi: 10.1109/ICASSP.
          <year>2018</year>
          .
          <volume>8462539</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Schreibman</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barnov</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gendelman</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tzirkel</surname>
            <given-names>E</given-names>
          </string-name>
          .
          <source>RTF Based LCMV Beamformer with Multiple Reference Microphones // Proc 2020 28th European Signal Processing Conference (EUSIPCO)</source>
          , Amsterdam, Netherlands,
          <year>2021</year>
          , pp.
          <fpage>181</fpage>
          -
          <lpage>185</lpage>
          , doi: 10.23919/Eusipco47968.
          <year>2020</year>
          .
          <volume>9287468</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Chazan</surname>
            <given-names>S. E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goldberger</surname>
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gannot</surname>
            <given-names>S.</given-names>
          </string-name>
          <article-title>DNN-Based Concurrent Speakers Detector and its Application to Speaker Extraction with LCMV Beamforming /</article-title>
          / Proc 2018 IEEE International Conference on Acoustics,
          <source>Speech and Signal Processing (ICASSP)</source>
          , Calgary, AB, Canada,
          <year>2018</year>
          , pp.
          <fpage>6712</fpage>
          -
          <lpage>6716</lpage>
          , doi: 10.1109/ICASSP.
          <year>2018</year>
          .
          <volume>8462407</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>[21] https://labrosa.ee.columbia.edu/projects/snreval/.</mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Araki</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Okada</surname>
            <given-names>M</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Higuchi</surname>
            <given-names>T</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ogawa</surname>
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nakatani</surname>
            <given-names>T.</given-names>
          </string-name>
          <article-title>Spatial correlation model based observation vector clustering and MVDR beamforming for meeting recognition /</article-title>
          <source>/ Proc 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)</source>
          , Shanghai, China,
          <year>2016</year>
          , pp.
          <fpage>385</fpage>
          -
          <lpage>389</lpage>
          , doi: 10.1109/ICASSP.
          <year>2016</year>
          .
          <volume>7471702</volume>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>