=Paper=
{{Paper
|id=Vol-3666/paper02
|storemode=property
|title=Edge computing applications: using a linear MEMS microphone array for UAV position detection through sound source localization
|pdfUrl=https://ceur-ws.org/Vol-3666/paper02.pdf
|volume=Vol-3666
|authors=Andrii V. Riabko,Tetiana A. Vakaliuk,Oksana V. Zaika,Roman P. Kukharchuk,Valerii V. Kontsedailo
|dblpUrl=https://dblp.org/rec/conf/doors/RiabkoVZKK24
}}
==Edge computing applications: using a linear MEMS microphone array for UAV position detection through sound source localization==
Edge computing applications: using a linear MEMS microphone array for UAV position detection through sound source localization Andrii V. Riabko1 , Tetiana A. Vakaliuk2,3,4,5 , Oksana V. Zaika1 , Roman P. Kukharchuk1 and Valerii V. Kontsedailo6 1 Oleksandr Dovzhenko Hlukhiv National Pedagogical University, 24 Kyivska Str., Hlukhiv, 41400, Ukraine 2 Zhytomyr Polytechnic State University, 103 Chudnivsyka Str., Zhytomyr, 10005, Ukraine 3 Institute for Digitalisation of Education of the NAES of Ukraine, 9 M. Berlynskoho Str., Kyiv, 04060, Ukraine 4 Kryvyi Rih State Pedagogical University, 54 Universytetskyi Ave., Kryvyi Rih, 50086, Ukraine 5 Academy of Cognitive and Natural Sciences, 54 Universytetskyi Ave., Kryvyi Rih, 50086, Ukraine 6 Inner Circle, Nieuwendijk 40, 1012 MB Amsterdam, Netherlands Abstract This study explores the use of a microphone array to determine the position of an unmanned aerial vehicle (UAV) based solely on the sound of its engines. The accuracy of localization depends crucially on the arrangement of the microphones. The study also considers a mathematical model of pulse density modulation for a digital MEMS microphone. It demonstrates the frequency dependence of the efficiency of a differential array of first-order microphones. Based on this frequency dependence of directivity and the instability model of the microphone parameters, a rational operating frequency range for the normal functioning of the microphone array can be established. The study proposes a model of a linear microphone array based on MEMS omnidirectional microphones. With a specific geometrical arrangement, this array produces a bidirectional pattern, which can be easily transformed into a unidirectional pattern using specialized algorithms or hardware (e.g., ADAU1761 codecs). Keywords edge computing, UAV, sound source localization, MEMS microphone, microphone array, frequency, directivity 1. Introduction Determining the position of a UAV (Unmanned Aerial Vehicle) by the sound of its engines can be important for several reasons. In military or security applications, being able to identify and locate UAVs by their engine sounds can help in detecting potential threats, including hostile drones or unauthorized surveillance. Sound-based localization can aid in the development of countermeasures to mitigate the risks posed by UAVs in sensitive areas. Sound-based UAV detection can complement existing air traffic management systems, providing additional situational awareness for managing airspace and preventing collisions with manned aircraft. In search and rescue operations or in case of lost or malfunctioning drones, sound-based tracking can assist in locating and recovering UAVs. In conservation efforts, it can help monitor UAVs used for illegal activities like poaching or wildlife disturbance. In urban areas or regions with dense UAV traffic, sound-based tracking can be useful for enforcing regulations related to UAV flight paths, altitudes, and no-fly zones. For protecting privacy, doors-2024: 4th Edge Computing Workshop, April 5, 2024, Zhytomyr, Ukraine " ryabko@meta.ua (A. V. Riabko); tetianavakaliuk@acnsci.org (T. A. Vakaliuk); ksuwazaika@gmail.com (O. V. Zaika); kyxap4yk1@ukr.net (R. P. Kukharchuk); valerakontsedailo@gmail.com (V. V. Kontsedailo) ~ http://irbis-nbuv.gov.ua/ASUA/0051396 (A. V. Riabko); https://acnsci.org/vakaliuk/ (T. A. Vakaliuk); http://pfm.gnpu.edu.ua/index.php/struktura1/2015-04-01-14-50-26 (O. V. Zaika); http://irbis-nbuv.gov.ua/ASUA/0076404 (R. P. Kukharchuk) 0000-0001-7728-6498 (A. V. Riabko); 0000-0001-6825-4697 (T. A. Vakaliuk); 0000-0002-8479-9408 (O. V. Zaika); 0000-0002-7588-7406 (R. P. Kukharchuk); 0000-0002-6463-370X (V. V. Kontsedailo) © 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR ceur-ws.org Workshop ISSN 1613-0073 Proceedings 14 sound-based detection can help identify UAVs flying near private properties, providing a means to take legal action against intrusive drones. Studying the acoustic signatures of UAVs can aid in research and development efforts to design quieter and more environmentally friendly drones. During natural disasters or emergencies, knowing the positions of UAVs, such as those used for aerial surveys or damage assessment, can assist in coordinating response efforts. Sound-based UAV detection can be employed in border control to monitor and respond to unauthorized drone incursions. As drone delivery and urban air mobility concepts develop, sound-based localization can contribute to managing UAV traffic in urban environments. While sound-based UAV localization offers several advantages, it also has limitations, such as accuracy challenges in noisy environments and the need for specialized equipment. Therefore, it is often used in conjunction with other tracking and detection methods, such as radar, visual recognition, and GPS, to provide comprehensive situational awareness and enhance safety and security in various applications. The goal of our work is to develop a software and hardware system for capturing hardware- synchronized sound using digital MEMS microphones (Microelectromechanical Systems, MEMS) for further use in sound source localization systems. This system is intended for further use in sound source localization systems, marking a significant advancement in the field of edge computing. 2. Theoretical background Over the past few decades, acoustic source localization has emerged as a focal point of interest within the research community [1, 2]. Most studies of sound source identification are based on the analysis of the physiological mechanism of human hearing [3, 4]. It is common practice to use arrays of microphones [5]. An actual problem is acoustic beam formation for sound source localization and its application [6]. Microphone array processing represents a well-established methodology employed in the estimation of sound source direction. In a groundbreaking contribution by Yamada et al. [7], they introduce an innovative approach referred to as Multiple Triangulation and Gaussian Sum Filter Tracking (MT-GSFT). This advanced technique adeptly derives the precise location of sound sources through triangulation, utilizing microphone arrays seamlessly integrated into a fleet of multiple drones [7]. The domain of speech signal processing encompasses several critical areas, and among them, multiple sound source localization (SSL) stands out as a notable and relevant field. A notable contribution to this field comes from Firoozabadi et al. [8], who introduced a two-step approach for the localization of multiple sound sources in three dimensions (3D). This method relies on the precise estimation of time delays (TDE) and strategically leverages distributed microphone arrays (DMA) to enhance the accuracy and effectiveness of the localization process [8]. Sasaki et al. [9] present a method designed to map the 3D coordinates of a sound source by leveraging data gathered from an array of microphones, with each microphone providing an autonomous directional estimate. Additionally, LiDAR technology is employed to create a comprehensive 3D representation of the surroundings and accurately determine the sensor’s position with six degrees of freedom (6-DoF). Catalbas et al. [10] conduct a comparative analysis, assessing the effectiveness of generalized cross- correlation techniques in contrast to noise reduction filters concerning the estimation of sound source trajectory. Throughout the entire movement, they calculate the azimuth angle between the sound source and the receiver. This calculation relies on the parameter of Interaural Time Difference (ITD) to determine the azimuth angle. They then evaluate the accuracy of the estimated delay using various types of Generalized Cross-Correlation (GCC) algorithms for comparison. It is possible for unmanned aerial vehicles (UAVs) to use audio information to compensate for poor visual information. Hoshiba et al. [11] developed a microphone array system built into the UAV to localize the sound source in flight. They developed the Spherical Microphone Array System (SMAS), consisting of a microphone array, a stable wireless network communication system, and intuitive visualization tools. Tachikawa et al. [12] introduced an innovative approach that involves estimating positions by utilizing 15 a modified variant of the convex clustering method in conjunction with sparse coefficients estimation. Additionally, they put forth a technique for constructing a well-suited monopole dictionary, which is based on coherence, ensuring that the convex clustering-based method can accurately estimate the distances of sound sources. The study involved conducting a series of numerical and measurement experiments aimed at assessing the effectiveness and performance of this novel methodology. When dealing with multiple sound sources, establishing a reliable data association between local- ization information and the corresponding sound sources becomes paramount for achieving optimal performance. To address the challenges posed by data association uncertainty, Wakabayashi et al. [13] extended the Global Nearest Neighbor (GNN) approach, introducing a modified version known as GNN-c, specifically tailored to meet the real-time and low-latency requirements of drone audio applications. The outcome of their efforts showcases a system capable of accurately estimating the positions of multiple sound sources, achieving an impressive accuracy level of approximately 3 meters. Many acoustic image-based sound source diagnosis systems suffer from spatial stationary limitations, making it challenging to integrate information from various capture positions, thereby leading to unreli- able and incomplete diagnostics. In their paper, Carneiro and Berry [14] introduce a novel measurement methodology called Acoustic Imaging Structure From Motion (AISFM). This approach utilizes a mobile spherical microphone array to create acoustic images through beamforming, seamlessly integrating data from multiple capture positions. Their method is not only proposed but also meticulously developed and rigorously validated, offering a promising solution to enhance the accuracy and comprehensiveness of sound source diagnostics. In a research conducted by Kita and Kajikawa [15] a sound source localization (SSL) technique is introduced, specifically designed for the localization of sources situated within structures, including mechanical equipment and buildings. The registration of acoustic signals with cross-shaped antennas is widely discussed in the literature [16]. Advanced signal processing methods involving multiple microphones can enhance noise resilience. However, as the quantity of microphones employed escalates, the computational overhead rises con- comitantly. This, in turn, curtails response time and hinders their extensive adoption across various categories of mobile robotic platforms [17]. Within the realm of robot audition, sound source localization (SSL) holds a pivotal role, serving as a fundamental component. SSL empowers a robotic platform to pinpoint the origin of sound using auditory cues exclusively. Its significance extends beyond mere sound localization, as it significantly influences other facets of robot audition, including source separation. Moreover, SSL contributes to elevating the quality of human-robot interaction by augmenting the robot’s perceptual prowess [18]. In general, machine learning is widely used in acoustics [19, 20]. In the realm of human-robot interaction, He et al. [21] have introduced a pioneering approach. Their proposal involves harnessing neural networks for the simultaneous detection and localization of multiple sound sources. This innovative method represents a departure from conventional signal processing techniques by offering a distinct advantage: it necessitates fewer stringent assumptions about the environmental conditions, thereby enhancing its adaptability and effectiveness [21]. Ebrahimkhanlou and Salamone [22] have put forth an advanced methodology for localizing acoustic emissions (AE) sources within metallic plates, especially those with intricate geometric features like rivet-connected stiffeners. This innovative approach leverages two deep learning techniques: a stack of autoencoders and a convolutional neural network (CNN), strategically employed to enhance the accuracy and precision of the localization process [22]. In their pioneering work, Adavanne et al. [23] have introduced an innovative solution – a convolu- tional recurrent neural network (CRNN) – designed to address the intricate task of joint sound event localization and detection (SELD) within three-dimensional (3-D) space. This method represents a significant advancement in the field, enabling the simultaneous identification and spatial localization of multiple overlapping sound events with remarkable precision. Let’s summarize the theoretical review. Localizing a sound source means determining the direction or location from which a sound is emanating. There are several algorithms and techniques used for 16 sound source localization, and the choice of method often depends on the specific application and available hardware. Here are some commonly used algorithms. Time Difference of Arrival (TDOA) is based on measuring the time it takes for a sound to reach multiple microphones. By comparing the time differences, it’s possible to triangulate the source’s position. Cross-correlation or beamforming techniques are often used to calculate the time differences accurately. Generalized Cross-Correlation (GCC) is a technique used in conjunction with TDOA. It involves cross-correlating the signals from two or more microphones to find the delay between them. GCC-PHAT (GCC with Phase Transform) is a commonly used variant that works well in reverberant environments. Steering Vector Methods are commonly used in microphone arrays or beamforming applications [24]. They estimate the direction of arrival (DOA) by analyzing the phase differences between signals received by different microphones. Popular algorithms include Multiple Signal Classification (MUSIC) and Estimation of Signal Parameters via Rotational Invariance Techniques (ESPRIT). Acoustic Intensity Methods measure the sound intensity at multiple microphone positions and use this information to estimate the source direction. The Steered Response Power (SRP) algorithm is an example of this approach. Machine learning and deep learning techniques, such as neural networks and support vector machines, can be used to train models for sound source localization. These models can take input from multiple microphones and learn to predict the source location based on training data. Particle filtering is a probabilistic method that estimates the source location using a Bayesian filtering approach. It is useful when dealing with complex and dynamic environments. Some methods use time-frequency analysis techniques like the Short-Time Fourier Transform (STFT) or Wavelet Transform to analyze the spectral content of audio signals and infer the source location. In mobile sound source localization, the Doppler effect can be used to estimate the source’s speed and direction based on the frequency shift in the received signal. Many practical systems use a combination of the above techniques to improve accuracy and robustness, especially in real-world scenarios with noise and reverberation. The choice of algorithm depends on factors like the number and arrangement of microphones, envi- ronmental conditions, computational resources, and the desired level of accuracy. Different applications, such as robotics, audio conferencing, surveillance, and hearing aids, may employ different algorithms tailored to their specific requirements. Determining the position of a UAV (Unmanned Aerial Vehicle) based solely on the sound of its engines can be challenging but is feasible using a combination of sound source localization techniques and signal processing. Here’s a high-level overview of the process: 1. Microphone Array Setup: Set up a microphone array on the ground. The microphones should be strategically placed to capture the UAV’s sound from different angles. The arrangement of microphones plays a crucial role in accurate localization. The response of microphone arrays depends, first of all, on the number of microphones working on the array [25]. 2. Sound Data Collection: Record the sound generated by the UAV’s engines as it flies overhead. Ensure that the recording system has a high sampling rate to capture the sound accurately. 3. Time Delay of Arrival (TDOA): Analyze the recorded audio data to calculate the time delay of arrival (TDOA) of the sound at each microphone. TDOA is the time difference between when the sound reaches different microphones. This information is critical for triangulation. 4. Triangulation: Use the TDOA data from multiple microphones to triangulate the UAV’s posi- tion. Several algorithms, such as multilateration or beamforming, can help estimate the UAV’s coordinates based on the TDOA information. 5. UAV Sound Signature: To improve accuracy, consider using machine learning techniques to create a database of UAV sound signatures. This involves training a model to recognize the unique sound characteristics of different UAVs. When a new sound recording is obtained, the model can help identify the specific UAV type. 6. Integration with Other Sensors: For real-time tracking, integrate sound-based localization with other sensors like GPS, radar, or visual cameras. This fusion of data sources can provide more accurate and robust positioning. 17 7. Calibration and Testing: Regularly calibrate and test the microphone array and signal processing algorithms to ensure accurate and reliable results. It’s important to note that the accuracy of sound-based UAV localization depends on various factors, including the UAV’s altitude, speed, engine type, and background noise. Additionally, environmental conditions, such as wind and temperature, can affect sound propagation and localization accuracy. Therefore, this method may work best in controlled environments or in conjunction with other tracking methods for enhanced precision and reliability. 3. Research methods The goal of our work is to develop a software and hardware system for capturing hardware-synchronized sound using digital MEMS microphones (Microelectromechanical Systems, MEMS) for further use in sound source localization systems. Despite the fact that the use of radar equipment has become part of everyday practice when monitoring UAVs, there is some interest in assessing the possibility of using airborne acoustic signals for this purpose. The above applies mainly to receiving hydroacoustic antennas, i.e. to conditions when the speed of the source is much lower than the speed of sound 𝑀 = 𝑣/𝑐 ≪ 1. In the case of receiving air-acoustic signals propagating at a speed of sound significantly lower than that of hydroacoustic signals in water and created by fairly fast moving sources (passenger cars on autobahns, racing cars, the movement of airliners along runways during takeoff and landing, UAVs), there is a different situation. Research into the features of recording these signals with phased arrays remains relevant, since on their basis data can be obtained on the current coordinates and speed of movement of a moving object. The purpose of this work is to analyze the angular dependencies in the signal at the output of a receiving air-acoustic antenna and those qualitative changes in their nature that are introduced due to a combination of factors such as the Doppler effect and the sharp directivity of the antenna array. A special case is considered, which is widespread in everyday practice, when the trajectory of an object is rectilinear, lies in a horizontal plane, close and parallel to the Earth’s surface, and the speed of its movement is constant. As previously stated, the arrangement of microphones plays a crucial role in accurate localization. The purpose of the study is to find the optimal configuration of a microphone array for localizing a moving sound source (UAV). Directivity is the sensitivity of a microphone to sound depending on the direction or angle from which the sound is coming. Directionality or sound pickup angle is considered to be the area of possible location of the sound signal source, within which there is no significant loss of microphone efficiency. Microphones use different directivity characteristics. They are most often depicted as polar diagrams. This is done to graphically display sensitivity variations around the microphone over a 360-degree range, where the microphone is the center of the circle and the angular reference point is placed in front of the microphone. The polar pattern shows how a microphone’s sensitivity to a sound signal depends on the location of its source. Microphone arrays are an array of several microphones combined by joint digital signal processing. Microphone arrays provide the following advantages over single-channel systems: 1) directionality of sound reception; 2) noise suppression of point sources; 3) suppression of non-stationary environmental noise; 4) partial weakening of reverberation; 5) the possibility of spatial localization of the sound source; 6) the ability to accompany a moving point sound source. A microphone array is one of the types of directional microphones, implemented as a set of sound receivers operating in concert (in phase or with certain phase delays). Geometrically, gratings can be implemented in different configurations – one-dimensional (linear, arc-shaped), two-dimensional (flat, spherical), three-dimensional, spiral, with uniform or non-equidistant pitch. The array’s radiation pattern is created by changing the ratio of phase delays for different channels (in the simplest case, an in-phase array with a fixed position of the main lobe; in more complex and expensive implementations, 18 a scanning system). The implementation of phase delays can be hardware (for example, on analog delay lines) or software (digital). The basic microphone array structures are Broadside and Endfire (figure 1). Figure 1: Basic array structures. These structures use omnidirectional microphones (microphones that, regardless of their orientation, receive signals from any direction). The figure 2 shows signal reception versus direction for various frequencies with a single omnidirectional microphone. For one microphone, frequency invariance is observed. Figure 2: Dependence of signal reception on direction by one omnidirectional microphone for frequencies 500 Hz, 1 and 5 kHz. The Broadside structure is an array of omnidirectional microphones positioned perpendicular to the direction of the desired signal. Such arrays have an axis of symmetry, relative to which the sound is released without attenuation both “in front” of the array and “behind”. Such structures are widely used in applications where sound pressure waves enter the sensor array from one side. Consider a Broadside structure consisting of two microphones spaced 7.5 cm apart. The minimum response is 19 observed when the signal is incident at an angle of 90∘ or 270∘ (in this case, the angle between the direction of the useful signal and the normal to the line of elements is taken as 0∘ ). But this response strongly depends on the frequency of the received signal. Theoretically, such a system has a perfect zero at a frequency of 2.3 kHz. Above this frequency, depending on the direction of arrival, there are zeros at other angles (figure 3). The microphone array shows a clear directional characteristic at 4 kHz, and at 1 kHz its pattern is essentially omnidirectional. As a result, at lower frequencies the array cannot achieve significant spatial filtering. Figure 3: Dependence of signal reception on direction by a Broadside structure of two omnidirectional micro- phones for frequencies of 1 kHz, 2 kHz, 3 kHz and 4 kHz. The Endfire structure consists of several microphones located in the direction of the useful acoustic signal. This design is called a differential array of microphones. The delayed signal from the first microphone is summed with the signal from the next microphone. To create a cardioid polar pattern, the signal from the rear microphones must be delayed by the same amount of time that the sound waves travel between the two microphone elements. Such structures are used to produce cardioid, hypercardioid or supercardioid directional response and theoretically completely eliminate sound incident on the array at an angle of 180∘ . A unidirectional microphone is more sensitive to sound coming from one direction and less sensitive to sounds from other directions. The most typical for such microphones is the cardioid characteristic, representing a peculiar diagram in the shape of a heart. At the same time, the peak of sensitivity is reached in the direction along the axis of the microphone, and the decline is in the opposite direction (figure 4). To generate a cardioid response in direction, the signal from the omnidirectional microphones must be delayed for a time equal to the propagation of the acoustic wave between the two elements. Developers of such systems have two degrees of freedom to change the output signal of the speaker system: changing the distance between microphones and changing the delay time. Figure 5 shows the signal reception versus direction for various frequencies by the Endfire structure with two elements and a distance between them of 2.1 cm. The distance between the microphones is crucial for the formation of a cardioid response. Figure 6 shows the same microphones, but placed at a distance of 15 cm. The structures considered have the following advantages and disadvantages. Advantages of Broad- side: flat geometry, simple processing implementation, ability to control the direction of the beam. Disadvantages of the Broadside: less off-axis rejection, close microphone spacing, and a large number of microphones needed to prevent spatial leakage. 20 Figure 4: Dependence of signal reception on a unidirectional microphone for frequencies of 500 Hz, 1 kHz and 2 kHz. Figure 5: Dependence of signal reception on direction by an Endfire structure of two omnidirectional microphones for frequencies from 1 to 10 kHz, which are located at a distance of 21 cm. Advantages of Endfire: Better off-axis suppression, smaller overall size. Disadvantages of Endfire: non-flat (volumetric) geometry, more complex processing, suppression of the useful signal in the low frequency range, the direction of the source of the useful signal must coincide with the axis of the microphone array; For two-dimensional gratings, beam formation is possible only in the horizontal direction (the grating array). To form a differential array of higher orders, you need to add additional microphones. Since the petals will deviate more back and to the side in the directional diagram, the distance between the microphones will have to be increased. The figure 7 shows an array of 4 microphones (third order), which forms 21 Figure 6: Dependence of signal reception on direction by an Endfire structure of two omnidirectional microphones for frequencies from 1 to 10 kHz, which are located at a distance of 15 cm. a supercardioid pattern. Consider how beam formation depends on the number of microphones and the distance between them. It is worth noting that the sensitivity and frequency response of all array microphones must be precisely matched. Figure 7: Differential array of 4 microphones (third order). Differential microphone arrays make it possible to obtain high directivity characteristics of the system with its small size. But with such a construction, the problem arises of a significant change in the characteristics of the entire system with a slight deviation of the parameters of an individual microphone from its nominal values. If this approach is used for critical applications, measures must be taken to reduce deviations of microphone parameters from nominal values. As stated earlier, a first-order 22 differential microphone array consists of two omnidirectional sensors separated by 𝑑 (figure 8). Figure 8: Structure of a first order differential microphone array. 4. Results When sound arrives from the main direction 𝜃 = 0, a delay appears between these sensors: 𝑑 𝜏𝐷 = , (1) 𝑐 where 𝑐 is the speed of sound. → − A plane wave, which is characterized by wave number 𝑘 , arrives at the input of the differential grating. Due to radial symmetry, the output signals of sensors 𝑋1 (𝜔) and 𝑋2 (𝜔) can be expressed → − by a function depending on the angle 𝜃 and frequency 𝜔. There is a relationship | 𝑘 |𝑑 = 𝑘𝑑 = 𝜔𝜏𝐷 between the wave number and the frequency of the signal. At the central point of the array, you can place a virtual microphone with an output signal 𝑋0 (𝜔). A plane wave incident at an angle 𝜃 with wave number causes 𝑘 = 2𝜋/𝜆 the appearance of signals at the output of microphones 𝑋1 and 𝑋2 : 𝑘𝑑 𝑘𝑑 𝑋1 (𝜔) = 𝑋0 (𝜔)𝑒𝑗 2 cos 𝜃 , 𝑋2 (𝜔) = 𝑋0 (𝜔)𝑒−𝑗 2 cos 𝜃 . (2) At the output of the differential lattice we get 1 𝑌𝐷 (𝜔) = (𝑋1 (𝜔) − 𝑋2 (𝜔)𝑒𝑗𝜔𝜏 . (3) 2 The directivity function of the differential array 𝐻𝐷 is the ratio of the signal at the output of the array 𝑌𝐷 (𝜔) to the signal at the output of the virtual microphone 𝑋0 (𝜔): (︂ (︂ )︂)︂ −𝑗 𝜔𝜏 𝑘𝑑 𝜏 𝐻𝐷 (𝜔, 𝜃) = 𝑗𝑒 2 sin + cos 𝜃 . (4) 2 𝜏𝐷 Usually very small values of 𝑘𝑑 ≪ 1 are considered, which makes it possible to use the approximation sin 𝛼 ≈ 𝛼. In this case, the idealized directivity function 𝐻𝐷 has the form: (︂ )︂ 𝑘𝑑 𝜏 𝐻𝐷 (𝜔, 𝜃) ≈ 𝐻𝐷 (𝜃) = 𝑗 ̃︀ + cos 𝜃 . (5) 2 𝜏𝐷 With this view, the main characteristics of differential microphone arrays are obvious: 1) the form of 𝐻̃︀ 𝐷 (𝜃) is determined by the expression 𝜏 /𝜏𝐷 + cos 𝜃, which does not depend on frequency; 2) due to the subtraction of the signal a phase shift occurs by 𝜋/2; 3) the frequency response of the directivity function 𝐻𝐷 (𝜔) has the form of a first-order high-pass filter. At low frequencies, the output signal 𝑌𝐷 (𝜔) becomes highly susceptible to any changes in the shape of the characteristic 𝐻𝐷 (𝜔). For this reason, the distance d should not be chosen too small, which may lead to a conflict with the condition 𝑘𝑑 ≪ 1. 23 The exact expression for the directivity function (4) contains a sine function that scales the amplitude. It is rational to limit the operating range of the differential grating in the low frequency range to the first maximum of the sine. This first maximum fixes the cutoff frequency 𝜔𝑐: 𝜋 𝜔𝑐 = . (6) 𝜏𝐷 + 𝜏 For low frequencies, the directivity characteristics are practically independent of frequency. However, as the frequency increases, the shape of the frequency response becomes more and more deformed. In addition, at some frequencies the signal is completely suppressed. In order to compensate for the high-frequency nature of the behavior 𝐻𝐷 (𝜔, 𝜃) it is necessary to develop a filter 𝑊𝑒𝑞 (𝜔). For the main direction 𝜃 = 0, the adjusted frequency response𝐻𝐷 (𝜔, 𝜃 = 0)𝑊𝑒𝑞 (𝜔) must be constant and equal to 0 dB, and for frequencies below 𝜔𝑐 : (︁1 )︁ , 0 < 𝜔 < 𝜔𝑐 , {︃ 𝜋𝜔 𝑊𝑒𝑞 (𝜔) = sin 2𝜔𝑐 (7) 1, in other cases. For low frequencies 𝜔 → 0, the filter gain 𝑊𝑒𝑞 has very large values. This means that any noise present in the input signal will be greatly amplified. The level of this noise is determined by the specific sensor. This circumstance limits the frequency range of the signal for processing using a differential microphone array. The directional properties of a microphone array are characterized by the directional coefficient (DI). It can be expressed as the ratio of the squared modulus of the directivity function in the main direction to the average value of the squared modulus in all directions: |𝐻(𝜔, 𝜃 = 0)|2 𝐷𝐼(𝜔) = . (8) 1 ∫︀2𝜋 ∫︀𝜋 4𝜋 |𝐻(𝜔, 𝜃)|2 sin 𝜃𝑑𝜃𝑑𝜙 0 0 Taking into account the exact expression for the directivity function (4), we can obtain a new expression for the dependence of the directivity on frequency: 2𝑠𝑖𝑛2 𝜔2 (𝜏𝐷 + 𝜏 ) (︀ )︀ 𝐷𝐼𝐷 (𝜔) = (9) 1 − 𝑠𝑖 (𝜔𝜏𝐷 ) cos (𝜔𝜏 ) where 𝑠𝑖(𝑥) = 𝑥1 sin(𝑥). The efficiency factor for low frequencies is obtained similarly to the result of approximation 𝐻̃︀ 𝐷 according to expression (5): 2 lim 𝐷𝐼𝐷 (𝜔) = 𝐻 ̃︀ 𝐷 = 3(𝜏𝐷 + 𝜏 ) . (10) 𝜔→0 3𝜏𝐷2 + 3𝜏 2 Let us study the influence of microphone parameter mismatch for first-order differential arrays. We use a model of instability of microphone parameters in the form of a transfer function 𝑀 = 𝑀𝑟𝑒𝑓 +Δ𝑀 . The nominal transfer function of the sensor 𝑀𝑟𝑒𝑓 in this case is normalized to the value 1. It is assumed that the deviation Δ𝑀 is an independent random variable with variance: 2 𝜎𝑀 = 𝐸{|Δ𝑀 |2 }, (11) where 𝐸{|Δ𝑀 |2 } expectation operator. Signals from two sensors in figure 8 will then be written as follows: ^ 1 (𝜔) = 𝑋0 (𝜔)(1 + Δ𝑀1 )𝑒𝑗 𝑘𝑑 𝑋 2 cos 𝜃 ^ 𝑘𝑑 , 𝑋 2 (𝜔) = 𝑋0 (𝜔)(1 + Δ𝑀2 )𝑒−𝑗 2 cos 𝜃 . (12) The directivity function 𝐻 ^ 𝐷 for a differential array, taking into account the instability of the micro- phone parameters, can be obtained similarly to expression (4). But now there are additional conditions 24 that depend on Δ𝑀𝑖 (𝑖 = 1, 2). For random numbers, the quadratic terms remain, and the linear ones are set to zero, so we get: 𝐸{|𝐻^ 𝐷 (𝜔, 𝜃)|2 } = |𝐻𝐷 (𝜔, 𝜃)|2 + 2𝜎𝑀 2 . (13) As a result, we can obtain a modified expression for DI: 2 𝜔 2 (︀ )︀ 2 sin (𝜏𝐷 + 𝜏 ) + 𝜎𝑀 ^ 𝐷 (𝜔)|} = 𝐸{|𝐷𝐼 2 2 (14) 1 − 𝑠𝑖 (𝜔𝜏𝐷 ) cos (𝜔𝜏 ) 𝜎𝑀 It is important to understand that in expression (13) the efficiency factor 𝐻𝐷 (𝜔, 𝜃) characterizes the behavior of the system at high frequencies. While the 𝑊𝑒𝑞 equalization filter takes into account the effects of microphone instability and enhances them for low frequencies. Thus, this work shows the dependence of the efficiency of a differential array of first-order micro- phones on frequency. It can be supplemented using a model of instability of microphone parameters at low frequencies. Based on the presented dependence of the directivity on frequency and the instability model of the microphone parameters, a rational operating frequency range for the normal functioning of the microphone array can be determined. The lower limit of this range is limited by the instability of the microphone parameters, and the upper cutoff frequency is determined by the geometry of the array 𝑑. Currently, there are many applications in which acoustic signals are processed. Microelectromechan- ical microphones (MEMS) are increasingly being used for these purposes. The use of such microphones allows the construction of differential microphone arrays. Microelectromechanical systems (MEMS) are a variety of microdevices of a wide variety of designs and purposes, in the production of which modified microelectronics technological techniques are used. Typically, all elements of such systems are placed on a common silicon base, the size of which is only a couple of millimeters. A MEMS microphone is an electro-acoustic device for converting sound vibrations into electrical waves, which is small enough to be installed in a tightly integrated product, for example: a smartphone, headset, speakerphone, laptop or any other device. There are two fundamentally important elements in such microphones: an integrated circuit (ASIC) and a MEMS sensor. It is the latter that ensures the capture and subsequent transmission of sound. The MEMS sensor itself consists of a flexible membrane and a rigidly fixed cover. Under the influence of air pressure, the membrane moves, changing the capacitance between the plates. This data is recalculated and output as an electrical signal to an integrated circuit. It is this signal that is converted into the sound that we hear. Thanks to their design, MEMS microphones have the following advantages. Greater resistance to noise, vibration and temperature changes due to the absence of unnecessary connecting elements. Multiple MEMS microphones can be combined together to create a single array. Thanks to capacitive technology, these microphone arrays can capture sound from a precisely defined direction, effectively canceling echoes and background noise. Unlike other small microphones, such as electrets, MEMS microphones include more additional elements, such as preamplifiers, various filters and analog-to-digital converters. This means greater functionality while maintaining microscopic dimensions. Possibility of mounting such devices on the board using soldering. Despite their many advantages, MEMS microphones are also not without their disadvantages. As we wrote above, MEMS microphones are often used as part of arrays, which increases the sound capture area, but at the same time reduces the service life of the devices. To work correctly, all microphones must work in unison, but the likelihood of one of them breaking is much higher than an individual device. Worse protection from moisture and dust than other microphones. Microphone arrays include two or more built-in microphones, to which is added a programmable microprocessor designed to continuously determine the primary source of audio input and optimally adjust the output to achieve the best sound quality. Let us highlight the most significant quality indicators of sound capture systems: • useful signal/noise ratio, where the useful signal is the sound of the drone engine, and the noise is background noise, the microphone’s own noise, and sounds from non-target sources; 25 • the shape of the radiation pattern and the ability of the system to change it depending on the environment; • ability to localize the source of a useful signal and measurement accuracy parameters. The most common way to build a signal capture unit is based on analog microphone arrays. A description of the problems that arise when developing analog microphone arrays, as well as the rationale for reducing the importance of the problem when using digital microphones in audio capture systems, is given in table 1. Table 1 Problems encountered in the development of analog microphone arrays and justification for the feasibility of using digital microphones. Problem when designing an analog ar- Rationale for using digital mirophones ray microphones Sharp increase in cost Lack of a large number of auxiliary analog components Reduced yield of suitable products due Reducing the total number of microcircuits and topological com- to the large number of components plexity leads to an increase in the percentage of usable products due to the general laws of statistics Increased development and debugging Implementation of algorithms in code and digital interface blocks, costs which allows you to attract developers with less qualifications and experience High sensitivity to electromagnetic ra- The use of digital components is less sensitive to static failures and diation and power quality degradation of power supply quality Increased production cycle Less topological complexity guarantees the ability to produce a product according to almost any modern technological standards, making the launch process faster and cheaper Increasing the testing cycle The digital implementation allows you to write synthetic tests and generate input signals in the same way. Digital generators are more flexible and low cost, and the testing and debugging process is reduced to working with code As an alternative to existing approaches that have the disadvantages outlined above, the authors of this work proposed to use an architecture built using digital MEMS microphones, which have become widespread recently. The meter for these microphones is located on-chip, so its digital output is minimally affected by the components that surround it. The most simple, inexpensive and perfect solution in terms of signal capture parameters was developed, which consists of using digital MEMS microphones and an Arduino microcomputer. When choosing a digital microphone for use in a linear differential microphone array, it is important to consider the following factors: • Sensitivity: The sensitivity of the microphone is a measure of how well it can convert sound waves into electrical signals. A higher sensitivity microphone will be able to pick up quieter sounds, but it may also be more susceptible to noise. • Signal-to-Noise Ratio (SNR): The SNR of the microphone is a measure of the ratio of the desired signal (sound) to the undesired signal (noise). A higher SNR microphone will have less noise, resulting in cleaner recordings. • Dynamic range: The dynamic range of the microphone is the range of sound pressure levels that it can accurately measure. A wider dynamic range microphone will be able to capture both very loud and very quiet sounds without distortion. • Linearity: The linearity of the microphone is a measure of how accurately it can reproduce the input signal. A more linear microphone will produce recordings that are more faithful to the original sound. In addition to the above factors, it is also important to consider the cost and availability of the microphone when making a selection. 26 Each digital MEMS microphone can be simplified into the model shown in figure 9. Input sound vibrations are converted through a MEMS membrane into a weak electrical signal, which is then fed to the input of amplifier A. The pre-amplified signal then passes through an analog low-pass filter (LPF), which is necessary to protect against aliasing. The final element of signal processing in the microphone is a 4th order Σ − Δ modulator, which converts the input analog signal into a one-bit digital stream. The frequency of data bits from the output of the Σ − Δ modulator is equal to the frequency of the input timing signal CLK and, as a rule, lies in the range from 1 to 4 MHz. Figure 9: A simple model of a digital MEMS microphone. In the time domain, the output of a Σ − Δ modulator is a jumbled collection of ones and zeros. However, if we assign a value of 1.0 to each high logical level of the microphone output, and a value of –1.0 to each low level and then perform a Fourier transform, we will obtain a spectrogram of the output data from the microphone. Let’s look at the pins of a digital microphone. VDD – microphone power supply, GND – Ground, CLK – input clock signal, synchronously with which the DATA line switches its DATA states. During one half of the CLK cycle this pin is in a high impedance state, and during the second half it serves as a pin for reading data from the Σ − Δ output of the microphone modulator. 𝐿/𝑅𝑆𝑒𝑙 – this pin is used to control switching of the DATA line. If 𝐿/𝑅𝑆𝑒𝑙 is connected to VDD, then after some time after detecting the rising edge of the CLK signal, the DATA pin goes into a high impedance state, and after the arrival of the falling edge of the CLK signal, the DATA pin is connected to the Σ − Δ output of the microphone modulator. If 𝐿/𝑅𝑆𝑒𝑙 is connected to GND, the edges of the CLK signal, along which the DATA line switches, are reversed (figure 10). Figure 10: The pins of a digital microphone. To isolate the audio frequency band signal, the data from the microphone must be filtered and resampled at a lower frequency (usually 50–128 times lower than the sampling frequency of the Σ − Δ modulator). A digital low-pass filter filters out external noise and the microphone’s own noise outside the operating band (𝑓 > 𝐹𝐶𝐿𝐾 /2𝑀 ) to protect against aliasing, and also makes it possible to reduce the data repetition rate. In figure 11 presents one of the possible options for processing a one-bit data stream from a microphone, implemented in software on a DSP or in hardware in audio codecs. Shown in figure 11, the sampling frequency compression circuit (compressor) lowers the sampling frequency due to the fact that from every 𝑀 samples of the filtered signal 𝑤(𝑚𝑀 ), 𝑀 ˘1 sample is discarded. The 27 input and output of the converter shown in figure 8 are related by the following expression: ∞ ∑︁ 𝑦(𝑚) = 𝑤(𝑚𝑀 ) = ℎ(𝑘)𝑥(𝑚𝑀 − 𝑘). (15) 𝑘=−∞ Figure 11: Signal conversion by Σ − Δ modulator. MEMS microphones have PDM (Pulse-Density Modulation, PDM) outputs. Pulse density modulation is a method of transmitting the relative change in signal per sample, which can be mathematically described by the formula 𝑥[𝑛] = −𝐴(−1𝑎[𝑛] ), (16) where 𝑥[𝑛] contains in each term the relative change in the signal in the form of 1 bit with a sign, which is specified by the transition. A negative increment is a transition from 1 to 0, a positive increment is from 0 to 1. Repeating ones increases the overall amplitude of the signal, and repeating zeros decreases (figure 12). Figure 12: Period of a sine wave per 100 samples. A mathematical model for pulse density modulation can be obtained using a delta-sigma modulator model. In the discrete frequency domain, the operation of a delta-sigma modulator can be described by the formula 𝑂(𝑧) = 𝐼(𝑧) + 𝐸(𝑧)(1 − 𝑧 −1 ), (17) where 𝑂(𝑧), 𝐼(𝑧) are the signal spectra at the input and output of the modulator; 𝐸(𝑧) is the sampling error of the delta-sigma modulator; 1 − 𝑧 −1 is high-pass filter. As a result of transforming the formula, we get 1 𝑂(𝑧) = 𝐸(𝑧)[𝐼(𝑧) − 𝑂(𝑧)𝑧 −1 )] . (18) 1 − 𝑧 −1 According to this formula, the error 𝐸(𝑧) reduces the value of the signal at the output 𝑂(𝑧) in the low-frequency region and increases it in the high-frequency region, as a result of which the quantization noise spectrum shifts predominantly to the high-frequency region. 28 Table 2 Array characteristics (ULA with 2 microphones). Array characteristic Value Array directivity 2.97 dBi at 0 Az; 0 El Array span x=0 m y=17 mm z=0 m Number of elements 2 HPBW 60.50∘ Az / 360.00∘ El FNBW 180.00∘ Az / -∘ El SLL - dB Az / - dB El Element polarization None Let 𝑖[𝑛] be a sample of the signal at the input of the modulator in the time domain, and 𝑜[𝑛] be a sample of the output signal, then, using the inverse 𝑧-transform, we can proceed to the expression 𝑜[𝑛] = 𝑖[𝑛] + 𝑒[𝑛] − 𝑒[𝑛 − 1], (19) where if𝑥[𝑛] ≥ 𝑒[𝑛 − 1]; {︂ 1 𝑜[𝑛] = (20) −1 if𝑥[𝑛] < 𝑒[𝑛 − 1]; 𝑒[𝑛] = 𝑜[𝑛] − 𝑖[𝑛] + 𝑒[𝑛 − 1]. (21) The signal from the output signal sample 𝑜[𝑛] is represented as 1 bit and takes values ±1, and is implemented so that the value of the current quantization error 𝑒[𝑛] is minimal. In this case, the quantization error 𝑒[𝑛] of each sample appears at the device input during the subsequent sample. When implementing frequency converters in software, a finite impulse response (FIR) filter or an Infinite impulse response (IIR) filter can be used as a digital LPF. Developers should be very careful when choosing the type of filter, its length and bit depth, since the performance of the entire system as a whole directly depends on this. A correctly calculated and implemented decimator (frequency converter) in some cases will significantly reduce the cost of products and increase its technical characteristics. As a second option, audio codecs adapted for this can be used to convert data from the output of a digital microphone, which will significantly reduce product development time. For example, Analog Devices offers the ADAU1361 and ADAU1761 codecs, which are suitable for the ADMP521 microphones. In our work we used a microphone ADMP521. However, the process of creating digital audio devices becomes simple in terms of hardware implementation and complex in terms of writing programs for the microcontrollers used. Next, we conducted a simulation and computational experiment of a uniform linear array of 2 omnidirectional microphones using Matlab Sensor Array Analyzer. The following model parameters were used. The distance between the microphones is 20 mm. The board has two MEMS microphones spaced 20 mm apart. This spacing is ideal for detecting acoustic events. Additionally, the 20 mm spacing is equivalent to 8 · 2.54 mm, which makes it suitable for DIP (Dual In-line Package) – a type of housing for microcircuits, electronic modules and some other electronic components. Experimentally, a distance of 0.017 m was determined for the formation of a bi-directional pattern. Next, we conducted a simulation and computational experiment of a uniform linear array of 2 omnidirectional microphones using Matlab Sensor Array Analyzer. The following model parameters were used. The distance between the microphones is 17 mm. The speed of sound is 343 m/s, the signal frequency is 10 kHz. As a result, we obtained the parameters listed in table 2. The Matlab script is listed below: % Create a Uniform Linear Array Object Array = phased.ULA(’NumElements’,2, ’ArrayAxis’,’y’); Array.ElementSpacing = 0.017; 29 Array.Taper = ones(1,2).’; % Create an omnidirectional microphone element Elem = phased.OmnidirectionalMicrophoneElement; Elem.FrequencyRange = [0 10000]; Array.Element = Elem; % Assign Frequencies and Propagation Speed Frequency = 10000; PropagationSpeed = 343; % Create Figure % Plot Array Geometry figure; viewArray(Array,’ShowNormal’,true, ’ShowTaper’,true,’ShowIndex’,’All’, ’ShowLocalCoordinates’,true,’ShowAnnotation’,true, ’Orientation’,[0;0;0]); % Find the weights w = ones(getNumElements(Array), length(Frequency)); % Plot 2d azimuth graph format = ’polar’; cutAngle = 0; plotType = ’Directivity’; plotStyle = ’Overlay’; figure; pattern(Array, Frequency, -180:180, cutAngle, ’PropagationSpeed’, PropagationSpeed, ’CoordinateSystem’, format ,’weights’, w, ’Type’, plotType, ’PlotStyle’, plotStyle); % Find the weights w = ones(getNumElements(Array), length(Frequency)); % Plot 2d elevation graph format = ’polar’; cutAngle = 0; plotType = ’Directivity’; plotStyle = ’Overlay’; figure; pattern(Array, Frequency, cutAngle, -90:90, ’PropagationSpeed’, PropagationSpeed, ’CoordinateSystem’, format ,’weights’, w, ’Type’, plotType, ’PlotStyle’, plotStyle); % Find the weights w = ones(getNumElements(Array), length(Frequency)); % Plot U Pattern format = ’uv’; plotType = ’Directivity’; plotStyle = ’Overlay’; figure; pattern(Array, Frequency, -1:0.01:1, 0, ’PropagationSpeed’, PropagationSpeed, ’CoordinateSystem’, format,’weights’, w, ’Type’, plotType, ’PlotStyle’, plotStyle); The resulting pattern has the shape of a bi-directional (figure 13). As you can see, the design of the grille allows you to create a grille with the main lobes directed at -90 and 90 degrees. To form a cardiode radiation pattern, as mentioned above, it is necessary to use delay-and-sum and filter-and-sum algorithms. The meaning of these algorithms is that microphone signals are added with different delays (different phase shifts), aligning the phases of signals coming from the selected direction (source 30 localization) for each frequency. In this case, the beamforming algorithm makes it possible to amplify the signals generated by sound coming from the selected direction, i.e. performs a kind of focusing of sounds. Figure 13: The geometry of the array and the directional diagram (linear array of 2 microphones). ADMP521 microphones were connected to the ADAU1761 codec in accordance with the technical specifications of both products (figure 14). A model was also created based on an array of four omnidirectional microphones located at a distance of 20 mm (figure 15). The following model parameters were used. The distance between the microphones is 17 mm. The speed of sound is 343 m/s, the signal frequency is 10 kHz. As a result, we obtained the parameters listed in table 3. Table 3 Array characteristics (ULA with 4 microphones). Array characteristic Value Array directivity 5.98 dBi at 0 Az; 0 El Array span x=0 m y=51 mm z=0 m Number of elements 4 HPBW 26.52∘ Az / 360.00∘ El FNBW 60.58∘ Az / -∘ El SLL 11.30 dB Az / - dB El Element Polarization None The Matlab script has the following form: % Create a Uniform Linear Array Object Array = phased.ULA(’NumElements’,4, ’ArrayAxis’,’y’); Array.ElementSpacing = 0.017; Array.Taper = ones(1,4).’; % Create an omnidirectional microphone element Elem = phased.OmnidirectionalMicrophoneElement; Elem.FrequencyRange = [0 10000]; Array.Element = Elem; 31 Figure 14: Connection diagram of ADMP521 microphones to ADAU1761 codec. % Assign Frequencies and Propagation Speed Frequency = 10000; PropagationSpeed = 343; % Create Figure % Plot Array Geometry figure; viewArray(Array,’ShowNormal’,true,’ShowTaper’,true,’ShowIndex’,’All’, ’ShowLocalCoordinates’,true,’ShowAnnotation’,true,’Orientation’,[0;0;0]); % Calculate Steering Weights Freq3D = 10000; % Find the weights w = ones(getNumElements(Array), length(Frequency)); % Plot 3d graph format = ’polar’; plotType = ’Directivity’; figure; pattern(Array, Freq3D , ’PropagationSpeed’, PropagationSpeed, ’CoordinateSystem’, format,’weights’, w(:,1), ’ShowArray’,false,’ShowLocalCoordinates’,true, ’ShowColorbar’,true,’Orientation’,[0;0;0],’Type’, plotType); % Find the weights w = ones(getNumElements(Array), length(Frequency)); 32 Figure 15: The geometry of the array and the directional diagram (linear array of 4 microphones). % Plot 2d azimuth graph format = ’polar’; cutAngle = 0; plotType = ’Directivity’; plotStyle = ’Overlay’; figure; pattern(Array, Frequency, -180:180, cutAngle, ’PropagationSpeed’, PropagationSpeed,’CoordinateSystem’, format ,’weights’, w, ’Type’, plotType, ’PlotStyle’, plotStyle); % Find the weights w = ones(getNumElements(Array), length(Frequency)); % Plot 2d elevation graph format = ’polar’; cutAngle = 0; plotType = ’Directivity’; plotStyle = ’Overlay’; figure; pattern(Array, Frequency, cutAngle, -90:90, ’PropagationSpeed’, PropagationSpeed,’CoordinateSystem’, format ,’weights’, w, ’Type’, plotType, ’PlotStyle’, plotStyle); So, during the computational experiment, we built 2 linear microphone arrays with bi-directionality. The directionality of these arrays can be easily converted to unidirectional (cardioid) using known algorithms or hardware (codecs). The tuning of the circuit to create cardioid directivity will be considered in further studies. 33 5. Discussion The following questions require additional discussion and clarification: the dependence of the azimuthal pattern of the proposed linear microphone arrays on the source frequency; the choice of analog or digital MEMS microphones and labor-intensiveness in the development of a microphone array; the use of directional microphones instead of omnidirectional; peculiarities of localization of a moving sound source (Doppler effect, reflection from obstacles, etc.); higher-order differential beam array formers; signal processing algorithms of microphone arrays. 6. Conclusions The study looks at setting up a microphone array to determine the position of a UAV (unmanned aerial vehicle) based solely on the sound of its engines. The location of the microphones plays a crucial role for accurate localization. A mathematical model of pulse density modulation of a digital MEMS microphone is also considered. This work shows the dependence of the efficiency of a differential array of first-order microphones on frequency. Based on the presented dependence of the directivity on frequency and the instability model of the microphone parameters, a rational operating frequency range for the normal functioning of the microphone array can be determined. A model of a linear microphone array based on MEMS omnidirectional microphones is proposed, which with a certain geometrical arrangement give a bi-directional pattern, which, in principle, can be easily transformed into a unidirectional one with the use of special algorithms or hardware (for example, ADAU1761 codecs). Refinement of the circuit to achieve cardioid directivity will be addressed in forthcoming research. 7. Author contributions Conceptualization, methodology – Andrii V. Riabko, Oksana V. Zaika; setting tasks, conceptual analysis – Tetiana A. Vakaliuk, Oksana V. Zaika; development of the model – Andrii V. Riabko, Valerii V. Kontsedailo; software development, verification – Andrii V. Riabko, Roman P. Kukharchuk; analysis of results, visualization – Roman P. Kukharchuk, Tetiana A. Vakaliuk; drafting of the manuscript – Valerii V. Kontsedailo, reviewing and editing – Tetiana A. Vakaliuk. All authors have read and approved the published version of this manuscript. References [1] M. Cobos, F. Antonacci, A. Alexandridis, A. Mouchtaris, B. Lee, A Survey of Sound Source Localization Methods in Wireless Acoustic Sensor Networks, Wireless Communications and Mobile Computing 2017 (2017) 3956282. doi:10.1155/2017/3956282. [2] A. R. Petrosian, R. V. Petrosyan, I. A. Pilkevych, M. S. Graf, Efficient model of PID controller of unmanned aerial vehicle, Journal of Edge Computing 2 (2023) 104–124. doi:10.55056/jec.593. [3] M. Risoud, J.-N. Hanson, F. Gauvrit, C. Renard, P.-E. Lemesre, N.-X. Bonne, C. Vincent, Sound source localization, European Annals of Otorhinolaryngology, Head and Neck Diseases 135 (2018) 259–264. doi:10.1016/j.anorl.2018.04.009. [4] W. A. Yost, M. T. Pastore, Y. Zhou, Sound Source Localization Is a Multisystem Process, Springer International Publishing, Cham, 2021, pp. 47–79. doi:10.1007/978-3-030-57100-9_3. [5] E. King, A. Tatoglu, D. Iglesias, A. Matriss, Audio-visual based non-line-of-sight sound source localization: A feasibility study, Applied Acoustics 171 (2021) 107674. doi:10.1016/j.apacoust. 2020.107674. [6] P. Chiariotti, M. Martarelli, P. Castellini, Acoustic beamforming for noise source localization – reviews, methodology and applications, Mechanical Systems and Signal Processing 120 (2019) 422–448. doi:10.1016/j.ymssp.2018.09.019. 34 [7] T. Yamada, K. Itoyama, K. Nishida, K. Nakadai, Sound Source Tracking by Drones with Microphone Arrays, in: 2020 IEEE/SICE International Symposium on System Integration (SII), 2020, pp. 796–801. doi:10.1109/SII46433.2020.9026185. [8] A. D. Firoozabadi, P. Irarrazaval, P. Adasme, D. Zabala-Blanco, P. Palacios-Játiva, H. Durney, M. Sanhueza, C. Azurdia-Meza, Three-dimensional sound source localization by distributed microphone arrays, in: 2021 29th European Signal Processing Conference (EUSIPCO), 2021, pp. 196–200. doi:10.23919/EUSIPCO54536.2021.9616326. [9] Y. Sasaki, R. Tanabe, H. Takemura, Probabilistic 3D sound source mapping using moving micro- phone array, in: 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2016, pp. 1293–1298. doi:10.1109/IROS.2016.7759214. [10] M. C. Catalbas, M. Yildirim, A. Gulten, H. Kurum, S. Dobrišek, Estimation of Trajectory and Location for Mobile Sound Source, International Journal of Advanced Computer Science and Applications 7 (2016). doi:10.14569/IJACSA.2016.070934. [11] K. Hoshiba, K. Washizaki, M. Wakabayashi, T. Ishiki, M. Kumon, Y. Bando, D. Gabriel, K. Nakadai, H. G. Okuno, Design of UAV-Embedded Microphone Array System for Sound Source Localization in Outdoor Environments, Sensors 17 (2017) 2535. doi:10.3390/s17112535. [12] T. Tachikawa, K. Yatabe, Y. Oikawa, 3D sound source localization based on coherence-adjusted monopole dictionary and modified convex clustering, Applied Acoustics 139 (2018) 267–281. doi:10.1016/j.apacoust.2018.04.033. [13] M. Wakabayashi, H. G. Okuno, M. Kumon, Drone audition listening from the sky estimates multiple sound source positions by integrating sound source localization and data association, Advanced Robotics 34 (2020) 744–755. doi:10.1080/01691864.2020.1757506. [14] L. Carneiro, A. Berry, Three-dimensional sound source diagnostic using a spherical microphone array from multiple capture positions, Mechanical Systems and Signal Processing 199 (2023) 110455. doi:10.1016/j.ymssp.2023.110455. [15] S. Kita, Y. Kajikawa, Fundamental study on sound source localization inside a structure using a deep neural network and computer-aided engineering, Journal of Sound and Vibration 513 (2021) 116400. doi:10.1016/j.jsv.2021.116400. [16] F. R. do Amaral, J. Rico, M. A. F. de Medeiros, Design of microphone phased arrays for acoustic beamforming, Journal of the Brazilian Society of Mechanical Sciences and Engineering 40 (2018) 354. doi:10.1007/s40430-018-1275-5. [17] F. Grondin, F. Michaud, Lightweight and optimized sound source localization and tracking methods for open and closed microphone array configurations, Robotics and Autonomous Systems 113 (2019) 63–80. doi:10.1016/j.robot.2019.01.002. [18] C. Rascon, I. Meza, Localization of sound sources in robotics: A review, Robotics and Autonomous Systems 96 (2017) 184–210. doi:10.1016/j.robot.2017.07.011. [19] M. J. Bianco, P. Gerstoft, J. Traer, E. Ozanich, M. A. Roch, S. Gannot, C.-A. Deledalle, Machine learning in acoustics: Theory and applications, The Journal of the Acoustical Society of America 146 (2019) 3590–3628. doi:10.1121/1.5133944. [20] H. Niu, Z. Gong, E. Ozanich, P. Gerstoft, H. Wang, Z. Li, Deep-learning source localization using multi-frequency magnitude-only data, The Journal of the Acoustical Society of America 146 (2019) 211–222. doi:10.1121/1.5116016. [21] W. He, P. Motlicek, J.-M. Odobez, Deep Neural Networks for Multiple Speaker Detection and Localization, in: 2018 IEEE International Conference on Robotics and Automation (ICRA), 2018, pp. 74–79. doi:10.1109/ICRA.2018.8461267. [22] A. Ebrahimkhanlou, S. Salamone, Single-Sensor Acoustic Emission Source Localization in Plate- Like Structures Using Deep Learning, Aerospace 5 (2018) 50. doi:10.3390/aerospace5020050. [23] S. Adavanne, A. Politis, J. Nikunen, T. Virtanen, Sound Event Localization and Detection of Overlapping Sources Using Convolutional Recurrent Neural Networks, IEEE Journal of Selected Topics in Signal Processing 13 (2019) 34–48. doi:10.1109/JSTSP.2018.2885636. [24] G. Chardon, Theoretical analysis of beamforming steering vector formulations for acoustic source localization, Journal of Sound and Vibration 517 (2022) 116544. doi:10.1016/j.jsv.2021. 35 116544. [25] B. da Silva, A. Braeken, K. Steenhaut, A. Touhafi, Design Considerations When Accelerating an FPGA-Based Digital Microphone Array for Sound-Source Localization, J. Sensors 2017 (2017) 6782176:1–6782176:20. doi:10.1155/2017/6782176. 36