Next Article in Journal
Validation of Zulu Watch against Polysomnography and Actigraphy for On-Wrist Sleep-Wake Determination and Sleep-Depth Estimation
Previous Article in Journal
Deep Learning Based Evaluation of Spermatozoid Motility for Artificial Insemination
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Binaural Heterophasic Superdirective Beamforming

1
Center of Intelligent Acoustics and Immersive Communications, Northwestern Polytechnical University, 127 Youyi West Road, Xi’an 710072, China
2
INRS-EMT, University of Quebec, 800 de la Gauchetiere Ouest, Montreal, QC H5A 1K6, Canada
3
Andrew and Erna Viterby Faculty of Electrical Engineering, Technion-Israel Institute of Technology, Technion City, Haifa 32000, Israel
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(1), 74; https://doi.org/10.3390/s21010074
Submission received: 31 October 2020 / Revised: 16 December 2020 / Accepted: 22 December 2020 / Published: 25 December 2020
(This article belongs to the Section Physical Sensors)

Abstract

:
The superdirective beamformer, while attractive for processing broadband acoustic signals, often suffers from the problem of white noise amplification. So, its application requires well-designed acoustic arrays with sensors of extremely low self-noise level, which is difficult if not impossible to attain. In this paper, a new binaural superdirective beamformer is proposed, which is divided into two sub-beamformers. Based on studies and facts in psychoacoustics, these two filters are designed in such a way that they are orthogonal to each other to make the white noise components in the binaural beamforming outputs incoherent while maximizing the output interaural coherence of the diffuse noise, which is important for the brain to localize the sound source of interest. As a result, the signal of interest in the binaural superdirective beamformer’s outputs is in phase but the white noise components in the outputs are random phase, so the human auditory system can better separate the acoustic signal of interest from white noise by listening to the outputs of the proposed approach. Experimental results show that the derived binaural superdirective beamformer is superior to its conventional monaural counterpart.

1. Introduction

Microphone arrays combined with proper beamforming methods have been used in a wide range of applications, such as hearing aids, smart headphones, smart speakers, voice communication, automatic speech recognition (ASR), human–machine interfaces, etc., to extract signals of interest from noisy observations. Many beamformers have been developed over the last few decades [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17], among which the so-called superdirective beamformer [9] is particularly attractive. It is derived by maximizing the directivity factor (DF), which is equivalent to maximizing the gain in signal-to-noise ratio (SNR) in diffuse noise, subject to the distortionless constraint at the endfire direction. So, this beamformer is more efficient than other fixed beamformers for suppressing noise, interference, and reflections incident from different directions in noisy and reverberant environments [14,18]. It also has a frequency-invariant beampattern if the sensor spacing is small, which is essential for acquiring high-fidelity broadband acoustic and speech signals.
However, there is one major drawback with the existing superdirective beamforming approach: the white noise amplification is very serious at low frequencies. As a consequence, the application of this beamformer requires well-designed arrays with high-quality microphones of extremely low self-noise level, e.g., at least below 0 dB(A) for second- and higher-order superdirective beamformers, which is difficult if not impossible to attain. This white noise amplification problem considerably limits the use of the superdirective beamformer in practical applications [18,19] and how to deal with this problem has become an important issue that has attracted a significant amount of research attention. A number of methods have been developed subsequently in the literature, including the so-called robust superdirective beamformer [9,20], the combined superdirective beamformer [21], the optimized superdirective beamformer [22], the subspace superdirective beamformer [11,23], and the reduced-rank superdirective beamformer [24]. While the problem is approached from different perspectives, the fundamental principle underlying those methods stays the same, i.e., making a compromise between the DF and the level of white noise gain (WNG) [10,20,22,25,26]. In other words, all those methods attempt to circumvent the white noise amplification problem by sacrificing the DF and, as a consequence, the resulting beamformer may no longer be superdirective.
In this work, we take a different avenue. Instead of sacrificing the DF to improve the WNG, we design the superdirective beamformer, which consists of two sub-beamformers, each generates an output. The two sub-beamformers are designed to be orthogonal to each other so that the acoustic signal of interest in the binaural outputs is in phase while the (amplified) white noise is random phase. This design is strongly motivated by studies and facts in psychoacoustics, which showed that the location (or direction) information of signals has a significant impact on speech intelligibility in the human auditory system.
Many experiments have been conducted to study the influence of the direction information of speech and white noise (especially at frequencies below 1 kHz) on speech intelligibility [27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48]. Briefly, the impact of the source direction on the perception of speech in the human binaural auditory system can be classified into two scenarios: in phase and out of phase; while the perception of noise can be divided into three scenarios: in phase, random phase, and out of phase, where in phase means that in every frequency the binaural (two-channel) signals have the same phase and out of phase means that in every frequency the phase difference between the binaural signals is exactly 180 . An illustration of the impact of binaural signals and noise phase on speech intelligibility, inspired from [49], is shown in Figure 1, where the leftmost column indicates the phase relationship of the white noise at the left and right channels, and the first row represents the phase relationship of binaural speech signals. The combination indicates the influence of interaural direction relations on the localization of speech and white noise in space. The circle shape indicates that the signal is concentrated in a limited area, in front of the head for the case of in phase or on either side of the ears for the case of random phase. The rectangle shape indicates that the signal spreads in the area behind both ears in the case of out of phase.
The blue circle indicates that the speech signals at the left and right channels are in phase but the white noise components at the two channels are in random phase, which is related to the proposed binaural superdirective beamformer.
Listening tests have been conducted to study the intelligibility of different phase combinations. It is confirmed that the phase combination of the desired speech signal and white noise has a significant influence on intelligibility. A list of the most common six scenarios, summarized in [49] and represented in Table 1, can be divided into three categories: antiphasic, heterophasic, and homophasic. Listening tests showed that the antiphasic category corresponds to the highest intelligibility, which can be 25 % higher than that of the homophasic category in low SNR scenarios. The intelligibility of the heterophasic category is also higher than that of the homophasic case though lower than that of the antiphasic scenario. Inspired by this, it is desirable to design beamformers that make the binaural outputs corresponding to the antiphasic or heterophasic cases.
In this paper, we use the interaural coherence (IC) to describe the auditory localization information. Consider a diffuse noise field; when the IC of the binaural signals reaches its maximum, i.e., 1, there is a precise region of the sound source, which is located in the middle of the head, i.e., the in-phase case; however, when the binaural signals are completely incoherent, i.e., the IC equals to 0, there are two independent sources at the two ears. This corresponds to the random phase case. In many approaches in the literature, binaural processing generates two collinear filters. The resulting output ICs for white (due to noise amplification) and diffuse noises are 1. So, both the signal of interest and the noise in the human auditory system are perceived to be in the same region. Consequently, our brain will have difficulties to separate them with a binaural presentation and intelligibility will certainly be affected. Apparently, the conventional monaural superdirective beamformer and the conventional collinear binaural processing belong to the homophasic case, which has the lowest intelligibility. To improve intelligibility, we propose a binaural superdirective beamformer by constructing two orthogonal filters, with which the IC for the white noise components is equal to zero while the IC for the desired signal components is equal to one in the binaural outputs. Consequently, with the proposed method, our auditory system can more easily distinguish between the signal of interest and the (amplified) white noise by listening to the binaural outputs, leading to improved intelligibility [50]. An illustration of the proposed binaural superdirective beamformer with a uniform linear array is shown in Figure 2, which can suppress spatial noise while seperating the desired signal and the white noise into different perception zone. Since the binaural superdirective beamformer developed in this paper corresponds to the heterophasic case [49] (see also Table 1), we name it binaural heterophasic superdirective beamformer. Note that the IC information has been used in the traditional binaural speech enhancement methods for binaural cues preservation [51,52]. However, the proposed binaural heterophasic superdirective beamformer uses the IC magnitude information in a very different way where the two orthogonal sub-beamformers are designed to minimize the IC magnitude of the white noise components while maximizing the IC magnitude of the diffuse noise components in the binaural outputs to achieve better perceptual separation of the signal of interest and white noise.
The rest of this paper is organized as follows. In Section 2, we present the signal model and formulate the problem. In Section 3, we briefly review the derivation of the conventional superdirective beamformer. In Section 4, we discuss binaural linear filtering and the associated performance measures. In Section 5, we derive the binaural heterophasic superdirective beamformer. Then, in Section 6, we present some experiments to validate the theoretical study. Finally, conclusions are presented in Section 7.

2. Signal Model and Problem Formulation

We consider a source signal of interest (plane wave), in the farfield, that propagates from the azimuth angle, θ , in an anechoic acoustic environment at the speed of sound, i.e., c = 340 m/s, and impinges on a uniform linear array (ULA) consisting of 2 M omnidirectional microphones. In this scenario, the corresponding steering vector (of length 2 M ) is [5]
d ω , θ = 1 e j ω τ 0 cos θ e j ( 2 M 1 ) ω τ 0 cos θ T ,
where j is the imaginary unit with j 2 = 1 , ω = 2 π f is the angular frequency, with f > 0 being the temporal frequency, τ 0 = δ / c is the delay between two successive sensors at the angle θ = 0 , with δ being the interelement spacing, and the superscript T is the transpose operator.
Assume that the desired signal comes from a specific direction θ = θ s . From the steering vector defined in (1), we can express the frequency-domain observation signal vector of length 2 M as [2]
y ω = Y 1 ω Y 2 ω Y 2 M ω T = x ω + v ω = d ω , θ s X ω + v ω ,
where Y m ω is the mth microphone signal, x ω = d ω , θ s X ω , X ω is the zero-mean source signal of interest, which is also called the desired signal, d ω , θ s is the signal propagation vector, which is same as the steering vector at θ = θ s , and v ω is the zero-mean additive noise signal vector defined similarly to y ω . We deduce that the 2 M × 2 M covariance matrix of y ω is
Φ y ω = E y ω y H ω = ϕ X ω d ω , θ s d H ω , θ s + Φ v ω = ϕ X ω d ω , θ s d H ω , θ s + ϕ V 1 ω Γ v ω ,
where the superscript H is the conjugate-transpose operator, E [ · ] denotes mathematical expectation, ϕ X ω = E X ω 2 is the variance of X ω , Φ v ω = E v ω v H ω is the covariance matrix of v ω , ϕ V 1 ω = E V 1 ω 2 is the variance of the noise, V 1 ω , at the first sensor, and Γ v ω = Φ v ω / ϕ V 1 ω is the pseudo-coherence matrix of the noise. We assume that noises at different sensors have the same variance.
In order to design the superdirective beamformer, we make two basic assumptions [9,19].
(i)
The sensor spacing, δ , is much smaller than the acoustic wavelength, λ = c / f , i.e., δ λ (this implies that ω τ 0 2 π ). This assumption is required so that the true acoustic pressure differentials can be approximated by finite differences of the microphones’ outputs.
(ii)
The desired source signal propagates from the angle θ s = 0 (endfire direction). Therefore, (2) becomes
y ω = d ω , 0 X ω + v ω ,
and, at the endfire, the value of the beamformer beampattern should always be equal to 1 (or maximal).
Our objective in this paper is to derive a binaural superdirective beamformer, which can take advantage of the human binaural auditory system to separate the desired speech signal from white noise so that the intelligibility of the beamformer’s output signals will be higher than that of the output of the conventional (monaural) superdirective beamformer. To that end, we will find two various and useful estimates of X ω , each for one of the binaural channels, so that along with our binaural hearing system, white noise amplification will be perceptually attenuated thanks to this binaural presentation.

3. Conventional Superdirective Beamformer

The conventional linear fixed beamforming technique is performed by applying a complex weight at the output of each microphone and then sum all the weighted outputs together to get an estimate of the source signal [2,19], i.e.,
Z ω = h H ω y ω = X ω h H ω d ω , 0 + h H ω v ω ,
where Z ω is the estimate of the desired source signal, X ω , and h ω is a spatial linear filter of length 2 M containing all the complex weights. We see from (5) that the distortionless constraint should be
h H ω d ω , 0 = 1 .
Now, we can define the directivity factor (DF) of the beamformer as [3,19,43]
D h ω = h H ω d ω , 0 2 1 2 0 π h H ω d ω , θ 2 sin θ d θ = h H ω d ω , 0 2 h H ω Γ d ω h ω ,
where Γ d ω = 1 2 0 π d ω , θ d H ω , θ sin θ d θ whose i j th ( i , j = 1 , 2 , , 2 M ) element is
Γ d ω i j = sin ω ( j i ) τ 0 ω ( j i ) τ 0 = sinc ω ( j i ) τ 0 .
The matrix Γ d ω can be viewed as the pseudo-coherence matrix of the spherically isotropic (diffuse) noise.
By taking into account the distortionless constraint in (6), the maximization of the DF in (7) leads to the well-known superdirective beamformer [9,53]:
h SD ω = Γ d 1 ω d ω , 0 d H ω , 0 Γ d 1 ω d ω , 0 ,
whose DF is
D h SD ω = d H ω , 0 Γ d 1 ω d ω , 0 ,
which, obviously, is maximal. Besides maximizing the DF, the other great advantage of h SD ω is that the corresponding beampattern is almost frequency invariant. However, white noise amplification is a tremendous problem. Consequently, the superdirective beamformer can only be used with a very small number of microphones and/or with regularization of the matrix Γ d ω , but this regularization affects the DF as well as the shape of the beampattern, which makes the beamformer more frequency dependent. Therefore, there is still a great interest to find new ideas to improve this superdirective beamformer.

4. Binaural Linear Filtering and Performance Measures

In this section, we explain binaural linear filtering in connection with fixed beamforming and propose some important performance measures in this context.
The extension of the conventional (monaural) fixed linear beamforming to the binaural case can be done by applying two complex-valued linear filters, h 1 ω and h 2 ω of length 2 M , to the observed signal vector, y ω , i.e.,
Z i ω = h i H ω y ω = X ω h i H ω d ω , 0 + h i H ω v ω , i = 1 , 2 ,
where Z 1 ω and Z 2 ω are two different estimates of X ω . The variance of Z i ω is then
ϕ Z i ω = h i H ω Φ y ω h i ω = ϕ X ω h i H ω d ω , 0 2 + h i H ω Φ v ω h i ω = ϕ X ω h i H ω d ω , 0 2 + ϕ V 1 ω h i H ω Γ v ω h i ω .
It is clear that the two distortionless constraints are
h i H ω d ω , 0 = 1 , i = 1 , 2 .
A very important performance measure is the input SNR, which can be obtained from (3), i.e.,
iSNR ω = ϕ X ω ϕ V 1 ω .
According to (12), the binaural output SNR can be defined as
oSNR h 1 ω , h 2 ω = ϕ X ω ϕ V 1 ω × i = 1 2 h i H ω d ω , 0 2 i = 1 2 h i H ω Γ v ω h i ω .
In the particular case of h 1 ω = i i and h 2 ω = i j , the binaural output SNR is equal to the input SNR, in which i i and i j are, respectively, the ith and jth columns of I 2 M (i.e., the 2 M × 2 M identity matrix). According to (14) and (15), the binaural SNR gain can be expressed as
G h 1 ω , h 2 ω = oSNR h 1 ω , h 2 ω iSNR ω = i = 1 2 h i H ω d ω , 0 2 i = 1 2 h i H ω Γ v ω h i ω .
From the above definition, the following two measures that are very helpful for binaural fixed beamforming can be deduced:
  • the binaural white noise gain (WNG):
    W h 1 ω , h 2 ω = i = 1 2 h i H ω d ω , 0 2 i = 1 2 h i H ω h i ω
  • and the binaural DF:
    D h 1 ω , h 2 ω = i = 1 2 h i H ω d ω , 0 2 i = 1 2 h i H ω Γ d ω h i ω ,
    where Γ d ω is defined in the previous section.
The beampattern is another fundamental performance measure for fixed beamformers. The binaural beampattern can be defined as
B h 1 ω , h 2 ω , θ 2 = 1 2 i = 1 2 h i H ω d ω , θ 2 .
In order to have two meaningful estimates of the desired signal, we are going to extensively exploit the interaural coherence (IC) of the noise. It is well known that, in a multi-source environment, the IC (or its modulus) is important for source localization since it is very strongly related to the two principal binaural cues, i.e., the interaural time difference (ITD) and interaural level difference (ILD), that the brain uses to localize sounds. Psychoacoustically, the localization performance decreases when the IC decreases [33]. Furthermore, the IC affects significantly the perception for acoustic field width.
Let A ω and B ω be two zero-mean complex-valued random variables. The coherence function (CF) between A ω and B ω is defined as
γ A B ω = E A ω B * ω E A ω 2 E B ω 2 ,
where the superscript * is the complex-conjugate operator. It is clear that 0 γ A B ω 2 1 . For any pair of sensors ( i , j ) , the input IC of the noise is simply the CF between V i ω and V j ω , i.e.,
γ V i V j ω = E V i ω V j * ω E V i ω 2 E V j ω 2 = i i T Φ v ω i j i i T Φ v ω i i × i j T Φ v ω i j = i i T Γ v ω i j i i T Γ v ω i i × i j T Γ v ω i j = γ i i ω , i j ω .
For white noise, the input IC is γ w ω = 0 , obviously. For diffuse noise, Γ v ω = Γ d ω , the input IC is Γ d ω i j .
Similarly, we can define the output IC of the noise as the CF between the filtered noises in Z 1 ω and Z 2 ω , i.e.,
γ h 1 ω , h 2 ω = h 1 H ω Φ v ω h 2 ω h 1 H ω Φ v ω h 1 ω × h 2 H ω Φ v ω h 2 ω = h 1 H ω Γ v ω h 2 ω h 1 H ω Γ v ω h 1 ω × h 2 H ω Γ v ω h 2 ω .
In the particular case of h 1 ω = i i and h 2 ω = i j , the input and output ICs are equal, i.e., γ i i ω , i j ω = γ h 1 ω , h 2 ω . It can be checked that the output ICs of white (with the same power) and diffuse noises can be presented as, separately,
γ w h 1 ω , h 2 ω = h 1 H ω h 2 ω h 1 H ω h 1 ω × h 2 H ω h 2 ω
and
γ d h 1 ω , h 2 ω = h 1 H ω Γ d ω h 2 ω h 1 H ω Γ d ω h 1 ω × h 2 H ω Γ d ω h 2 ω .
In many approaches in the literature, the two derived filters h 1 ω and h 2 ω are collinear, i.e.,
h 1 ω = ς ω h 2 ω ,
where ς ω 0 is a complex-valued number. In this case, one can check that γ h 1 ω , h 2 ω = γ w h 1 ω , h 2 ω = γ d h 1 ω , h 2 ω = 1 . Since the desired source signal is also fully coherent at any pair of sensors, both the desired signal and noise are perceived in the same region. As a result, our brain will have difficulties to separate them with binaural presentation and intelligibility will certainly be affected. For better separation between white noise and desired source, we should find orthogonal filters since, in this scenario, the output IC for white noise will be equal to 0, the same way is its corresponding input.

5. Binaural Heterophasic Superdirective Beamformer

In this section, we consider orthogonal binaural filters, i.e., h 1 H ω h 2 ω = 0 , since we want the output IC for white noise to be zero. We also want to maximize γ d h 1 ω , h 2 ω so that not only the signals of interest from a point source at the two binaural outputs are coherent, the diffuse (or any correlated) noise at the binaural outputs will be perceived as less diffuse as possible. For that, we will exploit the maximum modes of this CF.
The symmetric matrix Γ d ω can be diagonalized as [54]
U T ω Γ d ω U ω = Λ ω ,
where
U ω = u 1 ω u 2 ω u 2 M ω
is an orthogonal matrix (note that each eigenvector is normalized according to the sign of its first value in our application because each eigenvector of Γ d ω may have two opposite directions), i.e.,
U T ω U ω = U ω U T ω = I 2 M
and
Λ ω = diag λ 1 ω , λ 2 ω , , λ 2 M ω
is a diagonal matrix. The orthonormal vectors u 1 ω , u 2 ω , , u 2 M ω are the eigenvectors corresponding, respectively, to the eigenvalues λ 1 ω , λ 2 ω , , λ 2 M ω of the matrix Γ d ω , where λ 1 ω λ 2 ω λ 2 M ω > 0 .
It can be shown that the two orthogonal filters that maximize (24) are [55]
h 1 ω = u 1 ω + u 2 M ω 2 = q + , 1 ω h 2 ω = u 1 ω u 2 M ω 2 = q , 1 ω .
In this case, we get the first maximum mode of the CF:
γ d q + , 1 ω , q , 1 ω = λ , 1 ω ,
with corresponding vectors q + , 1 ω and q , 1 ω , where
λ , 1 ω = λ 1 ω λ 2 M ω λ 1 ω + λ 2 M ω = λ , 1 ω λ + , 1 ω .
Similarly, we find that all the M maximum modes of the CF are
γ d q + , m ω , q , m ω = λ , m ω ,
for m = 1 , 2 , , M , where
λ , m ω = λ m ω λ 2 M m + 1 ω λ m ω + λ 2 M m + 1 ω = λ , m ω λ + , m ω
and
q + , m ω = u m ω + u 2 M m + 1 ω 2 q , m ω = u m ω u 2 M m + 1 ω 2 .
It can be verified that
λ , 1 ω λ , 2 ω λ , M ω .
From (35), the two semi-orthogonal matrices ( 2 M × M ) can be written as
Q + ω = q + , 1 ω q + , 2 ω q + , M ω ,
Q ω = q , 1 ω q , 2 ω q , M ω ,
where
Q + T ω Q + ω = Q T ω Q ω = I M ,
Q + T ω Q ω = Q T ω Q + ω = 0 ,
with I M being the M × M identity matrix. It can be shown that
Q + T ω Γ d ω Q ω = Q T ω Γ d ω Q + ω = Λ ω ,
Q + T ω Γ d ω Q + ω = Q T ω Γ d ω Q ω = Λ + ω ,
where
Λ ω = 1 2 diag λ , 1 ω , λ , 2 ω , , λ , M ω ,
Λ + ω = 1 2 diag λ + , 1 ω , λ + , 2 ω , , λ + , M ω
are two diagonal matrices of size M × M .
Let N be a positive integer number with 2 N M (a different value of N gives a different degree of tradeoff between WNG and DF). We define the two semi-orthogonal matrices (of size 2 M × N ):
Q + , : N ω = q + , 1 ω q + , 2 ω q + , N ω ,
Q , : N ω = q , 1 ω q , 2 ω q , N ω .
In the rest, we consider orthogonal filters of the forms:
h 1 ω = Q + , : N ω h ¯ : N ω h 2 ω = Q , : N ω h ¯ : N ω ,
where
h ¯ : N ω = H ¯ 1 ω H ¯ 2 ω H ¯ N ω 0
is a complex-valued filter of length N. For this class of orthogonal filters, the output IC for diffuse noise is
γ d h 1 ω , h 2 ω = h ¯ : N H ω Λ , N ω h ¯ : N ω h ¯ : N H ω Λ + , N ω h ¯ : N ω = γ d h ¯ : N ω ,
where
Λ , N ω = 1 2 diag λ , 1 ω , λ , 2 ω , , λ , N ω ,
Λ + , N ω = 1 2 diag λ + , 1 ω , λ + , 2 ω , , λ + , N ω .
It can be shown that
1 γ h ¯ : 1 ω γ h ¯ : 2 ω γ h ¯ : M ω 0 .
With (47), the binaural WNG, DF, and power beampattern can be expressed as, respectively,
W h ¯ : N ω = h ¯ : N H ω C ω , 0 C H ω , 0 h ¯ : N ω 2 h ¯ : N H ω h ¯ : N ω ,
D h ¯ : N ω = h ¯ : N H ω C ω , 0 C H ω , 0 h ¯ : N ω 2 h ¯ : N H ω Λ + , N ω h ¯ : N ω ,
and
B h ¯ : N ω , θ 2 = h ¯ : N H ω C ω , θ C H ω , θ h ¯ : N ω 2 ,
where
C ω , θ = Q + , : N T ω d ω , θ Q , : N T ω d ω , θ
is a matrix of size N × 2 with the distortionless constraint being
C H ω , 0 h ¯ : N ω = 1 = 1 1 .
To fulfill this constraint, we must take N 2 .
Substituting (47) into (12) and using the distortionless constraint, the variance of Z i ω becomes
ϕ Z i ω = ϕ X ω + ϕ V 1 ω × h ¯ : N H ω Q ± , : N T ω Γ v ω Q ± , : N ω h ¯ : N ω ,
where Q ± , : N ω = Q + , : N ω for ϕ Z 1 ω and Q ± , : N ω = Q , : N ω for ϕ Z 2 ω . In the case of diffuse-plus-white noise, i.e., Γ v ω = Γ d ω + α I 2 M , where α is a parameter that determines the relative level between the diffuse and white noises, (58) simplifies to
ϕ Z i ω = ϕ X ω + ϕ V 1 ω × h ¯ : N H ω Λ + , N ω h ¯ : N ω + α h ¯ : N H ω h ¯ : N ω ,
showing that ϕ Z 1 ω = ϕ Z 2 ω . Again, using (57), we find that the cross-correlation between Z 1 ω and Z 2 ω is
ϕ Z 1 Z 2 ω = E Z 1 ω Z 2 * ω = ϕ X ω + ϕ V 1 ω h ¯ : N H ω Q + , : N T ω Γ v ω Q , : N ω h ¯ : N ω ,
whose form for diffuse-plus-white noise is
ϕ Z 1 Z 2 ω = ϕ X ω + ϕ V 1 ω h ¯ : N H ω Λ , N ω h ¯ : N ω ,
which, as expected, does not depend on the white noise. For Γ v ω = Γ d ω + α I 2 M , the output IC of the estimated signals is
γ Z 1 Z 2 ω = ϕ Z 1 Z 2 ω ϕ Z 1 ω ϕ Z 2 ω = iSNR ω + h ¯ : N H ω Λ , N ω h ¯ : N ω iSNR ω + h ¯ : N H ω Λ + , N ω h ¯ : N ω + α h ¯ : N H ω h ¯ : N ω .
We deduce that for large input SNRs, the localization cues of the estimated signals depend mostly on the ones of the desired signal, while for low SNRs, they depend mostly on the ones of the diffuse-plus-white noise.
One possible binaural superdirective beamformer can be obtained by minimizing the sum of the filtered diffuse noise signals subject to the distortionless constraint in (57), i.e.,
min h ¯ : N ω 2 h ¯ : N H ω Λ + , N ω h ¯ : N ω s . t . h ¯ : N H ω C ω , 0 = 1 T .
We easily get
h ¯ : N , BSD ω = Λ + , N 1 ω C ω , 0 C H ω , 0 Λ + , N 1 ω C ω , 0 1 1
and the corresponding binaural DF is
D h ¯ : N , BSD ω = 1 1 T C H ω , 0 Λ + , N 1 ω C ω , 0 1 1 .
Therefore, the proposed binaural superdirective beamformer is
h 1 , BSD ω = Q + , : N ω h ¯ : N , BSD ω h 2 , BSD ω = Q , : N ω h ¯ : N , BSD ω .
Note that another form of the binaural superdirective beamformer can be derived by maximizing the binaural DF in (18) subject to the distortionless constraints, which will be left to the reader’s investigation to make the paper concise.

6. Experiments and Analysis

In this section, we study the performance of the developed binaural heterophasic superdirective beamforming method and compare it to the monaural superdirective beamformer through experiments. For fair comparison, the orders of the binaural heterophasic and the monaural superdirective beamformers are set to the same number so the DF of the two beamformers would be similar.

6.1. Performance Analysis

We first evaluate the beampattern (given in (55)) of the binaural superdirective beamformer. A ULA is used with an interelement spacing equal to 1 cm. The beampatterns of the derived beamformer are plotted in Figure 3, where M = 6 , 8 , 10 , 12 , and N = M / 2 , at f = 1 kHz. Note that, given a ULA with M microphone sensors, one can design binaural superdirective beamformers of order from 1 to M / 2 . When the order increases, the DF becomes larger but the WNG becomes smaller. In this work, we only show the case with N = M / 2 in the simulation for maximum DF. Two-dimensional (2D) plots of the corresponding beampatterns are shown in Figure 4; one can observe that, in all cases, they are almost frequency invariant.
Next, We study the performance of the binaural heterophasic superdirective beamformer in terms of WNG and DF, according to (53) and (54), respectively. The results for the WNG and DF are plotted in Figure 5, where the first-, second-, third-, and fourth-order binaural superdirective beamformers are designed with M = 6 , 8 , 10 , 12 , respectively (this is the basic requirement for the design of the binaural heterophasic superdirective beamformer as shown in Section 5). One can see that the WNGs of binaural superdirective beamformer decrease with the order while the DFs of beamformers increase with the order. Besides, for each order, the DF does not change much with frequency, which is an important property for processing broadband signals like speech. Figure 6 plots the WNGs and DFs of the binaural superdirective beamformers versus parameter N, where 2 N M / 2 . This parameter is introduced to gain flexibility for achieving compromise between WNG and DF. As seen, the DF increases while the WNG decreases with the increase of the value of N. In practice, one can tune the parameter N according to the application requirement.
Next, we study the ICs of the binaural heterophasic and conventional superdirective beamformers under the same conditions according to (23) and (24). Figure 7 plots the output ICs of both beamformers as a function of frequency in white and diffuse noises, respectively. As seen, in the diffuse noise case, the ICs of both beamformers are equal to one within the studied frequency range; in the white noise case, the IC of the binaural superdirective beamformer is equal to zero while for the conventional superdirective beamformer it is equal to one. This means that in the two output signals of the binaural superdirective beamformer, the speech signal is completely coherent, while the white noise is completely incoherent; so, the output signals correspond to the heterophasic case as discussed in Section 1, in which the speech and white noise can be regarded as two separate direction sources in space.
Figure 8 plots the IC magnitude of the outputs of the binaural heterophasic superdirective beamformer, which is given in (62), versus frequency in different input SNR conditions. One can see that this IC increases with frequency. This is due to the fact that white noise amplification mainly happens at low frequencies. The output IC in white noise is zero, causing the low-frequency output IC of the entire signal to approach zero. This shows the impact of white noise amplification from the perspective of the output IC. For a fixed frequency, it is seen that the output IC increases with the input SNR, and it approaches one at a high input SNR. This can be easily explained: as the input SNR increases, the desired signal component dominates the beamforming output. The output IC of the desired signal is one, so the output IC of the two output signals in this condition also approaches one. Consequently, one can conclude that, for high input SNRs, the localization cues of the estimated signals depend mostly on the ones of the desired signal, while at low SNRs, they depend mostly on the ones of the noise.

6.2. Experiments in Real Environments

In this subsection, we evaluate the performance of the proposed binaural heterophasic superdirective beamformer in real acoustic environments. The experiments were conducted in a 10.5 × 6 × 2.8 m conference room. A ULA is used, which consists of 8 microphones, where the elements spacing is δ = 1.1 cm. The SNR of the microphones is 60 dB(A). A photo of the designed array and the experimental setup are shown in Figure 9. To make the experiments repeatable, we first used the microphone array to record sound signals from a loudspeaker located in the ULA’s endfire direction. Then, both the conventional and binaural superdirective beamformers were implemented to get the outputs.
Figure 10 plots the time-domain signals and their spectrograms of the output signals of the conventional and binaural superdirective beamformers. It is clearly seen that the outputs of both beamformers suffer from serious white noise amplification, where the desired signal is almost covered by the white noise. It is also seen from spectrograms that white noise amplification mainly occurs at low frequencies.
As emphasized previously, the main advantage of the proposed binaural superdirective beamforming method is to have the human binaural auditory system to better separate the signal of interest from white noise after beamforming. To confirm this, we performed some subjective listening experiments. Firstly, we obtained a series of output signals of the implemented conventional and binaural superdirective beamformers. Specifically, we extracted seven audio clips from the “Voice of America” with each of length of 20 s. After playing and recording through the loudspeaker and the microphone array shown in Figure 9, we use the two kind of beamformers to perform superdirective beamforming to obtain the output signals. Then, five subjects were asked to listen to the output signals and draw up the zones of the sound source and white noise on the horizontal plane within a predesigned circle. Finally, we extracted the images sketched by each subject and averaged them to get the experimental results (note that in these experiments, we selected five subjects from the CIAIC–Center of Intelligent Acoustics and Immersive Communications, who are experienced in acoustic analysis and can clearly distinguish auditory events). Here we only provide four different zones, i.e., front, back, left, and right sides, to ask the subjects to choose, where an illustration of the test is shown in Figure 11.
Figure 12 presents the auditory maps averaged from the five subjects from the horizontal plane. As can be seen, for the conventional superdirective beamformer, all signals (desired signal plus white noise) are perceived to be in the middle of the head, which corresponds to the homophasic case. Oppositely, for the binaural superdirective beamforming, the signal of interest is perceived to be in the median plane of the head while white noise is located in each side of the ear, which corresponds to the heterophasic case. As discussed previously and summarized in Table 1, the speech intelligibility in heterophasic case is higher than the homophasic case (approximately 4 dB higher [49]). Consequently, the proposed binaural superdirective beamformer has better intelligibility than the conventional one.

7. Conclusions

In this paper, we addressed the problem of superdirective beamforming with small-spacing microphone arrays. While it can achieve the maximum spatial gain to suppress acoustic noise, the traditional superdirective beamformer suffers from white noise amplification, which is particularly serious at low frequencies. Many methods were developed in the literature to deal with this problem, but they all pay a price of sacrificing the DF and the resulting beamformers may no longer be superdirective. Motivated by studies and facts in psychoacoustics, we developed in this paper a binaural heterophasic superdirective beamformer, which consists of two sub-beamforming filters, each for one of the binaural channels. These two sub-beamformers are constrained to be orthogonal to each other to minimize the IC of the white noise components in the binaural outputs while maximize the IC of the diffuse noise components. As a result, the signal of interest in the binaural superdirective beamformer’s outputs is in phase while the white noise is random in phase, so that the human auditory system is able to more easily separate the acoustic signal of interest from white noise by listening to the outputs of the proposed beamformer. Simulations and experiments were carried out to evaluate the performance of the proposed binaural superdirective beamformer. The results corroborate with the theoretical analysis and confirm that the binaural superdirective beamforming corresponds to the heterophasic case studied in psychoacoustics. Based on the listening tests shown in the psychoacoustic study, one can conclude that the improvement in intelligibility is expected to be 4 dB at low SNR conditions in accordance with the psychoacoustic experiments verified in the literature.

Author Contributions

Conceptualization, J.C. and J.B.; methodology, J.C. and J.B.; software, Y.W. and J.J.; validation, Y.W., J.J. and G.H.; formal analysis, Y.W., J.J. and G.H.; investigation, Y.W., J.J. and G.H.; writing—original draft preparation, J.C., J.B. and Y.W.; writing—review and editing, G.H., J.C. and J.B.; visualization, Y.W. and J.J.; supervision, J.C. and J.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Benesty, J.; Chen, J.; Pan, C. Fundamentals of Differential Beamforming; Springer: Berlin, Germany, 2016. [Google Scholar]
  2. Benesty, J.; Chen, J.; Huang, Y. Microphone Array Signal Processing; Springer: Berlin, Germany, 2008. [Google Scholar]
  3. Elko, G.W. Superdirectional microphone arrays. In Acoustic Signal Processing for Telecommunication; Springer: Berlin, Germany, 2000; pp. 181–237. [Google Scholar]
  4. Brandstein, M.; Ward, D. Microphone Arrays: Signal Processing Techniques and Applications; Springer: Berlin, Germany, 2001. [Google Scholar]
  5. Johnson, D.H.; Dudgeon, D.E. Array Signal Processing: Concepts and Techniques; PTR Prentice Hall: Englewood Cliffs, NJ, USA, 1993. [Google Scholar]
  6. Huang, G.; Benesty, J.; Chen, J. On the Design of Frequency-Invariant Beampatterns with Uniform Circular Microphone Arrays. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 1140–1153. [Google Scholar] [CrossRef]
  7. Huang, G.; Chen, J.; Benesty, J. Insights into Frequency-Invariant Beamforming with Concentric Circular Microphone Arrays. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 2305–2318. [Google Scholar] [CrossRef]
  8. Benesty, J.; Chen, J.; Cohen, I. Design of Circular Differential Microphone Arrays; Springer: Berlin, Germany, 2015. [Google Scholar]
  9. Cox, H.; Zeskind, R.M.; Kooij, T. Practical supergain. IEEE Trans. Acoust. Speech Signal Process. 1986, 34, 393–398. [Google Scholar] [CrossRef]
  10. Huang, G.; Benesty, J.; Cohen, I.; Chen, J. A simple theory and new method of differential beamforming with uniform linear microphone arrays. IEEE/ACM Trans. Audio Speech Lang. Process. 2020, 28, 1079–1093. [Google Scholar] [CrossRef]
  11. Li, C.; Benesty, J.; Huang, G.; Chen, J. Subspace superdirective beamformers based on joint diagonalization. In Proceedings of the IEEE ICASSP, Shanghai, China, 20–25 March 2016; pp. 400–404. [Google Scholar]
  12. Pan, C.; Chen, J.; Benesty, J. Design of robust differential microphone arrays with orthogonal polynomials. J. Acoust. Soc. Am. 2015, 138, 1079–1089. [Google Scholar] [CrossRef]
  13. Huang, G.; Chen, J.; Benesty, J. Design of planar differential microphone arrays with fractional orders. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 28, 116–130. [Google Scholar] [CrossRef]
  14. Huang, G.; Chen, J.; Benesty, J. A flexible high directivity beamformer with spherical microphone arrays. J. Acoust. Soc. Am. 2018, 143, 3024–3035. [Google Scholar] [CrossRef]
  15. Jin, J.; Chen, J.; Benesty, J.; Wang, Y.; Huang, G. Heterophasic Binaural Differential Beamforming for Speech Intelligibility Improvement. IEEE Trans. Veh. Technol. 2020, 69, 13497–13509. [Google Scholar] [CrossRef]
  16. Elko, G.W.; Meyer, J. Microphone arrays. In Springer Handbook of Speech Processing; Springer: Berlin, Germany, 2008; pp. 1021–1041. [Google Scholar]
  17. Lotter, T.; Vary, P. Dual-channel speech enhancement by superdirective beamforming. EURASIP J. Appl. Signal Process. 2006, 1, 063297. [Google Scholar] [CrossRef] [Green Version]
  18. Huang, G.; Benesty, J.; Chen, J. Design of robust concentric circular differential microphone arrays. J. Acoust. Soc. Am. 2017, 141, 3236–3249. [Google Scholar] [CrossRef] [Green Version]
  19. Benesty, J.; Chen, J. Study and Design of Differential Microphone Arrays; Springer: Berlin, Germany, 2012. [Google Scholar]
  20. Mabande, E.; Schad, A.; Kellermann, W. Design of robust superdirective beamformers as a convex optimization problem. In Proceedings of the IEEE ICASSP, Taipei, Taiwan, 19–24 April 2009; pp. 77–80. [Google Scholar]
  21. Berkun, R.; Cohen, I.; Benesty, J. Combined beamformers for robust broadband regularized superdirective beamforming. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 877–886. [Google Scholar] [CrossRef]
  22. Crocco, M.; Trucco, A. Design of robust superdirective arrays with a tunable tradeoff between directivity and frequency-invariance. IEEE Trans. Signal Process. 2011, 59, 2169–2181. [Google Scholar] [CrossRef]
  23. Huang, G.; Benesty, J.; Chen, J. Superdirective Beamforming Based on the Krylov Matrix. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 2531–2543. [Google Scholar] [CrossRef]
  24. Pan, C.; Chen, J.; Benesty, J. Reduced-order robust superdirective beamforming with uniform linear microphone arrays. IEEE/ACM Trans. Audio Speech Lang. Process. 2016, 24, 1548–1559. [Google Scholar] [CrossRef]
  25. Crocco, M.; Trucco, A. Stochastic and analytic optimization of sparse aperiodic arrays and broadband beamformers with robust superdirective patterns. IEEE Trans. Audio, Speech, Lang. Process. 2012, 20, 2433–2447. [Google Scholar] [CrossRef]
  26. Berkun, R.; Cohen, I.; Benesty, J. A tunable beamformer for robust superdirective beamforming. In Proceedings of the IEEE IWAENC, Xi’an, China, 13–16 September 2016; pp. 1–5. [Google Scholar]
  27. Blauert, J. Spatial Hearing: The Psychophysics of Human Sound Localization; MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
  28. Blauert, J.; Lindemann, W. Spatial mapping of intracranial auditory events for various degrees of interaural coherence. J. Acoust. Soc. Am. 1986, 79, 806–813. [Google Scholar] [CrossRef] [PubMed]
  29. Hirsh, I.J. The Influence of Interaural Phase on Interaural Summation and Inhibition. J. Acoust. Soc. Am. 1948, 20, 536–544. [Google Scholar] [CrossRef]
  30. Hirsh, I.J. The relation between localization and intelligibility. J. Acoust. Soc. Am. 1950, 22, 196–200. [Google Scholar] [CrossRef]
  31. Jeffress, L.A.; Blodgett, H.C.; Deatherage, B.H. Effect of interaural correlation on the precision of centering a noise. J. Acoust. Soc. Am. 1962, 34, 1122–1123. [Google Scholar] [CrossRef]
  32. Kollmeier, B.; Brand, T.; Meyer, B. Perception of speech and sound. In Springer Handbook of Speech Processing; Springer: Berlin, Germany, 2008; pp. 61–82. [Google Scholar]
  33. Zimmer, U.; Macaluso, E. High binaural coherence determines successful sound localization and increased activity in posterior auditory areas. Neuron 2005, 47, 893–905. [Google Scholar] [CrossRef] [Green Version]
  34. Miller, G.A.; Licklider, J.C. The intelligibility of interrupted speech. J. Acoust. Soc. Am. 1950, 22, 167–173. [Google Scholar] [CrossRef]
  35. Blauert, J. Sound localization in the median plane. Acta Acust. United Acust. 1969, 22, 205–213. [Google Scholar]
  36. Jeffress, L.A. A place theory of sound localization. J. Comp. Physiol. Psychol. 1948, 41, 35. [Google Scholar] [CrossRef] [PubMed]
  37. Sandel, T.; Teas, D.; Feddersen, W.; Jeffress, L. Localization of sound from single and paired sources. J. Acoust. Soc. Am. 1955, 27, 842–852. [Google Scholar] [CrossRef]
  38. Hirsh, I.J.; Pollack, I. The role of interaural phase in loudness. J. Acoust. Soc. Am. 1948, 20, 761–766. [Google Scholar] [CrossRef]
  39. Kock, W. Binaural localization and masking. J. Acoust. Soc. Am. 1950, 22, 801–804. [Google Scholar] [CrossRef]
  40. Jeffress, L.A.; Blodgett, H.C.; Sandel, T.T.; Wood, C.L., III. Masking of tonal signals. J. Acoust. Soc. Am. 1956, 28, 416–426. [Google Scholar] [CrossRef]
  41. Jeffress, L.A.; Robinson, D.E. Formulas for the coefficient of interaural correlation for noise. J. Acoust. Soc. Am. 1962, 34, 1658–1659. [Google Scholar] [CrossRef]
  42. Licklider, J.C.R.; Miller, G.A. The Perception of Speech. In Handbook of Experimental Psychology; Stevens, S.S., Ed.; Wiley: Hoboken, NJ, USA, 1951; pp. 1040–1074. [Google Scholar]
  43. Beranek, L.L. Acoustics; Acoustic Society of America: Woodbury, NY, USA, 1986. [Google Scholar]
  44. Gerald, K.; Colburn, H.S. Informational Masking in Speech Recognition; Springer: Cham, Switzerland, 2017. [Google Scholar]
  45. Zobel, B.H.; Wagner, A.; Sanders, L.D.; Ba¸skent, D. Spatial release from informational masking declines with age: Evidence from a detection task in a virtual separation paradigm. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 146, 548–566. [Google Scholar] [CrossRef] [Green Version]
  46. Moore, B.C. Effects of hearing loss and age on the binaural processing of temporal envelope and temporal fine structure information. Hear. Res. 2020, 107991. [Google Scholar] [CrossRef]
  47. Esther, S.; van de Par, S. The role of reliable interaural time difference cues in ambiguous binaural signals for the intelligibility of multitalker speech. J. Acoust. Soc. Am. 2020, 147, 4041–4054. [Google Scholar]
  48. Jutras, B.; Lagacé, J.; Koravand, A. The development of auditory functions. In Handbook of Clinical Neurology; Elsevier: Amsterdam, The Netherlands, 2020; pp. 143–155. [Google Scholar]
  49. Licklider, J.C.R. The Influence of interaural phase relations upon the masking of speech by white noise. J. Acoust. Soc. Am. 1948, 20, 150–159. [Google Scholar] [CrossRef]
  50. Moore, B.C. An Introduction to the Psychology of Hearing; Brill: Leiden, The Netherlands, 2012. [Google Scholar]
  51. Marquardt, D.; Hohmann, V.; Doclo, S. Interaural coherence preservation in multi-channel Wiener filtering-based noise reduction for binaural hearing aids. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 2162–2176. [Google Scholar] [CrossRef]
  52. Hadad, E.; Marquardt, D.; Doclo, S. Theoretical analysis of binaural transfer function MVDR beamformers with interference cue preservation constraints. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 2449–2464. [Google Scholar] [CrossRef]
  53. Cox, H.; Zeskind, R.; Owen, M. Robust adaptive beamforming. IEEE Trans. Acoust. Speech Signal Process. 1987, 35, 1365–1376. [Google Scholar] [CrossRef] [Green Version]
  54. Golub, G.H.; Loan, C.F.V. Matrix Computations, 3rd ed.; The Johns Hopkins University Press: Baltimore, MD, USA, 1996. [Google Scholar]
  55. Eaton, M.L. A maximization problem and its application to canonical correlation. J. Multivar. Anal. 1976, 6, 422–425. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Illustration of different phase scenarios and the influence of the interaural phase relations on the localization of speech and white noise in space. The circle shape means that the signal is concentrated in a limited area, in front of the head for the case of in phase or on either side of the ears for the case of random phase. The rectangle shape indicates that the signal spreads in the area behind the ears, which is related to the case of out of phase. The red dotted circle indicates that the speech signal and the white noise are both in phase, which is related to the monaural superdirective beamformer. The blue circle indicates that the speech signal is in phase while the white noise is in random phase, which is related to the binaural superdirective beamformer developed in this work. This figure is a modified version of the results in [49].
Figure 1. Illustration of different phase scenarios and the influence of the interaural phase relations on the localization of speech and white noise in space. The circle shape means that the signal is concentrated in a limited area, in front of the head for the case of in phase or on either side of the ears for the case of random phase. The rectangle shape indicates that the signal spreads in the area behind the ears, which is related to the case of out of phase. The red dotted circle indicates that the speech signal and the white noise are both in phase, which is related to the monaural superdirective beamformer. The blue circle indicates that the speech signal is in phase while the white noise is in random phase, which is related to the binaural superdirective beamformer developed in this work. This figure is a modified version of the results in [49].
Sensors 21 00074 g001
Figure 2. Illustration of the proposed binaural superdirective beamformer, which suppresses directional acoustic interference and noise and meanwhile separates the desired signal and the white noise into different zones.
Figure 2. Illustration of the proposed binaural superdirective beamformer, which suppresses directional acoustic interference and noise and meanwhile separates the desired signal and the white noise into different zones.
Sensors 21 00074 g002
Figure 3. Beampatterns of the binaural heterophasic superdirective beamformer with various numbers of microphones: (a) M = 6 , (b) M = 8 , (c) M = 10 , and (d) M = 12 . Conditions of simulation: f = 1 kHz, δ = 1 cm, and N = M / 2 .
Figure 3. Beampatterns of the binaural heterophasic superdirective beamformer with various numbers of microphones: (a) M = 6 , (b) M = 8 , (c) M = 10 , and (d) M = 12 . Conditions of simulation: f = 1 kHz, δ = 1 cm, and N = M / 2 .
Sensors 21 00074 g003
Figure 4. Beampatterns of the binaural heterophasic superdirective beamformer as a function of the frequency with different numbers of microphones: (a) M = 6 , (b) M = 8 , (c) M = 10 , and (d) M = 12 . Conditions of simulation: δ = 1.0 cm and N = M / 2 .
Figure 4. Beampatterns of the binaural heterophasic superdirective beamformer as a function of the frequency with different numbers of microphones: (a) M = 6 , (b) M = 8 , (c) M = 10 , and (d) M = 12 . Conditions of simulation: δ = 1.0 cm and N = M / 2 .
Sensors 21 00074 g004
Figure 5. WNGs and DFs of the 1st-, 2nd-, 3rd-, and 4th-order binaural heterophasic superdirective beamformers versus frequency: (a) WNGs and (b) DFs. The sensor spacing: δ = 1 cm.
Figure 5. WNGs and DFs of the 1st-, 2nd-, 3rd-, and 4th-order binaural heterophasic superdirective beamformers versus frequency: (a) WNGs and (b) DFs. The sensor spacing: δ = 1 cm.
Sensors 21 00074 g005
Figure 6. WNGs and DFs of the 1st-, 2nd-, 3rd-, and 4th-order binaural heterophasic superdirective beamformers versus parameter N: (a) WNGs and (b) DFs. Conditions of simulation: f = 1 kHz and δ = 1 cm.
Figure 6. WNGs and DFs of the 1st-, 2nd-, 3rd-, and 4th-order binaural heterophasic superdirective beamformers versus parameter N: (a) WNGs and (b) DFs. Conditions of simulation: f = 1 kHz and δ = 1 cm.
Sensors 21 00074 g006
Figure 7. Magnitude of the output ICs: (a) diffuse noise and (b) white noise. Conditions of simulation: M = 8 for the binaural superdirective beamformer, M = 3 for the conventional superdirective beamformer, and δ = 1 cm. Note that the output IC of the binaural superdirective beamformer in diffuse noise is frequency-dependent; it approaches 1 under 4 kHz but decreases as the frequency increases.
Figure 7. Magnitude of the output ICs: (a) diffuse noise and (b) white noise. Conditions of simulation: M = 8 for the binaural superdirective beamformer, M = 3 for the conventional superdirective beamformer, and δ = 1 cm. Note that the output IC of the binaural superdirective beamformer in diffuse noise is frequency-dependent; it approaches 1 under 4 kHz but decreases as the frequency increases.
Sensors 21 00074 g007
Figure 8. Output IC of the estimated signals (see (62)) as a function of frequency and input SNR. Conditions of simulation: M = 8 , N = 4 , α = 1 , and δ = 1 cm.
Figure 8. Output IC of the estimated signals (see (62)) as a function of frequency and input SNR. Conditions of simulation: M = 8 , N = 4 , α = 1 , and δ = 1 cm.
Sensors 21 00074 g008
Figure 9. A photo of the designed array and the experimental setup for evaluating the binaural and conventional superdirective beamformers: (a) photo of the designed eight-microphone array, (b) a close view photo of the experimental setup, and (c) a wide angle photo of the experimental setup.
Figure 9. A photo of the designed array and the experimental setup for evaluating the binaural and conventional superdirective beamformers: (a) photo of the designed eight-microphone array, (b) a close view photo of the experimental setup, and (c) a wide angle photo of the experimental setup.
Sensors 21 00074 g009
Figure 10. Monaural and binaural superdirective beamformers in a conference room: (a) output of the 2nd-order monaural superdirective beamformer and its spectrogram with M = 3 and δ = 1.1 cm, and (b) output of the 2nd-order binaural superdirective beamformer and its spectrogram with M = 8 and δ = 1.1 cm.
Figure 10. Monaural and binaural superdirective beamformers in a conference room: (a) output of the 2nd-order monaural superdirective beamformer and its spectrogram with M = 3 and δ = 1.1 cm, and (b) output of the 2nd-order binaural superdirective beamformer and its spectrogram with M = 8 and δ = 1.1 cm.
Sensors 21 00074 g010
Figure 11. An illustration of the auditory map for subjects (horizontal-plane). During the test, the subjects were asked to mark the areas according to the sound source location they heard through headphones.
Figure 11. An illustration of the auditory map for subjects (horizontal-plane). During the test, the subjects were asked to mark the areas according to the sound source location they heard through headphones.
Sensors 21 00074 g011
Figure 12. The average auditory map marked by the listening subjects: (a) monaural superdirective beamformer and (b) binaural superdirective beamformer. The blue waves refer to the region in which the desired speech is heard, and the region filled with disorderly dots refer to the region in which the white noise is heard. Conditions of experiment: M = 8 , N = 4 , and δ = 1.1 cm.
Figure 12. The average auditory map marked by the listening subjects: (a) monaural superdirective beamformer and (b) binaural superdirective beamformer. The blue waves refer to the region in which the desired speech is heard, and the region filled with disorderly dots refer to the region in which the white noise is heard. Conditions of experiment: M = 8 , N = 4 , and δ = 1.1 cm.
Sensors 21 00074 g012
Table 1. Different scenarios for intelligibility study based on phase relationship between speech and noise [49].
Table 1. Different scenarios for intelligibility study based on phase relationship between speech and noise [49].
ScenarioSpeechNoiseClass
1Out of phaseIn phaseAntiphasic
2In phaseOut of phaseAntiphasic
3In phaseRandom phaseHeterophasic
4Out of phaseRandom phaseHeterophasic
5In phaseIn phaseHomophasic
6Out of phaseOut of phaseHomophasic
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Y.; Chen, J.; Benesty, J.; Jin, J.; Huang, G. Binaural Heterophasic Superdirective Beamforming. Sensors 2021, 21, 74. https://doi.org/10.3390/s21010074

AMA Style

Wang Y, Chen J, Benesty J, Jin J, Huang G. Binaural Heterophasic Superdirective Beamforming. Sensors. 2021; 21(1):74. https://doi.org/10.3390/s21010074

Chicago/Turabian Style

Wang, Yuzhu, Jingdong Chen, Jacob Benesty, Jilu Jin, and Gongping Huang. 2021. "Binaural Heterophasic Superdirective Beamforming" Sensors 21, no. 1: 74. https://doi.org/10.3390/s21010074

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop