Professional Documents
Culture Documents
Prony
Prony
net/publication/358044797
CITATIONS READS
0 447
1 author:
Charles L. Byrne
University of Massachusetts Lowell
247 PUBLICATIONS 6,787 CITATIONS
SEE PROFILE
All content following this page was uploaded by Charles L. Byrne on 23 January 2022.
CHARLES L. BYRNE
1. Prony’s Problem
The problem Prony solved in [5] is to determine J and the complex num-
bers aj and γj from finitely many values of the function
J
X
(1.1) f (t) = aj etγj .
j=1
where
.
(2.6) sj = (1, eγj , e2γj , ..., eM γj ).
Therefore the row space of D is spanned by the vectors sj , j = 1, ..., J, and
the rank of D is at most J. Therefore there is a nonzero vector c that is
orthogonal to each of the vectors sj for which aj is not zero.
PRONY 3
3. An Illustration
For the sake of illustration we consider the case of M = 3 and N = 6,
with J ≤ M . Suppose that our data are the real numbers dn , n = 1, ..., 6.
The matrix D is now
d1 d2 d3 d4
D = d2 d3 d4 d5
d3 d4 d5 d6
and the matrix D† D = DT D is
DT D =
d21 + d22 + d23
d1 d2 + d2 d3 + d3 d4 d1 d3 + d2 d4 + d3 d5 d1 d4 + d2 d5 + d3 d6
d2 d1 + d3 d2 + d4 d3
d22 + d23 + d24 d2 d3 + d3 d4 + d4 d5 d2 d4 + d3 d5 + d4 d6
.
d3 d1 + d4 d2 + d5 d3 d3 d2 + d4 d3 + d5 d4 d23 + d24 + d25 d3 d4 + d4 d5 + d5 d6
d4 d1 + d5 d2 + d6 d3 d4 d2 + d5 d3 + d6 d4 d4 d3 + d5 d4 + d6 d5 d24 + d25 + d26
Note that, for each pair of indices u and v, we have (DT D)u,v a sum of
products for which u − v is constant. Note, however, that the sum is not
over all the pairs for which this is the case.
where the Aj are complex random variables. We can then view the entries
dn of the random data vector d as instances of a random variable Xn given
by
J
. X
(4.2) Xn = Aj enγj .
j=1
When we compare the mth row of the matrix S with the mth row of the
matrix D, as given by Equation (2.5), we see that the coefficients am j in
Equation (4.4) and (aj emγj ) in Equation (2.5) play similar roles. This will
help us when we modify Prony’s method to deal with noise.
We denote by Rs the correlation matrix for d in the noise-free case; that
is,
(4.5) (Rs )k,n = E(dk dn ).
1 †
Then Rs is approximated by M S S.
where the zn , n = 1, ..., N , are random variables that are independent with
respect to one another and with respect to the random variables Aj , with
means equal to zero and variances equal to σ 2 . This is often described
as data containing additive white noise. Then the expected value of dk dn
becomes
(5.2) (Rd )k,n = E(dk dn ) = (Rs )k,n + σ 2 δ(n − k),
where δ(n − k) = 0, if n 6= k, and equal to one, when n = k. We then have
Rd = Rs + σ 2 I, where I is the identity matrix.
1 †
When the data contains additive white noise the matrix M S S is a statis-
tical estimate of the matrix Rd . Consequently, the contribution of the noise
is primarily to increase the main diagonal of S † S. To the extent that the
matrix D† D is analogous to the matrix S † S we should expect the effect of
additive white noise to be primarily an increase in the values on the main
diagonal of D† D. This suggests that, instead of looking for a nonzero vector
c such that Dc = 0, we should select c to be an eigenvector of D† D corre-
sponding to the smallest eigenvalue of that matrix. It will then follow that
such a vector c should be (nearly) orthogonal to each of the vectors sj for
which aj 6= 0.
6.2. Prewhiten, then Match. As I discussed in [2], the best linear unbi-
ased estimator (BLUE) for estimating γ, when d is the random data vector
d = γs + z, Q = E(zz† ), and x is the vector of measurements, is
(6.4) γ̂ = 1/s† Q−1 s s† Q−1 x .
6.4. Capon’s Estimator. When N is not large and some of the θj are
close to one another the functions A(θ) and B(θ) may not be able to resolve
these closely spaced components of R [1]. To improve resolution we can turn
to high-resolution methods.
As we saw in [2], the BLUE is based on finding the vector b that minimizes
b† Qb, subject to b† e(θ) = 1. The problem with this approach is that we
cannot determine Q from measurements and only know Q from theoretical
models. Capon [4] suggests finding, for fixed θ, the vector h(θ) = h that
minimizes h† Rh, subject to h† e(θ) = 1. The vector h(θ) is then
1
(6.7) h(θ) = R−1 e(θ).
e(θ) R−1 e(θ)
†
The idea here is that the filter h(θ) suppresses every component of the data
that is not a multiple of e(θ). This includes, but is not limited to, the
background noise. Capon’s estimator is then the function of θ defined by
E(|h(θ)† d|2 ) and given by
. 1
(6.8) C(θ) = .
e(θ) R−1 e(θ)
†
When the fixed θ is not one of the signal parameters C(θ) is relatively
small. This leads to improved resolution, since when θ lies between two
actual signal parameters the value of C(θ) will typically be smaller than its
value at either of the two actual signal parameters. However, as I discussed
in [1], this improved resolution can be lost when the data is perturbed by
phase errors.
References
[1] Byrne, C. (2021) “Noise in high-resolution signal processing.” , posted on Research-
Gate July 20, 2021.
[2] Byrne, C. (2021) “Modified inner products in signal detection.” , posted on Research-
Gate August 6, 2021.
[3] Byrne, C. (2021) “Avoiding prewhitening through dimensionality reduction in array
processing.” , posted on ResearchGate August 9, 2021.
[4] Capon, J. (1969) “High-resolution frequency-wavenumber spectrum analysis.”Proc.
of the IEEE 57, pp. 1408–1418.
[5] Prony, G.R.B. (1795) “Essai expérimental et analytique sur les lois de la dilatabilité
de fluides élastiques et sur celles de la force expansion de la vapeur de l’alcool, à
différentes températures.”Journal de l’Ecole Polytechnique (Paris) 1(2), pp. 24–76.
[6] Therrien, C. (1992) Discrete Random Signals and Statistical Signal Processing. En-
glewood Cliffs, NJ: Prentice-Hall.