[go: up one dir, main page]

0% found this document useful (0 votes)
22 views2 pages

Random Data Analysis and Measurement Procedures 280 281

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
22 views2 pages

Random Data Analysis and Measurement Procedures 280 281

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 2

PROBABILITY DENSITY FUNCTION ESTIMATES 265

Equation (8.85),

V
"PW1«^ (· )
8 86

The constant c is dependent on the autocorrelation function of the data and the
sampling rate. For continuous bandwidth-limited white noise, experimental studies
indicate that c « 0.3. If the bandwidth-limited white noise is digitized to N=2BT
discrete values, the experimental studies indicate that c = 1.0, as would be expected
from the results of Equation (8.85).

8.3.3 Normalized rms Error

The total mean square error of the probability density estimate p(x) is the sum of
the variance defined in Equation (8.86) and the square of the bias defined in
Equation (8.82). That is,

c PJx)
2
W p"{x)
2

(p{x)-p{x)) 2
(8.87)
2BTW 24

Hence, the normalized mean square error is approximated by

W 4
J'{x)
ε \β{χ)
λ
(8.88)
2BTWp{x) 576 [p(x)

The square root gives the normalized rms error.


It is clear from Equation (8.88) that there are conflicting requirements on the
window width Win probability density measurements. On the one hand, a large value
of Wis desirable to reduce the random error. On the other hand, a small value of Wis
needed to suppress the bias error. However, the total error will approach zero as
Τ —> oo if W is restricted so that W —> oo and WT —> oo. In practice, values of
W< 0.2σ will usually limit the normalized bias error to less than 1%. This is true
χ

because of the p"(x) term in the bias portion of the error given by Equation (8.88),
Probability density functions of common (approximately Gaussian) random data do
not display abrupt or sharp peaks, which are indicative of a large second derivative.

8.3.4 Joint Probability Density Function Estimates

Joint probability density function estimates for a pair of sample time history records
x(t) and y(i) from two stationary (ergodic) random processes {x(t)} and (y(i)} may be
defined as follows. Analogous to Equation (8.66), let

P[ ,W ;y,W ]
X x y = ψ (8.89)

estimate the joint probability that x(t) is inside the interval W centered at x, while
x

simultaneously y(t) is inside the interval W centered at y. This is measured by the ratio
v

T y j T, where T represents the amount of time that these two events coincide in time
Xi x v
266 STATISTICAL ERRORS IN BASIC ESTIMATES

T. Clearly, T will usually be a function of both χ and y. This estimated joint


Xi>

probability will approach the true probability P[x, W ; y, W ] as Tapproaches infinity, x y

namely,

Ρ [χ, W ; y, W ] =
x y lim Ρ [χ, W ; y, W ] = x y lim ^ (8.90)

The joint probability density function p(x, y) is now defined by

p*,y)= hm . . . . . . — - = hm ———'-= hm p[x,y) (8.91)


_o
W i WxWy τ-^οο W Wy x

W ^Q y W,->0 W,-.0
Wy^O Wy->0

where

P[x,W ;y,Wy] J^y_


x =

^ W Wy X TW Wyx ^ '

Assume that W and W are sufficiently small that the bias errors are negligible.
x y

Then the mean square error associated with the estimate p(x, y) will be given by the
variance of the estimate. As for first-order probability density estimates, this quantity
is difficult to determine precisely by theoretical arguments alone. However, by using
the same heuristic arguments that produced Equation (8.86), a general form for the
variance can be approximated. Specifically, for the special case where x(t) and y (i) are
both bandwidth-limited white noise with identical bandwidfhs B,

Vai\p(x,y)] « ° ^ 2 p
(8.93)
\yy ,y)\ BTW W 2 x y
v ;

where c is an unknown constant.

8.4 CORRELATION FUNCTION ESTIMATES

Consider now two sample time history records x(t) and y(r) from two stationary
(ergodic) random processes {x{t)} and {y(r)}. The next statistical quantities of interest
are the stationary autocorrelation functions /?»(τ) and R (x) and the cross-correlation yy

function R^x). To simplify the following derivation, the mean values μ and μ will be χ ν

assumed to be zero. For continuous data, x(t) and y(r), which exist only over a time
interval Γ, the sample cross-correlation estimate Λ^,(τ) can be defined by

- * - f 1
x(t)y(t + x)dt 0<x<T
Τ-τ) 0

(8.94)
ι C t + x)dt -T<x<0
—γτ *(*Μ-

You might also like