Otokorelasyon fonksiyonu tamamen stokastik bir işlemi tarif ediyor mu?


31

Stokastik bir süreç otokorelasyon işleviyle tamamen tanımlanmış mı?

Eğer değilse, hangi ek özelliklere ihtiyaç duyulur?

Yanıtlar:


44

Stokastik bir sürecin tam bir tanımlaması ile ne kastedilmektedir? De, matematiksel olarak, rastgele oluşur topluluğudur { x ( t ) : t T }{X(t):tT} rastgele değişkenlerin, her bir zaman noktası için bir tt bir in göstergesi grubu TT genellikle, TT , tüm gerçek hattı veya pozitif gerçek bir çizgidir, ve tam bir açıklama araçlarının her bir tamsayı için n 1n1 ve n,n zaman anları t 1 , t 2 , ... , t , nTt1,t2,,tnT , biz (birleşik) dağılımları bilmek nn rastgele değişkenler X ( t 1 )X(t1) , X ( t 2 )X(t2) , ... , x ( t , n ),X(tn) . This is anmuazzambilgi miktarı: biz CDF'yi bilmemiz gerekir X ( t )X(t) her seferinde anlık için tt , (iki boyutlu) eklem CDF X ( t 1 )X(t1) ve X ( t 2 )X(t2) zamanın bütün seçimler için instantst 1t1 ve t 2t2 , ve (üç boyutlu) CDFs X ( t 1 )X(t1) , X ( t 2 )X(t2) , ve X, ( t 3 )X(t3) , vb vb vb

Doğal olarak insanlar daha basit açıklamalar ve daha kısıtlayıcı modeller aradılar. Bir sadeleştirme, işlem zaman orijinindeki bir değişiklik için değişmez olduğunda meydana gelir. Bunun anlamı şudur:

  • İşlemdeki tüm rastgele değişkenler aynı CDF'lere sahiptir: F x ( t 1 ) ( x ) = F X ( t 2 ) ( x )FX(t1)(x)=FX(t2)(x) tüm t 1 , t 2 içint1,t2 .
  • Herhangi iki rastgele değişkenler aynı ortak CDF sahip biraz zaman belirli bir miktarda ayrılmış başka bir rastgele değişken çifti ile ayrılmış , aynı zaman miktarı. Örneğin, rastgele değişkenler X ( t 1 )X(t1) ve X ( t 1 + τ )X(t1+τ) ile ayrılır ττ saniye, rasgele değişkenlerdir x ( t 2 )X(t2) ve X ( t 2 + τ )X(t2+τ) ve böylece K x ( t 1 ),X(t1+τ)(x,y)=FX(t2),X(t2+τ)(x,y)FX(t1),X(t1+τ)(x,y)=FX(t2),X(t2+τ)(x,y)
  • Any three random variables X(t1)X(t1), X(t1+τ1)X(t1+τ1), X(t1+τ1+τ2)X(t1+τ1+τ2) spaced τ1τ1 and τ2τ2 apart have the same joint CDF as X(t2)X(t2), X(t2+τ1)X(t2+τ1), X(t2+τ1+τ2)X(t2+τ1+τ2) which as also spaced τ1τ1 and τ2τ2 apart,
  • and so on for all multidimensional CDFs. See, for example, Peter K.'s answer for details of the multidimensional case.

Effectively, the probabilistic descriptions of the random process do not depend on what we choose to call the origin on the time axis: shifting all time instants t1,t2,,tnt1,t2,,tn by some fixed amount ττ to t1+τ,t2+τ,,tn+τt1+τ,t2+τ,,tn+τ gives the same probabilistic description of the random variables. This property is called strict-sense stationarity and a random process that enjoys this property is called a strictly stationary random process or, more simply, s stationary random process.

Note that strict stationarity by itself does not require any particular form of CDF. For example, it does not say that all the variables are Gaussian.

The adjective strictly suggests that is possible to define a looser form of stationarity. If the NthNth-order joint CDF of X(t1),X(t2),,X(tN)X(t1),X(t2),,X(tN) is the same as the NthNth-order joint CDF of X(t1+τ),X(t2+τ),,X(tN+τ)X(t1+τ),X(t2+τ),,X(tN+τ) for all choices of t1,t2,,tNt1,t2,,tN and ττ, then the random process is said to be stationary to order NN and is referred to as a NthNth-order stationary random process. Note that a NthNth-order stationary random process is also stationary to order nn for each positive n<Nn<N. (This is because the nthnth-order joint CDF is the limit of the NthNth-order CDF as NnNn of the arguments approach : a generalization of FX(x)=limyFX,Y(x,y)FX(x)=limyFX,Y(x,y)). A strictly stationary random process then is a random process that is stationary to all orders NN.

If a random process is stationary to (at least) order 11, then all the X(t)X(t)'s have the same distribution and so, assuming the mean exists, E[X(t)]=μE[X(t)]=μ is the same for all tt. Similarly, E[(X(t))2]E[(X(t))2] is the same for all tt, and is referred to as the power of the process. All physical processes have finite power and so it is common to assume that E[(X(t))2]<E[(X(t))2]< in which case, and especially in the older engineering literature, the process is called a second-order process. The choice of name is unfortunate because it invites confusion with second-order stationarity (cf. this answer of mine on stats.SE), and so here we will call a process for which E[(X(t))2]E[(X(t))2] is finite for all tt (whether or not E[(X(t))2]E[(X(t))2] is a constant) as a finite-power process and avoid this confusion. But note again that

a first-order stationary process need not be a finite-power process.

Consider a random process that is stationary to order 22. Now, since the joint distribution of X(t1)X(t1) and X(t1+τ)X(t1+τ) is the same as the joint distribution function of X(t2)X(t2) and X(t2+τ)X(t2+τ), E[X(t1)X(t1+τ)]=E[X(t2)X(t2+τ)]E[X(t1)X(t1+τ)]=E[X(t2)X(t2+τ)] and the value depends only on ττ. These expectations are finite for a finite-power process and their value is called the autocorrelation function of the process: RX(τ)=E[X(t)X(t+τ)]RX(τ)=E[X(t)X(t+τ)] is a function of ττ, the time separation of the random variables X(t)X(t) and X(t+τ)X(t+τ), and does not depend on tt at all. Note also that E[X(t)X(t+τ)]=E[X(t+τ)X(t)]=E[X(t+τ)X(t+ττ)]=RX(τ),

E[X(t)X(t+τ)]=E[X(t+τ)X(t)]=E[X(t+τ)X(t+ττ)]=RX(τ),
and so the autocorrelation function is an even function of its argument.

A finite-power second-order stationary random process has the properties that

  1. Its mean E[X(t)]E[X(t)] is a constant
  2. Its autocorrelation function RX(τ)=E[X(t)X(t+τ)]RX(τ)=E[X(t)X(t+τ)] is a function of ττ, the time separation of the random variables X(t)X(t) and X(t+τ)X(t+τ), and does not depend on tt at all.

The assumption of stationarity simplifies the description of a random process to some extent but, for engineers and statisticians interested in building models from experimental data, estimating all those CDFs is a nontrivial task, particularly when there is only a segment of one sample path (or realization) x(t)x(t) on which measurements can be made. Two measurements that are relatively easy to make (because the engineer already has the necessary instruments on his workbench (or programs in MATLAB/Python/Octave/C++ in his software library) are the DC value 1TT0x(t)dt1TT0x(t)dt of x(t)x(t) and the autocorrelation function Rx(τ)=1TT0x(t)x(t+τ)dtRx(τ)=1TT0x(t)x(t+τ)dt (or its Fourier transform, the power spectrum of x(t)x(t)). Taking these measurements as estimates of the mean and the autocorrelation function of a finite-power process leads to a very useful model that we discuss next.


A finite-power random process is called a wide-sense-stationary (WSS) process (also weakly stationary random process which fortunately also has the same initialism WSS) if it has a constant mean and its autocorrelation function RX(t1,t2)=E[X(t1)X(t2)]RX(t1,t2)=E[X(t1)X(t2)] depends only on the time difference t1t2t1t2 (or t2t1t2t1).

Note that the definition says nothing about the CDFs of the random variables comprising the process; it is entirely a constraint on the first-order and second-order moments of the random variables. Of course, a finite-power second-order stationary (or NthNth-order stationary (for N>2N>2) or strictly stationary) random process is a WSS process, but the converse need not be true.

A WSS process need not be stationary to any order.

Consider, for example, the random process {X(t):X(t)=cos(t+Θ),<t<}{X(t):X(t)=cos(t+Θ),<t<} where ΘΘ takes on four equally likely values 0,π/2,π0,π/2,π and 3π/23π/2. (Do not be scared: the four possible sample paths of this random process are just the four signal waveforms of a QPSK signal). Note that each X(t)X(t) is a discrete random variable that, in general, takes on four equally likely values cos(t),cos(t+π/2)=sin(t),cos(t+π)=cos(t)cos(t),cos(t+π/2)=sin(t),cos(t+π)=cos(t) and cos(t+3π/2)=sin(t)cos(t+3π/2)=sin(t), It is easy to see that in general X(t) and X(s) have different distributions, and so the process is not even first-order stationary. On the other hand, E[X(t)]=14cos(t)+14(sin(t))+14(cos(t))+14sin(t)=0

for every t while E[X(t)X(s)]=14[cos(t)cos(s)+(cos(t))(cos(s))+sin(t)sin(s)+(sin(t))(sin(s))]=12[cos(t)cos(s)+sin(t)sin(s)]=12cos(ts).
In short, the process has zero mean and its autocorrelation function depends only on the time difference ts, and so the process is wide sense stationary. But it is not first-order stationary and so cannot be stationary to higher orders either.

Even for WSS processes that are second-order stationary (or strictly stationary) random processes, little can be said about the specific forms of the distributions of the random variables. In short,

A WSS process is not necessarily stationary (to any order), and the mean and autocorrelation function of a WSS process is not enough to give a complete statistical description of the process.

Finally, suppose that a stochastic process is assumed to be a Gaussian process ("proving" this with any reasonable degree of confidence is not a trivial task). This means that for each t, X(t) is a Gaussian random variable and for all positive integers n2 and choices of n time instants t1, t2, ,tn, the N random variables X(t1), X(t2), ,X(tn) are jointly Gaussian random variables. Now a joint Gaussian density function is completely determined by the means, variances, and covariances of the random variables, and in this case, knowing the mean function μX(t)=E[X(t)] (it need not be a constant as is required for wide-sense-stationarity) and the autocorrelation function RX(t1,t2)=E[X(t1)X(t2)] for all t1,t2 (it need not depend only on t1t2 as is required for wide-sense-stationarity) is sufficient to determine the statistics of the process completely.

If the Gaussian process is a WSS process, then it is also a strictly stationary Gaussian process. Fortunately for engineers and signal processors, many physical noise processes can be well-modeled as WSS Gaussian processes (and therefore strictly stationary processes), so that experimental observation of the autocorrelation function readily provides all the joint distributions. Furthermore since Gaussian processes retain their Gaussian character as they pass through linear systems, and the output autocorrelation function is related to th input autocorrelation function as Ry=h˜hRX

so that the output statistics can also be easily determined, WSS process in general and WSS Gaussian processes in particular are of great importance in engineering applications.


Could you, please, comment on "White Noise" in that sense? By definition the Autocorrelation at τ=0 is the variance of the random variables. Does it mean that AWGN (Additive White Gaussian Noise) has infinite variance? I ask it because usually people write n(t) N(0,N0/2), is wrong? Should it be written n(t) N(0,δ(0)N0/2)? Thanks.
Royi

1
@Drazick Please ask a separate question.
Dilip Sarwate

1
This is a fantastic mini-course in the definition of stationary processes. I've never seen anything like it--laid out so methodically and clearly. Community Wiki?
abalter

@Dilip Sarwate Excuse me for my ignorance. In the example. Why is E[X(t)]=0 for all t ? Did you assume ergodicity? How did you derive the probability density function of X(t) from the probability density function of theta to compute the expected value? E[X(t)X(s)] = E[cos(t+theta)*cos(s+theta)] right? Which steps did you take to simplify this expression and get to what you wrote? Thanks
VMMF

1
@VMMF There is NO ergodicity used. X(t)=cos(t+Θ) is a discrete random variable because Θ is a discrete random variable and it takes on values ±cos(t) and ±sin(t) with equal probability 14. Ergo, E[X(t)]=0. X(t)X(s) takes on values cos(t)cos(s), (cos(t))(cos(s))=cos(t)cos(s), sin(t)sin(s) and (sin(t))(sin(s))=sin(t)sin(s) with equal probability 14. Hence, E[X(t)(X(s)]=12(cos(t)cos(s)+sin(t)sin(s))=12cos(ts). Hence,
Dilip Sarwate
Sitemizi kullandığınızda şunları okuyup anladığınızı kabul etmiş olursunuz: Çerez Politikası ve Gizlilik Politikası.
Licensed under cc by-sa 3.0 with attribution required.