[DetnEst] 10. Kalman Filters

KBC·2024년 12월 10일
0

Detection and Estimation

목록 보기
16/23

Introduction

  • In 1960, Rudolf Kalman developed a way to solve some of the practical difficulties that arise when trying to apply Weiner filters
  • The three keys to leading to the Kalman filter
    • Wiener filter : LMMSE of a signal (only for stationary scalar signals and noises)
    • Sequential LMMSE : sequentially estimate a fixed parameter
    • State-space models : dynamical models for varying parameters
  • Kalman Filter : sequential LMMSE estimation for a time-varying parameter vector, but the time variation is constrained to follow a state-space dynamical model(allow the unknown parameters to evolve in time)
  • Kalman filter is an optimal MMSE estimator for jointly Gaussian case, and is the optimal LMMSE estimator in general

Dynamic Signal Models

  • Time-varying DC level in WGN example
    x[n]=A[n]+w[n]x[n]=A[n]+w[n]
    • Need to estimate A[n],  n=0,1,,N1A[n],\;n=0,1,\cdots,N-1
    • If we model the A[n]A[n] as a sequence of unknown deterministic parameters,
      MVUE becomes A^[n]=x[n]\hat A[n]=x[n] with variance σ2\sigma^2 (not desirable)
      • must model the correlation among A[n]A[n]
      • A[n]A[n] is a realization of a random process : Bayesian approach
    • Let's denote the signal to be estimated as s[n]s[n] instead of θ[n]\theta[n]
    • Assuming that the mean of the signal is known, we can assume zero-mean signal and add the mean back later

  • A simple model to specify the correlation : first-order Gauss-Markov process
    s[n]=as[n1]+u[n],  n0s[n]=as[n-1]+u[n],\;n\geq0
    • u[n]u[n] : WGN with variance
    • σu2\sigma^2_u : driving or excitation noise
    • s[1]N(μs,σs2)s[-1]\sim\mathcal{N}(\mu_s,\sigma^2_s), independent of u[n],  n0u[n],\;n\geq0
    • s[n]s[n] : output of LTI system driven by u[n]u[n]
  • Dynamical or state model : The current output (s[n]s[n]) depends only on the state of the system at the previous time (s[n1]s[n-1]) and the current input (u[n]u[n])
  • In general,
    s[n]=an+1s[1]+k=0naku[nk]s[n]=a^{n+1}s[-1]+\sum^n_{k=0}a^ku[n-k]
    linear with s[1]s[-1] and u[nk]u[n-k]'s : Gaussian random process

  • Gaussian random process
    E(s[n])=an+1μsE(s[n])=a^{n+1}\mu_s
    • The covariance between s[m]s[m] and s[n]s[n] assuming mnm\geq n is
      cs[m,n]=E[(s[m]E(s[m]))(s[n]E(s[n]))]=E[(am+1(s[1]μs)+k=0maku[mk])(an+1(s[1]μs)+l=0nalu[nl])]=am+n+2σs2+k=0ml=0nak+lE(u[mk]u[nl])=am+n+2σs2+σu2k=0ma2kanmE(u[mk]u[nl])=σu2δ[l(nm+k)]=am+n+2σs2+σu2anmk=0nma2kcs[m,n]=cs[n,m]for m<nc_s[m, n] = \mathbb{E}\left[(s[m] - \mathbb{E}(s[m]))(s[n] - \mathbb{E}(s[n]))\right]\\[0.2cm] = \mathbb{E}\left[\left(a^{m+1}(s[-1] - \mu_s) + \sum_{k=0}^m a^k u[m-k]\right) \left(a^{n+1}(s[-1] - \mu_s) + \sum_{l=0}^n a^l u[n-l]\right)\right]\\[0.2cm] = a^{m+n+2}\sigma_s^2 + \sum_{k=0}^m \sum_{l=0}^n a^{k+l} \mathbb{E}(u[m-k] u[n-l])\\[0.2cm] = a^{m+n+2}\sigma_s^2 + \sigma_u^2 \sum_{k=0}^m a^{2k} a^{n-m}\\[0.2cm] E(u[m-k]u[n-l])=\sigma^2_u\delta[l-(n-m+k)]\\[0.2cm] = a^{m+n+2}\sigma_s^2 + \sigma_u^2 a^{n-m} \sum_{k=0}^{n-m} a^{2k}\\[0.2cm] \rightarrow c_s[m, n] = c_s[n, m] \quad \text{for } m < n
  • s[n]s[n] is not WSS, but as n,  for a<1n\rightarrow \infty,\;\text{for }|a|<1
    E(s[n])0,rss[k]=cs[m,n]mn=kσu2ak1a2\mathbb{E}(s[n]) \rightarrow 0, \quad r_{ss}[k] = c_s[m, n] \Big|_{m-n=k} \rightarrow \frac{\sigma_u^2 a^k}{1 - a^2}
    s[n]s[n] can be WSS for n>0n>0 if we choose μs,σs2\mu_s,\sigma^2_s carefully
  • Recursive form
    E(s[n])=aE(s[n1])+E(u[n])=aE(s[n1])\mathbb{E}(s[n]) = a \mathbb{E}(s[n-1]) + \mathbb{E}(u[n]) \\ = a \mathbb{E}(s[n-1])
var(s[n])=E[(s[n]E(s[n]))2]=E[(as[n1]+u[n]aE(s[n1]))2]=a2var(s[n1])+σu2\text{var}(s[n]) = \mathbb{E}\left[(s[n] - \mathbb{E}(s[n]))^2\right] \\ = \mathbb{E}\left[(as[n-1] + u[n] - a \mathbb{E}(s[n-1]))^2\right] \\ = a^2 \text{var}(s[n-1]) + \sigma_u^2

(mean and variance propagation eqns)

  • ppth-order Gauss-Markov process
    s[n]=k=1pa[k]s[nk]+u[n]s[n] = -\sum_{k=1}^p a[k] s[n-k] + u[n]
The state of the system at time n:{s[n],s[n1],s[n2],,s[np+1]}\text{The state of the system at time } n : \{s[n], s[n-1], s[n-2], \dots, s[n-p+1]\}
  • state vector
    s[n]=[s[np+1]s[np+2]s[n]]=[01000010a[p]a[p1]a[p2]a[1]][s[np]s[np+1]s[n1]]+[001]u[n]s[n] = \begin{bmatrix} s[n-p+1] \\ s[n-p+2] \\ \vdots \\ s[n] \end{bmatrix} = \begin{bmatrix} 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & \vdots & \ddots & \vdots \\ -a[p] & -a[p-1] & -a[p-2] & \cdots & -a[1] \end{bmatrix} \begin{bmatrix} s[n-p] \\ s[n-p+1] \\ \vdots \\ s[n-1] \end{bmatrix} +\begin{bmatrix} 0 \\ 0 \\ \vdots \\ 1 \end{bmatrix} u[n]
s[n]=As[n1]+Bu[n],n0(vector Gauss-Markov model)s[n] = \mathbf{A}s[n-1] + \mathbf{B}u[n], \quad n \geq 0 \quad (\text{vector Gauss-Markov model})

  • Generalized vector Gauss-Markov model
    s[n]=As[n1]+Bu[n],  n0s[n]=As[n-1]+Bu[n],\;n\geq0
    • State model, AA : state transition matrix p×pp\times p
      • B:p×rB : p\times r matrix
      • s[n]:p×1s[n] : p\times 1 signal vector
      • u[n]:r×1u[n] : r\times 1 driving noise vector
    • Statiscal assumptions
      1. u[n]u[n] is a vector WGN sequence, i.e. u[n]u[n] is a sequence of uncorrelated jointly Gaussian vectors with E[u[n]]=0E[u[n]]=0
        As a result,
        E(u[m]uT[n])=0,  mnE(u[n]uT[n])=QE(u[m]u^T[n])=0,\;m\neq n\\[0.2cm] E(u[n]u^T[n])=Q
      2. s[1]N(μs,Cs)s[-1]\sim \mathcal{N}(\mu_s,C_s) and s[1]s[-1] independent of u[n],  n0u[n],\;n\geq0
  • Two DC power supplies example
    • Two DC power supply outputs which are independent (in a functional sense)
      s1[n]=a1s1[n1]+u1[n],s2[n]=a2s2[n1]+u2[n]s_1[n] = a_1 s_1[n-1] + u_1[n], \quad s_2[n] = a_2 s_2[n-1] + u_2[n]
s1[1]N(μs1,σs12),s2[1]N(μs2,σs22)s_1[-1] \sim \mathcal{N}(\mu_{s_1}, \sigma_{s_1}^2), \quad s_2[-1] \sim \mathcal{N}(\mu_{s_2}, \sigma_{s_2}^2)
u1[n],u2[n]:WGN with variance σu12 and σu22u_1[n], u_2[n]: \text{WGN with variance } \sigma_{u_1}^2 \text{ and } \sigma_{u_2}^2
  • All random variables are independent of each other
    s[n]=[s1[n]s2[n]]=[a100a2][s1[n1]s2[n1]]+[1001][u1[n]u2[n]]s[n] = \begin{bmatrix} s_1[n] \\ s_2[n] \end{bmatrix} = \begin{bmatrix} a_1 & 0 \\ 0 & a_2 \end{bmatrix} \begin{bmatrix} s_1[n-1] \\ s_2[n-1] \end{bmatrix} +\begin{bmatrix}1 & 0 \\0 & 1\end{bmatrix}\begin{bmatrix} u_1[n] \\ u_2[n] \end{bmatrix}
E(u[m]uT[n])=[E(u1[m]u1[n])E(u1[m]u2[n])E(u2[m]u1[n])E(u2[m]u2[n])]=[σu1200σu22]δ[mn]\mathbb{E}(u[m] u^T[n]) = \begin{bmatrix} \mathbb{E}(u_1[m] u_1[n]) & \mathbb{E}(u_1[m] u_2[n]) \\ \mathbb{E}(u_2[m] u_1[n]) & \mathbb{E}(u_2[m] u_2[n]) \end{bmatrix} = \begin{bmatrix} \sigma_{u_1}^2 & 0 \\ 0 & \sigma_{u_2}^2 \end{bmatrix} \delta[m-n]
Q=[σu1200σu22],s[1]=[s1[1]s2[1]]N([μs1μs2],[σs1200σs22])\mathbf{Q} = \begin{bmatrix}\sigma_{u_1}^2 & 0 \\ 0 & \sigma_{u_2}^2\end{bmatrix}, \quad s[-1] = \begin{bmatrix}s_1[-1] \\ s_2[-1] \end{bmatrix} \sim \mathcal{N} \left( \begin{bmatrix}\mu_{s_1} \\ \mu_{s_2} \end{bmatrix}, \begin{bmatrix} \sigma_{s_1}^2 & 0 \\ 0 & \sigma_{s_2}^2 \end{bmatrix} \right)

  • Statistical properties of the vector Gauss-Markov model
    s[n]=An+1s[1]+k=0nAkBu[nk]s[n]=A^{n+1}s[-1]+\sum^n_{k=0}A^kBu[n-k]
    where A0=IA^0=I \rightarrow Gaussian random process, E(s[n])=An+1E(s[1])=An+1μsE(s[n])=A^{n+1}E(s[-1])=A^{n+1}\mu_s
    cs[m,n]=E[(s[m]E(s[m]))(s[n]E(s[n]))T]c_s[m, n] = \mathbb{E}\left[(s[m] - \mathbb{E}(s[m]))(s[n] - \mathbb{E}(s[n]))^T\right]
=E[(Am+1(s[1]μs)+k=0mAkBu[mk])(An+1(s[1]μs)+l=0nAlBu[nl])T]= \mathbb{E}\left[\left(A^{m+1}(s[-1] - \mu_s) + \sum_{k=0}^m A^k B u[m-k]\right) \left(A^{n+1}(s[-1] - \mu_s) + \sum_{l=0}^n A^l B u[n-l]\right)^T\right]
=Am+1CsAn+1T+k=0ml=0nAkBE(u[mk]u[nl]T)BTAlT= A^{m+1} C_s A^{n+1^T} + \sum_{k=0}^m \sum_{l=0}^n A^k B \mathbb{E}(u[m-k] u[n-l]^T) B^T A^{l^T}
=Am+1CsAn+1T+k=0mAkBQBTA(nm+k)T,for m<n= A^{m+1} C_s A^{n+1^T} + \sum_{k=0}^m A^k B \mathbf{Q} B^T A^{(n-m+k)^T}, \quad \text{for } m < n
C[n]=Cs[n,n]=An+1CsAn+1T+k=0nAkBQBTAkTC[n] = C_s[n, n] = A^{n+1} C_s A^{n+1^T} + \sum_{k=0}^n A^k B \mathbf{Q} B^T A^{k^T}
In recursive form: E(s[n])=AE(s[n1]),C[n]=AC[n1]AT+BQBT\text{In recursive form: } \mathbb{E}(s[n]) = A \mathbb{E}(s[n-1]), \quad C[n] = A C[n-1] A^T + B \mathbf{Q} B^T

  • The eigenvalues of AA must be less than 11 magnitude for a stable process
    • If so, as nn\rightarrow \infty
      E(s[n])=An+1μs0E(s[n])=A^{n+1}\mu_s\rightarrow 0
    • It can be shown that An+1CsAn+1T0A^{n+1}C_s{A^{n+1}}^T\rightarrow 0 so that
      C[n]C=k=0AkBQBTAkTC[n]\rightarrow C=\sum^\infty_{k=0}A^kBQB^T{A^k}^T
    • Also, for steady state, C[n]=C[n1]=CC[n]=C[n-1]=C in recursion,
      C=ACAT+BQBTC=ACA^T+BQB^T
    • It is also possible to make AA, BB and QQ time-dependent

Scalar Kalman Filter

  • Data Model(scalar Gauss-Markov signal model)
    • s[n]=as[n1]+u[n],  n0 (scalar state equation)s[n]=as[n-1]+u[n],\;n\geq0\text{ (scalar state equation)}
    • x[n]=s[n]+w[n] (scalar observation equation)x[n]=s[n]+w[n]\text{ (scalar observation equation)}
  • Assumptions
    • u[n]N(0,σu2),  u[n]u[n]\sim\mathcal{N}(0, \sigma^2_u),\;u[n] and u[m]u[m] indep. for mnm\neq n
    • w[n]N(0,σn2),  w[n]w[n]\sim\mathcal{N}(0, \sigma^2_n),\;w[n] and w[m]w[m] indep. for mnm \neq n (variance can change with time)
    • s[1]N(0,σs2)s[-1]\sim\mathcal{N}(0, \sigma^2_s), independent of u[n],  w[n],  n0u[n],\;w[n],\;n\geq0
  • Sequential MMSE estimator that estimates s[n]s[n] based on x[0],x[1],,x[n]x[0],x[1],\cdots,x[n] as nn increases : filtering
  • Goal : recursively compute
    s^[nn]=E(s[n]x[0],x[1],,x[n])\hat s[n|n]=E(s[n]|x[0],x[1],\cdots,x[n])
  • Minimize E[s[n]s^[nn])2]s^[nn]=E(s[n]x[0],x[1],,x[n])E[s[n]-\hat s[n|n])^2] \rightarrow \hat s[n|n]=E(s[n]|x[0],x[1],\cdots,x[n])
  • Since θ=s[n]\theta=s[n] and x=[x[0]    x[n]]T\text{x}=[x[0]\;\cdots\;x[n]]^T are jointly Gaussian
    (linear, identical to LMMSE estimator)
    s^[nn]=CθxCxx1x\hat s[n|n]=C_{\theta x}C^{-1}_{xx}\text{x}
  • For non-Gaussian case, it is still a LMMSE estimator
    \rightarrow similar to sequential LMMSE estimation with vector space approach
  • Let's denote X[n]=[x[0]    x[n]]TX[n]=[x[0]\;\cdots\;x[n]]^T to avoid confusion with vector observations
    x~[n]=x[n]x^[nn1] : innovations^[nn]=E(s[n]X[n1],x~[n])=E(s[n]X[n1])+E(s[n]x~[n])Orthogonalization=s^[nn1]+E(s[n]x~[n])  ()\tilde x[n]=x[n]-\hat x[n|n-1] \text{ : innovation}\\[0.2cm] \hat s[n|n]=E(s[n]|X[n-1],\tilde x[n])=E(s[n]|X[n-1])+E(s[n]|\tilde x[n])\\[0.2cm] \text{Orthogonalization}\\[0.2cm] =\hat s[n|n-1]+E(s[n]|\tilde x[n]) \; (*)

  • s^[nn1]=E(as[n1]+u[n]X[n1])=as^[n1n1]\hat s[n|n-1]=E(as[n-1]+u[n]|X[n-1])=a\hat s[n-1|n-1]
  • E(s[n]x~[n])E(s[n]|\tilde x[n]) : MMSE estimator, linear because of the zero mean assumptions
    E(s[n]x~[n])=K[n]x~[n]=K[n](x[n]x^[nn1])\mathbb{E}(s[n] | \tilde{x}[n]) = K[n] \tilde{x}[n] = K[n] \left( x[n] - \hat{x}[n | n-1] \right)
where K[n]=E(s[n]x~[n])E(x~2[n]),as θ^=CθxCxx1x\text{where } K[n] = \frac{\mathbb{E}(s[n] \tilde{x}[n])}{\mathbb{E}(\tilde{x}^2[n])}, \quad \text{as } \hat{\theta} = C_{\theta x} C_{xx}^{-1} x
x^[nn1]=s^[nn1]+w^[nn1]=s^[nn1]\hat{x}[n | n-1] = \hat{s}[n | n-1] + \hat{w}[n | n-1] = \hat{s}[n | n-1]
E(s[n]x~[n])=K[n](x[n]s^[nn1])\rightarrow \mathbb{E}(s[n] | \tilde{x}[n]) = K[n] \left( x[n] - \hat{s}[n | n-1] \right)
()s^[n]=s^[n1]+K[n](x[n]s^[nn1])(*) \quad \hat{s}[n] = \hat{s}[n-1] + K[n] \left( x[n] - \hat{s}[n | n-1] \right)
where s^[nn1]=as^[n1n1],K[n]=E(s[n](x[n]s^[nn1]))E((x[n]s^[nn1])2)\text{where } \hat{s}[n | n-1] = a \hat{s}[n-1 | n-1], \quad K[n] = \frac{\mathbb{E}\left(s[n](x[n] - \hat{s}[n | n-1])\right)}{\mathbb{E}\left((x[n] - \hat{s}[n | n-1])^2\right)}

  • To evaluate K[n]K[n], we have
  1. E[s[n](x[n]s^[nn1])]=E[(s[n]s^[nn1])(x[n]s^[nn1])]E[s[n](x[n]-\hat s[n|n-1])]=E[(s[n]-\hat s[n|n-1])(x[n]-\hat s[n|n-1])]
    since s^[nn1]\hat s[n|n-1], a linear combination of s[nk]s[n-k], are orthogonal to error
  2. E[w[n](s[n]s^[nn1])]=0E[w[n](s[n]-\hat s[n|n-1])] =0
    K[n]=E((s[n]s^[nn1])(x[n]s^[nn1]))E((s[n]s^[nn1]+w[n])2)K[n] = \frac{\mathbb{E}\left((s[n] - \hat{s}[n | n-1])(x[n] - \hat{s}[n | n-1])\right)}{\mathbb{E}\left((s[n] - \hat{s}[n | n-1] + w[n])^2\right)}
=E((s[n]s^[nn1])2)E((s[n]s^[nn1])2)+σn2=M[nn1]M[nn1]+σn2= \frac{\mathbb{E}\left((s[n] - \hat{s}[n | n-1])^2\right)}{\mathbb{E}\left((s[n] - \hat{s}[n | n-1])^2\right) + \sigma_n^2} = \frac{M[n | n-1]}{M[n | n-1] + \sigma_n^2}

where M[nn1]M[n|n-1] is MSE when s[n]s[n] is estimated based on previous data


M[nn1]=E[(s[n]s^[nn1])2]=E[(a(s[n1]s^[n1n1])+u[n])2]M[n | n-1] = \mathbb{E}\left[(s[n] - \hat{s}[n | n-1])^2\right] = \mathbb{E}\left[(a(s[n-1] - \hat{s}[n-1 | n-1]) + u[n])^2\right]
=a2M[n1n1]+σu2= a^2 M[n-1 | n-1] + \sigma_u^2
M[nn]=E[(s[n]s^[nn])2]=E[(s[n]s^[nn1]K[n](x[n]s^[nn1]))2]M[n | n] = \mathbb{E}\left[(s[n] - \hat{s}[n | n])^2\right] = \mathbb{E}\left[\left(s[n] - \hat{s}[n | n-1] - K[n](x[n] - \hat{s}[n | n-1])\right)^2\right]
=E[(s[n]s^[nn1])2]2K[n]E[(s[n]s^[nn1])(x[n]s^[nn1])]+K[n]2E[(x[n]s^[nn1])2]= \mathbb{E}\left[(s[n] - \hat{s}[n | n-1])^2\right] - 2 K[n] \mathbb{E}\left[(s[n] - \hat{s}[n | n-1])(x[n] - \hat{s}[n | n-1])\right] \\ + K[n]^2 \mathbb{E}\left[(x[n] - \hat{s}[n | n-1])^2\right]
=M[nn1]2K[n]M[nn1]+K[n]2(M[nn1]+σn2)= M[n | n-1] - 2 K[n] M[n | n-1] + K[n]^2 (M[n | n-1] + \sigma_n^2)
=M[nn1](1K[n])+K[n]2σn2= M[n | n-1] (1 - K[n]) + K[n]^2 \sigma_n^2
=(1K[n])M[nn1]= (1 - K[n]) M[n | n-1]
  • Summary
    • Prediction : s^[nn1]=as^[n1n1]\hat s[n|n-1]=a\hat s[n-1|n-1]
    • Minimum Prediction MSE : M[nn1]=a2M[n1n1]+σu2M[n|n-1]=a^2M[n-1|n-1]+\sigma^2_u
    • Kalman Gain : K[n]=M[nn1]M[nn1]+σn2K[n]=\frac{M[n|n-1]}{M[n|n-1]+\sigma^2_n}
    • Correction : s^[nn]=s^[nn1]+K[n](x[n]s^[nn1])\hat s[n|n]=\hat s[n|n-1]+K[n](x[n]-\hat s[n|n-1])
    • Minimum MSE : M[nn]=(1K[n])M[nn1]M[n|n]=(1-K[n])M[n|n-1]
  • The derivation was with μs=0\mu_s=0, but the same equations result for μs0\mu_s \neq 0
    • Initialization : s^[11]=E(s[1])=μs\hat s[-1|-1]=E(s[-1])=\mu_s and M[11]=σs2M[-1|-1]=\sigma^2_s
    • K[n](x[n]s^[nn1])K[n](x[n]-\hat s[n|n-1]) can be thought as the estimator of u[n]u[n]
    • s^[nn]=as^[n1n1]+u^[n]\hat s[n|n] = a\hat s[n-1|n-1] + \hat u[n] where u^[n]=K[n](x[n]s^[nn1])\hat u[n]=K[n](x[n]-\hat s[n|n-1])

Scalar Kalman Filter - Properties

  1. Kalman filter is an extension to the sequential MMSE estimator (LMMSE for non-Gaussian) where the unknown parameter evolves in time according to the dynamic model. The sequential MMSE estimator is a special case of Kalman filter for a=1a=1 and σu2=0\sigma^2_u=0
  2. No matrix inversions are required
  3. The Kalman filter is a time varying filter
    s^[nn]=as^[n1n1]+K[n](x[n]as^[n1n1])=a(1K[n])s^[n1n1]+K[n]x[n]\hat s[n|n]=a\hat s[n-1|n-1]+K[n](x[n]-a\hat s[n-1|n-1])\\[0.2cm] =a(1-K[n])\hat s[n-1|n-1] + K[n]x[n]
    : first order recursive filter with time varying coefficients
  4. The Kalman filter provides its own performance measure. Minimum Bayesian MSE is computed as an integral part of the estimator
  5. The prediction stage increases the error, while the correction stage decreases it. The minimum prediction MSE and the minimum MSE is shown
  6. Predictions is an integral part of the Kalman filter. One-step prediction is already computed. If we desire the best two-step prediction, we can let σn2\sigma^2_n \rightarrow \infty implying x[n]x[n] is noisy enough to ignore and obtain s^[n+1n1]=s^[n+1n]=as^[nn]=as^[nn1]=a2s^[n1n1]\hat s[n+1|n-1] = \hat s[n+1|n]=a\hat s[n|n]=a\hat s[n|n-1]=a^2\hat s[n-1|n-1]
    It can be generalized to the ll step predictor
  7. The Kalman filter is driven by the uncorrelated innovation sequence and in steady-state can also by viewed as a whitening filter
    s^[nn]=as^[n1n1]+K[n](x[n]s^[nn1])\hat s[n|n]=a\hat s[n-1|n-1]+K[n](x[n]-\hat s[n|n-1])
    If we think the innovation
    x~[n]=x[n]s^[nn1]\tilde x[n]=x[n]-\hat s[n|n-1]
    as the Kalman filter output, it is a whitening filter
  8. For non-Gaussian case, it is still the optimum LMMSE estimator

Vector Kalman Filter

  • The p×1p \times 1 signal vector s[n]s[n] evolves in time according to the Gauss-Markov model
    x[n]=A[n]s[n1]+B[n]u[n],  n0x[n]=A[n]s[n-1]+B[n]u[n],\;n\geq0
  • A[n],B[n]A[n],B[n] are known matrices of dimension p×pp \times p and p×rp \times r, respectively
  • u[n]u[n] has the PDF u[n]N(0,Q[n])u[n]\sim\mathcal{N}(0, Q[n]) are independent
    E(u[m]uT[n])=0,E(u[m]u^T[n])=0, for mn  (u[n]m \neq n\;(u[n] is vector WGN)
  • s[1]s[-1] has the PDF s[1]N(μs,Cs)s[-1]\sim\mathcal{N}(\mu_s,C_s) and is independent of u[n]u[n]

  • The M×1M\times 1 observation vectors x[n]\text{x}[n] are modeled by the Bayesian linear model
    x[n]=H[n]s[n]+w[n],  n0\text{x}[n]=H[n]s[n]+w[n],\;n\geq0
    • H[n]H[n] is a known M×pM\times p observations matrix (which may be time varying)
    • w[n]w[n] is an M×1M \times 1 observation noise vector with PDF
    • w[n]N(0,C[n])w[n]\sim\mathcal{N}(0, C[n]) are independent
      E(w[m]wT[n])=0,  for mnE(w[m]w^T[n])=0,\;\text{for }m\neq n
      (If C[n]C[n] does not depend on nn, w[n]w[n] would be vector WGN)

  • The MMSE estimator of s[n]s[n] based on x[0],x[1],,x[n]x[0],x[1],\cdots,x[n]
    s^[nn]=E(s[n]x[0],x[1],,x[n])\hat s[n|n]=E(s[n]|x[0],x[1],\cdots,x[n])
  • Prediction
    • Minimum Prediction MSE Matrix (p×pp \times p)
      M[nn1]=A[n]M[n1n1]AT[n]+B[n]Q[n]BT[n]M[n|n-1]=A[n]M[n-1|n-1]A^T[n]+B[n]Q[n]B^T[n]
    • Kalman Gain Matrix (p×Mp\times M) :
      M[nn]=(IK[n]H[n])M[nn1]M[n|n]=(I-K[n]H[n])M[n|n-1]
  • The recursion is initialized by s^[11]=μs\hat s[-1|-1]=\mu_s, and M[11]=CsM[-1|-1]=C_s

All Content has been written based on lecture of Prof. eui-seok.Hwang in GIST(Detection and Estimation)

profile
AI, Security

0개의 댓글