Approaches for the robustification of Kalman filters (Q2772096)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: Approaches for the robustification of Kalman filters |
scientific article; zbMATH DE number 1706941
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | Approaches for the robustification of Kalman filters |
scientific article; zbMATH DE number 1706941 |
Statements
18 February 2002
0 references
mIC-filters
0 references
Kalman filtering
0 references
robustness
0 references
asymptotic stationarity
0 references
posterior-mode estimators
0 references
influence curve
0 references
linear estimator
0 references
missings
0 references
0 references
0.95271665
0 references
0.9499228
0 references
0.94937253
0 references
0.93387604
0 references
0.9306831
0 references
0 references
Approaches for the robustification of Kalman filters (English)
0 references
This extensive work is a systematical and self-contained treatment of various robustification concepts for Kalman filtering. The requirement of robustness arises from practical applications where the filtering algorithm is corrupted by outliers. The robustification problem is complex due to different types of outliers: While the outliers in the observation noise produce a reduced impact of the observation in the prediction and correction steps, the outliers in the system noise give structural breaks which are to be recognized as fast as possible.NEWLINENEWLINENEWLINEThe first part of the work deals with the so-called rLS-, rIC-, and mIC-filters. These algorithms differ from the classical Kalman filter due to a robustified correction step. The author shows the asymptotic stationarity of these methods. A simulation study provides a comparison of the rLS- and rIC-filters with the posterior-mode estimators showing that the former perform better. The second part is devoted to statistical aspects of the EM-algorithm and its robustification. To get the local asymptotic approach, the \(L_2\)-differentiability is generalized giving the notions of the influence curve and of the linear estimator for a model with missings. Within this setting, a classical optimal one-step-algorithm and a robust one-step algorithm are presented.NEWLINENEWLINENEWLINEWe should emphasize the efforts to present the background needed for understanding the matter. The reader benefits permanently from the appendix, which provides a description of corresponding tools from probability theory, asymptotical statistics and other related fields.
0 references