The smoothing problem (not to be confused with smoothing in statistics, image processing and other contexts) is the problem of estimating an unknown probability density function recursively over time using incremental incoming measurements. It is one of the main problems defined by Norbert Wiener.[1][2] A smoother is an algorithm that implements a solution to this problem, typically based on recursive Bayesian estimation. The smoothing problem is closely related to the filtering problem, both of which are studied in Bayesian smoothing theory.
A smoother is often a two-pass process, composed of forward and backward passes. Consider doing estimation (prediction/retrodiction) about an ongoing process (e.g. tracking a missile) based on incoming observations. When new observations arrive, estimations about past needs to be updated to have a smoother (more accurate) estimation of the whole estimated path until now (taking into account the newer observations). Without a backward pass (for retrodiction), the sequence of predictions in an online filtering algorithm does not look smooth. In other words, retrospectively, it is as if we are using future observations for improving estimation of a point in past, when those observations about future points become available. Note that time of estimation (which determines which observations are available) can be different to the time of the point that the prediction is about (that is subject to prediction/retrodiction). The observations about later times can be used to update and improved the estimations about earlier times. Doing so leads to smoother-looking estimations (retrodiction) about the whole path.
Some variants include:[3]
The terms Smoothing and Filtering are used for four concepts that may initially be confusing: Smoothing (in two senses: estimation and convolution), and Filtering (again in two senses: estimation and convolution).
Smoothing (estimation) and smoothing (convolution) despite being labelled with the same name in English language, can mean totally different mathematical procedures. The requirements of problems they solve are different. These concepts are distinguished by the context (signal processing versus estimation of stochastic processes).
The historical reason for this confusion is that initially, the Wiener's suggested a "smoothing" filter that was just a convolution. Later on his proposed solutions for obtaining a smoother estimation separate developments as two distinct concepts. One was about attaining a smoother estimation by taking into account past observations, and the other one was smoothing using filter design (design of a convolution filter).
Both the smoothing problem (in sense of estimation) and the filtering problem (in sense of estimation) are often confused with smoothing and filtering in other contexts (especially non-stochastic signal processing, often a name of various types of convolution). These names are used in the context of World War 2 with problems framed by people like Norbert Wiener.[1][2] One source of confusion is the Wiener Filter is in form of a simple convolution. However, in Wiener's filter, two time-series are given. When the filter is defined, a straightforward convolution is the answer. However, in later developments such as Kalman filtering, the nature of filtering is different to convolution and it deserves a different name.
The distinction is described in the following two senses:
1. Convolution: The smoothing in the sense of convolution is simpler. For example, moving average, low-pass filtering, convolution with a kernel, or blurring using Laplace filters in image processing. It is often a filter design problem. Especially non-stochastic and non-Bayesian signal processing, without any hidden variables.
2. Estimation: The smoothing problem (or Smoothing in the sense of estimation) uses Bayesian and state-space models to estimate the hidden state variables. This is used in the context of World War 2 defined by people like Norbert Wiener, in (stochastic) control theory, radar, signal detection, tracking, etc. The most common use is the Kalman Smoother used with Kalman Filter, which is actually developed by Rauch. The procedure is called Kalman-Rauch recursion. It is one of the main problems solved by Norbert Wiener.[1][2] Most importantly, in the Filtering problem (sense 2) the information from observation up to the time of the current sample is used. In smoothing (also sense 2) all observation samples (from future) are used. Filtering is causal but smoothing is batch processing of the same problem, namely, estimation of a time-series process based on serial incremental observations.
But the usual and more common smoothing and filtering (in the sense of 1.) do not have such distinction because there is no distinction between hidden and observable.
The distinction between Smoothing (estimation) and Filtering (estimation): In smoothing all observation samples are used (from future). Filtering is causal, whereas smoothing is batch processing of the given data. Filtering is the estimation of a (hidden) time-series process based on serial incremental observations.
Original source: https://en.wikipedia.org/wiki/Smoothing problem (stochastic processes).
Read more |