- About this Journal ·
- Abstracting and Indexing ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Recently Accepted Articles ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents
The Scientific World Journal
Volume 2013 (2013), Article ID 878417, 15 pages
Analytical Analysis of Motion Separability
School of Aerospace, Mechanical and Manufacturing Engineering, RMIT University, Corner of Plenty Road and McKimmies Road, Bundoora, Victoria 3083, Australia
Received 29 August 2013; Accepted 26 September 2013
Academic Editors: S.-S. Liaw and Z. Yang
Copyright © 2013 Marjan Hadian Jazi et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Motion segmentation is an important task in computer vision and several practical approaches have already been developed. A common approach to motion segmentation is to use the optical flow and formulate the segmentation problem using a linear approximation of the brightness constancy constraints. Although there are numerous solutions to solve this problem and their accuracies and reliabilities have been studied, the exact definition of the segmentation problem, its theoretical feasibility and the conditions for successful motion segmentation are yet to be derived. This paper presents a simplified theoretical framework for the prediction of feasibility, of segmentation of a two-dimensional linear equation system. A statistical definition of a separable motion (structure) is presented and a relatively straightforward criterion for predicting the separability of two different motions in this framework is derived. The applicability of the proposed criterion for prediction of the existence of multiple motions in practice is examined using both synthetic and real image sequences. The prescribed separability criterion is useful in designing computer vision applications as it is solely based on the amount of relative motion and the scale of measurement noise.
Computer vision problems in general involve manipulation of complicated manifolds. However, due to the mathematical and computational complexities of finding solutions in those spaces, a large group of problems are solved via projections leading to approximate solutions found by solving systems of linear equations. For example, solutions of important computer vision problems such as optical flow , fundamental matrix , and parametric image segmentation of man-made objects  are commonly found by solving systems of linear equations. After decades of research in these areas, a rich collection of methods to both robustly and efficiently solve those problems is currently available [4, 5].
Substantial efforts are also devoted to solve the motion segmentation problem . A major shortcoming of those solutions stems from the fact that structures in visual data are not precisely defined. The segmentation methods are typically considered successful when those methods are able to partition data in a way that by visual inspection, segments are deemed to be part of an identifiable object. As such, there is no theory as yet to predict which two motions are separable for a given set of data or what would be the minimum relative velocity that would constitute another motion. This question has important engineering applications particularly for designing devices that use visual measurements of speed, such as visual traffic surveillance systems, as a source of information.
Generally speaking with any probability distribution with infinite tails such as commonly used Gaussian, there will always be a finite probability of misclassified data no matter what the separation. The “magic bullet” of a clean threshold where “it makes sense to declare two structures as separate and assign points, and on the other side of the threshold you can't distinguish” is incompatible with this model. The concept of separability itself is also not well defined and one can identify at least two different (but related) notions as outlined here.(1)Detectability: existence of two structures (motions) in a given data is detectable, but the detected structures are not necessarily distinguishable. (2)Separability: structures (motions) can be distinguished from each other and the segmentation of data between different structures can be performed up to a desired (given) level of reliability.
In engineering applications, the latter notion of separability is of significant interest. The characteristics of the structures are important and those characteristics cannot be measured unless the structures are well defined. The misclassification error of separating two models in general will be governed by fitting errors of the two putative models as well as the overlap of the distributions. Although the deciding line is always going to be somewhat arbitrary and dependent on what the engineering problem will tolerate, this paper makes an important contribution by quantifying the probability of misclassification. Our theoretical derivations also show that the separability problem is very complicated and development of an elegant solution that could predict separability in all cases does not appear to be straightforward. We have however been able to show how the Gaussian error of data relates to error of pairwise estimates of model parameters and how to quantify the exact overlap of their distributions. This at least enables us to identify the crossover point which is a useful guide for defining a structure and ultimately designing reliable equipment. We explain the use of crossover point in section 3-B, after we derive the above relationship.
Optical flow calculation has been one of the most studied problems of computer vision and its calculation [4, 7–10] and segmentation [11, 12] and the derivation of its confidence measures (error quantification) [13, 14] have been refined over several decades. However, the limits of using optical flow knowledge as a primary source of physical perception are yet to be established. More importantly, where there are at least two motions in a local area, it is currently not possible to predict how much difference between those motions is required for those motions to be separable.
For instance, as shown in Figure 1, one would need to know how much relative motion between different objects in this image would be required to reconstruct the scene geometry using motion information. In this figure, we have highlighted three local areas where two are on one column (having very similar motion) and the other on a different column. The question here is to ascertain how much difference in motion between these objects would be required to distinguish their motions in the given situation.
The issue of finding the limits of detection for near discontinuities in visual data was first discussed in . The work focused on range data segmentation and the separability criteria were only derived for special cases including parallel or crease structures. The above work was later extended to study the effect of consistency  and finite sample bias  of commonly used estimators on the separability of close structures. However, those works were also limited to the special case of parallel discontinuities only. Although the optical flow segmentation is a dual problem of the range segmentation discussed in , the underlying structures in motion segmentation are not limited to parallel structures and our aim here is to develop general analysis that is independent of the segmentation strategy.
To properly address the separability issue, we first need to have a precise definition of a structure. Having defined a structure, we then need to establish the general conditions for the separability of those structures for cases where nearby structures exist.
We focus on answering the above question for the two most similar motions in an image which their motion estimations are modeled as an instance of the optical flow problem. In essence, we aim to find a general condition for segmentation of motion when the two motions are modeled by optical flow constraints and formulated as a solution of a linear system of equations. The overall scope is therefore much broader and includes the ubiquitous problem of confirming the existence of multiple close solutions in a system of linear 2D equations in the presence of noise and outliers.
The rest of this paper is organized as follows. The motion separability problem is formulated in Section 2. A solution for the prediction problem is presented in Section 3 followed by the results of usability experiments using both synthetic and real data presented in Section 4. Section 5 concludes the paper.
2. Problem Formulation: Motion Separability
The current trend in optical flow estimation is to use nonparametric representations. Those methods (e.g., [4, 7–10]) often use either variational or discrete optimization methods to find solutions that show a degree of smoothness across motion boundaries by imposing certain regularization terms. These approaches involve tuning a large number of parameters that their optimum values depend on the structure of a scene. The analysis of a variety of optical flow estimation techniques has however shown that “only a small number of key choices produce statistically significant improvements” in the overall accuracy of those methods . More importantly, the above analysis has also shown that applying a median filter to intermediate flow values produces the most significant improvement. This implies that the appropriate separation of different motions is a key ingredient of the estimation process.
Since our aim here is to quantify the separability of two motions, we need to disentangle the optical flow estimation from the scene dependant implications of smoothness imposed implicitly by the nonparametric methods. To achieve this aim, we start the motion separability analysis by modeling the optical flow problem in its classical form presented in  without imposing an arbitrary smoothness across the motion boundaries. The local optical flow, without smoothness imposition, is modeled as a solution of a linear system: in which is a matrix with two columns, each containing the spatial derivatives ( and ) in directions of the velocity components , is a vector of associated temporal derivatives, and represents the noise . In this set up, if there are two motions, the above system of equations must be separable into two systems of linear equations and the separability would be a function of the difference of those motions. Intuitively, if the difference is very small, compared to the accuracy of estimation, the variations would be similar to noise and they would not be separable. Otherwise, it should be possible to separate those sets of equations.
Our aim here is to find the sufficient condition for the separability of two motions and as such, we use the most parsimonious motion model, constant flow, as it represents the least accurate (most inseparable) scenario. Models with more parameters (e.g., affine) are expected to produce better estimates of motions and therefore their separability condition is covered by the above model separability condition.
For a local area containing multiple motions, the system of (1) would contain a number of coherent subsystems which would appear as different clusters of cointersecting lines in the plane (see Figure 2(a)). To formulate the separability problem, without loss of generality, we can assume that there are only two motions in the area. The extent of the area is not fixed and the ones considered are the closest motions and therefore the hardest to distinguish from each other. If there are more motions, they are by definition further apart and would not affect the outcome. In this setting, the effect of other possible motions would be similar to gross outliers and are not expected to have a major impact on the separability issue as optical flow calculation methods are typically robust to the influences of outliers .
The motion separability problem, in its abstract form, is now represented by the problem of predicting the separability of two clusters of cointersecting lines based on two factors: the “distance” (the relative motion) between those clusters and the extent of the spread of constraint lines in each cluster (measurement noise). In its dual space, as shown in Figure 2(b), this problem is equivalent to the separability of two linear structures in a cloud of 2D points. To the best of our knowledge, currently there is no theory to predict the separability of those line clusters generally.
A common approach to tackle the above problem, in its dual form, is to use the Hough transform [19, 20]. In this approach, the Hough domain is first uniformly quantized into number of cells and each cell maintains a count of passing lines. The center of the cell with the largest number of counts is an estimate of the underlying structure parameters. To detect multiple structures, a threshold is specified and cells with counts exceeding a threshold are considered as putative structures. This approach has a major drawback stemming from the fact that the probability density function of the structure parameters has to be estimated using a discrete histogram as shown in Figure 2(f). As the figure shows, the interpretation of the histogram for the detection of different motions would heavily depend on the values of manually set thresholds.
To address this issue, Dahyot  has proposed a form of statistical kernel modeling for the Hough transform to estimate a continuous histogram. The proposed approach needs two pieces of information: the shape of a suitable kernel and the appropriate bandwidth. The overall approach is able to accurately estimate the parameters by locating the peaks of the histograms in cases where the structures are distinct. However, the choices of kernel shape and bandwidth have significant impact on the spread of the final pdf which is the key information for the separability analysis. As such, kernel density estimation appears not to be an appropriate tool for separability prediction.
To overcome the above problems, we propose to use a discrete part of the parameter space spanned by the exhaustive sampling of possible intersections as partly shown in Figure 2(e). To devise the separability condition for the optic flow problem, we derive the probability distribution of those samples as well as providing a precise definition for a structure (motion) in the space of those samples. The combination of these two pieces of information enables us to quantify the least amount of relative motion that would be separable for a given set of data. We will then show that those derivations are useful for predicting separability in practical situations.
3. Separability Prediction
An important aspect of replacing clusters of lines with all their intersection points (using existing parameter space discretization) is that the process changes the nature of the problem from finding the solution of an overdetermined system of linear equations (with multiple solutions) to a much simpler position estimation problem with multiple clusters. In the position estimation problem, the prediction of separability is fairly straight forward and the only information it requires is the definition of a cluster and the distributions of data points (line intersections).
Our practical intuition for solving this problem comes from our experiments with a variety of gray scale and color images commonly used for optical flow evaluations [21, 22]. We observed that in all of these images, the ratio of spatial derivatives of image intensity function (ratios of and for different pixels), to a large degree, is uniformly distributed over a fairly broad range of values. To provide some evidence, examples of the distribution of the ratios of the spatial derivatives of different image sequences in the Middlebury dataset are shown in Figure 3.
Although there might be small patches in some specific images in which the texture has a dominant direction (e.g. parallel strips), overall the commonly experienced textures have derivatives in all directions and very little can be generalized about the ratio of those derivatives. In fact, from an information point of view, assuming a uniform distribution is the least restrictive assumption one can make about the characteristics of those textures.
To establish the separability condition for linear structures (e.g., motions modeled by optical flow), we first replace all lines with their pairwise intersection points and then derive the probability distribution of those points by assuming that the ratio of spatial derivatives is uniformly distributed. The system of (1) is then replaced by a cluster of points and the problem of finding the best solution to the above system based on a measure of goodness (e.g., least squares) is transformed to an instance of a well-known position estimation problem.
3.1. Error Distribution
To derive the distribution of the intersections points (for an equation system with one solution), we denote the true velocity by , the distance between a point and the true velocity by , and the probability density function by . This distance represents the error in the direction that the motion separability is examined in. If one of the components of the two motions is separable, then those motions will also be separable. In our derivations, we perform the calculations for an arbitrary direction and the rationale is that the separability is always tested in a particular direction. For instance, if the optical flow calculation is used to monitor highway traffic, the separability of horizontal velocities of different vehicles is of interest to the system designer. The differences of velocities in other directions (e.g., vertical in this case) are of no significance. The design parameter for this system is the scale of measurement noise for one vehicle in that particular direction and it depends on many factors including the quality of cameras, vehicles textures, and visibility conditions. The above design parameter (scale of measurement noise) captures the overall effect of those factors.
It is important to note that for the sake of simplicity, we model the flow calculation and its errors in the classical regression framework rather than the more accurate geometric modeling framework [23, 24]. This simplification is justified because the regression model for optical flow calculations is almost as accurate as the geometric model as long as the optical flow constraints are not parallel (or near parallel) to the vertical axis . Although values of the ratio of spatial partial derivatives of image intensity function (ratios of and for different pixels) are between plus and minus infinity, values between plus and minus one span half the velocity space. In this half space, the lines are away from the vertical direction and the classical regression modeling is accurate. We conduct our calculations for the above interval and later extend those to the whole velocity space.
To develop a solution to the above separability problem, we first introduce and prove a proposition regarding the distribution of pairwise samples. We then use the result to define structure and devise a separability criterion. Suppose that ,…, are independent identically distributed random observations drawn from the model where(i) is normally distributed with zero mean and known variance (although noise of different optical flow constraints is likely to be correlated, this assumption simplifies the modeling (of an otherwise intractable problem) while its computational bias is shown to be relatively small [9, 25]. Our definition of a structure, provided later, is inherently robust to influence of large perturbations and would not include samples that might have been generated by degenerate constraints. Our experiments with both synthetic and real images have also shown that this assumption does not generate significant bias in the final results);(ii) is uniformly distributed.
Proposition 1. Under assumptions (i) and (ii), the estimation error (difference between the estimated value and the real value ) has the following distribution:
Proof. Using (2), the coordinates of the intersection of lines and are
and the estimation error for the th point (which is the intersection of and lines) can be written as
It is important to note that values of denote the slope of linear equations and values between represent constraints that cover half the space between and . Since the estimation error is not a function of the chosen coordinate system, the distribution of error should also be the same in the other half of space. We use this fact and simplify the derivations by first finding the distribution where . The estimation error distribution for other values of will be identical.
To derive the distribution of the , we denote and first find the distribution of . The distribution of is then calculated as the inverse of .
If the probability density function of a random variable is denoted by , the probability density function of is as follows :
Using the above equation, the relationship between distributions of and can now be written as
And since is assumed to have uniform distribution (i.e., half the space), the probability density function of can be calculated as
Substituting in (8), we have
To calculate the pdf of different terms in (5), we note that the probability density function of the product of two independent random variables and () is as follows :
Rewriting (5), we have and the probability density function of can be written by using (11), as
Substituting (10) in (13), we have
The integrand of the second integral is continuous in its domain and the integration result is as follows:
The first integral however has a discontinuity at zero and is generally intractable. Nonetheless, we have found a neat approximation to solve this integration problem. To outline our solution, we first denote the integrand by and note that the Dirac delta function can be written as
Combining (17) and (16), we have
We also note that, as it is shown in Figure 4, is almost zero everywhere except around the origin. Inspired by this and (18), we approximate this two-dimensional discontinuity by a multiplication of two delta functions in every dimension. We later show that this is a very accurate approximation in terms of the total sum of probabilities. Consequently, we assume and the result of the first integral in (14) can now be calculated as
To show that the above approximation is accurate in terms of the total probabilities, we first note that the sum of probabilities associated with the second integral in (14), which was calculated exactly, is since
Therefore, the sum of probabilities associated with the first integral in (14) must also be . Importantly, the above approximation satisfies this requirement as shown below:
Having calculated both integrals of (14), the probability density function of can be written as
To find the distribution of all residuals (12), using the law of total probability, we write and by defining and to be we can rewrite (24) as and comparing this with (14), we have
To derive the pdf of estimation errors given by (12), using the law of total probability, we can write Equation (10) shows that and () have the same distribution and is greater than , when . Using definition (25), the above equation can then be rewritten as
We note that since values are independent and identically distributed (iid) random variables, and can also be considered iid random variables. Therefore, the distribution of the is the convolution of the two distributions given by (27): This concludes the proof.
The numerically simulated overall errors and the shape of the above theoretical distribution are plotted in Figure 5. This figure shows the high accuracy of the above derivation in predicting the shape of errors distribution.
Using Proposition 1, we can model optical flow calculation as (2), where is the slope, is the vertical intercept of the th linear equation, and , are the image velocities. Thus, the estimation error of velocity can be calculated as (31).
3.2. Definition of a Structure
Definition of a structure is a cornerstone of any data segmentation solution. In classical statistics, the definition of a structure is simpler than computer vision domain as the data is assumed to have only one structure and the structure always has the majority of data . In that context, a structure in data space is defined as a majority of data satisfying in which is a measure of the goodness for a data point, is a constant (normally set between 2 to 3 based on the desired significance level of the Gaussian distribution), and is an estimate of the scale of measurement noise.
The above represents a circular definition as the in the data space is measured using some attributes of the structure that are being defined. To bypass this issue, as shown in Figure 6(a), the estimation and segmentation are commonly conducted using random or guided sampling  and a good (in a statistical sense) sample is used in place of the true model. As it was mentioned earlier, this complicates the analysis of separability as the segmentation results would depend on the method by which the problem is solved. For instance, in the Hough Space segmentation, the definition of a structure and accuracy of segmentation outcomes would depend on the used histogram bandwidth for which the appropriate value is not known a priori.
To address the above issue our analysis, as shown in Figure 6(b), is conducted in the space of all pairwise samples. Our earlier derivation of the probability distribution of those samples enables the development a precise definition for a structure in this space. This definition is the key to solving the separability problem. To develop a unique (noncircular) definition, we use two basic principles that broadly define a cluster in any space.
The first principle guiding our definition is that a structure should be represented by the largest possible set of samples to include all the attributes of the modeled quantity. Also, any putative structure must include more than half of the overall samples to ensure the uniqueness of the definition for a given set of samples.
The second principle guiding our definition is that the probability of a given amount of error for a structure in the sample space should be less than the probability of the measurement noise for the same error value. This means that any data should always be more probable given the true model than any other model (sample).
To demonstrate the application of the above principles for defining a structure, we first find the crossover points by equating the derived error pdf (31) with the Gaussian function representing the measurement noise distribution. The result shows that error probabilities of samples up to times the measurement noise scale is smaller than the measurement noise probabilities () for all measurement noise scales. This fact, for the case of , is demonstrated by plotting those functions in Figure 7. Furthermore, the sample size of the above group (with ) is significantly larger than half. Exact values of the sample size and its variance up to crossover points are calculated here: The above calculations show that a very large majority (above ) of samples provide a precise representation of the underlying structure. Also, this definition is in-line to its traditional counterpart in the data space as we have (from (32)): or simply .
A structure in sample space is then formally defined as a cluster of at least half the samples where all of its samples satisfy . It is important to note that the above definition includes a large number of samples (more than half of the combinations of all samples which is significantly larger than the number of observations) and by definition is robust to influence of outliers [16, 17]. Also, the definition is independent of the distribution of samples and only depends on the scale of measurement noise (a design parameter) in the data space.
3.3. Segmentation Feasibility
Having derived the probability density function of error samples and provided a precise definition of a structure in the sample space, we can now examine the feasibility of the segmentation of the system of linear equations representing optical flows in two coherent systems by a simple cluster analysis. The separability of two systems of linear equations with close solutions can now be examined by considering the separability of the sample distributions. We note that two structures, as defined earlier, are separable if the distance between their means is at least , where and are the standard deviations of their associated measurement noise, respectively.
This statement is simply explained by looking at its contradiction. If we assume that the means of the distributions are already known, it would still not be possible to segment the data cleanly unless the two distributions have no overlap. This implies that the distance between two means has to be at least equal to sum of the extent of those distributions. This presents a sufficient condition for the separability which is able to predict the motion separability using only the amount of relative motion and the scale of noise in the measurement data. It is important to note here that the condition explicitly assumes that there are two distinct motions and in contrast to  (and subsequent works that followed this), the flow is not assumed to be varying smoothly.
To examine the usability of proposed theory for motion separability prediction, results of several experiments on standard video sequences for optical flow analysis are discussed. First, a set of controlled experiments using synthetically generated texture were conducted to simulate the separability of similar motions. Both the amount of relative motion and scale of measurement noise were changed in those simulations and the effect of those on the separability of existing motions were analyzed. Then, the application of the proposed theory for prediction of different motions in various video sequences (e.g., from Middlebury) with multiple motions were examined. The result shows that the proposed criterion is capable of predicting the separability of the motion of different objects.
Calculation of image derivatives is an important aspect of optical flow calculation and there are different ways to ensure that image derivatives are not affected by noise and aliasing . Our experiments however showed that although using multiresolution or relatively sophisticated image interpolation techniques (similar to ones used in ) improves the appearance of final results, the conclusions remained unchanged. Consequently, and for the sake of simplicity, spatial and temporal derivatives in our experiments were all calculated using convolution with Gaussian filters with the standard deviation of to pixels in all directions.
The usability of the proposed theory for motion separability predictions is examined here using a sinusoidal synthetic image sequence . The texture of the image sequence is generated by the superposition of two sinusoidal moving plane waves. The central square of the image is stationary, while the surrounding pixels are manipulated to exhibit different constant velocities. Figure 8 demonstrates a sample frame of the image sequence and highlights (by white rectangles) locations of two patches of size on both moving and stationary parts. In this simulation, the scale of noise for both patches is the same and therefore the separability condition is .
To simulate the effect of noise on the separability predictions, the optical flow constraints for the previously mentioned areas were perturbed by additive Gaussian noise.
In the first simulation, the normalized histogram of all samples for both patches containing two motions for different scales of added noise is shown in Figures 9(a)–9(c). Using the above separability condition , we would predict that when is less than , the above motions are separable. To show the validity of this prediction, the normalized histograms for values less, at, and above are shown. Those figures show that for , two motions are clearly separable.
In the second simulation, the result of changing the relative velocity for a given amount of noise () is examined. The normalized histogram for velocities less than, at, and greater than the predicted values are also shown in Figures 9(d)–9(f). These figures again show that as long as then the two motions are separable.
4.2. Real Image Experiments
As for real data usability, at the beginning, we raised the question of how to predict the least amount of required relative motion between different objects in the Marbled Block sequence (shown in Figure 1) that would make those separable. The proposed theory is now able to predict the separability of different objects using their motion information. Considering that the scale of noise for calculation of local optical flow in this sequence is measured to be around pixels/frame, the proposed separability condition, , predicts that if the relative motion is greater than , those motions are separable. In Figure 1 the relative motion between the two highlighted areas that are located on two different columns is around and therefore we expect those to be separable. Both the theoretical and actual distributions of the flow samples for the combined data are shown in Figure 10(a). In contrast, the maximum relative motion of the areas on one column is around and since it is less than the above separability threshold, those motions are expected to be inseparable. Again, the theoretical and actual distributions of the flow samples for the combined data of the two highlighted areas on a single column are shown in Figure 10(b).
To demonstrate the application of the proposed separability prediction criterion, the separability of different moving objects in a number of image sequences that are commonly used for motion analysis including four from the Middlebury  (called Urban2, Mequon, Grove2 and RubberWhale) benchmarks was examined. In all of those sequences, three patches on two different moving objects were chosen and those are highlighted in part (a) of Figures 11, 12,13, and 14. Patches that are on one object and have very similar motions are not expected to be separable. However, the other patch which is on a different object and has sufficiently different motion is expected to be separable from either of those patches. For each image sequence, both the measured and analytical error probability distributions for two different patches of two different objects (shown in parts (b) and (c) of Figures 11–14) as well as distributions of the joint patches of the same and different objects (shown in parts (d) and (e) of Figures 11–14) were plotted. Those plots show that the analytically derived probability for different patches is close to their real values. The numerical results of the above experiments in terms of relative motion between different patches, average scale of noise for the image sequence, and the separability verdict of pairs of patches are provided in Table 1.
The validity of the above predictions provides evidence that the proposed theory is able to correctly predict the separability of motion for real world applications.
A new theoretical framework to predict the feasibility of optical flow segmentation is presented. The framework enables the theoretical derivation of the optical flow estimation error probability density function as well as a precise definition for a visual structure based on its motion. The combination of these two elements is used to develop a segmentation feasibility criterion that can predict the separability of multiple motions. Applications of the theoretical results for the prediction of the separability of multiple motions were examined using both synthetic and real image sequences. The result illustrates that the proposed criterion is able to correctly predict the separability in those cases.
This research was supported under Australian Research Council’s Discovery Projects funding scheme (DP0878801). The authors declare that there is no conflict of interests regarding the publication of this paper.
- J. L. Barron, D. J. Fleet, and S. S. Beauchemin, “Performance of optical flow techniques,” International Journal of Computer Vision, vol. 12, no. 1, pp. 43–77, 1994.
- R. I. Hartley and A. Zisserman, Multiple View Geometry in Computer Vision, Cambridge University Press, 2nd edition, 2004.
- N. Gheissari, A. Bab-Hadiashar, and D. Suter, “Parametric model-based motion segmentation using surface selection criterion,” Computer Vision and Image Understanding, vol. 102, no. 2, pp. 214–226, 2006.
- A. Bruhn, J. Weickert, and C. Schnörr, “Lucas/Kanade meets Horn/Schunck: combining local and global optic flow methods,” International Journal of Computer Vision, vol. 61, no. 3, pp. 1–21, 2005.
- A. Bab-Hadiashar and D. Suter, “Robust optic flow computation,” International Journal of Computer Vision, vol. 29, no. 1, pp. 59–77, 1998.
- T.-J. Chin, J. Yu, and D. Suter, “Accelerated hypothesis generation for multistructure data via preference analysis,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 34, no. 4, pp. 625–638, 2012.
- A. Bruhn, J. Weickert, T. Kohlberger, and C. Schnörr, “A multigrid platform for real-time motion computation with discontinuity-preserving variational methods,” International Journal of Computer Vision, vol. 70, no. 3, pp. 257–277, 2006.
- T. Kohlberger, C. Schnörr, A. Bruhn, and J. Weickert, “Domain decomposition for variational optical-flow computation,” IEEE Transactions on Image Processing, vol. 14, no. 8, pp. 1125–1137, 2005.
- D. Sun, S. Roth, and M. J. Black, “Secrets of optical flow estimation and their principles,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '10), pp. 2432–2439, San Francisco, Calif, USA, June 2010.
- D. Sun, B. Sudderth, and M. J. Black, “Layered segmentation and optical flow estimation over times,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2012.
- R. Vidal, Y. Ma, and S. Sastry, “Generalized principal component analysis (GPCA),” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 12, pp. 1945–1959, 2005.
- Y. Weiss and E. H. Adelson, “Unified mixture framework for motion segmentation: incorporating spatial coherence and estimating the number of models,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 321–326, June 1996.
- Y.-H. Kim and A. C. Kak, “Error analysis of robust optical flow estimation by least median of squares methods for the varying illumination model,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 28, no. 9, pp. 1418–1435, 2006.
- J. Kybic and C. Nieuwenhuis, “Bootstrap optical flow confidence and uncertainty measure,” Computer Vision and Image Understanding, vol. 115, no. 10, pp. 1449–1462, 2011.
- C. V. Stewart, “Bias in robust estimation caused by discontinuities and multiple structures,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 8, pp. 818–833, 1997.
- R. Hoseinnezhad and A. Bab-Hadiashar, “Consistency of robust estimators in multi-structural visual data segmentation,” Pattern Recognition, vol. 40, no. 12, pp. 3677–3690, 2007.
- R. Hoseinnezhad, A. Bab-Hadiashar, and D. Suter, “Finite sample bias of robust estimators in segmentation of closely spaced structures: a comparative study,” Journal of Mathematical Imaging and Vision, vol. 37, no. 1, pp. 66–84, 2010.
- E. P. Simoncelli, E. H. Adelson, and D. J. Heeger, “Probability distributions of optical flow,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 310–315, June 1991.
- A. Goldenshluger and A. Zeevi, “The Hough transform estimator,” Annals of Statistics, vol. 32, no. 5, pp. 1908–1932, 2004.
- R. Dahyot, “Statistical hough transform,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 31, no. 8, pp. 1502–1509, 2009.
- S. Baker, D. Scharstein, J. P. Lewis, S. Roth, M. J. Black, and R. Szeliski, “A database and evaluation methodology for optical flow,” International Journal of Computer Vision, vol. 92, no. 1, pp. 1–31, 2011.
- M. Otte and H. Nagel, “Optical flow estimation: advances and comparisons,” in Proceedings of the Computer Vision (ECCV '94), pp. 49–60, 1994.
- K. Kanatani, Statistical Optimization for Geometric Computation: Theory and Practice, vol. 18, 1996.
- J. Bigun, G. H. Granlund, and J. Wiklund, “Multidimensional orientation estimation with applications to texture analysis and optical flow,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 13, no. 8, pp. 775–790, 1991.
- N. Ohta, “Optical flow detection using a general noise model for gradient constraint,” in Computer Analysis of Images and Patterns, vol. 1296 of Lecture Notes in Computer Science, pp. 669–676, 1997.
- V. K. Rohatgi, An Introduction to Probability Theory Mathematical Statistics, 1967.
- P. J. Rousseeuw and A. M. Leroy, Robust Regression and Outlier Detection, 1987.
- E. P. Simoncelli, “Design of multi-dimensional derivative filters,” in Proceedings of the IEEE International Conference Image Processing (ICIP '94), vol. 1, pp. 790–794, November 1994.