Link to paper

Pretty cool paper. While both typical assumptions of optical flow methods are rather well-motivated, there are important violations of both constraints that happen in the real world, both of which I’ve been thinking a lot about. There are

  1. Violations of brightness constancy, which happen with “transparency, specular reflections, shadows, or fragmented occlusion (for example, looking through the branches of the tree)“. I think this is a really underappreciated point in unsupervised optical flow.
  2. Violations of smoothness which happen at object boundaries, at which the optical flow is discontinuous. Currently, smoothness objectives create oversmoothing (as seen above).

How do we address these problems? Intuitively, it would make sense to model these explicitly and then try to reduce these. For the first, we can imagine perhaps prioritizing hue over other components (an idea I’ve been thinking about as I’ve been worrying about this), in order to reduce the penalty from shadows. For the second, we could create layered images in which the constraint is not violated within each image. However, these both suck, and they’re imperfect patches of the problem. (For each additional method of violation, you’d have to come up with another way to fix it, adding complexity to your model.)

Instead, Black & Anandan propose modifying the energy functional to be robust to outliers. As both constraints are satisfied in most of the image, we can model all exceptions, no matter their physical origin, as outliers (of different kinds). By increasing robustness to outliers, we can ideally minimize the impact of all of these simultaneously.

They use ideas from robust estimation to do this. The actual way they do this is stupid simple; they replace the losses within the H&S or L&K methods with truncated quadratics, or other robust functions such as these below:

The key features of these robust estimation functions are that as , . Note that the constraint is on the partial derivative of , not itself (a fact which I find surprising); i.e. in plain language, at some point, increasing the error doesn’t increase the loss much further. (The right-hand functions are proportional to .)

I think a lot of the time they consider locally constant/affine flow functions, so L&K might be closer to what they’re doing, but their loss function looks a lot like that of H&S.

The loss objective they use is

where is a robust function. They demonstrate the high performance of this function in the face of discontinuities, specular reflections, fragmented occlusions, etc.

Tags

Energy or loss minimization Robust estimation Robustness to noise Truncated quadratic objectives Multi-scale pyramids, coarse-to-fine Local methods NOT deep learning Old school Optical flow Smoothness constraints Unsupervised Affine, parameterized optical flow

Cited

Cited By

Return: index