Vincent Auvray - Contributions
(Back)
Quick insight of the transparent motion estimation
Three main approaches are proposed in the litterature to estimate transparent motions:
Partial adaptation of classical motion estimators (Black and Anandan, Irani and Peleg), but for partial transparencies only.
Detection of planes in the 3D Fourier space (Shizawa and Mase, Pingault and Pellerin, Stuke and Aach),
but it must assume that motions are translationals over dozen of frames.
Transparent motion estimation in the direct space (Pingault and Pellerin, Stuke and Aach, Toro), which relies on an
equation equivalent for the brightness consistancy. For two layers:
It assumes that the motions are constant over two successive time intervals.
Our contributions belong to this last class of approches. (More)
Our contributions in bitransparent configuration
In a first time, we have developed a synthetic image generation process which follows the physics of X-Ray
image formation. That allows us to compare our estimates to a known ground truth.
(More)
We have developped three generations of motion estimators dedicated to the bitransparency. This refers to a transparency
made of two layers over the image. (And two layers only!)
Our first contribution (presented to ICIP'05) did not allow us to reach an estimation accuracy
good enough on clinical sequences with a realistic noise.
This is why we imagined a second algorithm, presented at MICCAI'05. To be as robust to noise
as possible, we tried to constraint the problem as much as we could. Many real exams observations convinced us that the anatomical
motions (heart beating, lungs dilation, diaphragm translation) could be modeled with affine displacement fields.
This approach allows for precisions of 0.6 and 2.8 pixels on images typical for diagnostic and interventional images respectively.
We finally proposed in ICIP'06 an evolution of this algorithm, reaching precisions of 0.6
pixels on diagnostic images and 1.2 on fluoroscopic images, and presented real examples.
(More)(Demos)
Estimated motion fields on a part of cardiac exam in situation of bi-transparency.
Our contributions in bidistributed transparency
Our contribution until that point estimated the transparent motions present in a sequence showing one configuration only:
two layers having coherent motions were present over the whole image. Most of the contribution on this topic refer to this
situation.
Now, the real images are more complex: they contain more that two layers overall, but rarely more than two locally. Which is why
we have introduced the concept of bidistributed transparency to account for images that can be segmented in areas
containing one or two layers.
Fluoroscopic exam image, that can be segmented in an area heart+lungs, heart+diaphragm, heart+spine, etc.
Our observation of real sequences showing that a large majority of clinical exams results in images of this kind, we have developed
an approach to process them, which will be presented in ICIP'06. The idea is to separate
the input sequence in its different two-layers area, and to process them with the previous method.
More precisely, we propose a joint motion-based segmentation and motion estimation technique, presented in a Markovian
formalism. To determine both the velocity fields and the corresponding blocks labeling, we minimize an energy made of a data-driven
term based on the fundamental equation of the transparent motion presented above, and a smoothing term. It is iteratively minimized
using the Iterative Reweighted Least Square method to estimate robustly the parameters of the motion model when the labels are fixed,
and the ISM technique to estimate the labels once the motion parameters are fixed. Such a method is fast and reliable if correctly initialized.
To do so, we resort to a transparent block-matching followed by the fields extraction with the Hough transform.
This algorithm leads to precisions of 0.6 pixels on diagnostic images and 1.2 on fluoroscopic images.
Two images for a fluoroscopic sequence, the estimated motions (the field corresponding to the background is worth
zero and thus invisible) and the resulting segmentation.
Processing of a video sequence in a situation of bi-distributed transparency. Top left: the segmentation in layers,
top right the estimated motions, and on the bottom the compensated difference images.
We present above an example of estimation on clinical images, and a sequence of results on a video sequence with transparency.
It pictures a corn-flakes box reflected on the frame covering a Mona-Lisa patchwork. In both cases, segmentations and estimations
match the observation. We also present the sequence of compensated differences on the video sequence. It proves that T=the motion
estimation is correct since the corresponding layer disappears.
(Back)
We also present a demo, showing the different stages of motion estimation.
Application to denoising
We also need to know how to use this motion information to denoise our sequences. We have proposed an original framework to do so
in MICCAI'05.
. It yielded a very disappointing asymptotic deoising of 20%.
(More)
We have therefore developed the so-called hybrid filters, which decide locally between different possible motion compensations:
Transparent motion compensation when both layers are textured.
Compensation of the motion of the more textured layer in any other case (one layer only is textured, or both layers are homogeneous). This way,
we preserve the useful information without impairing the achievable denoising power.
This approach can be used both in purely temporal denoising filter, as well as in spatio-temporal denoising filters. In the former case, we achieve
a denoising of about 50% (in standart deviation), without useful information blurring (submitted to ISBI'07).
Fluoroscopic sequence processing with the temporal hybrid filter (left) and a classical adaptative temporal filter (right).
The heart is much more contrasted on the left for an equivalent denoising.
Back