In:
Human Brain Mapping, Wiley, Vol. 36, No. 4 ( 2015-04), p. 1365-1380
Abstract:
Registration performance can significantly deteriorate when image regions do not comply with model assumptions. Robust estimation improves registration accuracy by reducing or ignoring the contribution of voxels with large intensity differences, but existing approaches are limited to monomodal registration. In this work, we propose a robust and inverse‐consistent technique for cross‐modal, affine image registration. The algorithm is derived from a contextual framework of image registration. The key idea is to use a modality invariant representation of images based on local entropy estimation, and to incorporate a heteroskedastic noise model. This noise model allows us to draw the analogy to iteratively reweighted least squares estimation and to leverage existing weighting functions to account for differences in local information content in multimodal registration. Furthermore, we use the nonparametric windows density estimator to reliably calculate entropy of small image patches. Finally, we derive the Gauss–Newton update and show that it is equivalent to the efficient second‐order minimization for the fully symmetric registration approach. We illustrate excellent performance of the proposed methods on datasets containing outliers for alignment of brain tumor, full head, and histology images. Hum Brain Mapp 36:1365–1380, 2015 . © 2014 Wiley Periodicals, Inc .
Type of Medium:
Online Resource
ISSN:
1065-9471
,
1097-0193
Language:
English
Publisher:
Wiley
Publication Date:
2015
detail.hit.zdb_id:
1492703-2
Permalink