Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Image registration
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
{{Short description|Mapping of data into a single system}} {{Confused|Printing registration}} [[Image:Registrator Demo2.png|thumb|right|Registering and summing multiple exposures of the same scene improve signal to noise ratio, allowing one to see things previously impossible to see. In this picture, the distant Alps are made visible, although they are tens of kilometers into the haze.]] '''Image registration''' is the process of transforming different sets of data into one coordinate system. Data may be multiple photographs, data from different sensors, times, depths, or viewpoints.<ref>[http://portal.acm.org/citation.cfm?id=146374 Lisa Gottesfeld Brown, A survey of image registration techniques (''abstract''), ACM Computing Surveys archive, volume 24, issue 4, December 1992), pages 325 - 376]</ref> It is used in [[computer vision]], [[medical imaging]],<ref>[http://www.nature.com/nmeth/journal/v8/n6/full/nmeth.1602.html biological imaging and brain mapping]</ref> military [[automatic target recognition]], and compiling and analyzing images and data from satellites. Registration is necessary in order to be able to compare or integrate the data obtained from these different measurements. == Algorithm classification == === Intensity-based vs feature-based === Image registration or image alignment algorithms can be classified into intensity-based and feature-based.<ref name="AG">A. Ardeshir Goshtasby: [http://www.wiley.com/WileyCDA/WileyTitle/productCd-0471649546.html 2-D and 3-D Image Registration for Medical, Remote Sensing, and Industrial Applications], Wiley Press, 2005.</ref> One of the images is referred to as the ''target'', ''fixed'' or ''sensed'' image and the others are referred to as the ''moving'' or ''source'' images. Image registration involves spatially transforming the source/moving image(s) to align with the target image. The reference frame in the target image is stationary, while the other datasets are transformed to match to the target.<ref name="AG"/> Intensity-based methods compare intensity patterns in images via correlation metrics, while feature-based methods find [[correspondence problem|correspondence]] between image features such as points, lines, and contours.<ref name="AG"/> Intensity-based methods register entire images or sub-images. If sub-images are registered, centers of corresponding sub images are treated as corresponding feature points. Feature-based methods establish a correspondence between a number of especially distinct points in images. Knowing the correspondence between a number of points in images, a geometrical transformation is then determined to map the target image to the reference images, thereby establishing point-by-point correspondence between the reference and target images.<ref name="AG"/> Methods combining intensity-based and feature-based information have also been developed.<ref name="PapademetrisJackowski2004">{{cite book|last1=Papademetris|first1=Xenophon|title=Medical Image Computing and Computer-Assisted Intervention – MICCAI 2004|last2=Jackowski|first2=Andrea P.|last3=Schultz|first3=Robert T.|last4=Staib|first4=Lawrence H.|last5=Duncan|first5=James S.|chapter=Integrated Intensity and Point-Feature Nonrigid Registration|volume=3216|year=2004|pages=763–770|issn=0302-9743|doi=10.1007/978-3-540-30135-6_93|series=Lecture Notes in Computer Science|isbn=978-3-540-22976-6}}</ref> === Transformation models === Image registration algorithms can also be classified according to the transformation models they use to relate the target image space to the reference image space. The first broad category of transformation models includes [[linear transformation]]s, which include rotation, scaling, translation, and other affine transforms.<ref>http://www.comp.nus.edu.sg/~cs4243/lecture/register.pdf {{Bare URL PDF|date=March 2022}}</ref> [[Linear transformation]]s are global in nature, thus, they cannot model local geometric differences between images.<ref name="AG"/> The second category of transformations allow 'elastic' or 'nonrigid' transformations. These transformations are capable of locally warping the target image to align with the reference image. Nonrigid transformations include [[radial basis functions]] ([[Thin plate spline|thin-plate]] or surface splines, [[Radial basis functions#Types|multiquadrics]], and [[support (mathematics)#Compact support|compactly-supported transformations]]<ref name="AG"/>), physical continuum models (viscous fluids), and large deformation models ([[diffeomorphism]]s). Transformations are commonly described by a parametrization, where the model dictates the number of parameters. For instance, the translation of a full image can be described by a single parameter, a translation vector. These models are called parametric models. Non-parametric models on the other hand, do not follow any parameterization, allowing each image element to be displaced arbitrarily.<ref>{{cite journal|last1=Sotiras|first1=A.|last2=Davatzikos|first2=C.|last3=Paragios|first3=N.|title=Deformable Medical Image Registration: A Survey|journal=IEEE Transactions on Medical Imaging|date=July 2013|volume=32|issue=7|pages=1153–1190|doi=10.1109/TMI.2013.2265603|pmid=23739795|pmc=3745275|url=https://hal.inria.fr/hal-00858737/document}}</ref> There are a number of programs that implement both estimation and application of a warp-field. It is a part of the [[Statistical parametric mapping|SPM]] and [[AIR (program)|AIR]] programs. === Transformations of coordinates via the law of function composition rather than addition === Alternatively, many advanced methods for spatial normalization are building on structure preserving transformations [[homeomorphism]]s and [[diffeomorphism]]s since they carry smooth submanifolds smoothly during transformation. Diffeomorphisms are generated in the modern field of [[Computational anatomy|Computational Anatomy]] based on flows since diffeomorphisms are not additive although they form a group, but a group under the law of function composition. For this reason, flows which generalize the ideas of additive groups allow for generating large deformations that preserve topology, providing 1-1 and onto transformations. Computational methods for generating such transformation are often called [[Large deformation diffeomorphic metric mapping|LDDMM]]<ref>{{Cite book|url=https://books.google.com/books?id=8WdlWJepgWMC|title=Brain Warping|last=Toga|first=Arthur W.|date=1998-11-17|publisher=Academic Press|isbn=9780080525549|language=en}}</ref><ref>{{Cite web|url=https://utah.pure.elsevier.com/en/publications/landmark-matching-on-brain-surfaces-via-large-deformation-diffeom-2|title=Landmark matching on brain surfaces via large deformation diffeomorphisms on the sphere — University of Utah|website=utah.pure.elsevier.com|access-date=2016-03-21|archive-url=https://web.archive.org/web/20180629235930/https://utah.pure.elsevier.com/en/publications/landmark-matching-on-brain-surfaces-via-large-deformation-diffeom-2|archive-date=2018-06-29|url-status=dead}}</ref><ref>{{Cite journal|url=https://www.researchgate.net/publication/220660081|title=Computing Large Deformation Metric Mappings via Geodesic Flows of Diffeomorphisms|journal=International Journal of Computer Vision|volume=61|issue=2|pages=139–157|doi=10.1023/B:VISI.0000043755.93987.aa|access-date=2016-03-21|year=2005|last1=Beg|first1=M. Faisal|last2=Miller|first2=Michael I.|last3=Trouvé|first3=Alain|last4=Younes|first4=Laurent|s2cid=17772076}}</ref><ref>{{Cite journal|last1=Joshi|first1=S. C.|last2=Miller|first2=M. I.|date=2000-01-01|title=Landmark matching via large deformation diffeomorphisms|journal=IEEE Transactions on Image Processing|volume=9|issue=8|pages=1357–1370|doi=10.1109/83.855431|issn=1057-7149|pmid=18262973|bibcode=2000ITIP....9.1357J}}</ref> which provide flows of diffeomorphisms as the main computational tool for connecting coordinate systems corresponding to [[Computational anatomy#The metric on geodesic flows of landmarks, surfaces, and volumes within the orbit|the geodesic flows of Computational Anatomy]]. There are a number of programs which generate diffeomorphic transformations of coordinates via [[Diffeomorphic Mapping in Computational Anatomy|diffeomorphic mapping]] including MRI Studio<ref>{{Cite web|url=https://www.mristudio.org|title=MRI Studio}}</ref> and MRI Cloud.org<ref>{{Cite web|url=https://mricloud.org/|title=MRICloud Brain Mapping}}</ref> === Spatial vs frequency domain methods === Spatial methods operate in the image domain, matching intensity patterns or features in images. Some of the feature matching algorithms are outgrowths of traditional techniques for performing manual image registration, in which an operator chooses corresponding [[Feature (computer vision)|control points]] (CP) in images. When the number of control points exceeds the minimum required to define the appropriate transformation model, iterative algorithms like [[RANSAC]] can be used to robustly estimate the parameters of a particular transformation type (e.g. affine) for registration of the images. Frequency-domain methods find the transformation parameters for registration of the images while working in the transform domain. Such methods work for simple transformations, such as translation, rotation, and scaling. Applying the [[phase correlation]] method to a pair of images produces a third image which contains a single peak. The location of this peak corresponds to the relative translation between the images. Unlike many spatial-domain algorithms, the phase correlation method is resilient to noise, occlusions, and other defects typical of medical or satellite images. Additionally, the phase correlation uses the [[fast Fourier transform]] to compute the cross-correlation between the two images, generally resulting in large performance gains. The method can be extended to determine rotation and scaling differences between two images by first converting the images to [[log-polar coordinates]].<ref>{{cite journal|author=B. Srinivasa Reddy|author2=B. N. Chatterji|title=An FFT-Based Technique for Translation, Rotation and Scale-Invariant Image Registration|journal=IEEE Transactions on Image Processing|volume=5|issue=8|pages=1266–1271|date=Aug 1996|doi=10.1109/83.506761|pmid=18285214|bibcode=1996ITIP....5.1266R |s2cid=6562358}}</ref><ref>Zokai, S., Wolberg, G., [https://www-cs.ccny.cuny.edu/~wolberg/pub/tip05.pdf "Image Registration Using Log-Polar Mappings for Recovery of Large-Scale Similarity and Projective Transformations"]. ''IEEE Transactions on Image Processing'', vol. 14, No. 10, October, 2005.</ref> Due to properties of the [[Fourier transform]], the rotation and scaling parameters can be determined in a manner invariant to translation. === Single- vs multi-modality methods === Another classification can be made between single-modality and multi-modality methods. Single-modality methods tend to register images in the same modality acquired by the same scanner/sensor type, while multi-modality registration methods tended to register images acquired by different scanner/sensor types. Multi-modality registration methods are often used in [[medical imaging]] as images of a subject are frequently obtained from different scanners. Examples include registration of brain [[Computed tomography|CT]]/[[MRI]] images or whole body [[Positron emission tomography|PET]]/[[Computed tomography|CT]] images for tumor localization, registration of contrast-enhanced [[Computed tomography|CT]] images against non-contrast-enhanced [[Computed tomography|CT]] images<ref>{{Cite journal |last1=Ristea |first1=Nicolae-Catalin |last2=Miron |first2=Andreea-Iuliana |last3=Savencu |first3=Olivian |last4=Georgescu |first4=Mariana-Iuliana |last5=Verga |first5=Nicolae |last6=Khan |first6=Fahad Shahbaz |last7=Ionescu |first7=Radu Tudor |title=Cy ''Tran'': A cycle-consistent transformer with multi-level consistency for non-contrast to contrast CT translation |journal=Neurocomputing |year=2023 |volume=538 |page=126211 |doi=10.1016/j.neucom.2023.03.072 |arxiv=2110.06400 |s2cid=257952429 }}</ref> for segmentation of specific parts of the anatomy, and registration of [[ultrasound]] and [[Computed tomography|CT]] images for [[prostate]] localization in [[radiotherapy]]. === Automatic vs interactive methods === Registration methods may be classified based on the level of automation they provide. Manual, interactive, semi-automatic, and automatic methods have been developed. Manual methods provide tools to align the images manually. Interactive methods reduce user bias by performing certain key operations automatically while still relying on the user to guide the registration. Semi-automatic methods perform more of the registration steps automatically but depend on the user to verify the correctness of a registration. Automatic methods do not allow any user interaction and perform all registration steps automatically. === Similarity measures for image registration === Image similarities are broadly used in [[medical imaging]]. An image [[similarity measure]] quantifies the degree of similarity between intensity patterns in two images.<ref name="AG"/> The choice of an image similarity measure depends on the modality of the images to be registered. Common examples of image similarity measures include [[cross-correlation]], [[mutual information]], sum of squared intensity differences, and ratio image uniformity. Mutual information and normalized mutual information are the most popular image similarity measures for registration of multimodality images. Cross-correlation, sum of squared intensity differences and ratio image uniformity are commonly used for registration of images in the same modality. Many new features have been derived for cost functions based on matching methods via [[Computational anatomy|large deformations]] have emerged in the field [[Computational anatomy|Computational Anatomy]] including [[Computational anatomy#Measure matching: unregistered landmarks|Measure matching]] which are pointsets or landmarks without correspondence, [[Computational anatomy#Curve matching|Curve matching]] and [[Computational anatomy#Surface matching|Surface matching]] via mathematical [[Current (mathematics)|currents]] and varifolds. == Uncertainty == There is a level of [[uncertainty]] associated with registering images that have any spatio-temporal differences. A confident registration with a measure of uncertainty is critical for many [[change detection]] applications such as medical diagnostics. In [[remote sensing]] applications where a digital image pixel may represent several kilometers of spatial distance (such as NASA's [[LANDSAT]] imagery), an uncertain image registration can mean that a solution could be several kilometers from ground truth. Several notable papers have attempted to quantify uncertainty in image registration in order to compare results.<ref name="simonson">Simonson, K., Drescher, S., Tanner, F., A Statistics Based Approach to Binary Image Registration with Uncertainty Analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 29, No. 1, January 2007</ref><ref name="Domokos">Domokos, C., Kato, Z., Francos, J., Parametric estimation of affine deformations of binary images. Proceedings of IEEE [[International Conference on Acoustics, Speech, and Signal Processing]], 2008</ref> However, many approaches to quantifying uncertainty or estimating deformations are computationally intensive or are only applicable to limited sets of spatial transformations. == Applications == [[File:Mni-autoreg 03-registered.png|thumb|Registration of two [[MRI]] images of the brain]] Image registration has applications in remote sensing (cartography updating), and computer vision. Due to the vast range of applications to which image registration can be applied, it is impossible to develop a general method that is optimized for all uses. [[Medical imaging|Medical image]] registration (for data of the same patient taken at different points in time such as change detection or tumor monitoring) often additionally involves ''elastic'' (also known as ''nonrigid'') registration to cope with deformation of the subject (due to breathing, anatomical changes, and so forth).<ref>{{Cite journal |last1=Zhao |first1=Shengyu |last2=Lau |first2=Tingfung |last3=Luo |first3=Ji |last4=Chang |first4=Eric I-Chao |last5=Xu |first5=Yan |year=2020 |title=Unsupervised 3D End-to-End Medical Image Registration With Volume Tweening Network |url=https://ieeexplore.ieee.org/document/8889674 |journal=IEEE Journal of Biomedical and Health Informatics |volume=24 |issue=5 |pages=1394–1404 |doi=10.1109/JBHI.2019.2951024 |pmid=31689224 |arxiv=1902.05020 |s2cid=61153704 |issn=2168-2208}}</ref><ref>{{Cite arXiv |last1=Chen |first1=Junyu |last2=He |first2=Yufan |last3=Frey |first3=Eric C. |last4=Li |first4=Ye |last5=Du |first5=Yong |date=2021-04-13 |title=ViT-V-Net: Vision Transformer for Unsupervised Volumetric Medical Image Registration |class=eess.IV |eprint=2104.06468 }}</ref><ref>{{Cite book |last1=Burduja |first1=Mihail |last2=Ionescu |first2=Radu Tudor |title=2021 IEEE International Conference on Image Processing (ICIP) |chapter=Unsupervised Medical Image Alignment with Curriculum Learning |year=2021 |chapter-url=https://ieeexplore.ieee.org/document/9506067 |pages=3787–3791 |doi=10.1109/ICIP42928.2021.9506067|arxiv=2102.10438 |isbn=978-1-6654-4115-5 |s2cid=231986287 }}</ref> Nonrigid registration of medical images can also be used to register a patient's data to an anatomical atlas, such as the [[Talairach coordinates|Talairach]] atlas for neuroimaging. In other cases, nonrigid registration is explicitly not utilized since rigid registration methods preserve the underlying geometry, e.g., in [[inner ear]] imaging.<ref name="DOI10.1038/s41598-025-90842-2">J. Gerb, V. Kirsch, E. Kierig, Thomas Brandt, Marianne Dieterich, Rainer Boegle: ''Optimizing spatial normalization of multisubject inner ear MRI: comparison of different geometry-preserving co-registration approaches.'' In: ''Scientific Reports.'' 2025, Band 15, Nummer 1 {{DOI|10.1038/s41598-025-90842-2}}.</ref> In [[Radiation therapy]] rigid image registration (RIR) is a fundamental element in most imaging software systems. It involves aligning images by applying translational and rotational adjustments, up to a six degrees of freedom—three for translation along the x, y, and z axes, and three for rotation about these axes. RIR aligns images appropriately using these six parameters, allowing for precise treatment planning and delivery.<ref>{{Cite journal |last=Yuen |first=Johnson |last2=Barber |first2=Jeffrey |last3=Ralston |first3=Anna |last4=Gray |first4=Alison |last5=Walker |first5=Amy |last6=Hardcastle |first6=Nicholas |last7=Schmidt |first7=Laurel |last8=Harrison |first8=Kristie |last9=Poder |first9=Joel |last10=Sykes |first10=Jonathan R. |last11=Jameson |first11=Michael G. |date=2020 |title=An international survey on the clinical use of rigid and deformable image registration in radiotherapy |url=https://aapm.onlinelibrary.wiley.com/doi/10.1002/acm2.12957 |journal=Journal of Applied Clinical Medical Physics |language=en |volume=21 |issue=10 |pages=10–24 |doi=10.1002/acm2.12957 |issn=1526-9914 |pmc=7075391 |pmid=32915492}}</ref> In [[astrophotography]], image alignment and stacking are often used to increase the signal to noise ratio for faint objects. Without stacking it may be used to produce a timelapse of events such as a planet's rotation of a transit across the Sun. Using control points (automatically or manually entered), the computer performs transformations on one image to make major features align with a second or multiple images. This technique may also be used for images of different sizes, to allow images taken through different telescopes or lenses to be combined. In [[cryo-TEM]], instability causes specimen drift and many fast acquisitions with accurate image registration is required to preserve high resolution and obtain high signal to noise images. For low SNR data, the best image registration is achieved by cross-correlating all permutations of images in an image stack.<ref name=Savitsky>{{cite journal|title=Image registration of low signal-to-noise cryo-STEM data|year=2018|author1=Savitsky |author2=El Baggari |author3=Clement |author4=Hovden |author5=Kourkoutis |journal=Ultramicroscopy|volume=191|pages=56–65|doi=10.1016/j.ultramic.2018.04.008|pmid=29843097|arxiv=1710.09281|s2cid=26983019}}</ref> Image registration is an essential part of panoramic image creation. There are many different techniques that can be implemented in real time and run on embedded devices like cameras and camera-phones. == See also == {{Portal|Free and open-source software}} * [[Computational Anatomy]] * [[Correspondence problem]] * [[Digital image correlation and tracking]] * [[Georeferencing]] * [[Image correlation]] * [[Image rectification]] * [[Inverse consistency]] * [[Point set registration]] * [[Rubbersheeting]] * [[Spatial normalization]] * [[Spatial verification]] ==References== <references/> ==External links== {{Commons category|Image registration}} *Richard Szeliski, [http://research.microsoft.com/apps/pubs/default.aspx?id=75695 Image Alignment and Stitching: A Tutorial]. Foundations and Trends in Computer Graphics and Computer Vision, 2:1-104, 2006. * B. Fischer, J. Modersitzki: [http://iopscience.iop.org/article/10.1088/0266-5611/24/3/034008/meta Ill-posed medicine – an introduction to image registration]. Inverse Problems, 24:1–19, 2008 * Barbara Zitová, Jan Flusser: [http://library.utia.cas.cz/prace/20030125.pdf Image registration methods: a survey]. Image Vision Comput. 21(11): 977-1000 (2003). * C. Je and H.-M. Park. [https://dx.doi.org/10.1016/j.image.2013.04.002 Optimized Hierarchical Block Matching for Fast and Accurate Image Registration]. Signal Processing: Image Communication, Volume 28, Issue 7, pp. 779–791, August, 2013. * [http://www.mathworks.com/help/images/registering-multimodal-mri-images.html Registering Multimodal MRI Images] using [[Matlab]]. * [http://elastix.isi.uu.nl/ elastix] {{Webarchive|url=https://web.archive.org/web/20120419053714/http://elastix.isi.uu.nl/ |date=2012-04-19 }}: a toolbox for rigid and nonrigid registration of images. * [http://sourceforge.net/projects/niftyreg niftyreg]: a toolbox for doing near real-time robust rigid, affine (using block matching) and non-rigid image registration (using a refactored version of the free form deformation algorithm). * [http://www.mathworks.com/discovery/image-registration.html Image Registration techniques using MATLAB] * [https://www.robots.ox.ac.uk/~vgg/software/image-compare/ Image Compare] application automatically compares a pair of images and highlights their differences. This application runs in desktop and mobile phone browsers without requiring installation. {{DEFAULTSORT:Image Registration}} [[Category:Computer vision]] [[Category:Medical imaging]]
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)
Pages transcluded onto the current version of this page
(
help
)
:
Template:Bare URL PDF
(
edit
)
Template:Cite arXiv
(
edit
)
Template:Cite book
(
edit
)
Template:Cite journal
(
edit
)
Template:Cite web
(
edit
)
Template:Commons category
(
edit
)
Template:Confused
(
edit
)
Template:DOI
(
edit
)
Template:Doi
(
edit
)
Template:Portal
(
edit
)
Template:Short description
(
edit
)
Template:Sister project
(
edit
)
Template:Webarchive
(
edit
)