Abstract
Purpose
The registration of a 3D atlas image to 2D radiographs enables 3D preoperative planning without the need to acquire costly and highdose CTscans. Recently, many deeplearningbased 2D/3D registration methods have been proposed which tackle the problem as a reconstruction by regressing the 3D image immediately from the radiographs, rather than registering an atlas image. Consequently, they are less constrained against unfeasible reconstructions and have no possibility to warp auxiliary data. Finally, they are, by construction, limited to orthogonal projections.
Methods
We propose a novel endtoend trainable 2D/3D registration network that regresses a dense deformation field that warps an atlas image such that the forward projection of the warped atlas matches the input 2D radiographs. We effectively take the projection matrix into account in the regression problem by integrating a projective and inverse projective spatial transform layer into the network.
Results
Comprehensive experiments conducted on simulated DRRs from patient CT images demonstrate the efficacy of the network. Our network yields an average Dice score of 0.94 and an average symmetric surface distance of 0.84 mm on our test dataset. It has experimentally been determined that projection geometries with 80\(^{\circ }\) to 100\(^{\circ }\) projection angle difference result in the highest accuracy.
Conclusion
Our network is able to accurately reconstruct patientspecific CTimages from a pair of nearorthogonal calibrated radiographs by regressing a deformation field that warps an atlas image or any other auxiliary data. Our method is not constrained to orthogonal projections, increasing its applicability in medical practices. It remains a future task to extend the network for uncalibrated radiographs.
Introduction
Radiography or Xray imaging is the most common imaging procedure for many orthopaedic interventions thanks to its ability to visualise internal structures with a relatively low radiation dose and low acquisition cost. Apart from diagnosis, it is a valuable imaging technique for intraoperative guidance and postoperative evaluation. It also plays a crucial role in preoperative surgical planning and the selection of the right implants. In case of total hip arthroplasty surgeries, for example, it has been shown that the proper positioning and orientation of the acetabular component largely determines the functional outcome of the implant [1, 2]. Thereby, it is essential that parameters such as the centre of rotation of the hip joint, leg length, and hip offset remain preserved after the surgery and thus are correctly assessed on the radiographs.
Although many surgical planning tools rely on twodimensional (2D) radiographs, their clinical interpretation can be hampered by overlapping structures and magnification effects. The assessment from radiographs can also be influenced by the patient’s positioning. To avoid the difficulties associated with 2D projections, threedimensional (3D) computed tomography (CT) images are preferred for surgical planning because they are less ambiguous [3]. They also allow to study the cortical and cancellous bone, in addition to the outer bone surface [4]. CTbased planning, however, is associated with higher radiation doses and to far more expensive image acquisitions. Previous research has therefore suggested the reconstruction of a patientspecific 3D model from two or more 2D radiographs by registering a 3D CT atlas image to 2D radiographs, referred to as 2D/3D registration [4].
Related work
Recently, deeplearning (DL) methods have been proposed that reconstruct a 3D image from 2D radiographs by means of a neural network that encodes the 2D radiographs into a latent variable which is decoded into a 3D CT volume [5,6,7,8]. Compared to 3D/3D registrations, these networks need to bridge between the different dimensionalities in the encoder and decoder, which can be done by reshaping the 2D feature maps [5] or by treating the feature channel as the third spatial dimension [6]. Others exploit the orthogonality between biplanar projections by copying each feature map along a different dimension [7]. The X2CTGAN network uses two different mechanisms to bridge the dimensionalities [8]. They apply a fully connected layer on the flattened latent variable, before applying a nonlinear activation function and reshaping it into a 3D feature map. For the skip connections, they apply 2D convolutions on the 2D feature maps, which are then copied along the third axis and fed into a 3D convolutional layer.
In this paper, we propose a novel atlasbased 2D/3D registration network that estimates a registration field based on a pair of calibrated radiographs. The main contributions of our proposed method are as follows:

It follows a registration approach instead of a reconstruction approach, by regressing a deformation field which can be used to warp an atlas or any auxiliary data like segmentation maps. This avoids an additional segmentation step to extract a surface model.

It decomposes the total registration function into an affine and a local part in order to reduce restrictions on the orientation of input data.

It is not restricted to orthogonal projections, unlike other DLmethods in the literature. To this end, we propose an invProST layer to better combine bidirectional feature maps, as an extension to [9].

It is validated on simulated digitally reconstructed radiographs (DRRs) from a large collection of patient CT images, and compared to other registration approaches in the literature [7, 10].
Methodology
Registration network architecture
Overview of network
The registration network, shown in Fig. 1, estimates a registration field \(\Psi \) that maps the atlas image V (with associated label map S) to the moving image space, such that the forward projection of the warped atlas, \(V\circ \Psi \), matches the input radiographs \(I_i\), with \(i \in \{\text {AP}, \text {LAT}\}\). The registration field \(\Psi \) can be decomposed into an affine transformation T and a local backwards deformation field \(\phi \). Both transformations are separately regressed by two sequential network modules and composed at the end of the network to yield the total deformation field \(\Psi = \phi + T \circ \phi \), which is used to warp the atlas image.
Projective spatial transform layer
The projective spatial transformer (ProST), introduced by Gao et al. [9], simulates a 2D perspective projection \({\hat{I}} \in \mathrm{I\!R}^{S_x\times S_y }\) from a 3D volume V by sampling this volume at grid locations \(G \in \mathrm{I\!R}^{S_x\times S_y \times K }\). The grid consists of K sampling points, uniformly distributed along each ray connecting the Xray source location to each pixel of the 2D detector. This canonical grid can be transformed by an affine transformation \(T_{geom}\) in order to represent the actual projection geometry. This projection geometry transformation \(T_{geom}\) is known for calibrated radiographs and serves as input parameter to the network. The 3D volume V can be interpolated at the transformed grid positions \(T_{geom}(G)\) to obtain an Xray beamaligned image volume \(V_{beam} \in \mathrm{I\!R}^{S_x\times S_y \times K }\) in the beamspace:
The conebeam projection is then obtained by integration along each ray, which is equivalent to a “parallel projection” of the interpolated volume:
Affine registration module
The affine registration network consists of two ProST layers which project the 3D atlas image along the AP and lateral direction. The ProST output \({\hat{I}}_i\) and the input radiograph \(I_i\) are concatenated into a 2channel 2D image and fed into a 2D encoder, corresponding to the ith projection direction. Each of the five encoder levels consists of a strided convolution, a batchnormalisation layer and a leaky rectified linear activation unit (LeakyReLU). Each level reduces the spatial size of the feature map by a factor two and doubles the number of features. At each encoder level, the AP and lateral features (and the preceding combined features) are concatenated and convolved.
The accumulated 2D feature map at the last encoder level is flattened and fed into a dense layer which regresses the seven parameters of the affine transformation T between the floating image and the atlas. The bias and kernel weights of the dense layer are initialised by zero and a narrow normal distribution, respectively, such that the initial affine transformation during training is close to identity. A spatial transform layer warps the atlas image V by the affine transformation T [11], before being fed to the local registration network.
Local registration module
The local registration network consists of two separated Unetshaped networks, each associated with a different projection direction. Each Unetshaped network is composed of a 2D encoder and 3D decoder and is preceded by a ProST layer that projects the affine transformed atlas image. Each level of the 2D encoder consists of a strided and nonstrided 2D convolution. By consequence, each level halves the spatial size and doubles the number of features of the feature maps. After each 2D convolution, a batch normalisation and a LeakyReLU activation are applied. The last 2D feature map is copied \(M=4\) times along the first dimension to obtain a 3D feature map.
The spatial dimensions of the 3D feature map are increased by the 3D decoder, while reducing the number of features as follows: [64, 32, 32, 16, 16, 16]. Each decoding step applies a 3D convolution with stride one and a LeakyReLU activation, followed by upsampling the feature map by a factor of two. The 3D feature maps are defined in the beamspace, which gives a natural meaning to the above operations. While stacking 2D maps corresponds to increasing the number of sampling points per ray, the upsampling also increases the number of rays.
The network has skip connections between the 2D encoder and 3D decoder at each resolution level of the Ushaped network in order to recover spatial information loss that might have happened during downsampling. Along each skip connection, the 2D feature maps are copied along the first dimension a number of times, such that its shape corresponds to the 3D decoded feature map’s shape. After copying the feature map, the skip connection applies a 3D convolution, a batch normalisation, and a LeakyReLU activation to the feature map.
invProST
The decoded 3D feature maps are defined in the beamspace and need to be converted to physical space to align them with each other before combining them. Therefore, we apply an “invProST” layer to the 3D feature maps, which samples the feature maps at locations \(G^{1}\):
with \(G^{1}\) the canonical sampling coordinates in the beam space, which are determined by the length of the rays connecting the source location with each voxel and by the intersection point of those rays with the detector plane.
It can be verified that successively applying the ProST of Eq. (1) and invProST of Eq. (3) on an image volume V results in approximately the same image V apart from interpolation approximations. Only voxels in the original image that fall outside the conebeam become zero in the final image.
After the invProST layer, the output tensors of the AP and the lateral network branches can be combined by concatenation, and convolved into a 3channel tensor which is interpreted as a stationary velocity field. This velocity field is integrated by a “scaling and squaring”method to obtain a diffeomorphic deformation field \(\phi \) [11, 12].
Semisupervised learning
The training of the network is semisupervised, which means that the training of the network relies on auxiliary data. In our experiment, the segmentation labels of the groundtruth CT volumes were used to mask the image volumes before feeding them into the network, as we are only interested in reconstructing the femur bone from the radiograph images.
The registration quality of the endtoend network during training and validation is quantified by a loss function that consists of a normalised crosscorrelation (NCC) function between the groundtruth image volume \(V^f\) and the warped atlas image V after the affine and local registration. Furthermore, it contains a regularisation term on the smoothness of the local deformation field:
The hyperparameter \(\delta =0.01\) balances the contribution between the smoothness term and the image similarity loss. Note that the loss function does not include the label maps anymore, as the images themselves are already masked.
Experiments
In this section, we evaluate the performance of our network. “Experimental settings” section discusses the generation of the different datasets and provides details on the evaluation and training procedure. “Experimental results” section presents the registration results for AP and lateral radiographs while comparing with other methods. We also report the sensitivity to inaccurate input parameters and the accuracy for nonorthogonal projections. An ablation study is presented in “Ablation study”.
Experimental settings
CTdata preprocessing and augmentation
A total of 315 angioCT images were acquired and split into a training set of 235 subjects, a validation set of 40 subjects for model selection, and a test set of 40 subjects used to report performance. From each CTimage, the left and right femurs were extracted and rotated to a reference system that aligns the anteriorposterior and lateral views of that femur with the x and yaxis of the image. The femur reference frame of each image was defined based on the neck and shaft axis of the femur. To allow some pose variation around this canonical reference pose, we applied random affine transformations to the image with strict constraints. The randomised angles were allowed within a range of 10\(^{\circ }\) extension/flexion, 10\(^{\circ }\) abduction/adduction and 10\(^{\circ }\) internal/external rotation.
After transforming the images to a pose that is close to that of the reference, the images were cropped around the femoral heads and resized in order to maintain the highest resolution as possible. The left femur images were flipped to resemble right ones. The final CT volumes have a size equal to \((192 \times 128 \times 192)\), and a resolution of \((0.664 \times 0.664 \times 1)\) mm\(^3\). Each image has a corresponding segmentation map S, obtained by graphcut segmentation method followed by manual corrections [13].
Generating DRR
Digitally reconstructed radiographs (DRR) were simulated from the femurcentred CT volumes by DeepDRR software [14]. DRRs were created with an image size of (422 \(\times \) 640), and downsampled to (160 \(\times \) 224) to fit the network’s input size. The sourcedetector distance and the isocentre distance of the projection geometry were fixed to 1000 mm and 925 mm, respectively. Two different datasets of DRRs were generated:

A dataset with orthogonal projections. The projection geometry was fixed to provide lateral and AP projections. The acquisition geometry corresponding to this dataset resembles best the experimental settings in the literature.

A dataset with generalised projection geometries. Projection matrices were parameterised by the left/right anterior oblique (LAO/RAO) angle \(\theta \), which was randomly varied between − 30\(^{\circ }\) and \(+\) 30\(^{\circ }\), around the perfect lateral and AP view. The craniocaudal angle was set to a constant value of 0\(^{\circ }\). Different combinations of LAO/RAO angles were made for biplanar experiments.
For both datasets, the CT label maps were projected along with the CT images to obtain a 2D labelmap for the DRRs. The DRRs were masked by these labelmaps before feeding them into the network. Note that other structures, in front and behind the femur, are still visible in the masked DRRs.
Evaluation metrics
The registration accuracy of the network is evaluated by means of the Dice score and the Jacard coefficient, which measure the overlap between the warped atlas label map and the groundtruth label map [15]:
We also report the average symmetric surface distance (ASSD), which measures the average geometric distance between the groundtruth and registered bone surfaces. The similarity between the warped atlas image and the groundtruth image volume is quantified by the structural similarity index (SSIM), which takes the luminance, contrast and structure into account. As our method is a registration method, its ability to estimate the right intensity values of the image volume is limited. It can only warp an atlas with fixed intensity values.
Training details
We implemented our network by using the TensorFlow library. The network was trained for 300 epochs on a NVIDIA Tesla A100 graphics card. The model requires 18.7 GB of memory when being trained with a batch size equal to one, and has a computational complexity of 722 GFLOPS. The lossfunction was minimised using the Adam optimizer, with the learning rate set to \(10^{5}\).
Experimental results
Comparison with other methods
This section describes the results of the registration to AP and lateral DRRs, by our proposed network and by two other networks for comparison. The evaluation metrics are listed in Table 1. Figure 2 illustrates the qualitative performance of the network by some registration examples.
The first comparison method registers a Bsplinebased statistical deformation model (SDM) to a pair of radiographs by regressing its principal component weights [10]. This is a deeplearning implementation of the classical method of Yu et al.(2017) [4]. The SDM guarantees plausible shapes and provides smoother deformation fields than our proposed method, as can be seen in Fig. 2. Nevertheless, it is outperformed by our method in terms of registration accuracy (\(p=10^{30}\)), as reported in Table 1. This indicates that the constraint on the deformation field by the SDM is too strong to correct for smallscale deformations. The lower SSIM value is due to the different atlas image being used for the SDMbased method. This atlas has an average intensity profile which cancels out more subtle local intensity variations.
The second comparison method is a reimplementation of the work of Kasten et al. [7], in which the 3D binary labelmap of the femur is immediately regressed from the biplanar radiographs, without deforming an atlas image. This method achieves a larger Dice score than our method (\(p=4\cdot 10^{4}\)), but lacks information about the internal structures. As it does not regress the 3D intensity values, the problem is considerably simplified.
Figure 2 shows a good alignment for our method between the input DRRs and the simulated perspective projections of the registered atlas images, including the cortical bone. The geometric distance error between the estimated and groundtruth surface model highlights the lesser trochanter as a challenging region to register accurately for all methods, while global structures like the femoral neck and shaft are more accurately reconstructed.
Sensitivity to inaccurate input
Our network requires calibrated radiographs as input, meaning that the corresponding projection matrix, parameterised by the intrinsic and extrinsic parameters, needs to be known. However, the orientation of an imaging system, like a Carm system, can never exactly be determined in practice, especially if both projections are taken at different times and the patient moves in between both acquisitions. In this experiment, we study how the uncertainty on the LAO/RAO projection angle affects the registration accuracy for projections which are in reality orthogonal. Figure 3 shows the evaluation metrics with respect to the difference between the groundtruth and input projection angle. For a discrepancy of 5\(^{\circ }\), the average dice score gets reduced from 0.94 to 0.90.
Generalised projection geometries
We retrained and evaluated the registration network on the DRR dataset with generalised projection angles. Instead of perfect AP and lateral DRRs, projections were randomly generated in a range of 60\(^{\circ }\) around the AP and lateral views. By training the network on such generalised dataset, the network can be reused for any projection geometry.
The overall average dice score on the generalised validation dataset (\(N=2880\)) equals \(0.923\pm 0.033\). Figure 4 shows the median Dice scores for different combinations of LAO/RAO projection angles. The Dice score is maximal for nearorthogonal projection geometries, where the angle between both projection directions is between 80\(^{\circ }\) and 110\(^{\circ }\). It is interesting to note that projections do not necessarily need to correspond to perfect AP and lateral views.
Ablation study
To study the effectiveness of individual components in our registration network, we retrained our network, omitting some modules. We used the same dataset as in “Experimental results” section for training, validation, and testing. The evaluation metrics, listed in Table 2, are compared to the original results of “Experimental results” section by means of a twosided paired ttest.
Effectiveness of affine network structure
In this experiment, the affine network of “Affine registration module” section was modified by removing the intermediate concatenations of AP and lateral feature maps. Instead, they were only combined at the end of the affine module, right before regressing the affine parameters. While the affine initialisation is significantly worsened by this, the local registration remains unaffected. It shows that the local registration has a large enough capture range to correct for variations left unseen by the affine initialisation.
Effectiveness of skipconnections
Removing the skip connections in the local network significantly reduces the registration accuracy \((p=10^{3})\). Secondly, it also increases the training time from 300 to 700 epochs, especially due to the slower training of the affine network. The mismatch in learning rate between the affine and local network can be explained by the vanishing gradient problem. In deep neural networks, the gradient might become very small for the early layers in the network, resulting in a negligible parameter update. The skip connections provide an alternative path to backpropagate the lossfunction, which is essential for updating the early network layers.
Effectiveness of two separate 3D decoders
Instead of treating the AP and lateral feature maps separately by two distinct encoderdecoder modules, this network variation combines both feature maps at each level of the 2D encoder, similar to the affine network structure, and only contains one 3D decoder. Skip connections are included between the combined 2D feature maps and 3D decoder. The affine registration module remains the same as depicted in Fig. 1. The results in Table 2 show a highly significant reduction in the affine and local registration accuracy, indicating the preference to decode the 3D feature maps for each projection direction separately.
Effectiveness of invProST layer
The invProST layer is responsible for spatially aligning the decoded 3D feature maps into a common coordinate system, before regressing the deformation field. If the invProST layer is left out and the 3D feature maps are directly concatenated instead, the registration accuracy is significantly reduced \((p<10^{16})\).
Discussion
In this work, we presented a DLmodel for 2D/3D registration, which substantially differs from other DLmethods in the literature. Instead of directly reconstructing the 3D image volume from a pair of DRRs, like in the model of Kasten et al. [7], our network estimates a deformation field that can warp an atlas to the floating space. This has the advantage that large deformations and unlikely shapes can be penalised. Secondly, the estimated deformation field can also be used to warp auxiliary data like label maps. Finally, our network is not restricted to perfect AP and lateral projections.
The comprehensive experiments performed on simulated DRRs from patient CT images show the efficacy of our registration method. The network achieves an average Dice score of 0.94 on the test dataset with orthogonal AP and lateral radiographs. While these biplanar views are the standard in musculoskeletal imaging, the acquisition of perfectly orthogonal AP and lateral radiographs is not always achievable in medical practice. Occasionally, instead of horizontal lateral projections, other lateral views, like the frogleg or Judet view, are sometimes preferred, depending on the underlying disorder [16]. It was experimentally determined that our method still achieves satisfying results for projection geometries deviating from orthogonality by up to \(\pm 10^{\circ }\).
The pair of radiographs that serves as input to our network needs to be calibrated, meaning that the intrinsic and extrinsic parameters of the projection matrix need to be known for both images. The deep network of Gao et al. [9] allow uncalibrated radiographs as input. Their network learns a convex similarity metric with respect to the pose parameters, which is close to the square of geodesic distances in SE(3). In the application phase, this convex similarity function can be optimised over the pose parameters by a conventional gradient descent method. It remains a topic of further research to implement this approach to our network in order to enable uncalibrated radiographs as input and to increase the applicability of the network for medical practices.
Furthermore, the input radiographs to our network need to have the femur masked out. While manually annotating the contours would be a subjective and timeconsuming task, automatic methods are proposed in the literature to obtain accurate femoral segmentation maps from radiographs [17]. Selection of the region of interest and segmentation would also be an important preprocessing step for registration of more complex anatomical structures.
Conclusion
This paper presents a novel endtoend 2D/3Dregistration network that registers a 3D atlas image to a pair of radiographs. The network regresses a pose similarity transform and a dense deformation field for local shape variations. It effectively accounts for the projection matrix through a projective and inverseprojective spatial transform layer. The experiments show an average Dice score of 0.94 and an average symmetric surface distance of 0.84 mm on the test dataset, which illustrate the effectiveness of our network.
References
Marcovigi A, Ciampalini L, Perazzini P, Caldora P, Grandi G, Catani F (2019) Evaluation of native femoral neck version and final stem version variability in patients with osteoarthritis undergoing robotically implanted total hip arthroplasty. J Arthroplasty 34(1):108–115
Lass R, Kubista B, Olischar B, Frantal S, Windhager R, Giurea A (2014) Total hip arthroplasty using imageless computerassisted hip navigation: a prospective randomized study. J Arthroplasty 29(4):786–791
Knafo Y, Houfani F, Zaharia B, Egrise F, ClercUrmès I, Mainard D (2019) Value of 3D preoperative planning for primary total hip arthroplasty based on biplanar weightbearing radiographs. BioMed Res Int 2019:1–7
Yu W, Tannast M, Zheng G (2017) Nonrigid freeform 2D–3D registration using a Bsplinebased statistical deformation model. Pattern Recogn 63:689–699
Shen L, Zhao W, Xing L (2019) Patientspecific reconstruction of volumetric computed tomography images from a single projection view via deep learning. Nat Biomed Eng 3(11):880–888
Henzler P, Rasche V, Ropinski T, Ritschel T (2018) Singleimage tomography: 3D volumes from 2D cranial Xrays. Comput Graph Forum 37:377–388
Kasten Y, Doktofsky D, Kovler I (2020) Endtoend convolutional neural network for 3D reconstruction of knee bones from biplanar Xray images. In: International Workshop on Machine Learning for Medical Image Reconstruction, pp 123–133
Ying X, Guo H, Ma K, Wu J, Weng Z, Zheng Y (2019) X2CTGAN: reconstructing CT from biplanar Xrays with generative adversarial networks. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10619–10628
Gao C, Liu X, Gu W, Killeen B, Armand M, Taylor R, Unberath M (2020) Generalizing spatial transformers to projective geometry with applications to 2D/3D registration. In: International conference on medical image computing and computer assisted intervention (MICCAI), pp 329–339
Van Houtte J, Gao X, Sijbers J, Zheng G (2021) 2D/3D registration with a statistical deformation model prior using deep learning. In: 2021 IEEE EMBS international conference on biomedical and health informatics (BHI), pp 1–4
Dalca AV, Balakrishnan G, Guttag J, Sabuncu MR (2018) Unsupervised learning for fast probabilistic diffeomorphic registration. In: International conference on medical image computing and computer assisted intervention (MICCAI), pp 729–738
Arsigny V, Commowick O, Pennec X, Ayache N (2006) A logeuclidean framework for statistics on diffeomorphisms. In: International conference on medical image computing and computer assisted intervention (MICCAI), pp 924–931
Krčah M, Székely G, Blanc R (2011) Fully automatic and fast segmentation of the femur bone from 3DCT images with no shape prior. In: 2011 IEEE international symposium on biomedical imaging: from nano to macro, pp 2087–2090
Unberath M, Zaech, JN, Lee SC, Bier B, Fotouhi J, Armand M, Navab N (2018) DeepDRR–a catalyst for machine learning in fluoroscopyguided procedures. In: International conference on medical image computing and computer assisted intervention (MICCAI), pp 98–106
Yeghiazaryan V, Voiculescu ID (2018) Family of boundary overlap metrics for the evaluation of medical image segmentation. J Med Imag 5(1):015006
Kruithoff ML, McGrory BJ (2021) The value of the direct lateral hip radiograph in an adult reconstruction practice. Arthroplasty Today 8:29–34
Xie W, Franke J, Chen C, Grützner PA, Schumann S, Nolte LP, Zheng G (2014) Statistical modelbased segmentation of the proximal femur in digital anteroposterior (AP) pelvic radiographs. Int J Comput Assist Radiol 9(2):165–176
Acknowledgements
GZ is supported by the National Key R&D Program of China via project 2019YFC0120603 and by the Natural Science Foundation of China via project U20A20199. JVH is supported for this research by the Research Foundation in Flanders (FWOSB 1S63918N). EA is supported by a senior clinical investigator fellowship of the Research Foundation Flanders (FWO). JS and JVH also acknowledge the Flemish Government under the “Onderzoeksprogramma Artificiele Intelligentie (AI) Vlaanderen” programme.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflict of interests
The authors declare that they have no competing interests.
Ethics approval
The study was reviewed and approved by the local ethics committee under reference B670201111480.
Informed Consent
Written informed consent was obtained from each patient before participation in the study.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Van Houtte, J., Audenaert, E., Zheng, G. et al. Deep learningbased 2D/3D registration of an atlas to biplanar Xray images. Int J CARS 17, 1333–1342 (2022). https://doi.org/10.1007/s11548022025863
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11548022025863
Keywords
 Deep learning
 Digitally reconstructed radiographs
 Xray imaging
 2D/3D image registration
 Image warping
 Atlas image