Next Article in Journal
Integrating the OHIF Viewer into XNAT: Achievements, Challenges and Prospects for Quantitative Imaging Studies
Previous Article in Journal
Influence of Spatial Resolution and Compressed SENSE Acceleration Factor on Flow Quantification with 4D Flow MRI at 3 Tesla
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Anatomy-Aware Inference of the 3D Standing Spine Posture from 2D Radiographs †

1
Department of Computer Science, Technical University of Munich, 85748 Garching, Germany
2
Department of Neuroradiology, Klinikum rech der Isar, 81675 Munich, Germany
3
Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
4
A.A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
5
Department of Quantitative Biomedicine, University of Zurich, 8006 Zurich, Switzerland
6
Institute of Computer Graphics and Vision, Graz University of Technology, 8010 Graz, Austria
7
School of Computer Science, University of Auckland, Auckland 1010, New Zealand
*
Author to whom correspondence should be addressed.
This paper is an extended version of our conference paper: Bayat, A.; Sekuboyina, A.; Paetzold, J.C.; Payer, C.; Stern, D.; Urschler, M.; Kirschke, J.S.; Menze, B.H. Inferring the 3D standing spine posture from 2D radiographs. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; pp. 775–784.
These authors contributed equally to this work.
Tomography 2022, 8(1), 479-496; https://doi.org/10.3390/tomography8010039
Submission received: 6 December 2021 / Revised: 30 January 2022 / Accepted: 3 February 2022 / Published: 11 February 2022

Abstract

:
An important factor for the development of spinal degeneration, pain and the outcome of spinal surgery is known to be the balance of the spine. It must be analyzed in an upright, standing position to ensure physiological loading conditions and visualize load-dependent deformations. Despite the complex 3D shape of the spine, this analysis is currently performed using 2D radiographs, as all frequently used 3D imaging techniques require the patient to be scanned in a prone position. To overcome this limitation, we propose a deep neural network to reconstruct the 3D spinal pose in an upright standing position, loaded naturally. Specifically, we propose a novel neural network architecture, which takes orthogonal 2D radiographs and infers the spine’s 3D posture using vertebral shape priors. In this work, we define vertebral shape priors using an atlas and a spine shape prior, incorporating both into our proposed network architecture. We validate our architecture on digitally reconstructed radiographs, achieving a 3D reconstruction Dice of 0.95 , indicating an almost perfect 2D-to-3D domain translation. Validating the reconstruction accuracy of a 3D standing spine on real data is infeasible due to the lack of a valid ground truth. Hence, we design a novel experiment for this purpose, using an orientation invariant distance metric, to evaluate our model’s ability to synthesize full-3D, upright, and patient-specific spine models. We compare the synthesized spine shapes from clinical upright standing radiographs to the same patient’s 3D spinal posture in the prone position from CT.

1. Introduction

A biomechanical load analysis of the spine in an upright standing position is highly warranted in various spine disorders to understand their cause and guide therapy [1]. Typical approaches for load estimation either use a computational shape model of the spine for all patients or obtain a subject-specific spine model from a 3D imaging modality such as magnetic resonance imaging (MRI) or computed tomography (CT) [2]. Even though MRI and CT images can capture 3D anatomical information, they need the patient to be in a prone or supine position (lying flat on a table) during imaging. Nevertheless, to analyze the spinal alignment in a physiologically upright standing position under weight bearing, orthogonal 2D plain radiographs (as depicted in Figure 1) are the de facto choice. A combination of both these worlds is of clinical interest to fully assess the true biomechanical situation, that is, to capture the patient-specific complex pathological spinal arrangement in a standing position with full 3D information [2,3,4].
As crucial spatial information is lost when projecting a 3D object in only two 2D planes [5], a random object cannot reliably be reconstructed from two orthogonal projections. However, the spine follows strong anatomical rules, that are repeated only with slight variations in any patient. Typical projections—that is, lateral and AP radiographs—cover most of these variations, both on a local (per vertebra) and global (overall spinal alignment) level.
The literature offers a wealth of registration-based methods [4,6,7,8,9,10] for relating 2D radiographs to 3D CT or MR images. In [4], the authors employ coarse manual registration for aligning 3D data to 2D sagittal radiographs for the lumbar vertebrae. Similarly, in [6], manually annotated vertebral bodies on 2D radiographs are used to measure the vertebral orientations in an upright (standing) position. Such solutions are time-consuming, labor-intensive, and are vulnerable to error. Note that both works only employ sagittal reformations to position the vertebrae. They thus ignore coronal reformation containing significant information about the spine’s natural curvature, especially in abnormal cases. Other approaches for this purpose introduce an automated 3D–2D spine registration algorithm [7,8], wherein authors suggest a multi-stage registration method by introducing a comparison metric for a CT projection and a radiograph. Since this metric is parameter-heavy and hand-crafted, the generalizability and inference speeds of these methods are limited. In [9,10], the 3D shape of the spine is reconstructed using a biplanar X-ray device called ‘EOS’. The advantage of the system is the low radiation dose required and that both projections are acquired simultaneously, allowing for a direct spatial correspondence between the two planes. Hindering its applicability is the high device cost and thus the lack of its presence in clinical routines.
The problem of reconstructing 3D shapes from 2D images has recently been explored using deep learning methods. For example, in [11], adversarial training is used to synthesize 3D CT images given two orthogonal radiographs. For a scale of spinal scans, this method is memory intensive and also fails to synthesize smaller 3D anatomies such as the vertebrae. Note that this method has been evaluated on digitally reconstructed radiographs (DRR) only, thus requiring further validation for its clinical usage. In [12,13], the authors design a model to generate a 3D shape given multiple arbitrary view 2D images. However, the input images include only the object of interest without background, which is not applicable to medical images like spinal radiographs. In our previous work on inferring the 3D standing spine posture [14], we introduce a new deep neural network architecture termed TransVert, to combine the 2D orthogonal image information and reconstruct a 3D shape. Since the vertebrae are heavily occluded by soft tissue and ribs in the radiographs, the reconstruction is challenging and in some cases the output shape is far from a vertebral shape.
In machine learning, shape priors help to reduce the search space of possible solutions, improving the accuracy and plausibility of solutions [15]. Priors are particularly effective when data are unclear, corrupt, with low signal-to-noise ratio or when training data are scarce [15,16]. The idea of incorporating shape priors into the neural network models has been explored by [15,17,18,19,20]. Most of the ideas proposed for using shape priors in deep neural networks are for a single object, while the spine shape is more complicated as it is made of multiple objects (vertebrae) connected to each other. The spine and vertebrae shapes are deformed, subject to some constraints of human anatomy. Thus, defining a prior for explaining intervertebral constraints and the spinal shape is crucial here.
A few works have modeled the spinal shape [10,21,22,23,24]. For instance, in [24] the authors proposed an automatic framework that segments vertebrae from arbitrary CT images with a complete spine model. They first scanned a commercially available plastic phantom to generate the template. Next, they manually registered it to ten actual full spine scans. The authors learned a statistical shape model of the spine in [23] by independently studying three models for cervical, thoracic and lumbar regions. Thus, their models do not learn the shape correlations across the full spine [21]. In [22], the authors propose a parametric model of the spine, which is computed using statistical inferences, image analysis techniques and manual rigid registration. In [10], the authors used a PCA-based model to decompose the spine shape into the spinal curve and local shape of vertebrae.
Besides statistical shape models [10,21,23,24], other probabilistic models, such as probabilistic atlas [25], graph models [26], Hidden Markov Models [27] and hierarchical models [28,29,30] have also been proposed. For instance, in [25], the authors proposed a probabilistic atlas of the spine. By co-registering 21 CT scans, a probability map is created which can be used to segment and detect the vertebrae with a special focus on ribs suppression. In [26] the authors proposed a probabilistic graphical model for the location and identification of the vertebrae in MR images. In both cases full spines were observed at training time. However, none of the works mentioned above are an end-to-end machine learning treatment of the problem of the reconstruction of 3D shapes of spines from 2D images.
We introduce an anatomy-aware deep neural network for fusing two orthogonal 2D radiographs to generate a 3D spine model. We define this 3D shape synthesis as a hybrid registration problem in which the network estimates vector fields to deform vertebral shape templates to achieve shape synthesis. We apply our model to clinical radiographs subject to noise and heavy tissue overlay. We incorporate shape priors into the model to enhance robustness against such artefacts in real-world data. We propose a training approach using synthetically generated radiographs from 3D CT, with full supervision from the CT’s 3D vertebral masks.

2. Materials and Methods

For synthesizing 3D data from 2D information, the following requirements are desired: First, for efficient recovery of 3D shape information from sagittal and coronal projections, the network needs to integrate the information from these projections appropriately. Second, recovering 3D shapes from 2D projections is inherently an ill-posed problem, requiring incorporation of prior knowledge. This knowledge includes vertebral shapes and the shape of the spine (spinal curvature).
Our work is mainly based on the TransVert network [14]. The TransVert network is designed for inferring the 3D standing spinal posture from 2D radiograph. Here, we overview the TransVert model to better explain our proposed model. We will introduce the TransVert+ model to address the limitations of TransVert.

2.1. TransVert

TransVert inputs are the sagittal and coronal vertebral image patches and their corresponding annotation images. The annotation images indicate the vertebra-of-interest (VOI). Given the four 2D inputs, the model predicts the vertebra’s full-body 3D shape, y , as a discrete voxel-map:
y = G ( x s , x c , y s , y c ) ,
where G denotes the mapping performed by TransVert, x s and x c denote the 2D vertebral sagittal and coronal reformations; y s and y c denote the corresponding VOI annotations. The VOI-annotation image is obtained by placing a disc of radius 1 mm around the vertebral centroid. Figure 2 depicts orthogonal input image patches with corresponding centroid annotations, indicating the vertebra of interest in that patch. In [14], other annotation choices (vertebral body and full vertebral masks) are analyzed in several experiments. The TransVert model is trained on sagittal and coronal digitally reconstructed radiographs (DRR) created from CT images. The model is supervised by the corresponding CT images’ voxel-level, vertebral segmentation masks. Since DRRs are similar in appearance to real radiographs, a model trained on DRRs can be deployed on clinical radiographs.
Figure 3 overviews the TransVert architecture [14]. TransVert consists of three blocks: a 2D sagittal encoder, a 2D coronal encoder, and a 3D decoder. In [14] we introduced a ‘map&fuse’ mechanism to combine the three blocks. The map&fuse block is designed to map 2D representation of the sagittal and coronal views into intermediate 3D latent representations and fuse them into a single 3D representation. The decoder block maps the 3D latent representation to a 3D shape in voxel space.

2.2. TransVert+: Anatomy-Aware TransVert

The TransVert model [14] performs 2D to 3D translation vertebra by vertebra. While this results in acceptable vertebral shape reconstruction, the model still is not aware of the spinal shape and relative position of vertebrae, which might result in inconsistency in the reconstructed spinal shape. To overcome this problem, we propose incorporating the spinal curve and vertebral shapes as prior. Instead of regressing the shapes directly, our model deforms an atlas to the desired vertebral shape.
Our network estimates the vector field that deforms a discrete atlas to the desired vertebral shape. We incorporate a spine atlas as a shape prior into our network architecture, to enforce vertebral shape constraints. We also define the spine curvature by atlas vertebral centroids, for enforcing the shape of the spine. In addition to image data, we include the vertebral labels as an additional annotation attached to the network input as depicted in Figure 2. We take a registration approach for shape synthesis and propose the TransVert+ network architecture.
TransVert+ takes five inputs, Equation (2), of which four are 2D images, the same as TransVert’s inputs, depicted in Figure 2. The fifth input is a vector of floats denoting the coordinates of the vertebral centroids ( C v ) in a global spine coordinate system, for providing the model with a holistic view of the spinal curve for more consistency. We desire a function G that outputs the vertebra’s full-body 3D shape, y , which is represented as a discrete voxel-map by deforming the vertebral shape templates ( y t ):
y = G ( x s , x c , y s , y c , C v ) y t .
We formulate this shape synthesis as a registration problem in which vertebral shape templates are deformed to obtain desired shapes. This includes both global affine transformations (scaling and 3D rotation for each vertebra, considering the global spinal shape) and local deformations on the vertebral surface. We denote them as two sub-tasks,
G ( x s , x c , y s , y c , C v ) = G a + G d ,
G a = A f f i n e D e c o d e r ( x s , x c , y s , y c , C v ) ,
G d = D e f o r m a b l e D e c o d e r ( x s , x c , y s , y c ) ,
where G in Equation (3) denotes the mapping performed by TransVert+. The transformation G is separated into affine ( G a ) and deformable ( G d ) transformations, explained in Equations (4) and (5) respectively. When inferring the affine transformation the entire spinal shape is considered, while for inferring the deformable transformation individual vertebral shape features are taken into account. We will elaborate more on this in the next sections.
Ideally, training the TransVert+ model requires radiograph images and their corresponding ‘real world’ standing 3D spine models. However, this correspondence does not exist. It is, in effect, the issue we aim to address. Thus, TransVert+ is trained on sagittal and coronal digitally reconstructed radiographs (DRR) generated from prone CT images with supervision from the voxel-level, vertebral segmentation masks of the corresponding CT images.
Generating 3D shapes from 2D information is an ill-posed problem. We model this task as a registration problem. Given 2D orthogonal information we estimate vector fields to deform shape templates to match the target 3D shape. We introduce a novel deep neural architecture to fuse the 2D information from orthogonal views and estimate the vector field required to maximize the alignment of the deformed template and the target shape. More specifically, to infer 3D spinal shapes, given 2D orthogonal radiographs and vertebral centroid coordinates in the global coordinate system, our model predicts a vector field to deform each vertebral shape template to match the target. The vector fields are predicted by considering each vertebral shape locally and also considering the spinal curvature globally. We show that incorporating the global spinal shape information improves the performance of the model.
Our model is composed of two encoders for each view, an affine decoder and a deformable decoder. In the following sections, we describe the model architecture and training scheme in detail.

2.3. Network Architecture

Figure 4 demonstrates a block diagram of the model and its subnetworks. The model is composed of two encoders, one for each view (a sagittal encoder and a coronal encoder) and two decoders (an affine decoder and a deformable decoder). The input to the encoders is the image patch of the vertebra from the radiograph and the centroid annotation on the vertebra of interest. The features extracted using the encoders are concatenated to the global centroid coordinates and fed to the affine decoder (to estimate the affine transformation parameters for each vertebra) and the deformable decoder (which is a fully convolutional network to estimate the deformation for each voxel). Next, the affine and deformable vector fields are summed to produce the final displacement field, which is used to warp the template and produce the 3D shape model. Finally, a 3D model of the spine can be generated by stacking the predicted 3D vertebrae shapes at their corresponding 3D centroid locations.

2.3.1. Sagittal and Coronal Encoders

The architecture of sagittal and coronal encoders are designed differently. In Figure 5, the architecture for each view is visualized. Each encoder is designed to reconstruct the missing third dimension of the 2D input. Therefore, they contain anisotropic convolutions with the longer side along the dimension that needs to be expanded. For instance, for a coronal input the anterior–posterior dimension needs to be expanded. For the same reason, the convolutional strides and padding directions are orthogonal for each of the views. We empirically observed that, employing the ‘squeeze and excitation’ block in the network results in a better performance than a naive fusion by concatenating the multiple channels. As input to the encoders, the vertebral images and VOI-annotations are combined using a ‘squeeze and excitation’ block (depicted in red) [31]. In our previous work [14] we showed that, using encoders for each view with anisotropic convolutions, outperforms using a single encoder or fusing input 2D data by outer product of the orthogonal 2D images.

2.3.2. Affine Decoder

Since the vertebrae are not completely visible in the radiographs, generating the 3D spine model given only the encoded features for each vertebra separately could lead to inaccurate orientation.
Including the vertebral centroid coordinates in the global coordinate system provides the model with a holistic shape of the spine. For example, Figure 1 demonstrates lateral and anterior–posterior (AP) view radiographs of three different patients. The distance between the centroids determines the scale. If one fits a curve to all of the vertebral centroids, the orientation of each vertebra should be almost perpendicular to the curve at each vertebral centroid. Although defining the orientation of the vertebrae and the other constraints, such as inter-vertebral distance, needs accurate vertebral landmark detection on radiographs, we postulate a Multi Layer Perceptron (MLP) which estimates the 3D affine parameters, given only the vertebral centroids and the information extracted by the encoders.
Since the features calculated by the encoders for data samples in a batch are independent of each other, we need an intra-batch fusion mechanism to give the model a holistic view of the features from different spinal regions. We assume that the vertebrae in a batch are from the same spine and in order (we train the model in this way). For intra-batch fusion, we flatten the features from both encoders and also the down-scaled vertebral shape templates and concatenate them with the vertebral centroid coordinates, for all vertebrae. These inputs are fed into our affine subnetwork, which is an MLP.
For each vertebra, the affine subnetwork estimates four parameters: the scaling factor (S), and the rotation about each axis ( θ x , θ y , θ z ). After the transformation matrix is created, the corresponding affine vector field is calculated. We do not include translation, since the input image patches are extracted in such a way that the vertebral centroid is at a fixed location in the image patch.
θ x , θ y , θ z , S = A f f i n e D e c o d e r ( x s , x c , y s , y c , C v ) ,
T a f f i n e = R ( θ x ) × R ( θ y ) × R ( θ z ) × T ( S ) ,
R ( θ x ) = 1 0 0 0 0 cos ( θ x ) sin ( θ x ) 0 0 sin ( θ x ) cos ( θ x ) 0 0 0 0 1
R ( θ y ) = cos ( θ y ) 0 sin ( θ y ) 0 0 1 0 0 sin ( θ y ) 0 cos ( θ y ) 0 0 0 0 1
R ( θ z ) = cos ( θ z ) sin ( θ z ) 0 0 sin ( θ z ) cos ( θ z ) 0 0 0 0 1 0 0 0 0 1
T ( S ) = S 0 0 0 0 S 0 0 0 0 S 0 0 0 0 1
A f f i n e v f = τ ( T a f f i n e ) ,
where the A f f i n e v f is the affine vector field and τ ( ) is the function to generate an affine vector field given the transformation matrix.

2.3.3. Deformable Decoder

In parallel with affine parameter estimation for rough alignment between the template and target shapes, we synthesize the finer details. We design a subnetwork termed the “deformable decoder” for estimating local shape deformations. The deformable decoder is a fully convolutional network. The inputs to this network are the intermediate 3D latent representation calculated by the encoders with the down-scaled (the same size as 3D latent representation) shape template as an extra channel. Including the shape template as another input channel helps the model to infer the differences between the template and target shapes. The deformable decoder’s output is a 3D vector field with target volume resolution. Contrary to the affine decoder, which estimates the affine transformation parameters based on the entire batch, the deformable decoder is focused on single data samples in the batch, to consider the finer shape details for each sample,
D e f o r m a b l e v f = D e f o r m a b l e D e c o d e r ( x s , x c , y s , y c ) .

2.4. Learning

We train the model using the 1 distance between the deformed templates and the target vertebral shapes in the voxel space. To regularize the deformations, we incorporate a smoothing term in the cost function. Solely using a regression loss leads to convergence to a local optimum in which a mean (or median) shape is predicted, especially in the highly varying regions of the vertebra such as the vertebral processes. We have,
L t o t a l = α p L 1 + α s L s m o o t h
L 1 = | | y G ( x s , x c , y s , y c , y t , C v ) y t | | 1
L s m o o t h = 1 X . Y . Z 0 X 0 Y 0 Z | | G ( x s , x c , y s , y c , y t , C v ) | | 1 d x d y d z ,
where α p and α s are weights for the network prediction and smoothness terms respectively, and are fixed to α p = 10 and α s = 0.1. Note that y contains an integer value of { 0 , i } , where i { 8 , 9 , · · · , 24 } represents the vertebral index from T1 to L5. Constraining the network to predict the vertebral index implicitly requires it to learn relating the vertebral index to the shape as a prior.
In order to impose a constraint for locally smooth deformations and a minimum displacement solution for our registration problem, we add L s m o o t h to penalize the 1 n o r m of the deformation field.
The network was implemented with the Pytorch framework on a Quadro P6000 GPU. It was trained until convergence using the Adam optimizer [32] with an initial learning rate of 0.0001.

2.5. Data

Recall that TransVert+ works with two data modalities: it is trained on DRRs extracted from CT images and is deployed on clinical radiographs.

2.5.1. CT data

We employed two sets of data: First, a public dataset for lung nodule detection with 800 chest CT scans [33], and second, an in-house dataset with 154 spinal CT scans. Overall, we work with ∼12 K vertebrae split 5:1 forming the training and validation set and report 5-fold cross-validated results. Of note, ref. [33] is a lung-centred dataset, thus consisting of few lumbar vertebrae.
Usually in spine CT scans [34] some parts of the ribs and tissues in distance from the spine are excluded. Thus, these scans could not be used for generating DRRs that are similar in appearance to real radiographs. However, using lung CT scans result in better DRRs.
All of spinal CT scans were resampled to 1 mm resolution. The CT scans of our in-house dataset had ground truth segmentation and the CT scans in [33] were segmented using [35]. Next, an experienced neuro-radiologist approved the generated masks to consider only accurate ones for the study. Consequently, we excluded 50 cases from [33].
We employ a ray-casting approach [36] to construct DRRs from CT scan. In this method, we define lines from the radiation source (focal point) to every single pixel on the DRR image and calculate the integral of the CT intensities over these lines. In this simulation we assign ( 180 cm ) and ( 150 cm ) to the radiation source-to-detector distance and the source-to-object distance parameters respectively. Examples of patches extracted from DRRs are illustrated in Figure 6 and Figure 7. Also, a complete example of orthogonal DRRs is shown in Figure 8.
Once the sagittal and coronal DRRs are generated, the inputs for TransVert+ are constructed by extracting image patches of size 64 × 64 around each vertebral centroid. Similarly, the VOI-annotations were also extracted automatically from the projected segmentation mask.

2.5.2. Clinical Radiographs

We validate TransVert+ on clinical, standing radiographs (pairs of lateral and anterior-posterior projections) acquired from 30 patients. Before deploying our TransVert+ model on the clinical radiographs, we resampled all radiographs to 1 mm resolution.
Image acquisition parameters such as the source-to-detector and source-to-object distances were similar to those used for DRR generation. We employed [37,38] to automatically generate the vertebral annotations on both views.

2.5.3. Image Normalization

We trained TransVert+ on DRRs and tested it on real clinical radiographs. To enable a transfer of learning between these modalities, we need to normalize the intensities to a similar range. Therefore, we employ z-score normalization, i.e., I = ( I μ I ) / σ I , where μ I and σ I are the mean and standard deviation of the image I , respectively.

2.6. Metrics

The Dice score is the most popular measurement for evaluating segmentation accuracy and measures the overlap between two binary images. However, the Dice score is a poor measure of segmentation accuracy when the shapes to be compared are not “blob-like”.
Another popular metric for segmentation evaluation is the Hausdorff distance. The Hausdorff distance between two segmentations represented as surface meshes is the maximum distance from vertices on the first mesh to the vertices on the second mesh.
The measurements, such as the Dice score, average many local errors measured at each voxel. Thus, a segmentation that is largely correct with a few major shape errors will have the same score as a segmentation that is only slightly wrong everywhere.
As mentioned before, we do not have the 3D ground truth for the clinical radiograph of a patient. We can acquire the 3D CT segmentation of the same patient. However, because of difference in spinal posture, the vertebral orientations are different and we cannot compare the 3D vertebrae reconstructed from clinical radiographs to the ones from CT using Dice score or Hausdorff distance. Thus, we desire a rotation-invariant metric to evaluate the model performance on clinical radiographs, for each individual vertebra.
The normalized weighted spectral distance (nWESD) [39,40] is a global shape measure based on heat trace analysis via the Laplace operator. The eigenvalues of the Laplacian of a shape are strongly connected to the shape’s geometric properties, such as its volume, surface area and mean curvature. The Laplace spectrum is invariant to isometric transformations (rigid transformations), and changes continuously as a shape’s boundary is transformed. Thus, because of rotational invariance, we can use this metric for comparing the 3D vertebral shapes reconstructed from the patient’s clinical radiographs to the vertebral shapes from their 3D CT segmentations.
The weighted spectral distance (WESD) between two binary segmentations Ω λ and Ω ξ is defined as:
ρ ( Ω λ , Ω ξ ) = [ n = 1 ( | λ n ξ n | λ n ξ n ) p ] 1 / p ,
where λ n and ξ n denote eigenvalues of the segmentations Ω λ and Ω ξ , and p R with p > d / 2 , d indicates the dimensionality of the binary segmentations.
The normalized WESD (nWESD) is derived using the fact that WESD converges as N (even though each eigenvalue spectrum is divergent) and is bounded above by W ( Ω λ , Ω ξ ) (for details, see [39]). Therefore, the (finite) nWESD ρ ¯ N ( Ω λ , Ω ξ ) can be defined as:
ρ ¯ N ( Ω λ , Ω ξ ) = ρ N ( Ω λ , Ω ξ ) W ( Ω λ , Ω ξ ) [ 0 , 1 )
ρ N ( Ω λ , Ω ξ ) = [ n = 1 N ( | λ n ξ n | λ n ξ n ) p ] 1 / p .

2.7. Experiments

In order to analyze the contribution of various architectural components of the TransVert+ (fusion of sagittal and coronal views with shape prior, the affine decoder and a convolutional decoder to estimate the displacement field for each vertebrae) and to validate its performance on clinical radiographs, we propose three sets of experiments.
(1)
We conducted an ablative study on the architectural choices including three steps: First, we evaluated the performance of the model working only with the affine decoder. Second, we trained the model solely with the deformable decoder to estimate the deformation fields. Third, we repeat the performance of our previous model TransVert, in which we did not incorporate the shape priors, but estimated the vertebral shapes directly in the voxel space without registration. Finally, we give results for TransVert+.
(2)
We deployed the model on clinical radiographs to reconstruct 3D standing spine postures.
(3)
We evaluated the model performance on clinical radiographs quantitatively. The performance evaluation in various settings was compared by computing the Dice coefficient, Hausdorff distance and normalized weighted spectral distance (nWESD) between the predicted 3D vertebral mask and the ground truth CT mask, where appropriate.

3. Results

The results of the ablation study are reported in Table 1. As we expected, the affine results are not better than TransVert, since the model roughly aligns the shape template to the target and the results lack shape details. Training the model with a deformable decoder performs better than the affine model. The deformable model achieves an average Dice score of 0.9405 , close to that of the TransVert model ( 0.9426 ). In the Hausdorff distance metric, TransVert performs better but for the nWESD metric, the deformable model achieves a better score. Finally, our main TransVert+ model outperformed all of the previous models in all metrics.
Figure 6 depicts coronal and sagittal image patches from three vertebrae, mid-slice of the resulting estimated displacement field and vertebral shape. In the estimated displacement field the brighter colors indicate greater displacements. Figure 7 shows an example point cloud (with 2048 points) from the predicted and ground truth shapes, along with a point-wise Chamfer distance map. Observe that the vertebra’s posterior region (vertebral process) is hardly visible in the image inputs. In spite of this, TransVert+ was able to reconstruct the 3D shape of vertebral processes. To calculate the Chamfer distance map for each point in the reconstructed vertebra’s point cloud, the nearest point in the ground truth point cloud is found and the square of distance is depicted. Figure 8 illustrates a 3D spine reconstruction based on 2D DRRs.
For a detailed comparison of the methods, in Figure 9 we report the mean Dice score, Hausdorff distance and nWESD distance for each vertebra. In general, TransVert+ outperforms almost all other models for each vertebra. According to this figure, the model performance drops in all metrics for lumbar vertebra.

3.1. 2D-to-3D Translation in Clinical Radiographs

Figure 10 depicts the results of deploying TransVert+ for reconstructing the 3D, patient-specific posture of the upright standing spine. As stated, there is no 3D ground truth spinal model for the clinical radiographs. Observe the matched reconstruction of the 3D spine posture to the spinal posture in radiographs.

3.2. Quantitative Evaluation of Performance on Clinical Radiographs

We quantitatively evaluated the performance of our model on clinical radiographs in two patients for whom we have CT scans in addition to orthogonal standing radiographs. Using our TransVert model and the TransVert+ model proposed in this work, we generated the 3D shapes of the vertebrae given the 2D clinical radiographs. Then we compared each vertebra to the corresponding one from the CT scan segmentation masks, which refer to the same object but in a different orientation. Conventional metrics like the Dice score or Hausdorff are not applicable in this case, but we can use the nWESD metric as it is invariant to rigid transformations (including rotation). The resulting nWESD scores are demonstrated in Figure 11, where the mean of nWESD was 0.13 and 0.15 for TransVert+ and TransVert respectively. Similarly, the standard deviation of nWESD was 0.08 for TransVert+ and 0.11 for TransVert. These scores are calculated on a group of 32 vertebrae. Despite better values for TransVert+, the difference was not significant.
To appreciate the difference of spine posture in the standing and lying down positions, we illustrate the two cases we used in this experiment in Figure 12. The spine postures reconstructed from radiographs in upright standing position are depicted on the left and the ones from CT are depicted on the right side of the figure.

4. Discussion

The approach we proposed could be used to simulate the spinal posture in a standing position under weight bearing. In Table 1 we showed that our approach could improve overall model performance in terms of Dice score, Hausdorff distance and nWESD. However, more work is required to improve the results and overcome the limitations.
First, as explained in the results section (Figure 9) the model performance drops for lumbar vertebrae in all models because of few lumbar vertebrae in the lung-centered dataset we used. Second, the sagittal and coronal view radiographs might not be perfectly perpendicular in clinical settings; whereas, we generated DRRs with orthogonal projection angles.
Training the model with a dataset containing more scans with lumbar vertebrae, could improve the model performance in the lumbar region. As depicted in Figure 12, we compared the standing and lying down spine postures of two patients. Acquiring more annotated pairs of CT scans and radiographs of the same patient could enrich the results and provide more data for validation. We use a relatively large-scale dataset for training our model. Although the size and anatomical variations in patients helps the model learn the shapes and deformations, we believe that data augmentation by adding a small randomness to the projection angle in the simulated radiographs could teach the model this randomness in clinical radiographs.

5. Conclusions

We introduced TransVert+, a neural network architecture to reconstruct a full 3D spinal model from 2D orthogonal radiographs by deforming vertebral shapes. We proposed a supervised approach to train the model on synthetic data (DRRs) and transfer the trained model to radiographs. We demonstrated an improved performance of the model after incorporating shape priors into the model and separating the registration problem into affine and deformable registrations. We improved the state-of-the-art in three different metrics for this task. Finally, we successfully validated our model on real-world clinical radiographs and quantitatively compared the results to CT segmentations of the same patient for the first time.

Author Contributions

Conceptualization, A.B. and A.S.; methodology, A.B., D.F.P. and A.S.; software, A.B., D.F.P., C.P., D.S. and M.U.; validation, A.B., J.S.K. and B.H.M.; formal analysis, A.B., A.S. and B.H.M.; investigation, J.S.K. and B.H.M.; resources, J.S.K. and B.H.M.; data curation, A.B. and J.S.K.; writing—original draft preparation, A.B., D.F.P. and A.S.; writing—review and editing, J.S.K., B.H.M. and M.U.; visualization, A.B. and A.S.; supervision, J.S.K. and B.H.M.; project administration, A.B.; funding acquisition, J.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the European Research Council (ERC) under the European Union’s ‘Horizon 2020’ research & innovation programme (GA637164–iBack–ERC–2014–STG).

Institutional Review Board Statement

This study was performed in line with the principles of the Declaration of Helsinki. Approval was granted by the Ethics Committee of the School of Medicine, Technical University of Munich (19_84-19 S).

Informed Consent Statement

Patient consent was waived. Since public datasets were used and in-house datasets were anonymized and approved by the Ethics Committee of the School of Medicine, Technical University of Munich.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank Polina Golland and Justin Solomon. We also acknowledge support of NIH NIBIB NAC P41EB015902 and Philips Inc.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DRRDigitally Reconstructed Radiographs
MRIMagnetic Resonance Imaging
CTComputed Tomography
nWESDnormalized Weighted Spectral Distance
VOIVertebra Of Interest
GTGround Truth

References

  1. Dreischarf, M.; Shirazi-Adl, A.; Arjmand, N.; Rohlmann, A.; Schmidt, H. Estimation of loads on human lumbar spine: A review of in vivo and computational model studies. J. Biomech. 2016, 49, 833–845. [Google Scholar] [CrossRef] [PubMed]
  2. Akhavanfar, M.; Kazemi, H.; Eskandari, A.; Arjmand, N. Obesity and spinal loads: A combined MR imaging and subject-specific modeling investigation. J. Biomech. 2018, 70, 102–112. [Google Scholar] [CrossRef] [PubMed]
  3. El Ouaaid, Z.; Shirazi-Adl, A.; Plamondon, A. Effect of changes in orientation and position of external loads on trunk muscle activity and kinematics in upright standing. J. Electromyogr. Kinesiol. 2014, 24, 387–393. [Google Scholar] [CrossRef] [PubMed]
  4. Eskandari, A.; Arjmand, N.; Shirazi-Adl, A.; Farahmand, F. Subject-specific 2D/3D image registration and kinematics-driven musculoskeletal model of the spine. J. Biomech. 2017, 57, 18–26. [Google Scholar] [CrossRef]
  5. Löffler, M.T.; Jacob, A.; Scharr, A.; Sollmann, N.; Burian, E.; El Husseini, M.; Sekuboyina, A.; Tetteh, G.; Zimmer, C.; Gempt, J.; et al. Automatic opportunistic osteoporosis screening in routine CT: Improved prediction of patients with prevalent vertebral fractures compared to DXA. Eur. Radiol. 2021, 31, 6069–6077. [Google Scholar] [CrossRef]
  6. Bauer, S.; Hausen, U.; Gruber, K. Effects of individual spine curvatures—A comparative study with the help of computer modelling. Biomed. Eng. Tech. 2012, 57, 132–135. [Google Scholar] [CrossRef]
  7. Ketcha, M.; De Silva, T.; Uneri, A.; Jacobson, M.; Goerres, J.; Kleinszig, G.; Vogt, S.; Wolinsky, J.; Siewerdsen, J. Multi-stage 3D–2D registration for correction of anatomical deformation in image-guided spine surgery. Phys. Med. Biol. 2017, 62, 4604. [Google Scholar] [CrossRef]
  8. De Silva, T.; Uneri, A.; Ketcha, M.; Reaungamornrat, S.; Kleinszig, G.; Vogt, S.; Aygun, N.; Lo, S.; Wolinsky, J.; Siewerdsen, J. 3D–2D image registration for target localization in spine surgery: Investigation of similarity metrics providing robustness to content mismatch. Phys. Med. Biol. 2016, 61, 3009. [Google Scholar] [CrossRef] [Green Version]
  9. Humbert, L.; De Guise, J.A.; Aubert, B.; Godbout, B.; Skalli, W. 3D reconstruction of the spine from biplanar X-rays using parametric models based on transversal and longitudinal inferences. Med. Eng. Phys. 2009, 31, 681–687. [Google Scholar] [CrossRef]
  10. Aubert, B.; Vazquez, C.; Cresson, T.; Parent, S.; de Guise, J.A. Toward automated 3D spine reconstruction from biplanar radiographs using CNN for statistical spine model fitting. IEEE Trans. Med. Imaging 2019, 38, 2796–2806. [Google Scholar] [CrossRef]
  11. Ying, X.; Guo, H.; Ma, K.; Wu, J.; Weng, Z.; Zheng, Y. X2CT-GAN: Reconstructing CT from biplanar X-rays with generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–17 June 2019; pp. 10619–10628. [Google Scholar]
  12. Xie, H.; Yao, H.; Sun, X.; Zhou, S.; Zhang, S. Pix2vox: Context-aware 3d reconstruction from single and multi-view images. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019; pp. 2690–2698. [Google Scholar]
  13. Xie, H.; Yao, H.; Zhang, S.; Zhou, S.; Sun, W. Pix2Vox++: Multi-scale context-aware 3D object reconstruction from single and multiple images. Int. J. Comput. Vis. 2020, 128, 2919–2935. [Google Scholar] [CrossRef]
  14. Bayat, A.; Sekuboyina, A.; Paetzold, J.C.; Payer, C.; Stern, D.; Urschler, M.; Kirschke, J.S.; Menze, B.H. Inferring the 3D standing spine posture from 2D radiographs. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Lima, Peru, 4–8 October 2020; pp. 775–784. [Google Scholar]
  15. Lee, M.C.H.; Petersen, K.; Pawlowski, N.; Glocker, B.; Schaap, M. TETRIS: Template transformer networks for image segmentation with shape priors. IEEE Trans. Med. Imaging 2019, 38, 2596–2606. [Google Scholar] [CrossRef] [Green Version]
  16. Sokooti, H.; De Vos, B.; Berendsen, F.; Lelieveldt, B.P.; Išgum, I.; Staring, M. Nonrigid image registration using multi-scale 3D convolutional neural networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 11–13 September 2017; pp. 232–239. [Google Scholar]
  17. Milletari, F.; Rothberg, A.; Jia, J.; Sofka, M. Integrating statistical prior knowledge into convolutional neural networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Quebec City, QC, Canada, 11–13 September 2017; pp. 161–168. [Google Scholar]
  18. Oktay, O.; Ferrante, E.; Kamnitsas, K.; Heinrich, M.; Bai, W.; Caballero, J.; Cook, S.A.; De Marvao, A.; Dawes, T.; O’Regan, D.P.; et al. Anatomically constrained neural networks (ACNNs): Application to cardiac image enhancement and segmentation. IEEE Trans. Med. Imaging 2017, 37, 384–395. [Google Scholar] [CrossRef] [Green Version]
  19. Balakrishnan, G.; Zhao, A.; Sabuncu, M.R.; Guttag, J.; Dalca, A.V. Voxelmorph: A learning framework for deformable medical image registration. IEEE Trans. Med. Imaging 2019, 38, 1788–1800. [Google Scholar] [CrossRef] [Green Version]
  20. Nosrati, M.S.; Hamarneh, G. Incorporating prior knowledge in medical image segmentation: A survey. arXiv 2016, arXiv:1607.01092. [Google Scholar]
  21. Meng, D.; Keller, M.; Boyer, E.; Black, M.; Pujades, S. Learning a statistical full spine model from partial observations. In Proceedings of the International Workshop on Shape in Medical Imaging, Lima, Peru, 4–8 October 2020; pp. 122–133. [Google Scholar]
  22. Gajny, L.; Ebrahimi, S.; Vergari, C.; Angelini, E.; Skalli, W. Quasi-automatic 3D reconstruction of the full spine from low-dose biplanar X-rays based on statistical inferences and image analysis. Eur. Spine J. 2019, 28, 658–664. [Google Scholar] [CrossRef] [Green Version]
  23. Mirzaalian, H.; Wels, M.; Heimann, T.; Kelm, B.M.; Suehling, M. Fast and robust 3D vertebra segmentation using statistical shape models. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 3379–3382. [Google Scholar]
  24. Klinder, T.; Ostermann, J.; Ehm, M.; Franz, A.; Kneser, R.; Lorenz, C. Automated model-based vertebra detection, identification, and segmentation in CT images. Med. Image Anal. 2009, 13, 471–482. [Google Scholar] [CrossRef]
  25. Ruiz-España, S.; Domingo, J.; Díaz-Parra, A.; Dura, E.; D’Ocón-Alcañiz, V.; Arana, E.; Moratal, D. Automatic segmentation of the spine by means of a probabilistic atlas with a special focus on ribs suppression. Preliminary results. In Proceedings of the 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25–29 August 2015; pp. 2014–2017. [Google Scholar]
  26. Schmidt, S.; Kappes, J.; Bergtholdt, M.; Pekar, V.; Dries, S.; Bystrov, D.; Schnörr, C. Spine detection and labeling using a parts-based graphical model. In Proceedings of the Biennial International Conference on Information Processing in Medical Imaging, Kerkrade, The Netherlands, 2–6 July 2007; pp. 122–133. [Google Scholar]
  27. Glocker, B.; Feulner, J.; Criminisi, A.; Haynor, D.R.; Konukoglu, E. Automatic localization and identification of vertebrae in arbitrary field-of-view CT scans. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nice, France, 1–5 October 2012; pp. 590–598. [Google Scholar]
  28. Seifert, S.; Barbu, A.; Zhou, S.K.; Liu, D.; Feulner, J.; Huber, M.; Suehling, M.; Cavallaro, A.; Comaniciu, D. Hierarchical parsing and semantic navigation of full body CT data. In Proceedings of the Medical Imaging 2009: Image Processing; International Society for Optics and Photonics: Bellingham, WA, USA, 2009; Volume 7259, p. 725902. [Google Scholar]
  29. Zhan, Y.; Maneesh, D.; Harder, M.; Zhou, X.S. Robust MR spine detection using hierarchical learning and local articulated model. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nice, France, 1–5 October 2012; pp. 141–148. [Google Scholar]
  30. Cai, Y.; Osman, S.; Sharma, M.; Landis, M.; Li, S. Multi-modality vertebra recognition in arbitrary views using 3D deformable hierarchical model. IEEE Trans. Med. Imaging 2015, 34, 1676–1693. [Google Scholar] [CrossRef]
  31. Roy, A.G.; Navab, N.; Wachinger, C. Concurrent spatial and channel ‘squeeze & excitation’in fully convolutional networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; pp. 421–429. [Google Scholar]
  32. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  33. Armato, S.G., III; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef]
  34. Liebl, H.; Schinz, D.; Sekuboyina, A.; Malagutti, L.; Löffler, M.T.; Bayat, A.; El Husseini, M.; Tetteh, G.; Grau, K.; Niederreiter, E.; et al. A computed tomography vertebral segmentation dataset with anatomical variations and multi-vendor scanner data. Sci. Data 2021, 8, 284. [Google Scholar] [CrossRef]
  35. Sekuboyina, A.; Husseini, M.E.; Bayat, A.; Löffler, M.; Liebl, H.; Li, H.; Tetteh, G.; Kukačka, J.; Payer, C.; Štern, D.; et al. VerSe: A vertebrae labelling and segmentation benchmark for multi-detector CT images. Med. Image Anal. 2021, 73, 102166. [Google Scholar] [CrossRef]
  36. Staub, D.; Murphy, M.J. A digitally reconstructed radiograph algorithm calculated from first principles. Med. Phys. 2013, 40, 011902. [Google Scholar] [CrossRef] [Green Version]
  37. Bayat, A.; Sekuboyina, A.; Hofmann, F.; El Husseini, M.; Kirschke, J.S.; Menze, B.H. Vertebral Labelling in Radiographs: Learning a Coordinate Corrector to Enforce Spinal Shape. In Proceedings of the International Workshop and Challenge on Computational Methods and Clinical Applications for Spine Imaging, Shenzhen, China, 17 October 2019; pp. 39–46. [Google Scholar]
  38. Sekuboyina, A.; Rempfler, M.; Valentinitsch, A.; Menze, B.H.; Kirschke, J.S. Labeling Vertebrae with Two-dimensional Reformations of Multidetector CT Images: An Adversarial Approach for Incorporating Prior Knowledge of Spine Anatomy. Radiol. Artif. Intell. 2020, 2, e190074. [Google Scholar] [CrossRef]
  39. Konukoglu, E.; Glocker, B.; Criminisi, A.; Pohl, K.M. WESD–Weighted Spectral Distance for Measuring Shape Dissimilarity. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2284–2297. [Google Scholar] [CrossRef] [Green Version]
  40. Konukoglu, E.; Glocker, B.; Ye, D.H.; Criminisi, A.; Pohl, K.M. Discriminative segmentation-based evaluation through shape dissimilarity. IEEE Trans. Med. Imaging 2012, 31, 2278–2289. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Lateral and anterior-posterior (AP) view radiographs of three patients with spinal curvature annotation. Considering the vertebral centroid coordinates and spinal curvature facilitates determining the scale and the vertebral orientation.
Figure 1. Lateral and anterior-posterior (AP) view radiographs of three patients with spinal curvature annotation. Considering the vertebral centroid coordinates and spinal curvature facilitates determining the scale and the vertebral orientation.
Tomography 08 00039 g001
Figure 2. Vertebral image patches with corresponding annotations. The network inputs are 2D orthogonal view vertebrae patches and the centroid indicates the vertebra of interest.
Figure 2. Vertebral image patches with corresponding annotations. The network inputs are 2D orthogonal view vertebrae patches and the centroid indicates the vertebra of interest.
Tomography 08 00039 g002
Figure 3. Overview of 2D image to 3D shape translation in TransVert. The network inputs are 2D orthogonal view vertebrae patches and the centroid indicating the vertebra of interest.
Figure 3. Overview of 2D image to 3D shape translation in TransVert. The network inputs are 2D orthogonal view vertebrae patches and the centroid indicating the vertebra of interest.
Tomography 08 00039 g003
Figure 4. Architecture of TransVert+. Our model is composed of sagittal and coronal 2D encoders, an affine 3D decoder and a deformable 3D decoder. The down-scaled shape templates are concatenated with the features extracted by encoders and fed to the decoders. The centroid coordinates in the global spine coordinate system are concatenated with the affine decoder feature maps. The ⊗, ⊕ and © operators represent warping, addition and concatenation, respectively.
Figure 4. Architecture of TransVert+. Our model is composed of sagittal and coronal 2D encoders, an affine 3D decoder and a deformable 3D decoder. The down-scaled shape templates are concatenated with the features extracted by encoders and fed to the decoders. The centroid coordinates in the global spine coordinate system are concatenated with the affine decoder feature maps. The ⊗, ⊕ and © operators represent warping, addition and concatenation, respectively.
Tomography 08 00039 g004
Figure 5. Architecture of the orthogonal encoders, which employ anisotropic convolutions, with an anisotropy along the dimensions that need to be expanded. We use ‘squeeze and excitation’ blocks (depicted in red) to fuse the image features and annotation features.
Figure 5. Architecture of the orthogonal encoders, which employ anisotropic convolutions, with an anisotropy along the dimensions that need to be expanded. We use ‘squeeze and excitation’ blocks (depicted in red) to fuse the image features and annotation features.
Tomography 08 00039 g005
Figure 6. Visualization of coronal and sagittal image patches from three vertebrae. First and second columns are the coronal and sagittal image patches, third column shows one slice of the predicted deformation field and last column is the corresponding slice in the resulting shape.
Figure 6. Visualization of coronal and sagittal image patches from three vertebrae. First and second columns are the coronal and sagittal image patches, third column shows one slice of the predicted deformation field and last column is the corresponding slice in the resulting shape.
Tomography 08 00039 g006
Figure 7. Shape modelling with TransVert+ on DRRs: The first column indicates the image input. The second and third columns visualize the ground truth (GT) vertebral mask, and the fourth visualizes the predicted 3D shape model. The last column shows an overlaid Chamfer distance map between point clouds of GT and prediction.
Figure 7. Shape modelling with TransVert+ on DRRs: The first column indicates the image input. The second and third columns visualize the ground truth (GT) vertebral mask, and the fourth visualizes the predicted 3D shape model. The last column shows an overlaid Chamfer distance map between point clouds of GT and prediction.
Tomography 08 00039 g007
Figure 8. Orthogonal DRRs of a patient and reconstructed 3D spine model. DRRs are generated from CT scans and our model is trained to reconstruct the 3D spine shape from DRRs.
Figure 8. Orthogonal DRRs of a patient and reconstructed 3D spine model. DRRs are generated from CT scans and our model is trained to reconstruct the 3D spine shape from DRRs.
Tomography 08 00039 g008
Figure 9. Vertebra-wise comparison of different architectures (Affine, Deformable, TransVert and TransVert+) using Dice scores, Hausdorff and nWESD distances. A higher Dice score indicates better performance. A lower Hausdorff and nWESD distance indicates better performance.
Figure 9. Vertebra-wise comparison of different architectures (Affine, Deformable, TransVert and TransVert+) using Dice scores, Hausdorff and nWESD distances. A higher Dice score indicates better performance. A lower Hausdorff and nWESD distance indicates better performance.
Tomography 08 00039 g009
Figure 10. Full 3D spine models: 3D patient-specific spine models constructed from real clinical radiographs. Each sagittal and coronal view radiograph pair is from a different patient.
Figure 10. Full 3D spine models: 3D patient-specific spine models constructed from real clinical radiographs. Each sagittal and coronal view radiograph pair is from a different patient.
Tomography 08 00039 g010
Figure 11. Comparing vertebral shapes to the ones predicted from radiographs using TransVert (Blue) and TransVert+ (Orange) using the nWESD metric. The mean nWESD metric for TransVert+ is lower than TransVert. For nWESD, lower is better.
Figure 11. Comparing vertebral shapes to the ones predicted from radiographs using TransVert (Blue) and TransVert+ (Orange) using the nWESD metric. The mean nWESD metric for TransVert+ is lower than TransVert. For nWESD, lower is better.
Tomography 08 00039 g011
Figure 12. Comparing 3D reconstructions of the standing spinal posture from clinical radiographs (left, green) to the spinal posture of the same patient in lying-down position from CT imaging (right, yellow) in two different patients. In the upright-standing posture the spine is under natural weight bearing, which leads to a different spine curvature.
Figure 12. Comparing 3D reconstructions of the standing spinal posture from clinical radiographs (left, green) to the spinal posture of the same patient in lying-down position from CT imaging (right, yellow) in two different patients. In the upright-standing posture the spine is under natural weight bearing, which leads to a different spine curvature.
Tomography 08 00039 g012
Table 1. Architectural ablation study: The performance progressively improves with the addition of each component. While TransVert outperforms the separate Affine and Deformable TransVert models in terms of Dice score and Hausdorff distance, including both of the Affine and the Deformable decoders in the TransVert+ model performs better than TransVert.
Table 1. Architectural ablation study: The performance progressively improves with the addition of each component. While TransVert outperforms the separate Affine and Deformable TransVert models in terms of Dice score and Hausdorff distance, including both of the Affine and the Deformable decoders in the TransVert+ model performs better than TransVert.
SetupDiceHausdorff (mm)nWESD
AffTransVert0.884712.400.3181
DefTransVert0.94059.080.2430
TransVert0.94267.930.2779
TransVert+0.95105.700.1797
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Bayat, A.; Pace, D.F.; Sekuboyina, A.; Payer, C.; Stern, D.; Urschler, M.; Kirschke, J.S.; Menze, B.H. Anatomy-Aware Inference of the 3D Standing Spine Posture from 2D Radiographs. Tomography 2022, 8, 479-496. https://doi.org/10.3390/tomography8010039

AMA Style

Bayat A, Pace DF, Sekuboyina A, Payer C, Stern D, Urschler M, Kirschke JS, Menze BH. Anatomy-Aware Inference of the 3D Standing Spine Posture from 2D Radiographs. Tomography. 2022; 8(1):479-496. https://doi.org/10.3390/tomography8010039

Chicago/Turabian Style

Bayat, Amirhossein, Danielle F. Pace, Anjany Sekuboyina, Christian Payer, Darko Stern, Martin Urschler, Jan S. Kirschke, and Bjoern H. Menze. 2022. "Anatomy-Aware Inference of the 3D Standing Spine Posture from 2D Radiographs" Tomography 8, no. 1: 479-496. https://doi.org/10.3390/tomography8010039

Article Metrics

Back to TopTop