Abstract
Objective
Microwave lung ablation (MWA) is a minimally invasive and inexpensive alternative cancer treatment for patients who are not candidates for surgery/radiotherapy. However, a major challenge for MWA is its relatively high tumor recurrence rates, due to incomplete treatment as a result of inaccurate planning. We introduce a patient-specific, deep-learning model to accurately predict post-treatment ablation zones to aid planning and enable effective treatments.
Materials and methods
Our IRB-approved retrospective study consisted of ablations with a single applicator/burn/vendor between 01/2015 and 01/2019. The input data included pre-procedure computerized tomography (CT), ablation power/time, and applicator position. The ground truth ablation zone was segmented from follow-up CT post-treatment. Novel deformable image registration optimized for ablation scans and an applicator-centric co-ordinate system for data analysis were applied. Our prediction model was based on the U-net architecture. The registrations were evaluated using target registration error (TRE) and predictions using Bland-Altman plots, Dice co-efficient, precision, and recall, compared against the applicator vendor’s estimates.
Results
The data included 113 unique ablations from 72 patients (median age 57, interquartile range (IQR) (49–67); 41 women). We obtained a TRE ≤ 2 mm on 52 ablations. Our prediction had no bias from ground truth ablation volumes (p = 0.169) unlike the vendor’s estimate (p < 0.001) and had smaller limits of agreement (p < 0.001). An 11% improvement was achieved in the Dice score. The ability to account for patient-specific in-vivo anatomical effects due to vessels, chest wall, heart, lung boundaries, and fissures was shown.
Conclusions
We demonstrated a patient-specific deep-learning model to predict the ablation treatment effect prior to the procedure, with the potential for improved planning, achieving complete treatments, and reduce tumor recurrence.
Clinical relevance statement
Our method addresses the current lack of reliable tools to estimate ablation extents, required for ensuring successful ablation treatments. The potential clinical implications include improved treatment planning, ensuring complete treatments, and reducing tumor recurrence.
Key Points
-
Reliable tools to predict the extent of ablation treatments are currently lacking.
-
Our novel patient-specific deep-learning algorithm was shown to predict ablation zones with higher accuracy and less bias compared to the currently used estimates provided by applicator vendor.
-
Our method for ablation prediction allows for real-time clinical deployment, with potential for improved treatment planning and reduced tumor recurrence.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Microwave lung ablation (MWA) is used to eradicate tumors in the lung [1], and offers a low morbidity and inexpensive alternative to surgical resection [2]. Additionally, MWA shows no detectable long-term effects on pulmonary function, making it an attractive alternative to stereotactic body radiation therapy [3]. However, widespread adoption of MWA has been hindered by higher local recurrence rates [4], which result from incomplete ablations with insufficient tumor margins [5].
Achieving adequate margins (> 5 mm) with MWA is challenging [5]. The ablation device vendors provide a chart of the expected ablation/treatment zone dimensions as a function of power and duration (henceforth referred to as the vendor model), which is used for pre-procedure planning. The vendor model is based on ablations of ex-vivo animal organs. However, clinical ablation zones are affected by nearby vessels and airways that act as heat sinks, pleural boundaries that cause distortions, and complex lung deformations due to patient position and breathing. Observed ablation zones demonstrate marked deviation from the vendor models in size and shape [6,7,8]. There is wide variability in ablation zones even for the same power and duration settings [6, 9, 10]. As a result, the treatment plan based on the vendor model may significantly differ from the actual treatment, potentially leading to incomplete thermal destruction of tumors and failure to establish an adequate margin.
Biophysical models that attempt to simulate ablation zones by solving partial differential equations governing the physical processes during MWA [11, 12] are limited in their ability to account for patient-specific tissue properties, lack clinical validation, and are too computationally demanding for clinical application. In a recent study relating “antenna work” (power x duration) with ablation zone volume, the authors noted the wide range of volumes for similar work and emphasized the importance of accounting for patient-specific characteristics [10].
In this work, we introduce a new, data-driven approach for ablation prediction. We hypothesize that pre-procedure computerized tomography (CT) scan intensity features contain patient-specific characteristics that can inform the ablation zone size and shape, and along with the ablation settings, are predictive of the final ablation zone extents. We treat the problem as that of learning a deep-neural network-based parameterized function that takes the pre-procedure CT, the position of the applicator, and ablation power and duration as input and predicts the post-procedure ablation zone as output. Our goal is to predict the final ablation zone at the first follow-up post-procedure scan.
Materials and methods
Data
The data was obtained as part of an institutional review board approved (IRB/Privacy Board-A, protocol 17-353)Footnote 1 retrospective study of patients who underwent MWA at our institution between 01/2015 and 01/2019. Exclusion criteria included the use of a probe other than NeuWave PR (Ethicon US), multiple probes or multiple burns performed at the same site, two or more adjacent sites with overlapping ablation zones, or if background lung parenchyma could not be differentiated from the ablation zone, resulting in 113 unique ablations from 72 patients (see Fig. 1). Pre-procedure and 1-month follow-up post-procedure CT scans were processed. MWA power and duration were obtained from radiology reports.
Pre-processing pipeline
Figure 2b, c provide an overview of the pre-processing steps in our pipeline. The tumor and ablation zone were manually segmented in the pre and follow-up scans, respectively (Fig. 2b), by a radiologist with > 15 years of experience [13]. Contours were drawn marking the boundary of the tumor and ablation zone on axial CT image slices in 3D Slicer software (v4.11.1) [14], and further edited on coronal and sagittal views. As ablation zones can have internal regions of high or low Hounsfield units (HU), all pixels within the drawn boundary were assumed to be part of the segmentation independent of the HU. The applicator position was defined by the radio-lucent tract visible in the ablation zone on the follow-up scan (Fig. 2c).
For the vendor model, we used the vendor’s user interface software (v3.1.0), which provided dimensions of expected ablation ellipsoids for different powers (35 W to 65 W in 5 W increments) and durations (1 to 10 min in 1 min increments). Ablation zone dimensions were linearly interpolated for power and duration in between these specifications.
We applied a novel deformable image registration methodology for ablation procedures, utilizing the free-form deformation model based on B-spline transformation [15] to register the pre and follow-up scans (Fig. 2b). An initial rigid registration was performed for coarse alignment followed by a coarse-to-fine multi-resolution deformable registration where the result of each resolution acted as input to the next [16]. We constrained the registration by focusing on the regions surrounding the tumor and ablation zone and applied a rigidity penalty on the tumor to avoid large tumor deformations. The registration optimization used mutual information to measure the similarity between images and restricted excessive bending of the transform to avoid unnatural deformations via regularization. We followed Klein et al to set various registration parameters [17]. More details on the registration methodology can be found in Supplemental G. All registrations were performed using the SimpleElastix image registration library (v0.9.1) [18, 19].
To enable comparison of ablation zones at different positions/orientations across patients, we defined an applicator-centric co-ordinate system (ACCS) [20], an oblique 3D co-ordinate grid oriented along the applicator and centered at the applicator tip (Fig. 2c) for all computation. The ACCS grid size was set to 64 × 64 × 64 mm3 with a sampling rate of 1 mm along each dimension. Numpy (v1.21.2) [21], SciPy (v1.7.3) [22], and NiBabel (v3.2.1) [23] Python libraries were used for data processing.
Model
Our model was a fully convolutional neural network based on the U-Net [24] architecture (Fig. 2d and Supplemental A). The U-net consists of (1) the encoder that reduces the dimensionality of the input, aggregating semantic information (what is in the image), (2) the decoder that combines the encoded semantic information with higher resolution spatial information from initial encoding layers via skip connections (where in the image) and up-samples them to the original size to produce output. The first input channel was the 3D pre-CT scan. The ablation power/duration is represented by the dimensions of the vendor ellipsoid, which was hence used as the 2nd input channel after placing at the applicator position.
Training
The segmented ablation zones on follow-up scans were used as ground truth for training since they are the clinical standard for assessing lung ablation zones and treatment margins [25, 26]. We utilized the sum of cross entropy and the Dice co-efficient between the predicted and ground truth as the loss function to train the network [27]. We performed nested cross-validation [28] for training and testing (Supplemental B): (1) In the outer loop, we performed 7-fold cross-validation (CV) [29] resulting in 7 train-test (TT) CV splits, where, in each split, 6 folds were used for model training and the remaining 7th fold (unseen during training) used for testing. The results on the test folds from all the 7 TT splits were averaged to compute the final test set performance [29, 30]. (2) In the inner loop, the training data in each TT split (6 folds) was further split into 5 train-validation (TV) CV splits, where, in each TV split, the model was trained on 4 folds and tuned on the remaining 5th validation fold held out during training. The purpose of these further CV splits was to train different models on each of the 5 TV splits of the training data from a single TT split, which were then applied on the corresponding TT split test fold and the results were ensembled by averaging their softmax outputs. Hence, we trained 5 different models from the training data of each TT split, resulting in a total of 7 × 5 = 35 models on the whole dataset. We followed Isensee et al for the training methodology (Supplemental B; https://github.com/MIC-DKFZ/nnUNet) [27, 31]. The code for pre-processing and model training can be accessed at https://github.com/tenres/Abl-Pred.
Statistics
Image registrations were evaluated using the target registration error (TRE; Supplemental C) [32]. Volumes and shapes of our prediction and vendor model were compared with ground truth ablation zones using Bland-Altman plots [33]. Paired and two sample t-tests were used to test for differences between the ground truth and vendor model. Levene’s test was used to test the differences in the limits of agreement (LOA) [34]. The degree of overlap between prediction and ground truth was computed using the Dice co-efficient [35], precision, and recall [36] (Supplemental D). The mean and 95% confidence intervals (CIs) were computed [29], which were compared against that of the vendor model. Statsmodel (v0.13.5) [37], and SciPy (v1.7.3) [22] Python libraries were used for all analyses.
Results
After exclusions, there were 72 participants (median age 57 years, 47–69 (interquartile range, IQR); 41 women) with 113 MWA procedures (Fig. 1). Cases with poor TRE (> 2 mm) were excluded from training/testing, resulting in 52 ablations from 40 patients (Fig. 1). Table 1 lists tumor, cancer, and CT imaging details. There were 17 ablations performed using the median power of 65 watts (range 20–65 watts), and duration of 5 min (range 1–10 min), and the variability in their ablation zone volumes (same power/duration) is shown in Fig. 2a. The TRE on the final 52 cases is shown in Fig. 4a with a median of 0.88, IQR (0.50–1.31). The inference time of our algorithm was a fraction of a second.
Bias and variability of volume and shape for our prediction and vendor models
Bland-Altman plots comparing the volume and shape characteristics of our prediction and vendor model against the true ablation zone are shown in Fig. 3. The volume bias (Fig. 3a) of our prediction (–780), was lower compared to the vendor model (2410) and this difference was significant (p < 0.001). Assuming differences from ground truth to be normally distributed (y-axis), the bias of our prediction was not different from the line of equality with ground truth (0 bias) (p = 0.169); whereas, the vendor model bias was significant, p < 0.001. The range in the LOA of our model was smaller than the vendor model and this difference was significant (p < 0.001). The sphericity (Fig. 3b) and surface-to-volume ratio (Fig. 3c) biases of our prediction were lower compared to the vendor model (0.15 vs. 0.22 and –0.12 vs. –0.24, p < 0.001). The volume and shape differences demonstrate better agreement of our prediction with the true ablation zone compared to vendor.
Dice score, precision, recall
We calculated the Dice scores of our prediction and vendor model with the true ablation zones (Fig. 4b and Table 2). The median Dice scores on the test set were 0.62 ± 0.12 (CI: 0.56–0.64) for our model compared with 0.56 ± 0.16 (CI: 0.49–0.59) for the vendor model, demonstrating an 11% improvement over the vendor model. The vendor model also showed a broader IQR consistent with the large LOA range. We saw improved precision of our model with median scores of 0.65 ± 0.22 (CI: 0.60–0.75) compared with 0.43 ± 0.23 (CI: 0.39–0.54) for the vendor model (Fig. 4c and Table 2). The vendor model demonstrated better recall with median scores of 0.89 ± 0.14 (CI: 0.79–0.89) compared with 0.70 ± 0.22 (CI: 0.60–0.73) for our model.
Effect of patient-specific local anatomy
To better understand the quantitative improvement in Dice score, we evaluated our prediction and vendor model in their ability to account for patient-specific local anatomy. Several scenarios with two examples each are presented.
Heat-sink effects
Figure 5a demonstrates local heat-sink effects from adjacent vessels and airways that locally modulate the shape of the ablation zone. Our prediction does not extend into the vessels or airways, similar to the ground truth. In contrast, the vendor model overestimates the ablation zone and extends into the vessels as well as the background lung parenchyma. Figure 5b demonstrates global heat-sink effects on ablation zone size and shape. In example 1 (left), multiple small and medium-sized vessels (blue arrows) are seen in the cross-section coming in and out of the plane with a cumulative effect of decreasing the overall ablation zone size. Our model closely follows the true ablation zone whereas the vendor overestimates the back and sides of the applicator. Notably, the effects of vessels are not distributed symmetrically around the needle. In example 2 (right), multiple vessels (blue arrows) are now seen in the plane. Again, our prediction conforms to the asymmetric narrowing and decreased size of the ablation zone. Figure 4e demonstrates this effect quantitatively, where our model has a higher Dice overlap with true ablation compared to vendor when close to large vessels that cause heat-sink effects, whereas is similar to vendor when away from large vessels.
Borders
Figure 5c demonstrates the effects of fissures, and faint radio-dense lines demarcated by blue arrows on post images. Our prediction follows the fissure, whereas the vendor model extends beyond. Figure 5d shows the effect of chest-wall boundaries, which our prediction adheres to closely, whereas the vendor model extends beyond the boundary in example 2. Figure 6a shows mediastinal borders where our prediction adheres to the local border whereas the vendor underestimates (example 1) or overestimates (example 2) the ablation zone.
Ablation shape
The overall shape of the ablation zone is affected by lung anatomy. Figure 6b example 1 shows asymmetric narrowing of the ablation zone to which our model adheres closely but which the vendor model overestimates. In example 2, the ablation zone appears rotated relative to the vendor model, possibly due to the adjacent vascular structures, which our model is able to predict. Figure 6c example 1 shows our prediction narrowing only at the rear end of the ablation, similar to ground truth observation. More examples in 3D are provided in Supplemental H.
Failures
We also identified categories of scenarios in which our model consistently failed to predict the ablation zone. Figure 6c example 2 and Fig. 5c example 2 (green arrow) show cases where the ablation zone contains a large cavity, which our model excludes. In Fig. 6d example 1, our prediction extends too far and wide along the back of the probe. A close review of intra-procedure ablation images revealed that the lung was torqued and compressed intra-operatively during applicator positioning. Example 2 shows confounding background anatomy where our model incorrectly extends beyond the ablation zone (towards the left), adhering to scars that have an ablation border-like appearance.
Discussion
We have presented a data-driven deep-learning model to predict lung ablation zones as they appear on the follow-up scan based on pre-procedure imaging and ablation parameters. Compared with existing vendor models, we demonstrated higher Dice scores, decreased volume variability, and decreased bias in volume and shape from true ablation. Our model is also able to better account for the effects of patient-specific local anatomy on the ablation zone including heat-sink effects, organ/lobes boundaries, and variations in the shape of the ablation zone. Moreover, we note that the improvement in Dice’s performance is likely an underestimate (see Supplemental E).
Clinically, the tradeoff between precision and recall is significant. A bias towards over-estimating the ablation zone increases recall but may cause the operator to under-treat the tumor, leading to local recurrence. Whereas a bias towards under-estimating the ablation zone can increase precision but may result in overly aggressive treatment and potential complications (Supplemental F). The low precision and high recall of the vendor model show a striking mismatch that translates clinically to over-estimation of the ablation zone, incomplete treatment, and high local recurrence. Our model showed no bias, a marked improvement in precision at the cost of some decrease in recall, potentially averting tumor under-treatment, local recurrence, and repeat procedures. Depending on location and proximity to critical structures, one could tune the threshold to modulate the precision-recall tradeoff.
Our approach offers several advantages over biophysical ablation modeling. We directly incorporate patient-specific observed data via CT scans whereas biophysical models depend on multiple tissue properties that are impractical to measure [11, 38]. We directly optimize the likelihood of the observed ablation zones on follow-up CT [29], the standard time point for treatment verification, therefore incorporating post-ablation processes such as tissue contraction and healing [39, 40]. Whereas biophysical models lack such feedback and only model thermal dose and ensuing cell death. Our algorithm is faster (inference takes a fraction of a second) making it suitable for clinical implementation [11].
There are several limitations of this work. Our dataset is small, which limits the ability of our current model to generalize to uncommon ablation zones. This can be overcome with more data. The model does not include intra-operative changes that likely account for multiple failed cases discussed before. The model also assumes the accurate position of the applicator, which can be difficult to ascertain from follow-up imaging in some cases. The model is trained to predict the ablation zone on the one-month follow-up scan, which correlates with necrotic tissue [41]. Moreover, the one-month follow-up scan has been shown to be a clinically relevant surrogate for necrosis in studies that correlate ablation zone margin with clinical outcomes [4, 39, 40] and hence is a clinically relevant endpoint for prediction. Notably, common sources of noise on early post-procedure imaging such as hemorrhage, edema, and pneumothorax resolve by the standard of care one-month follow-up scan. We exclude ablations with multiple burns or multiple applicators, which limits the generalizability of the model. However, given enough input data, we believe the model will be able to accommodate all these factors as well. Our model does not directly incorporate histology information and it is possible that histology may affect heat. We note however that histology has been shown to be encoded in CT data and therefore may be indirectly encoded in the current model [42]. Incorporating additional relevant, non-imaging input data may be considered in future work.
Our method has several important potential clinical applications [43, 44]. Pre-procedure planning with our model can be used to establish applicator trajectory, power, and duration. These could be tuned to optimize the margin with the goal of decreasing local recurrence and is safer and cost-effective compared to intra-operative biopsy for margin control [45]. Further studies are needed to establish margin assessment using our tool. The model can be used to identify “weak points” in the ablation that may require additional treatment. The rapidity of prediction allows for real-time clinical tuning by the operator. By pre-defining a minimum ablation margin around the nodule, our model can predict a set of “optimal solutions” from which the operator can select. Finally, our approach can be applied more broadly to other organs and other treatment modalities.
In conclusion, we have presented a patient-specific data-driven deep-learning approach to predict the post-procedure lung ablation zone. The work demonstrates a novel application of deep learning for a common yet unresolved clinical problem with improvement over the current standards and the best results to date to our knowledge. We anticipate generalizations of our approach to have broad applications to treatment planning and prediction in oncology.
Notes
Study was determined as low risk; the patient consent requirement was waived.
Abbreviations
- ACCS:
-
Applicator-centric co-ordinate system
- CI:
-
Confidence interval
- CT:
-
Computerized tomography
- CV:
-
Cross validation
- HU:
-
Hounsfield unit
- IQR:
-
Interquartile range
- kVp:
-
Kilovoltage peak
- LLL:
-
Left lower lobe
- LOA:
-
Limits of agreement
- LUL:
-
Left upper lobe
- METS:
-
Metastatic disease
- MWA:
-
Microwave lung ablation
- RLL:
-
Right lower lobe
- RML:
-
Right middle lobe
- RUL:
-
Right upper lobe
- TRE:
-
Target registration error
- TT:
-
Train-test
- TV:
-
Train-validation
References
Simon CJ, Dupuy DE, Mayo-Smith WW (2005) Microwave ablation: principles and applications. Radiographics 25:S69–S83
Kwan SW, Mortell KE, Hippe DS, Brunner MC (2014) An economic analysis of sublobar resection versus thermal ablation for early-stage non–small-cell lung cancer. J Vasc Interv Radiol 25:1558–1564
Binkley MS, Shrager JB, Chaudhuri A et al (2016) Time course and predictive factors for lung volume reduction following stereotactic ablative radiotherapy (SABR) of lung tumors. Radiat Oncol 11:1–8
Healey TT, March BT, Baird G, Dupuy DE (2017) Microwave ablation for lung neoplasms: a retrospective analysis of long-term results. J Vasc Interv Radiol 28:206–211
Gao S, Stein S, Petre EN et al (2018) Micropapillary and/or solid histologic subtype based on pre-treatment biopsy predicts local recurrence after thermal ablation of lung adenocarcinoma. Cardiovasc Intervent Radiol 41:253–259
Dev A, Keshavamurthy KN, Salkin R et al (2021) Quantitative analysis of tissue contraction and volume variability of microwave lung ablation zones. J Vasc Interv Radiol 33:S58–S59
Glidden DT, Cook M, Collins S et al (2014) Clinically observed ablation volumes as compared to vendor specified volumes: the emperorʼs new clothes. J Vasc Interv Radiol 3:S76
Blackmon SH, Sterner RM, Eiken PW et al (2021) Technical and safety performance of CT-guided percutaneous microwave ablation for lung tumors: an ablate and resect study. J Thorac Dis 13:6827
Dev A, Keshavamurthy KN, Salkin R et al (2022) Abstract No. 124 Quantitative analysis of tissue contraction and volume variability of lung microwave ablation zones. J Vasc Interv Radiol 33:S58–S59
Huber TC, Miller G, Patrie J, Angle JF (2021) Relationship of antenna work and ablation cavity volume following percutaneous microwave ablation of hepatic tumors. J Vasc Interv Radiol 32:536–543
Prakash P (2010) Theoretical modeling for hepatic microwave ablation. Open Biomed Eng J 4:27
Chiang J, Wang P, Brace CL (2013) Computational modelling of microwave tumour ablations. Int J Hyperthermia 29:308–317
Hershman M, Yousefi B, Serletti L et al (2021) Impact of interobserver variability in manual segmentation of non-small cell lung cancer (NSCLC) applying low-rank radiomic representation on computed tomography. Cancers (Basel) 13:5985
Fedorov A, Beichel R, Kalpathy-Cramer J et al (2012) 3D Slicer as an image computing platform for the Quantitative Imaging Network. Magn Reson Imaging 30:1323–1341
Rueckert D, Sonoda LI, Hayes C, Hill DL, Leach MO, Hawkes DJ (1999) Nonrigid registration using free-form deformations: application to breast MR images. IEEE Trans Med Imaging 18:712–721
Lester H, Arridge SR (1999) A survey of hierarchical non-linear medical image registration. Pattern Recognit 32:129–149
Klein S, Staring M, Murphy K et al (2009) Elastix: a toolbox for intensity-based medical image registration. IEEE Trans Med Imaging 29:196–205
Marstal K, Berendsen F, Staring M, Klein S (2016) SimpleElastix: A user-friendly, multi-lingual library for medical image registration. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops. IEEE. pp. 134–142
Yaniv Z, Lowekamp BC, Johnson HJ, Beare R (2018) SimpleITK image-analysis notebooks: a collaborative environment for education and reproducible research. J Digit Imaging 31:290–303
Keshava KN, Kimia BB, Cook M et al (2015) A methodology to analyze treatment zone geometry and variability of percutaneous thermal ablation. In: Energy-based Treatment of Tissue and Assessment VIII. SPIE, pp. 240–247
Harris, Millman CR, der Walt KJ, van SJ et al (2020) Array programming with NumPy. Nature 585:357–362. https://doi.org/10.1038/s41586-020-2649-2
Virtanen P, Gommers R, Oliphant TE et al (2020) SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat Methods 17:261–272. https://doi.org/10.1038/s41592-019-0686-2
Brett M, Markiewicz CJ, Hanke M et al (2024) nipy/nibabel: 5.2.1. Zenodo https://doi.org/10.5281/zenodo.10714563
Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5–9, 2015, Proceedings, Part III 18. pp. 234–241
Yan P, Tong A, Nie X, Ma M (2021) Assessment of safety margin after microwave ablation of stage I NSCLC with three-dimensional reconstruction technique using CT imaging. BMC Med Imaging 21:96
Chheang S, Abtin F, Guteirrez A et al (2013) Imaging features following thermal ablation of lung malignancies. In: Seminars in interventional radiology. pp. 157–168
Isensee F, Jaeger PF, Kohl SAA et al (2021) nnU-Net: a self-configuring method for deep learning-based biomedical image segmentation. Nat Methods 18:203–211
Stone M (1974) Cross-validatory choice and assessment of statistical predictions. J R Stat Soc Series B (Methodological) 36:111–147
Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT press
Blum A, Kalai A, Langford J (1999) Beating the hold-out: bounds for k-fold and progressive cross-validation. In: Proceedings of the twelfth annual conference on computational learning theory, Association for Computing Machinery, New York, July 1999
Chen L-C, Papandreou G, Kokkinos I et al (2017) Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFS. IEEE Trans Pattern Anal Mach Intell 40:834–848
Fitzpatrick JM, West JB (2001) The distribution of target registration error in rigid-body point-based registration. IEEE Trans Med Imaging 20:917–927
Bland JM, Altman D (1986) Statistical methods for assessing agreement between two methods of clinical measurement. The lancet 327:307–310
Bland JM (2023) Is there a method I can use to assess the differences in limits of agreement between groups? In: https://www-users.york.ac.uk/~mb55/meas/comp_loa.htm
Dice LR (1945) Measures of the amount of ecologic association between species. Ecology 26:297–302
Powers DMW (2020) Evaluation: from precision, recall and F-measure to ROC, informedness, markedness and correlation. J Mach Learn Technol 2:37–63
Seabold S, Perktold J (2010) Statsmodels: econometric and statistical modeling with python. 9th Python in Science Conference, SCRIP, Austin, 28 June–3 July 2010
Sebek J, Albin N, Bortel R, Natarajan B, Prakash P (2016) Sensitivity of microwave ablation models to tissue biophysical properties: a first step toward probabilistic modeling and treatment planning. Med Phys 43:2649–2661
Ahmed M, Solbiati L, Brace CL et al (2014) Image-guided tumor ablation: standardization of terminology and reporting criteria—a 10-year update. Radiology 273:241–260
Moussa AM, Ziv E, Solomon SB, Camacho JC (2019) Microwave ablation in primary lung malignancies. Semin Intervent Radiol 36:326–333
Yamamoto A, Nakamura K, Matsuoka T et al (2005) Radiofrequency ablation in a porcine lung model: correlation between CT and histopathologic findings. AJR Am J Roentgenol 185:1299–1306
Perez-Johnston R, Araujo-Filho JA, Connolly JG et al (2022) CT-based radiogenomic analysis of clinical stage I lung adenocarcinoma with histopathologic features and oncologic outcomes. Radiology 303:664–672
Dupuy DE, Fernando HC, Hillman S et al (2015) Radiofrequency ablation of stage IA non–small cell lung cancer in medically inoperable patients: results from the American College of Surgeons Oncology G roup Z 4033 (Alliance) trial. Cancer 121:3491–3498
Yan P, Tong A, Nie X, Ma M (2021) Assessment of safety margin after microwave ablation of stage I NSCLC with three-dimensional reconstruction technique using CT imaging. BMC Med Imaging 21:1–11
Vasiniotis Kamarinos N, Vakiani E, Gonen M et al (2022) Biopsy and margins optimize outcomes after thermal ablation of colorectal liver metastases. Cancers (Basel) 14:693
Funding
This research was funded in part through the NIH/NCI Cancer Center Support Grant P30 CA008748.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Guarantor
The scientific guarantor of this publication is Krishna Nand Keshava Murthy.
Conflict of interest
The authors of this manuscript declare relationships with the following companies: Ethicon.
Statistics and biometry
Two of the authors have significant statistical expertise. No complex statistical methods were necessary for this paper.
Informed consent
Written informed consent was waived by the Institutional Review Board.
Ethical approval
The MSKCC Institutional Review Board approval was obtained.
Study subjects or cohorts overlap
No study subjects or cohorts have been previously reported.
Methodology
-
Retrospective
-
Observational
-
Performed at one institution
Additional information
Publisher’s Note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Keshavamurthy, K.N., Eickhoff, C. & Ziv, E. Pre-operative lung ablation prediction using deep learning. Eur Radiol (2024). https://doi.org/10.1007/s00330-024-10767-8
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s00330-024-10767-8