Introduction

When sand casted parts come out of the mold they have to be cleaned in a fettling process to remove sprues, runners, risers, and flashing. Flashing, also called burrs, is the material left on the cast part in the separation plane between the two molds. Figure 1 shows a cast part with risers and flashing. Fettling of cast parts is important to ensure that the part meets its design requirements. Removal of sprues, risers, and runners can be done in a cutting process that leaves a burr to be removed. The burrs from the cutting and flashing are removed in a deburring process. Deburring can be very challenging and are mostly done manually (Aertbeliën, 2009) where the workers are exposed to high noise and vibration levels. It is therefore desirable to automate the deburring process.

Detecting and measuring the size of burrs to adjust the feed rate and registration of CAD model with the workpiece-scan for creating a deburring tool path are some of the essential steps in automated deburring process. A deburring process is traditionally planned on either a reference CAD model of reference workpiece (Onstein et al., 2020), this includes both the tool path and machining parameters. The challenge with automating deburring of cast parts is that both the workpiece and the burrs vary in shape and size due to the solidifying in the casting process. During solidifying, uneven shrinkage and deformation can occur. The uneven shrinkage and deformation can be controlled by optimizing the casting process, but it cannot be completely avoided (Huang et al., 2021). To formalize this observation, we define the domain gap as a measure of the discrepancy between the CAD (source) and target (scan) domains. This means that the deburring process has to be robust to domain gap, i.e. corresponding CAD and scan geometric variations.

The conventional approach to burr localization (Huang et al., 2021) follows a two-step coarse and fine point cloud registration approach where the relative positions of workpieces and CAD models must be accurate. However, such approach are sensitive to variation in point density, missing data (partial scans) and noise in the scan. Furthermore, registration of the workpieces and CAD models fails due to domain gap issues.

Fig. 1
figure 1

Cast part with riser/feeder and flashing

In this paper, we propose a machine learning approach that is robust to missing data in the scans, view point, noise, rotation and scale. Our approach is able to learn features of the CAD model/scan without any labeled data through a series of augmentations to the CAD models. Note that the our approach uses only CAD models for training and is able to generalize to scans during testing. Furthermore, using the learned features we show a burr localization approach that is able to estimate burr height similar to high resolution CT scans with Z-statistic value (\(z=0.279\)) supporting the hypothesis that the CT measured and our estimate burr height distributions are similar. This paper is structured as follows. “Background” section briefly reviews previous works on deburring and burr measurement. In “Methods” section we present the proposed burr detection and quantization approach, and the data collection setup. In “Results” section we present evaluation of the CAD-scan registration and burr size estimation. Finally, in “Discussion and conclusion” section we conclude the paper with discussion and future work.

Background

Scan-to-CAD registration for tool path planning

The deburring process consists of, among other steps, detecting and quantifying the burrs, generating a deburring tool path and optimising machining parameters. Because of the geometric variations caused by deformations and uneven shrinkage during solidifying, the deburring process has to be adapted to each individual workpiece. The deburring tool path is traditionally generated off-line based on the CAD model using a CAM software. This reference path has to be registered onto the workpiece, where the goal is to find the corresponding relationship between two point sets and compute an appropriate transformation (Du et al., 2015). The most widely used method for registration of 3D shapes is the iterative closest point (ICP) algorithm (Besl & McKay, 1992).

Kosler et al. (2016) propose a method using ICP to register a 3D scan with the CAD model to adapt a reference tool path, generated using off-line teaching, to the workpiece. Song and Song (2013) propose a method that generates a tool path using CAM software and then correct the path based on teaching points using an ICP-based contour matching algorithm. These two ICP-based methods assume that the deformations caused by the solidification can be ignored because it is not significant for the overall geometry.

A method that consider the deformations are presented in Villagrossi et al. (2017). A reference tool path is generated trough teaching on reference workpiece. The workpiece to be deburred and the reference are both scanned and registered using ICP to get a rough alignment. To compensate for the deformations, a set of control points are taught on the reference part in an area where burrs are highly unlikely. The same control points are found on the workpiece, and the local deformations are computed. The reference tool path is compensated based on the computed deformations. Kuss et al. (2016) propose a method where a series of CAD models are generated based on the dimensional tolerances of the workpiece CAD model. Registration using ICP is then performed on the scan of the workpiece and the set of CAD models. The CAD model with the best fit is used for generating a tool path using CAM software. Both proposed methods work well for simple geometries, but both will be challenging and time consuming for more complex geometries.

Béarée et al. (2011) and Huang et al. (2021) propose methods that segment the part and do registration on each sub-segment. The tool path is then adapted to each segment. Both methods are based on the assumption that the deformations within one segment is negligible.

Adjusting feed rate

A burr can be defined as undesirable or unwanted projections of the material formed as the result of a manufacturing process (Aurich et al., 2009). The shape and size of a burr can be described in different ways, it can be defined by its longitudinal and cross-sectional profile including thickness, height, and radius. The (ISO 13715:2019, 2020) standard use only one parameter to describe the burr. This is the height from the intended geometry and top of the burr as shown in Fig. 2.

Fig. 2
figure 2

External edge geometry as described in ISO 13715:2019 (2020)

During deburring, a set of machining parameters has to be set. One of these are feed rate. An approach to tuning this parameter is to choose a feed rate that is slow enough to be able to remove the largest burr. This will most likely increase the cycle time. If, on the other hand, the feed rate is too high, the tool can be unable to remove the burr. Methods have been proposed to adjust the feed rate based on burr height (Lai et al., 2018; Xiong et al. 2018), where the feed rate is adjusted based on given burr height thresholds.

Burr measurement

There is a large number of burr measuring and detection methods available. The most appropriate system depend on the application, requested accuracy, and the burr values to be measured. Burr measurement systems can be divided into two main categories: in-process and out of process (Franke et al., 2010). Out of process can be divided into two categories: with contact and contactless. Contact methods include stylus methods and contactless methods include optical and electro-mechanical solutions. In-process methods include process monitoring, force and sound emission analysis.

Contact methods are only suitable for measuring burr height and can be limited due to workpiece material stiffness. Various optical systems for measuring burrs are available. These include camera systems, microscopes and lasers. The measured values from the optical systems are analyzed using specific measurement software. To characterize non-uniform burrs, several measurement are necessary.

A thorough comparison of burr measurement method formed by drilling and milling is presented in Franke et al. (2010). The results of the test show variations in the measurements, especially when the burrs are formed in an angle. It is concluded that very high caution is necessary for the comparison of burr measurements.

A burr measurement method based on burr surface area is proposed by Bahçe and Özdemir (2021), where they aim at evaluating burr height, arc length, area, and the geometrical characteristics of the burr. The proposed method capture a 3D scan of the burr and fit a cubic Bézier curve around the outline of the burr. All burr parameters are gathered from this curve.

Tellaeche and Arana (2016) use ICP to register a scan from a 3D vision systems with the CAD model. The matching is used to characterize different type of burrs.

Fig. 3
figure 3

3D image acquisition setup with rotating workspace and high-resolution 3D camera at different heights

Summary

In summary, local CAD-scan registration approaches such as ICP, require a good initial guess of the transformation between CAD and scan for convergence. Furthermore, earlier works assume nearly identical CAD and scan ignoring the deformation caused by solidification process (domain gap). Proposed methods that consider the deformations are time consuming and not applicable for complex geometries.

Compared to earlier works, our CAD-Scan registration approach is robust to deformation caused by the solidification process, differences between CAD and scan, and relies on global registration unlike ICP. Moreover, burr measurement is based on local smoothness of scan surface by reconstructing burr free scan which do not assume exact match between CAD and scan. Common geometric features between CAD and scan that simulate domain differences are learned on the CAD model using sparse convolutional neural networks (CNN). In the next section, we present the proposed approach to learning features that are robust to domain gap between CAD and scan followed by burr measurement.

Methods

In the process of developing an automatic burr detection and deburring path planning system, having to align the reference CAD model to scan is the first step. Compared to CAD models, scans exhibit camera noise, missing data, manufacturing artifacts and geometric variations. Owing to the irregularity and variations between the reference CAD model and scan, it is difficult to align synthetic data to scan - which complicates burr measurement. To circumvent this problem, this paper proposes a deep learning model that extracts robust features from reference CAD model and scans that is robust CAD-to-scan domain gap, as well as variations in viewpoint, noise, rotation and scale. Figure 4 shows our pipeline for automated burr detection. Our acquisition setup captures high-resolution scans of the parts, with efficient alignment of registration ground truth (see “3D data acquisition” section). CNN features are extracted from a pre-trained model for the scan and the reference CAD model. Traditional CNN techniques are mainly applied to data with a structured grid, point cloud, on the other hand consists of sparse and unordered set of 3D points. These properties of point clouds make it difficult to use traditional CNN architectures for point cloud processing. Therefore, pre-trained model is based on a sparse CNN deep learning model (Choy et al., 2019) trained only on CAD models (see “Training on CAD models” section). Finally, given the approximate alignment between the reference CAD model and scan, a burr detection approach (see “Burr detection and measurement” section) is proposed to quantify the height of the burr as shown in Fig. 4.

3D data acquisition

Our data acquisition setup consists of a turntable and a high-resolution 3D camera, to capture accurate point clouds of the parts and acquire ground truth registration data in an efficient manner. The sensor used in our setup is a Zivid OneFootnote 1, which is an RGB-D camera based on structured light, with \(1920\times 1200\) pixels and a Field-of-View (FOV) of \(40^\circ \times 26^\circ \). The acquisitions were performed with High Dynamic Range (HDR) imaging with 5 different exposure times from 40 to 100 ms, at a distance of about 0.5 m, which gives a nominal depth resolution of 0.18 mm.

Each workpiece is placed on a turntable with a dark background and ArUco markers along the edges, and recorded with 8 different rotation angles and 3 different camera heights (see Fig. 3). For each height, the camera tilt is adjusted to keep the part in view. This gives 24 pointclouds (scans) per workpiece, resulting in 480 total number of scans that are automatically co-registered during post-processing, using ArUco marker detection from OpenCV. However, in some of the scans, the ArUco markers were not detected with OpenCV, therefore, we are not able to establish accurate ground truth pose for these parts. After removing parts without ground truth pose, we are left with 285 scans in total. In this way, we only need one manual registration per workpiece for ground truth generation.

The acquired dataset consists of 4 parts with CAD models and 20 different physical workpieces in total (10 part ID 1, (7 with burr, 3 without burr), 2 from part ID 2, (1 with burr, 1 without burr), 7 from part ID 3, (5 with burr, 2 without burr), 1 from Part ID 4, (without burr)).

Fig. 4
figure 4

Overview of the proposed approach: The dataset contains multiple CAD models and their corresponding scans taken from different view angles. To align CAD with the scan, DPFH features are extracted using pre-trained sparse CNN network. The extracted DPFH feature for both scan and CAD is used for global feature matching(alignment) with RANSAC. Finally, burr height is estimated by taking the distance between surface reconstructed scan and the input scan (“Burr detection and measurement” section)

Training on CAD models

Supervised training of deep neural networks requires large amounts of labeled data. This complicates their application to domains where training data is scarce and/or the process of collecting new datasets is laborious and expensive. Therefore, we aim to train an sparse CNN model using only CAD models which can be generalized to work on Zivid scans (more details on the network architecture in “Data pre-processing and training” section). We used CAD models as it is easy to generate in large numbers and geometric variations. As we wish to train solely on CAD models and have generalized performance on Zivid scans, the domain gap between CAD and scan need to be small as possible. Consequently, we seek to emulate the real domain by applying a series of augmentations to the CAD models. Moreover, as we have limited training samples the augmentations themselves increases our training samples to an arbitrary amount.

In addition to traditional augmentations such as point jittering, random rotation and translation, we simulate a casting burr randomly, generate a random viewpoint of the cloud and apply scaling which affects point density. The input to the random view-point generation (rendering) is the CAD model of the workpiece and then it is rendered from desired virtual view points. The generated point-view point can be controlled by randomly changing the camera position and orientation.

Given the CAD models, it is known a priori the possible location where a burr could form. Therefore, using the separation plane of the parts, we can simulate burrs by extruding additional points in this plane. By calculating the outward facing normals of the CAD point cloud, we know that point will extrude from the separation plane along the direction of the normal. Consequently, we produce extra points along the separation plane as

$$\begin{aligned} p_{B} = \vec {p}_i + \Delta L \vec {n}_i , \quad \forall i \in S \end{aligned}$$
(1)

where \(p_{B}\) is a set of added burr points, \(p_i\) are points in the separation plane S and \(\vec {n}_i\) is the associated outward facing normal of the point. A burr thickness of L can be achieved by iteratively adding burr points, \(p_{B}\), with increasing \({\Delta L}_j\) until \(\sum _j {\Delta L}_j = L\).

Our approach to scan-to-CAD alignment relies on the fact that geometric feature of CAD model ideally be locally invariant to different data augmentation that simulate scan(view point, point density, noise, jittering, etc) and can easily be computed (Khoury et al., 2017). However, traditional Fast Point Feature Histogram (FPFH) (Rusu et al., 2009) features are sensitive to such data augmentation and results in an incorrect alignment between CAD and scan. Therefore, we aim to make the FPFH descriptor insensitive to such data augmentation within the point neighbourhoods. To make traditional FPFH features robust against noise, occlusion and etc, we train a sparse CNN (Choy et al., 2019) that predicts a robust FPFH feature hereinafter referred to as DPFH (a Deep learning based FPFH feature descriptor). The learned features are robust to the transformation that simulate scan and the model is trained using solely the CAD models. The pseudocode for training DPFH feature network is summarized in Algorithm  1.

Fig. 5
figure 5

Sample augmentations. In a the original point cloud of the CAD model is shown. The burr locations are colored red. In b a synthetic burr is added by adding points along the casting plane. In c the points that are not visible for a given camera angle and height are removed. Further, random jittering is applied d different scaling is applied (affecting point density) as in e and random rotation and translation is applied (f) (Color figure online)

figure a

Burr detection and measurement

Based on the DPFH features we get an initial proposed registration, in terms of rotation \(R_0\) and translation \(T_0\), using RANSAC with parameters as in Choi et al. (2015). In order to achieve accurate burr measurement, the alignment of scan and CAD has to be close to perfect. This is implausible as burrs and other deficiencies in the scan will affect the resulting registration. Consequently, in order achieve accurate burr measurement further processing is needed. Approximate location of burr relative to CAD model is pre-computed (ROI on CAD) since casting process is known and a burr is formed in the separation plane between cope and drag. As casting burrs are mostly present in the separation plane, we have a priori information regarding where the deficiencies are most likely present in the CAD model, shown in Fig. 5a Therefore, we can extract points in a region of interest (ROI) as;

$$\begin{aligned} p_{C,\mathrm {ROI}}&= p_{C,i} \quad \forall i \in \mathrm {ROI} \end{aligned}$$
(2)
$$\begin{aligned} p_{S,\mathrm {ROI}}&= p_{S,i} \quad \forall i \quad \mathrm {where} \quad \big \Vert p_{S,i}- p_{C,\mathrm {ROI}} \big \Vert < \epsilon . \end{aligned}$$
(3)

here, \(p_C\) and \(p_S\) describe the points in the CAD and scan point clouds respectively. The \(\mathrm {ROI}\) is extracted as a thick slab in the separation plane in the CAD model. The points in the scan that are within the \(\mathrm {ROI}\) are chosen as all points that have a distance norm to its closest neighbour in the CAD \(\mathrm {ROI}\) below some threshold \(\epsilon \).

figure b
Fig. 6
figure 6

Burr measurement. In a rough alignment of CAD and scan is shown. The prior information on burr locations are colored red. In b the region of interest interest (ROI) is extracted along the burr location. In c a difference of normals is calculated for ROI region and based on difference of normals, burr points are removed and the surface is reconstructed in d. e shows the final detected burr. The burr height is represented as bright red (large burr) and dark (no burr). (Best viewed in color) (Color figure online)

Given the ROI, difference of normals (Ioannou et al., 2012) are computed. The intuition behind using difference of normals here is that if the direction of the two surface normals computed at different radius is nearly identical, then the structure of the surface does not change significantly. By contrast, if there are burrs on the surface, the direction of the two estimated normals are likely to vary by a larger margin. Figure 6c shows difference of normals along the ROI region. Therefore, a burr is detected by thresholding the difference of normals above some \(\epsilon \).

To simulate a burr-free scan, Poisson surface reconstruction (Kazhdan et al., 2006) is applied after removing the detected burr points. The Poisson surface reconstruction method solves a regularized optimization problem to obtain a smooth scan surface without burrs. Finally, the burr height is approximated by taking the distance between original and burr free surface reconstructed scan. The complete overview of burr detection and measurement approach is given in Algorithm 2.

Data pre-processing and training

The network is trained only on CAD models. We converted the workpiece CAD model from STEP 3D file to point cloud by sampling 1M points to generate point clouds with uniform density. Two out of four CAD models are used for training. Note that none of 285 scans are used in the training. We use a Sparse Residual U-Net (Sparse CNN, Choy et al., 2019) architecture in this work. It is a 34-layer U-Net architecture that has an encoder network of 21 convolution layers and a decoder network of 13 convolution and deconvolution layers. It follows the 2D ResNet basic block design and each conv/deconv layer in the network is followed by Batch Normalization and ReLU activation. The overall Sparse CNN architecture has 37.85M parameters. The Sparse CNN architecture was originally designed in Choy et al. (2019) that achieved significant improvement over prior methods on the challenging ScanNet (Dai et al., 2017) semantic segmentation benchmark. We train the sparse CNN network on a single RTX A6000 with batch size 32 and voxel-size 1cm. Adam is used with a learning rate \(1e^{-3}\). We trained the network for 500 epochs, and observed that the loss converged after approximately 100 epochs. The network outputs the estimated DPFH feature, \({\hat{y}}\), and the final loss function has the following form.

$$\begin{aligned} {\mathscr {L}}_{KL}(y,{\hat{y}}) = \frac{1}{M}\sum _{i=1}^{M}{y}_i \log {\frac{{y}_i}{{\hat{y}}_i}} \end{aligned}$$
(4)

where M is the total number of points. For registration, we used global RANSAC registration based on feature matching and using \(r_n\) = 3 correspondences.

Results

We analyze the proposed approach pipeline in two scenarios: CAD to scan registration where we estimate an SE(3) transformation between CAD models and the Zivid scans, and burr size estimation which generates a per point estimation of burr height for all viewing angles of the workpiece.

Scan-to-CAD registration evaluation

For global registration, we use RANSAC (feature matching). For each iteration of RANSAC, we select N random points from the CAD point cloud. The nearest neighbor in the DPFH feature space is queried to find their corresponding points in the scan point cloud. We examine if aligned point clouds are near together to quickly discard erroneous matches (4cm). A transformation matrix is computed using the remaining points. RANSAC is conducted for 500 iterations (Open3d.org, 2015).

Table 1 Rotation and translation error average over scans with burrs (B) and feeder(P)
Table 2 Rotation and translation error average of scans after manual de-burring

The performance of CAD to scan registration algorithm is evaluated under different views of the workpieces using the Translation Error (TE) and Rotation Error (RE) metrics that measures the deviations between the predicted and and ground truth pose as defined in Elbaz et al. (2017). Given the ground truth rotation \(\mathbf{R}\) and translation \(\mathbf{T}\) of each object, the TE and RE are defined as follows:

$$\begin{aligned} \begin{aligned} TE&= \Vert \tilde{\mathbf{T}}- \mathbf{T}\Vert \\ RE&= arccos\left( \frac{trace({{\tilde{\mathbf{R}}}^T} \mathbf {R})-1}{2}\right) \\ \end{aligned} \end{aligned}$$
(5)

where \(\tilde{\mathbf{T}}\) and \(\tilde{\mathbf{R}}\) denote the estimated translation vector and translation matrix, respectively. Note in this experiment we train the encoder network only on CAD models and do not fine-tune on the other datasets.

We report the registration results on the dataset, which contains 4 different CAD models as discussed in “3D data acquisition” section. We measure translation error (TE) and rotation error (RE) as defined in Eq. 5, and accuracy. Accuracy is the ratio of successful scan to CAD registrations and we define a registration to be successful if its rotation error and translation error are smaller than predefined thresholds. Average TE and RE are computed for different views workpiece of each part.

We compare robustness of learned features, using the proposed DPFH for registration with the classical FPFH as baseline. The result is separately shown for physical parts with burr Table 1 and without burr Table 2 for clarity. In Tables 1 and  2, we measure accuracy with the TE threshold 10mm which is practical with the voxelization process, and RE threshold 10 degrees which is typically a good initial value for registration refinement. For this experiment we used 4 CAD models (parts) and 285 scans taken from different view points. Our learned features outperforms the baseline on accuracy by a large margin (91% vs. 56%) and achieves the lowest translation (5.05mm vs. 23.52) and rotation error (\(12.82^\circ \) vs \(55.96^\circ \)) consistently on both scans with burr and no-burr, and CAD models. In Table  3, we show the RANSAC registration fitness for both DPFH and FPFH by taking the ratio of the inlier correspondences to the number of points in the Zivid scans. Higher RANSAC registration fitness value indicates DPFH feature gives a higher matching points for CAD to scan registration task. We additionally show qualitative results in Fig. 7

Table 3 Registration Fitness (higher is better)
Fig. 7
figure 7

Qualitative comparison of alignments on three different CAD models. Our approach to learning feature by mimicking real scans produce reliable correspondences, which coupled with RANSAC algorithm, produces significantly more accurate alignments

Fig. 8
figure 8

CT scan. The raw CT scan is shown in (a). In b, the scan is registered with the CAD model. The colors indicate the difference between CAD and scan. The height of the burr is measured as the distance from the CAD model and the top of the burr (c). A burr region was measured for comparison with the Zivid scans (d) (Color figure online)

Fig. 9
figure 9

Comparison of ground truth burr height as explained in “Burr size evaluation data” section and the estimated burr height using our approach

Fig. 10
figure 10

Estimated burr height for region of interest and histogram of burr (325 sample points on the scan)

Burr size evaluation data

Although quantitative evaluation on the Zivid scans serves the purpose of quantifying the burr quantization algorithm, it is important to compare the estimated burr size with a high resolution industrial burr size measuring device. An industrial computed tomography (CT) scanning machine (Zeiss Metrotom 1500) is used to scan the workpiece and produce a volumetric representation of the scanned object. Volume Graphics Software was used to analyze the scans, register the parts, and measure the burr height (Fig. 8).

Burr detection and size estimation evaluation

In the following, we compare our approach to estimate burr height with a high resolution industrial CT scanning machine. The workpiece materials in this study is shown in Fig. 9. Using CT scanning machine as described in “Burr size evaluation data” section, burr height of discrete 100 points were measured. Using CT scanning machine, the burr height is measured to be \(0.79 \pm 0.17\) mm for the sampled discrete points. Figure 10 shows the region of interest and the distribution of burr height as estimated using our approach. To compare our estimate with high resolution CT scan result, we conducted a two sample z-test. The result shows the proposed approach provides similar result with the industrial standard burr height estimation technique (\(z=0.279\)).

Fig. 11
figure 11

Result on missing/noisy data: The ID2 cad model is shown a and b shows the RGB image of one of the physical workpieces of ID2 part standing on rotating workspace with camera facing the backside of the workpiece, see Fig. 3. In c we show the scanned point cloud with sensor noise and missing parts with the detected burr. Note that with the learned features, we are able to register the clean CAD model with the noisy scan shown in c and detect burrs

Discussion and conclusion

In this paper, we investigated burr size estimation and registration problem using 3D vision sensors. For the CAD models of the workpieces, the average translation and rotation errors of CAD models to real Zivid scans are shown in Tables 1 and 2 . The result reveals that the maximum translation error occurs for a workpiece 2P-1. This could be due the fact that the scans of 2P-1 have extra feeding component that is not part of the objects CAD model. Furthermore, workpieces that are symmetric and lack geometric variations (e.g. 3B-2) are challenging as it is not possible to find unique points for registration resulting in a large rotation error (Choy et al., 2020). While the focus of this work is mainly on burr height estimation and registration given the CAD and scan, the approach could further be improved by including an algorithm for CAD model retrieval (finding the most similar model). One extension of the proposed method would be to retrieve the CAD model given the scan. This could be accomplished by designing a neural network architecture that is specifically trained on shape similarity between scan and CAD geometry.

Our primary contribution is to provide a high-resolution characterization of burrs using 3D vision sensors. To do so, we developed a deep learning model that is trained unsupervised only on CAD models by mimicking real scans. The developed model is robust to differences between CAD and real world scans such as variations to point density as well as missing data in the scans. Figure 11 shows a side by side comparison of CAD model, RGB image and captured point cloud highlighting the domain gap between CAD and real scans. The proposed method is general in that it works with multiple CAD models and able to handle variations in physical parts due to solidifation process. Therefore, applicable in industrial settings. Together with registration, our burr height estimation approach is able to estimate burr height similar to high resolution CT scans with Z-statistic (\(z=0.279\)).