Skip to main content
  • Original research
  • Open access
  • Published:

Sequential deep learning image enhancement models improve diagnostic confidence, lesion detectability, and image reconstruction time in PET

Abstract

Background

Investigate the potential benefits of sequential deployment of two deep learning (DL) algorithms namely DL-Enhancement (DLE) and DL-based time-of-flight (ToF) (DLT). DLE aims to enhance the rapidly reconstructed ordered-subset-expectation-maximisation algorithm (OSEM) images towards block-sequential-regularised-expectation-maximisation (BSREM) images, whereas DLT aims to improve the quality of BSREM images reconstructed without ToF. As the algorithms differ in their purpose, sequential application may allow benefits from each to be combined. 20 FDG PET-CT scans were performed on a Discovery 710 (D710) and 20 on Discovery MI (DMI; both GE HealthCare). PET data was reconstructed using five combinations of algorithms:1. ToF-BSREM, 2. ToF-OSEM + DLE, 3. OSEM + DLE + DLT, 4. ToF-OSEM + DLE + DLT, 5. ToF-BSREM + DLT. To assess image noise, 30 mm-diameter spherical VOIs were drawn in both lung and liver to measure standard deviation of voxels within the volume. In a blind clinical reading, two experienced readers rated the images on a five-point Likert scale based on lesion detectability, diagnostic confidence, and image quality.

Results

Applying DLE + DLT reduced noise whilst improving lesion detectability, diagnostic confidence, and image reconstruction time. ToF-OSEM + DLE + DLT reconstructions demonstrated an increase in lesion SUVmax of 28 ± 14% (average ± standard deviation) and 11 ± 5% for data acquired on the D710 and DMI, respectively. The same reconstruction scored highest in clinical readings for both lesion detectability and diagnostic confidence for D710.

Conclusions

The combination of DLE and DLT increased diagnostic confidence and lesion detectability compared to ToF-BSREM images. As DLE + DLT used input OSEM images, and because DL inferencing was fast, there was a significant decrease in overall reconstruction time. This could have applications to total body PET.

Introduction

Positron Emission Tomography (PET) is an established non-invasive image modality that allows for visualisation of metabolic processes in the body, particularly in the settings of oncology [1], cardiology [2], and neurology [3]. Although PET is renowned for its sensitivity, spatial resolution is relatively low, and images can be limited by noise [4]. Many improvements have been made to PET instrumentation to address these limitations; these include hardware-based solutions such as optimisations of scintillators, photodetectors, extending the axial field-of-view, and implementation of time-of-flight technology (ToF). ToF detectors measure the annihilation photon arrival times with sufficient precision to constrain the location of the annihilation along the line of response, which leads to an increased signal-to-noise ratio in the resulting PET image [5,6,7,8,9,10].

Software-based solutions to improve PET image quality include advances in system modelling and in reconstruction algorithms. Iterative reconstruction methods, such as the ordered subset expectation maximisation (OSEM) algorithm [11], are frequently used for PET reconstructions. Since image noise increases with the number of iterations, OSEM is often stopped before approaching convergence with additional filters then applied to suppress noise [12]. In contrast, the block sequential regularised expectation maximisation (BSREM) algorithm inherently controls noise allowing for preservation of image quality at effective convergence [13]. BSREM uses a regularisation (or penalty) factor that steers image updates away from noisier solutions. Bayesian reconstructions such as BSREM have gained widespread acceptance but are more computationally expensive to date and have not supplanted OSEM as an industry standard, being unavailable from some vendors. The substantial hardware and software improvements that have been progressively realised since the advent of PET- Computed Tomography (CT) have combined to give improved diagnostic confidence and lesion detectability [14, 15], while facilitating shorter imaging times and reductions in injected activities [16].

In recent years, Machine Learning (ML) applications in PET have produced promising results. ML has been applied in multiple aspects of PET, including image reconstruction, image acquisition, attenuation/scatter correction and quantitative imaging. Applications frequently use deep learning (DL) network structures such as convolutional neural networks (CNNs), generative adversarial networks (GANs) and U-nets. While improved time resolution can be achieved with recent developments of ToF detectors, DL has been used to improve image quality of non-ToF scanners by training algorithms to reconstruct images with ToF-like benefits without ToF information [17, 18]. Assessments of image quality from the DL-based time-of-flight (DLT) algorithm were similar or superior to the images reconstructed using the native ToF information [17]. Multiple studies are providing support towards the deployment of DL to estimate the interaction energy, position, and arrival time of incident photons. However, this comes with a variety of difficulties such as achieving the required training data through extensive experimental measurements.

Given the ability of solving complex inverse problems, DL algorithms are suitable for various image reconstruction or image processing applications in PET. There are different approaches to solve the complex problems: one approach is to use DL to generate standard quality images from low-quality images. In this approach, algorithms are trained to improve images acquired from shorter acquisitions or lower injected activity scans, which are enhanced towards images of standard quality. A Deep Learning Enhancement (DLE) algorithm for 18F-FDG PET scans used deep neural networks to enhance rapidly reconstructed OSEM images to make them BSREM-like. Using the DLE algorithm, the computational reconstruction time was significantly reduced, and the image quality was maintained [19]. Another approach is to use DL to reconstruct PET images from sinograms [20, 21]. Other approaches, such as ‘hybrid domain learning’ rely on both ML approaches as well as analytical solutions to reach an optimal image [22].

The aim of the current study is to investigate the potential benefits of sequential deployment of two deep learning (DL) algorithms namely DL Enhancement (DLE) and DL-based time-of-flight (DLT).

Such (mis-) use brings risks of unexpected or undesired outcomes, but with potential for significant benefits in terms of image quality; we aimed to explore both aspects. DLE aims to enhance the rapidly reconstructed OSEM images towards BSREM images, whereas DLT aims to improve the quality of BSREM images reconstructed without ToF towards ToF-BSREM. As the algorithms differ in their purpose, the sequential application may allow the benefits from each to be combined.

Materials and methods

Reconstructions

PET raw data (i.e., sinograms) were reconstructed using five combinations of algorithms as shown in Table 1.

Table 1 Deep learning algorithm combinations explored in this study with brief descriptions

Reconstruction one (ToF-BSREM) is the gold standard, all subsequent reconstructions were compared against it. ToF-OSEM is the expected input for DLE, as used in reconstruction two (ToF-OSEM + DLE) and previously evaluated [19]. Reconstructions three (OSEM + DLE + DLT), four (ToF-OSEM + DLE + DLT) and five (ToF-BSREM + DLT) apply the DL algorithms beyond their intended use. In these cases, the input image being provided to the DL algorithm potentially has characteristics or features that are outside the boundaries of the training data used in creation of the DL model.

DLE is a 3D residual convolutional coder-decoder (U-Net) network developed and implemented in Pytorch. The model was trained in supervised sessions, where it mapped low-contrast high-noise OSEM PET images to low noise high contrast BSREM images. In the supervised sessions, the output OSEM + DLE was compared to a target patch BSREM and based on the result of the smooth L1 loss function the trainable parameters were updated. The DLE architecture is composed of convolutional layers (using 3 × 3 × 3 kernels), batch normalisation (BN), 3D max pooling layers and tri-linear up-sampling layers, skip and residual connections and leaky rectified linear unit (ReLU) activation functions. DLE used 510 [18F]-FDG PET/CT scans from six sites equipped with D710 and DMI, of these scans, n = 480 were used for training, n = 15 for validation and n = 25 for testing. The OSEM reconstructions were of matrix size 256 × 256, field-of-view 700 mm, voxel size 2.7 × 2.7 × (2.8 or 3.7) mm3 and 2 iterations, 34 and 24 subsets for Discovery MI and 710 scanners, respectively, with PSF and standard z-filter [19].

DLT is similar to DLE’s network that is a 3D residual U-Net developed and implemented in Pytorch. The model was trained in supervised sessions, where DLT compared the predicted ToF-BSREM images to target ToF-BSREM using the MSE loss function. For DLT, a total of 273 [18F]-FDG PET/CT scans were used from six sites equipped with DMI scanners only, split into training (n = 208), validation (n = 15), and testing (n = 50) sets. The matrix size of each reconstructed image was of 256 × 256 and field-of-view of 700 mm (x–y pixel size: 2.73 mm, slice thickness: 2.79 mm) [17].

The manufacturer’s Bayesian penalised-likelihood reconstruction algorithm was used to reconstruct all BSREM images with a fixed regularization factor, β = 400 [13]. All OSEM reconstructions used two iterations with 32 (D710) or 34 (DMI) subsets, no in-plane (x-y) post filter, and with the manufacturer’s standard z-filter which is the anticipated (trained) input for the DLE model. Out of the three DLT models available (low, medium, high) with names that describe the strength with which the models were trained to transform non-ToF BSREM images to their target ToF BSREM, the current study used ‘high’ consistently. This was based on previous reader preference [17], and to provide greater changes to the images thus increasing the likelihood of detectable differences in the current study while limiting the number of variables under examination.

Patient selection

The study made use of 40 whole-body 18F-FDG PET-CT scans. 20 scans were performed sequentially on a Discovery 710 and 20 on 25 cm axial field of view Discovery MI (both GE HealthCare). All scans were free breathing, and ungated. For each patient, a whole-body helical CT was performed for PET attenuation correction using 100–120 kVp, 150–200 mAs. For the DMI subjects, the range in activity was (mean ± std MBq) (391.4 ± 99.0 MBq), the patient size range was BMI (26.9 ± 5.6 kg/m2), FDG uptake time range was (80.1 ± 21.2 min) and the acquisition time per bed position was two minutes per bed position. For the D710 subjects, the range in activity was (309.0 ± 77.6 MBq). The patient size range was (27.2 ± 6.8 kg/m2), the FDG uptake time range was (89.7 ± 7.9 min) and all scans were three minutes per bed position.

Clinical and quantitative evaluation

Two experienced radiologists, reader 1 (K.M.B. 20 years board certified in clinical radiology and nuclear medicine) and reader 2 (P.A.F. 19 years board certified in clinical radiology and nuclear medicine), blinded to method of reconstructions, rated the images on a Likert scale (5 best) based on diagnostic confidence, lesion detectability and image quality. The Likert scale used was 0 (non-diagnostic), 1 (poor,), 2 (satisfactory), 3 (good), 4 (very good), and 5 (excellent) as in previous work [17]. Inter-reader agreement was determined using Intraclass Correlation Coefficient (ICC) (two-way random effects model) carried out in SPSS 29. To check for differences across groups a Friedman test was carried out. When significant differences were found, Wilcoxon signed-ranks test with Bonferroni post-hoc were performed for pair-wise comparisons with a significance threshold of p < 0.05.

Radiologist K.M.B identified lesions that were subtle and/or small and recorded the SUVmax of these lesions. To assess noise, the standard deviation between voxels within a 30 mm spherical VOI was calculated using VOIs placed in normal lung and liver. Group-wise differences were calculated using the Kruskal-Wallis test. When significant differences were found, Wilcoxon signed-ranks test with Bonferroni post-hoc were performed for pair-wise comparisons with a significance threshold of p < 0.05. Friedman and Wilcoxon-signed ranks tests were carried out in Python 3.11.

Results

Figures 1 and 2 display two example performances of the five reconstructions on two subjects. In Fig. 1, the images show that when DLE is applied the liver appears smoother, whereas when DLT is applied the lesion is more conspicuous, though the image may suffer from higher noise. In reconstruction four, where DLE + DLT is applied sequentially to ToF-OSEM, improvements due to both DL algorithms are visible. In Fig. 2, the lesion indicated by the arrow is small and of low uptake. The lesion is most visible in reconstructions one, four and five. This visibility is also reflected in the SUVmax values, where reconstructions five and four were highest and second highest respectively. The results for all lesion SUVmax values are summarised in Figs. 3 and 4.

Fig. 1
figure 1

Deep Learning Enhancement (DLE) + Deep Learning Time-of-flight (DLT) applications to a test subject with BMI 19.4 kg/m2, with an injected activity of 229 MBq FDG, scanned on a D710 PET-CT scanner. The subject is a male patient, staging scan for relapsed, high grade, transformed, follicular, non-Hodgkin’s lymphoma, with nodal and peritoneal disease. Axial PET images and the SUVmax of a tiny peritoneal nodule posterior to the right lobe of the liver (red arrow) are demonstrated. All images use an SUV scale of 0–6

Fig. 2
figure 2

Deep Learning Enhancement (DLE) + Deep Learning Time-of-flight (DLT) applications to a test subject with BMI 27.4 kg/m2, with an injected activity of 419.0 MBq FDG, scanned on a DMI PET-CT. The subject is a female patient, staging scan for breast carcinoma, with left axillary and a solitary, approximately 7 × 4 mm left supraclavicular node. Axial PET images and the SUVmax of the left supraclavicular fossa node (red arrow) are demonstrated. All images use an SUV scale of 0–6

Fig. 3
figure 3

Quantitative performance of five algorithms on lesion SUVmax scanned on a D710 PET-CT scanner. All p-values are calculated using the Wilcoxon signed-rank test and Bonferroni-corrected. P-values are not displayed if larger than 0.05. All lesions chosen were extreme cases where lesions were small (sub-cm)

Fig. 4
figure 4

Quantitative performance of five algorithms on lesion SUVmax scanned on a DMI PET-CT scanner. All p-values are calculated using the Wilcoxon signed-rank test and Bonferroni-corrected. P-values are not displayed if larger than 0.05. All lesions chosen were extreme cases where lesions were small (sub-cm)

Figures 3 and 4 display the quantitative performance of reconstruction methods 2–5 for SUVmax relative to method one as a percentage difference. SUVmax values for BSREM reconstructions were included in both Figures to display the impact of ToF-information on SUVmax. Figure 3 shows the results obtained for 28 challenging lesions from the D710 scans. Figure 4 shows the results obtained for 24 challenging lesions from the DMI scans. Error bars represent the standard error of the mean and p-values, calculated using the Wilcoxon-signed ranks test and corrected for multiple comparisons, are displayed if smaller than the significance level 0.05. Figure 3 shows that in the D710 data set, there is no significant difference between method one and methods two and three. A statistically significant increase in SUVmax is found in reconstructions four and five compared to one. Similarly, to the D710 data set, reconstruction five outperforms all other methods.

Figures 5 and 6 displays the quantitative performance of reconstructions 1–5 for noise and SUVmean in the lung and liver. Error bars represent the standard error of the mean, p-values calculated using the Wilcoxon-signed ranks test, and corrected for multiple comparisons, are displayed if smaller than the significance level 0.05. Figure 5 shows the results obtained from the D710 data set whereas Fig. 6 shows the results obtained from the DMI data set. All reconstructions in the DMI and D710 dataset gave similar SUVmean values for the uniform liver and lung region (Kruskal-Wallis p > 0.05). In both D710 data and DMI data, noise in the liver was significantly reduced when DLE and DLT was applied sequentially. No significant difference was found in lung noise in the DMI data set (Kruskal-Wallis p > 0.05).

Fig. 5
figure 5

Quantitative performance of five reconstructions of DLE/DLT combinations. Tested by drawing a 30 mm spherical VOI in both the lung and liver and measuring standard deviation between voxels and SUVmean for 20 patients scanned on a GE HealthCare Discovery 710 scanner

Fig. 6
figure 6

Quantitative performance of five reconstructions of DLE/DLT combinations. Tested by drawing a 30 mm spherical VOI in both the lung and liver and measuring standard deviation between voxels and SUVmean for 20 patients scanned on the GE DMI 10 scanner

Clinical reading

Tables 2 and 3 show the scores for reconstruction methods 1–5 Table 2 shows the scores for the D710 data set and Table 3 shows the scores the DMI data set. No images were scored as non-diagnostic. The values in bold are the highest score for each metric. Table 2 shows a preference for reconstructions four and five in lesion detectability and diagnostic confidence. Whereas reconstructions two and four were preferred for image quality. A similar pattern is displayed in the DMI dataset in Table 3 where there is a preference for reconstructions four and five in lesion detectability and diagnostic confidence and a preference for reconstruction two and four for image quality. Both tables show an overall preference for reconstruction four which deploys DLE + DLT on ToF-OSEM.

Table 2 Clinical image scoring from two readers of 20 whole-body FDG scans acquired on the D710 scanner. p-values were calculated using the Wilcoxon signed-rank test. The values in bold represent the best (highest) score for each metric
Table 3 Clinical image scoring from two readers of 20 whole-body FDG scans acquired on a DMI scanner. p-values were calculated using the Wilcoxon signed-rank test. The values in bold represent the best (highest) score for each metric

Discussion

The present study investigated the potential benefits of sequential deployment of two deep learning (DL) algorithms: DLE and DLT. Five reconstruction algorithms were tested on 40 FDG scans from two different scanners. Reconstruction four (ToF-OSEM + DLE + DLT) was the optimal combination out of the five reconstructions. It produced larger decrease in liver noise, significant increase in SUVmax and it was the overall preference in the clinical reading.

Reconstruction four (ToF-OSEM + DLE + DLT) and reconstruction five (ToF-BSREM + DLT) demonstrated a larger increase in lesion SUVmax relative to the gold standard reconstruction one (ToF-BSREM). Reconstruction five demonstrated the largest increase in SUVmax, particularly in the DMI data. This gain in SUVmax was reflected in the DMI clinical scoring, where reconstruction five and four scored highest and second highest for both lesion detectability and diagnostic confidence with small differences between their respective average reader scores (0.1 and 0.02 respectively). However, reconstruction five images were reported to suffer from noise, thus lowering the image quality score to third lowest. Regarding image quality, reconstruction four scored the highest. In the D710 data, where the gain in SUVmax between reconstructions four and five are similar, there is a preference for reconstruction four in both lesion detectability and diagnostic confidence due to less noise. Reconstruction two scored highest for image quality with reconstruction four being a close second, both reconstructions include DLE, further showcasing the algorithm’s denoising abilities. Furthermore, both DL algorithms have been trained on DMI data which has better ToF timing resolution, therefore, the benefits of applying DL may be higher for D710 input image data.

In addition to improving the lesion detectability, diagnostic confidence, and image quality, using DLE has the advantage of reducing the computational burden of the reconstructions. For example, on the DMI scanner, reconstruction time of BSREM and OSEM algorithms are 3.1 and 1.1 min per bed position respectively, as previously reported [19].

Whereas the addition of DLE for processing a whole-body scan is only  5s [19]. Therefore, DLE reconstructed images take a third of the reconstruction time and are comparable to BSREM images. This large saving in reconstruction time could significantly affect the feasibility of algorithm deployment and routine clinical use, especially in the case of larger field of view cameras where reconstruction times need to keep pace with higher patient throughput. The reconstruction times for the patients from each type of reconstruction, reconstructed using a GE research system, is provided in the supplementary Table S1.

There were some necessary limitations to this study, which aimed not for clinical validation but for rapid exploration of algorithm performance. The number of readers was only two, and there were only a moderate number of exams included in the study (40). Five combinations of algorithms were assessed, resulting in 400 whole body FDG PET/CT reviews in total. We considered this to be reasonable to meet the study aims but it is not exhaustive. Other image reconstruction and DL combinations are possible, we noted that the successive application of DLT led to noticeable image artifacts. No reconstruction artefacts were noted by the radiologists within any of the images presented in this paper, where DLT and DLE were combined. We sequentially applied DLT to ToF-BSREM images six times, results for lung and liver noise as well as lesion SUVmax can be found in supplementary Figures S1 and S2. Example images can be found in supplementary Figure S3 where image artefacts are visible following multiple applications of DLT.

The input images were of standard injected activity (or acquisition time), and we did not assess performance in the case of lower quality input images. Furthermore, only two PET-CT scanner models were included in the assessment and while this gives an indication that the results may be generalisable, similar evaluations using additional scanner models would give greater assurance in this regard. Images were assessed in terms of several quantitative metrics and scored by two readers, but the study could not assess performance for lesion detection or the rates of false positives which would be important considerations prior to any clinical use.

We recognise that in this study there were limitations with lack of testing regarding quantitative accuracy and bias caused by deep learning. Extensive evaluation in phantoms was beyond this study, this decision was influenced by the fact that DL models are trained on human data, and simple phantoms might not accurately represent what is encountered in clinical practice. For the purposes of this study, the images were compared to our gold standard, ToF BSREM which has well characterized quantitative performance shown in previous publications [13]. The aim of this study was to demonstrate the potential of the sequential network application with comparison to ToF-BSREM images. The natural progression from this initial study would be to develop and train a network that incorporates aspects of DLE and DLT, at which point it would be feasible (and necessary) to test for quantitative accuracy (for example through inserted lesions) and biases.

This study demonstrated the potential for the sequential application of DLE and DLT to FDG scans. We aimed to explore this possibility and to openly identify any significant image artefacts or anomalies that might be produced. We did not find any notable such features in the images that were assessed; on the contrary, we found a reader preference for the combined use of the DL algorithms, and this was supported by several quantitative measures. From this single and moderate-sized study we certainly do not advocate the clinical use of the sequential application of the two DL algorithms. We do however consider this study may help uncover new directions for the use of Artificial Intelligence (AI) in PET. Specifically, if these two DL algorithms were able to provide image improvements when combined and used beyond their intended purpose, it is reasonable to postulate that a dedicated algorithm (either single- or two-stage) can be developed with an intended purpose similar to combining DLE and DLT and using similar architecture. Such an algorithm would be expected to outperform the sequential (mis-) use of DLE and DLT. The positive results obtained in this study thus motivate further work on the topic.

Conclusion

The present study investigated the potential benefits of the sequential deployment of two deep learning algorithms: DLE and DLT. Our quantitative results show that, with ToF-OSEM as input, the sequential application of DLE and DLT can reduce noise in the liver and increase SUVmax of small target abnormalities significantly. This is reflected in the clinical reading where ToF-OSEM + DLE + DLT scored highest or second highest for lesion detectability and diagnostic confidence. The deployment of DLE may allow for faster scans and more rapid reconstruction times which could be of particular use for larger axial field of view cameras. Insight from this investigation can help guide future developments of AI and its use in PET.

Data availability

Data is available under reasonable request to the corresponding author.

References

  1. Langer A. A systematic review of PET and PET/CT in oncology: a way to personalize cancer treatment in a cost-effective manner? BMC Health Serv Res. 2010;10:283. https://doi.org/10.1186/1472-6963-10-283.

    Article  PubMed  PubMed Central  Google Scholar 

  2. Slomka P, Berman DS, Alexanderson E, Germano G. The role of PET quantification in cardiovascular imaging. Clin Transl Imaging. 2014;2(4):343–58. https://doi.org/10.1007/s40336-014-0070-2.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Tai YF, Piccini P. Applications of positron emission tomography (PET) in neurology Journal of Neurology. Neurosurg Psychiatry. 2004;75:669–76. https://doi.org/10.1136/jnnp.2003.028175.

    Article  CAS  Google Scholar 

  4. Moses WW. Fundamental limits of spatial resolution in PET. Nucl Instrum Methods Phys Res A. 2011;648(Supplement 1):S236–40. https://doi.org/10.1016/j.nima.2010.11.092.

  5. Karp JS, Surti S, Daube-Witherspoon ME, Muehllehner G. Benefit of time-of-flight in PET: experimental and clinical results. J Nucl Med. 2008;49(3):462–70. https://doi.org/10.2967/jnumed.107.044834.

    Article  PubMed  Google Scholar 

  6. Conti M, Eriksson L, Rothfuss H, Melcher CL. Comparison of fast scintillators with TOF PET potential. IEEE Trans Nucl Sci. June 2009;56(3):926–33. https://doi.org/10.1109/TNS.2008.2009446.

  7. Maurizio, Conti. Why is TOF PET reconstruction a more robust method in the presence of inconsistent data? IOP Publishing Phys Med Biol. 2011;56:155–68. https://doi.org/10.1088/0031-9155/56/1/010h.

    Article  Google Scholar 

  8. Surti S. Update on time-of-flight PET imaging. J Nucl Med. 2015;56(1):98–105. https://doi.org/10.2967/jnumed.114.145029.

    Article  ADS  PubMed  Google Scholar 

  9. Surti S, Karp JS. Advances in time-of-flight PET. Phys Med. 2016;32(1):12–22. https://doi.org/10.1016/j.ejmp.2015.12.007.

    Article  PubMed  PubMed Central  Google Scholar 

  10. Vandenberghe S, Mikhaylova E, D’Hoe E, Mollet P, Karp JS. Recent developments in time-of-flight PET. EJNMMI Phys. 2016;3(1):3. https://doi.org/10.1186/s40658-016-0138-3.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  11. Sangtae Ahn SG, Ross E, Asma J, Miao X, Jin L, Cheng SD, Wollenweber, Ravindra M, Manjeshwar. Quantitative comparison of OSEM and penalized likelihood image reconstruction using relative difference penalties for clinical PET. Phys Med Biol 60 5733. https://doi.org/10.1088/0031-9155/60/15/5733.

  12. Lantos J, Mittra ES, Levin CS, Iagaru A, Standard. OSEM vs. regularized PET image reconstruction: qualitative and quantitative comparison using phantom data and various clinical radiopharmaceuticals. Am J Nucl Med Mol Imaging. 2018;8(2):110–8.

    CAS  PubMed  PubMed Central  Google Scholar 

  13. Teoh EJ, McGowan DR, Macpherson RE, Bradley KM, Gleeson FV. Phantom and clinical evaluation of the bayesian penalized Likelihood Reconstruction Algorithm Q.Clear on an LYSO PET/CT system. J Nucl Med. 2015;56(9):1447–52. https://doi.org/10.2967/jnumed.115.159301.

    Article  CAS  PubMed  Google Scholar 

  14. Vandenberghe S, Mikhaylova E, D’Hoe E et al. Recent developments in time-of-flight PET. EJNMMI Phys 3, 3 (2016). https://doi.org/10.1186/s40658-016-0138-3.

  15. Aide, N., Lasnon, C., Kesner, A. et al. New PET technologies ? embracing progress and pushing the limits. Eur J Nucl Med Mol Imaging 2021;48:2711?2726. https://doi.org/10.1007/s00259-021-05390-4.

  16. Tonghe Wang, Yang Lei, Yabo Fu, Walter J. Curran, Tian Liu, Jonathon A. Nye, Xiaofeng Yang,Machine learning in quantitative PET: A review of attenuation correction and low-count image reconstruction methods, Physica Medica, 2020;76:294-306, ISSN 1120-1797, https://doi.org/10.1016/j.ejmp.2020.07.028.

  17. Mehranian A, Wollenweber SD, Walker MD, Bradley KM, Fielding PA, Huellner M, Kotasidis F, Su KH, Johnsen R, Jansen FP, McGowan DR. Deep learning-based time-of-flight (ToF) image enhancement of non-ToF PET scans. Eur J Nucl Med Mol Imaging. 2022;49(11):3740–9. https://doi.org/10.1007/s00259-022-05824-7.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  18. Berg E, Simon R, Cherry. Using convolutional neural networks to estimate time-of-flight from PET detector waveforms. 2018 Phys Med Biol 63 02LT0. https://doi.org/10.1088/1361-6560/aa9dc5.

  19. Mehranian A, Wollenweber SD, Walker MD, Bradley Kotasidis F, Jansen FP, McGowan DR. Image enhancement of whole-body oncology [18F]-FDG PET scans using deep neural networks to reduce noise. Eur J Nucl Med Mol Imaging. 2022;49(2):539–49. https://doi.org/10.1007/s00259-021-05478-x.

    Article  CAS  PubMed  Google Scholar 

  20. Häggström I, Schmidtlein CR, Campanella G, Fuchs TJ. DeepPET: a deep encoder-decoder network for directly solving the PET image reconstruction inverse problem. Med Image Anal. 2019;54:253–62. https://doi.org/10.1016/j.media.2019.03.013.

    Article  PubMed  PubMed Central  Google Scholar 

  21. Zhu B, Liu JZ, Cauley SF, Rosen BR, Rosen MS. Image reconstruction by domain-transform manifold learning. Nature. 2018;555(7697):487–92. https://doi.org/10.1038/nature25988.

    Article  ADS  CAS  PubMed  Google Scholar 

  22. Ravishankar S, Ye JC, Fessler JA. Image Reconstruction: From Sparsity to Data-Adaptive Methods and Machine Learning, in Proceedings of the IEEE, 2020;108(1):86-109, https://doi.org/10.1109/JPROC.2019.2936204.

Download references

Acknowledgements

We acknowledge the advice given on manuscript preparation by Dr Floris Jansen (GE HealthCare). We acknowledge the assistance with data processing by Manjeet Kuhar (Oxford University Hospitals).

Funding

This work is supported by GE HealthCare and the National Consortium of Intelligent Medical Imaging (NCIMI) through the Industry Strategy Challenge Fund, Innovate UK Grant 104,688. DRM is supported by the Cancer Research UK National Cancer Imaging Translational Accelerator (C34326/A28684 and C42780/A27066).

Author information

Authors and Affiliations

Authors

Contributions

MD wrote the initial draft manuscript and analysed the data. MD, KMB, MDW, DRM designed the study. KMB and PAF scored the clinical images. MD, AM, KMB, MDW, PAF, SDW, RJ, DRM interpreted the results and edited the manuscript. All authors have read and approved the final manuscript.

Corresponding author

Correspondence to Daniel R. McGowan.

Ethics declarations

Ethics approval

Patient data was used in this project in the form of anonymised retrospectively reconstructed image datasets. This practice does not require informed consent nor was specific ethics approval required. Study was performed in accordance with the ethical stands as laid down in the 1964 Declaration of Helsinki and its later amendments.

Consent to participate

Not applicable.

Consent for publication

Not applicable.

Competing interests

Oxford University Hospitals NHS FT has a research agreement with GE HealthCare. AM, SDW, RJ are employees of GE HealthCare.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Electronic supplementary material

Below is the link to the electronic supplementary material.

Supplementary Material 1

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dedja, M., Mehranian, A., Bradley, K.M. et al. Sequential deep learning image enhancement models improve diagnostic confidence, lesion detectability, and image reconstruction time in PET. EJNMMI Phys 11, 28 (2024). https://doi.org/10.1186/s40658-024-00632-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s40658-024-00632-4

Keywords