Skip to main content

Deep learning model for automatic image quality assessment in PET

Abstract

Background

A variety of external factors might seriously degrade PET image quality and lead to inconsistent results. The aim of this study is to explore a potential PET image quality assessment (QA) method with deep learning (DL).

Methods

A total of 89 PET images were acquired from Peking Union Medical College Hospital (PUMCH) in China in this study. Ground-truth quality for images was assessed by two senior radiologists and classified into five grades (grade 1, grade 2, grade 3, grade 4, and grade 5). Grade 5 is the best image quality. After preprocessing, the Dense Convolutional Network (DenseNet) was trained to automatically recognize optimal- and poor-quality PET images. Accuracy (ACC), sensitivity, specificity, receiver operating characteristic curve (ROC), and area under the ROC Curve (AUC) were used to evaluate the diagnostic properties of all models. All indicators of models were assessed using fivefold cross-validation. An image quality QA tool was developed based on our deep learning model. A PET QA report can be automatically obtained after inputting PET images.

Results

Four tasks were generated. Task2 showed worst performance in AUC,ACC, specificity and sensitivity among 4 tasks, and task1 showed unstable performance between training and testing and task3 showed low specificity in both training and testing. Task 4 showed the best diagnostic properties and discriminative performance between poor image quality (grade 1, grade 2) and good quality (grade 3, grade 4, grade 5) images. The automated quality assessment of task 4 showed ACC = 0.77, specificity = 0.71, and sensitivity = 0.83, in the train set; ACC = 0.85, specificity = 0.79, and sensitivity = 0.91, in the test set, respectively. The ROC measuring performance of task 4 had an AUC of 0.86 in the train set and 0.91 in the test set. The image QA tool could output basic information of images, scan and reconstruction parameters, typical instances of PET images, and deep learning score.

Conclusions

This study highlights the feasibility of the assessment of image quality in PET images using a deep learning model, which may assist with accelerating clinical research by reliably assessing image quality.

Peer Review reports

Background

Molecular imaging of positron emission tomography/computed tomography (PET/CT) has played an important role in nuclear medicine, such as noninvasive tumor diagnostic staging [1], efficacy evaluation [2], and research and development of new drugs [3]. However, high noise levels, missing or incomplete data, motion artifacts, inadequate preparation of patients, intravenous injection failure, improper placement, and scanning equipment miss-calibration might lead to the poor data quality of images and wrong conclusions [4], and the typical images with insufficient level quality are shown in Fig. 1. Therefore, the clinical image quality control of PET/CT is essential for excluding the clinical images with poor quality resulting from any problematic processes and avoiding bias in nuclear medicine medical quality management.

Fig.1
figure 1

Typical 18F-FDG PET/CT images with insufficient level quality: a imaging of severe brown fat, b intravenous extravasation, c mismatch from external patient movement, d tissue over-attenuation from improper placement

The current traditional clinical image quality evaluations in PET/CT rely on the subjective evaluation with naked eyes following the suggestions in literature [5,6,7], in which mean liver standardized uptake value normalized by lean body mass (SULs) are expected to be within 1.0 to 2.2 (and mean liver standardized uptake values (SUVs) within 1.3 to 3.0) [5], and blood pool SUL measurements are expected to close 1.2 (and blood pool SUVs around 1.6) [5,6,7]. However, visual judgment is time-consuming, which is impractical for the evaluation of large-batch images[8]. For artificial quality control, some mistakes in settings of acquisition parameters are easy to be ignored. For example, the statistics are reduced, caused by the shortening of acquisition time [9]. Moreover, there is still a lack of objective and unified standards for artificial visual evaluation, and level differences in doctors can easily lead to evaluation bias. Therefore, a robust, minimally biased, and fully automated PET/CT QA protocol is urgently needed. Convolutional neural networks (CNN) are a good option for the automatic medical image QA domain since they can robustly learn features without knowing a priori.

It is reported that the deep learning CNN technology has been used in different image quality automatically assessment systems, including optical coherence tomography (OCT) images [10], retinal images [11], diabetic retinopathy screening [12], high-frequency ultrasound images [13], CT images [14] and 3D T1-weighted brain MRI images [15]. However, there is still limited research on DL in the automatic control of clinical image quality for PET/CT. In a study, two different CNN algorithms were combined to assess spatial misalignment compared to a standard template, and the signal-to-noise ratio (SNR) difference compared to 200 static quality controlled 18F-fluorodopa (FDOPA) PET brain images from three different PET/CT scanners, in which 100% accurate QA classification was reported [16]. However, the 18F-FDOPA PET brain images are invalid to the quality control for full-body scans, and 18F-FDOPA is at the preclinical stage rather than 18F-FDG practical in clinical application. Another preprint shows Elisabeth took EARL standards as a reference and first reported CNN to determine the image quality of a PET torso image, but it is worth mentioning that they utilized only 2D slices instead of 3D volumes of the whole.

The quality evaluation of clinical images in PET/CT is one of the main tasks of the National Nuclear Medicine Quality Control Center in China, where the fully automatic quality control protocol for clinical images in PET/CT is still not available. To solve this problem, we carried out a series of studies on the quality evaluation of PET/CT clinical images. The contributions of this work are summarized as follows. First, the important characteristic parameters were extracted from the original data, and a quality control process was designed. Then, based on the principle of quintuples and the visual judgment of clinical doctors, a deep learning model for automatic quality control for clinical images in PET/CT was constructed, which has the characteristics of multiple parameters generated and large quantities of data required to avoid over-fitting during the training process. This approach pays attention to automated quality assessment based on deep learning, which can automatically assess the quality of PET images, and aims to contribute to the improvement of workflow, better optimization of image acquisition, and enhancing physician efficiency. Finally, combined with the DL-based image quality assessment model, the final QA report will include basic image information, scan and reconstruction parameters, a typical PET imaging example, and deep learning score.

Methods

Study design

The workflow of this study is presented in Fig. 2. This study included four major parts: (i) image acquisition, (ii) image preprocessing, (iii) modal training and cross-validation, and (iv) evaluation.

Fig.2
figure 2

The workflow for this study

Images

In this project, PET images previously obtained during the clinical assessment of various diseases were used to create the datasets utilized for this study. This project contains 89 PET images in clinical routine examination recruited retrospectively from the Peking Union Medical College Hospital (PUMCH) in China. The images were taken using a whole-body PET/CT scanner (Polostar NM680, SinoUnion Healthcare, Beijing, China), and the scanning conditions and parameters are set to be consistent. The PET scanning protocol details were listed as follows. The imaging agent 18F-fluorodeoxyglucose (FDG) was produced by PUMCH, with a PH value is about 7.0 and radiochemical purity > 95%. All image acquisition was carried out in a resting state in a quiet and dimly lit room. According to their weight, the patients were given 0.15 mci/kg intravenous injections of 18F-FDG for a calm rest for 45–60 min. PET scan used 5–6 beds, 2 min per bed position, and the layer thickness was 5 mm. The PET images were attenuated by CT data and reconstructed by the ordered-subsets expectation maximization (OSEM) algorithm with 10 subsets and 3 iterations, and 4.5-mm full width at half maximum (FWHM) Gaussian post-filtering. The matrix size of all PET reconstructions was 192 × 192, with a pixel size of 3.15 mm × 3.15 mm. Finally, the reconstructed PET images were transmitted to the post-processing platform.

A total of 71 PET images were selected for the training set for this model. The validation set was composed of 18 images for evaluating model performance. The ground-truth quality for images was conducted by two radiologists (with longer than 10 years of experience) with disagreements resolved by a third independent expert rad and classified into one of five grades (grade 1, grade 2, grade 3, grade 4, and grade 5). Grade 5 is the best image quality. This reference standard complies with the situation assessed by the following quality criteria based on 5-point Likert scales [17,18,19] and is detailed in Table 1. The number of images from grade 1 to grade 5 was 21(23.6%), 23(25.8%), 12(13.5%), 16(18.0%), 17 (19.1%). Typical examples from each grade are shown in Fig. 3.

Table 1 A description of quality annotation standards
Fig.3
figure 3

Examples of images with different qualities, including: a grade 1, b grade 2, c grade 3, d grade 4, and e grade 5

Image preprocessing

Image preprocessing included several operations and was conducted using Python (version 3.8.8) and MONAI (version 0.8.1) software. Firstly, Z -scores were used to normalize the scores, which had a normal distribution (99% of data had Z -scores between − 2 and 2). Secondly, all images were resampled to 128 × 128 × 256 pixels and 256 × 256 × 256 pixels, which were exported as input images. Thirdly, to increase the amount of training data, we used some transformations of the original training images [20]. Horizontal mirroring, rotations through random angles (± 10 degrees), gamma correction, and elastic deformation were used to produce new synthetic images [21, 22]. In addition, each PET image was resampled to isotropic spacing using linear interpolation to perform the model training or testing [23].

Deep learning

Deep learning (DL) is a relatively new approach and is one branch of machine learning. DL has been ubiquitously applied in medical image analysis. The purpose of this study is to solve a five-class polyp classification problem. While, due to the very limited number of each grade, it is likely to cause bias and affect the accuracy of DL. Therefore, we merge data from five grades into two-class.

Finally, we generated four tasks, taking into account the resample of images and the quantity of each grade. Detailed information about each specific task in this study was described in Table 2.

Table 2 Grouping criteria for each task

Dense convolutional network (DenseNet)

DenseNet connects each layer to every other layer in a feed-forward fashion. Thus, DenseNet has several compelling advantages: they alleviate the vanishing gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters [24]. The overall architecture is shown in Fig. 4. A deep DenseNet with three dense blocks, where each dense block was multiple stacks of convolution, batch normalization, and ReLU activation layers. The layers between two adjacent blocks are referred to as transition layers and change feature map sizes via convolution and pooling.

Fig.4
figure 4

An illustration of the architecture of the deep learning-based model. Residual blocks for a DenseNet and b Dense Block

Model training

This model also uses a 5 × 5 convolutional kernel, which has been shown to improve performance. The deep learning model was trained on the framework of Pytorch 1.10.1 version and an NVIDIA Quadro P3200 graphics card (NVIDIA, Santa Clara, CA) with 5 GB memory. A fivefold cross-validation with 80% training and 20% validation from each independent training was chosen to avoid bias in the data set (Additional file 1: Table S1 lists the distribution of data in training and validation sets for each fold of the cross-validation. Table 3 shows the image quality distribution for each of the different tasks using fivefold cross-validation in the training set and validation set.). The whole process is repeated 5 times such that all folds are used in the testing phase, and the average performance on the testing folds is computed as an unbiased estimate of the overall performance of the model, as shown in Fig. 5. In addition, the training loss function was binary cross-entropy. The optimizer was the Adam optimizer with a learning rate of \({10}^{-5}\). We set training iteration as 250 and run 1000 epochs on both the training set and testing set. We select polyLR as the learning rate, whose initial learning rate is 0.01.

Table 3 Distribution of image quality in training and validation sets for different tasks
Fig.5
figure 5

Training and validation workflow of each DL model for grade classification

Model performance

The automated classification in this study contains two classes, and we tested different multi-label classification algorithms. In this setting, the model classification performance is assessed in each class. Two-class model performance was determined by measuring the sensitivity (Sen), specificity (Spe), and accuracy (ACC) of all the introduced approaches. These parameters were calculated using Eqs. [25,26,27,28], where TP, TN, FP, and FN represent true positive, true negative, false positive, and false negative. Evaluation metrics were defined as follows:

$$\mathrm{a}\text{ccuracy} = \frac{\text{TP+TN}}{\text{TP+TN+FP+FN}}$$
(1)
$$\mathrm{s}\text{pecificity} = \frac{\text{TN}}{\text{TN+FP}}$$
(2)
$$\mathrm{s}\text{ensitivity} = \frac{\text{TP}}{\text{TP+FN}}$$
(3)

Performance was also evaluated via a receiver operating characteristic (ROC) curve. The area under the ROC Curve (AUC) is a general measure of the accuracy of a diagnostic test [29].

PET QA report

An image QA tool was developed combining the DL-based image quality evaluation model. First, the PET/CT images were inputted into the QA tool, where the image statistics are gathered for statistics control, and registration is performed to examine the motion artifacts in the images. Next, the input images were evaluated using the deep learning PET image quality assessment model. Finally, the report will give the results of the image quality assessment (Good or Poor). The final output includes basic image information, scan and reconstruction parameters, typical instances of PET images, and deep learning score. The basic information in the output report includes patient information (such as age, gender, height, and weight), examination information (such as drug injected, time of injection, the dose of injection, time of injection), and equipment.

Results

Deep learning performance

The programming and computations were performed using a computer with a CPU of an Intel(R) Xeon(R) Silver 4110 @ 2.10 GHz processor and Tesla V100 GPU support. The proposed method was tested on PET images with four different tasks. The performance indexes in the evaluation of each task were reported based on the sensitivity, specificity, ACC, and AUC. We train all tasks by using the same training and testing strategy.

Figure 6 displays the ACC, AUC, sensitivity, and specificity for the four tasks over the fivefold cross-validation experiment in the training and testing data set. It can be seen that task 1 has the best overall performance (ACC = 0.87, Spe = 0.85, Sen = 0.90, AUC = 0.94), followed by task 4 in the training set. However, the specificity of task 4 was relatively poor at 0.71. Performances of task 3 were similar to task4, even the ACC, AUC, and sensitivity can outperform them on task 4. While the specificity for task 3 is far too low at 0.54. In terms of grouping conditions, tasks 3 and 4 both classify grades 1 and 2 as a group with poor image quality, and classify grades 3, 4, and 5 as a group with good image quality. Therefore, we think that perhaps this way of grouping images may have relatively low specificity. With the results of the validation group, we found that only the performance of task 4 can still maintain satisfactory results, and even improve the specificity to 0.8. Disappointingly, the performance of task 1 on the validation set dropped overall (ACC = 0.74, Spe = 0.77, Sen = 0.75, AUC = 0.79), and the model was relatively less stable. Based on the above results, we believe that task 4 is the model with the best comprehensive performance among the 4 tasks.

Fig.6
figure 6

ACC, AUC, sensitivity and specificity results over fivefold cross validation experiment in training (a) and testing (b). For task1, in the train set, AUC = 0.94, ACC = 0.87, specificity = 0.86, and sensitivity = 0.90. In the test set, AUC = 0.79, ACC = 0.74, specificity = 0.75, and sensitivity = 0.77. For task 2, in the train set, AUC = 0.51, ACC = 0.45, specificity = 0.35, and sensitivity = 0.62. In the test set, AUC = 0.51, ACC = 0.58, specificity = 0.45, and sensitivity = 0.73. For task 3, in the train set, AUC = 0.83, ACC = 0.74, specificity = 0.54, and sensitivity = 0.95. In the test set, AUC = 0.79, ACC = 0.67, specificity = 0.42, and sensitivity = 0.95. For task 4, in the train set, AUC = 0.86, ACC = 0.77, specificity = 0.71, and sensitivity = 0.83. In the test set, AUC = 0.91, ACC = 0.85, specificity = 0.79, and sensitivity = 0.91

Task 4 classifies the poor and optimal image quality successfully. Therefore, we describe here in detail the indicators of task 4, including all results in fivefold cross-validation. During five cross-fold validation, the best task was the task4 which reported an average AUC performance of 0.86, with a standard deviation of 0.06. The best-performing model of task 4 was used for subsequent analysis and had an AUC on the internal test set of 0.92. Task 4 showed a sensitivity of 0.91 and specificity of 0.80 for distinguishing between poor image quality (grade 1, grade 2) and optimal quality (grade 3, grade 4, grade 5) images. The overall accuracy of our classifier was calculated to be 0.85 for poor image quality images versus optimal quality images. The performance of each fold was reported in Table 4. The performance of the task4 was analyzed using ROC (Fig. 7), and the confusion is shown in Fig. 8.

Table 4 Performance of different folds for task 4
Fig.7
figure 7

ROC curve for the training and validation set of task 4 for grade classification

Fig.8
figure 8

Confusion matrices and classification performance measures in task 4

Case report

The PET QA report can be obtained after the user inputs PET images. Typical classification results of fundus images are shown in Fig. 9, which shows the report of a poor quality of image generated by listmode data in reconstruction with one-third statistics. This reduction leads to an ultimately poor image quality assessed by a senior physician, which is consistent with the quality control report automatically generated by the quality rating system, confirming the usefulness of the PET image quality rating system. The typical instances of PET images and deep learning score in the case report can intuitively feedback the quality of images for the physicians who take charge of patient management. The scan needs to be repeated if the deep learning score shows poor result, such as Fig. 9. Moreover, the basic image information and scan and reconstruction parameters would guide the technicians to avoid the same failure in the repeated scan.

Fig.9
figure 9

Case from PET QA Report where the poor quality of image assessed failed

Discussion

Since the quality of PET images is essential for further accurate data analysis, in this study, an architecture based on a deep DenseNet was evaluated for the assessment of image quality in PET images. We compared four tasks and verified their ability to assess the quality of PET images, respectively. Finally, we found that task 4 achieved the best performance in identifying poor image quality (grade 1, grade 2) versus optimal quality (grade 3, grade 4, grade 5) images where the images with grade 3 were also confirmed by three physicians and they were qualified in disease diagnosis. Task4 can provide rapid image classification and clinically relevant image features that can be used to provide feedback on image quality. Furthermore, the proposed deep learning model demonstrates the ability to classify images into two specified quality grades, which can be further applied to a quality control system to assist the automatic recognition of poor-quality images in the future. Thus, we developed a QA tool that uses the aforementioned deep learning method. The PET QA report obtained after inputting PET images can describe the information related to image quality in detail, which could help doctors to have a more comprehensive understanding of image quality.

Quality control of all medical images is critical, including PET images. To date, the research on the quality control of medical images mostly focuses on the ultrasound image, retinal images, ultra-widefield fluorescein angiography (UWFA), or OCT [10, 15, 16, 30,31,32]. There are few reports on quality control of PET images and implementing PET quality control platforms. Currently, PET image quality was assessed by manually drawing a sphere of hepar and comparing the mean liver intensity value across patients. However, this manual method consumes a lot of handling manpower and time. Herein, we propose a novel method for PET image quality assessment based on deep learning. Although only a few studies have addressed attempted to achieve PET image quality assessment based on deep learning, their inadequacies and shortcomings make them difficult to exert greater value in clinical applications. Elisabeth et al. [33] performed two CNNs trained to automatically identify EARL compliant images and separate if an image is meeting older or newer EARL standards. In their study, the two-dimensional image slices were used as input to the CNN and not the 3D information of the whole image. As we all know, the number of training data was enlarged by using 2D slices, thereby possibly increasing the classification performance of deep learning models. However, compared with 3D information of the image, 2D slices lose a lot of image-related information, which will lead to deviations in the accuracy of image quality assessment results. In order to avoid this situation, our study chose to input the 3D data into the model utilizing as much useful information as possible to complete the image quality assessment. In addition, we input the 3D whole-body PET image into the model, which can reflect the overall quality of the image compared with the local image. Thomas et al. [34] developed an automated pipeline for user-friendly and reproducible analysis of images with the aim of automating all processing steps up to the statistical analysis of measures derived from the final output images. Unfortunately, this study only analyzed brain images. In addition, the validation of each radiotracer accuracy was performed with differing ROI and using different methods for calculating parametric values. These differences mean that it is not possible to quantitatively compare their method accuracy for each radiotracer. This is enough to prove that better generalization can only be obtained by analyzing all aspects of the image rather than just a certain part. In further work, in addition to deep learning scoring, our study will comprehensively consider various factors and design a working flow of PET clinical imaging quality control to provide a step-by step evaluation of each key information point for physicians.

However, there were several important limitations of the present study that should be acknowledged. First, all whole-body PET images were obtained on a single PET-CT scanner. In the future, the reproducibility of this deep learning model should be tested across different scanners. Second, the data for this study were collected from the headless upper body. Therefore, PET data, including brain scans, will be considered in the follow-up investigation. Finally, the number of PET images in this study is relatively small, which limited the model’s ability to incorporate more data from either cohort. What’s more, there remains considerable data bias between different image quality grades. We did not validate our model in an external validation dataset. Additionly, image quality ratings are only made by three senior physicians, which would bias the results of assessment. To further refine our model and test the efficacy, we next will collaborate with clinics and other hospitals to collect more images to increase the number of PET images per grade, and also collaborate with more experienced physicians to make the rating more credible. In addition, when the number of images is large enough, we will try to perform five classifications and strive to achieve a more detailed division of image quality. We believe that concerted efforts in terms of data quantity and quality are needed in Densely Connected Convolutional Networks to make our deep learning model successful. Further, the value of this model for PET image quality assessment will be verified in clinical practice.

Conclusion

In conclusion, this study highlights the feasibility of the assessment of image quality in PET images using a deep learning model. This method not only provides automated image selection for clinical PET image review but also provides feedback to image quality, which may assist with accelerating clinical research by reliably assessing image quality.

Availability of data and materials

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Abbreviations

QA:

Quality assessment

DL:

Deep learning

PUMCH:

Peking union medical college hospital

DenseNet:

Dense convolutional network

ACC:

Accuracy

ROC:

Receiver operating characteristic curve

AUC:

Area under the ROC Curve

PET/CT:

Positron emission tomography/computed tomography

SUV:

Standardized uptake value

SUL:

Standardized uptake value normalized by lean body mass

CNN:

Convolutional neural network

OCT:

Optical coherence tomography

SNR:

Signal-to-noise ratio

FDOPA:

Fluorodopa

FDG:

Fluorodeoxyglucose

OSEM:

Ordered-subsets expectation maximization

FWHM:

Full width at half maximum

TP:

Ture positive

TN:

Ture negative

FP:

False positive

FN:

False negative

Sen:

Sensitivity

Spe:

Specificity

UWFA:

Ultra-widefield fluorescein angiography

References

  1. Antoch G, Vogt FM, Freudenberg LS, Nazaradeh F, Goehde SC, Barkhausen J, Dahmen G, Bockisch A, Debatin JF, Ruehm SG. Whole-body dual-modality PET/CT and whole-body MRI for tumor staging in oncology. JAMA. 2003;290:3199–206.

    Article  CAS  PubMed  Google Scholar 

  2. Gallamini A, Zwarthoed C, Borra A. Positron emission tomography (PET) in oncology. Cancers. 2014;6:1821–89.

    Article  PubMed  PubMed Central  Google Scholar 

  3. Boss DS, Olmos RV, Sinaasappel M, Beijnen JH, Schellens JHM. Application of PET/CT in the development of novel anticancer drugs. Oncologist. 2008;13:25–38.

    Article  CAS  PubMed  Google Scholar 

  4. Tahmassebi A, Ehtemami A, Mohebali B, Gandomi A, Pinker K, Meyer-Baese A. Big data analytics in medical imaging using deep learning. SPIE Defense + Commercial Sensing. SPIE; 2019.

  5. Boktor RR, Walker G, Stacey R, Gledhill S, Pitman AG. Reference range for intrapatient variability in blood-pool and liver SUV for 18F-FDG PET. J Nucl Med. 2013;54:677–82.

    Article  CAS  PubMed  Google Scholar 

  6. Wahl RL, Jacene H, Kasamon Y, Lodge MA. From RECIST to PERCIST: evolving considerations for PET response criteria in solid tumors. J Nucl Med. 2009;50(Suppl 1):122S-S150.

    Article  CAS  PubMed  Google Scholar 

  7. Meignan M, Barrington S, Itti E, Gallamini A, Haioun C, Polliack A. Report on the 4th international workshop on positron emission tomography in lymphoma held in menton, France, 3–5 October 2012. Leuk Lymphoma. 2014;55:31–7.

    Article  PubMed  Google Scholar 

  8. Alfaro-Almagro F, Jenkinson M, Bangerter NK, Andersson JLR, Griffanti L, Douaud G, Sotiropoulos SN, Jbabdi S, Hernandez-Fernandez M, Vallee E, Vidaurre D, Webster M, McCarthy P, Rorden C, Daducci A, Alexander DC, Zhang H, Dragonu I, Matthews PM, Miller KL, Smith SM. Image processing and quality control for the first 10,000 brain imaging datasets from UK Biobank. Neuroimage. 2018;166:400–24.

    Article  PubMed  Google Scholar 

  9. Van Essen DC, Ugurbil K, Auerbach E, Barch D, Behrens TEJ, Bucholz R, Chang A, Chen L, Corbetta M, Curtiss SW, Della Penna S, Feinberg D, Glasser MF, Harel N, Heath AC, Larson-Prior L, Marcus D, Michalareas G, Moeller S, Oostenveld R, Petersen SE, Prior F, Schlaggar BL, Smith SM, Snyder AZ, Xu J, Yacoub E. The human connectome project: a data acquisition perspective. Neuroimage. 2012;62:2222–31.

    Article  PubMed  Google Scholar 

  10. Wang J, Deng G, Li W, Chen Y, Gao F, Liu H, He Y, Shi G. Deep learning for quality assessment of retinal OCT images. Biomed Opt Express. 2019;10:6057–72.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Chalakkal RJ, Abdulla WH, Thulaseedharan SS. Quality and content analysis of fundus images using deep learning. Comput Biol Med. 2019;108:317–31.

    Article  PubMed  Google Scholar 

  12. Yu F, Sun J, Li A, Cheng J, Wan C, Liu J, editors. Image quality classification for DR screening using deep learning. In: 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); 2017.

  13. Czajkowska J, Juszczyk J, Piejko L, Glenc-Ambroży M. High-frequency ultrasound dataset for deep learning-based image quality assessment. Sensors. 2022;22:1478.

    Article  PubMed  PubMed Central  Google Scholar 

  14. Gao Q, Li S, Zhu M, Li D, Bian Z, Lyu Q, Zeng D, Ma J, editors. Blind CT Image Quality Assessment via Deep Learning Framework. In: 2019 IEEE Nuclear Science Symposium and Medical Imaging Conference (NSS/MIC); 2019.

  15. Bottani S, Burgos N, Maire A, Wild A, Ströer S, Dormont D, Colliot O. Automatic quality control of brain T1-weighted magnetic resonance images for a clinical data warehouse. Med Image Anal. 2022;75: 102219.

    Article  PubMed  Google Scholar 

  16. Pontoriero AD, Nordio G, Easmin R, Giacomel A, Santangelo B, Jahuar S, Bonoldi I, Rogdaki M, Turkheimer F, Howes O, Veronese M. Automated data quality control in FDOPA brain PET imaging using deep learning. Comput Methods Programs Biomed. 2021;208: 106239.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Sonni I, Park S, Baratto L, Hatami N, Davidzon G, Srinivas S, Gambhir S, Iagaru A. Initial experience with a SiPM-based PET/CT scanner: influence of acquisition time on image quality. J Nuclear Med. 2017;58:1369.

    Google Scholar 

  18. Lenga L, Czwikla R, Wichmann JL, Leithner D, Albrecht MH, Booz C, Arendt CT, Yel I, D’Angelo T, Vogl TJ, Martin SS. Dual-energy CT in patients with colorectal cancer: Improved assessment of hypoattenuating liver metastases using noise-optimized virtual monoenergetic imaging. Eur J Radiol. 2018;106:184–91.

    Article  PubMed  Google Scholar 

  19. van Sluis J, Boellaard R, Somasundaram A, van Snick PH, Borra RJH, Dierckx RAJO, Stormezand GN, Glaudemans AWJM, Noordzij W. Image quality and semiquantitative measurements on the biograph vision PET/CT system: initial experiences and comparison with the biograph mCT. J Nucl Med. 2020;61:129–35.

    Article  PubMed  Google Scholar 

  20. Peskett TR, Rau F, O’Driscoll J, Patani R, Lowe AR, Saibil HR. A liquid to solid phase transition underlying pathological huntingtin Exon1 aggregation. Mol Cell. 2018;70:588-601.e6.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  21. Urban G, Tripathi P, Alkayali T, Mittal M, Jalali F, Karnes W, Baldi P. Deep learning localizes and identifies polyps in real time with 96% accuracy in screening colonoscopy. Gastroenterology. 2018;155:1069-78.e8.

    Article  PubMed  Google Scholar 

  22. Min J, Im H, Allen M, McFarland PJ, Degani I, Yu H, Normandin E, Pathania D, Patel JM, Castro CM, Weissleder R, Lee H. Computational optics enables breast cancer profiling in point-of-care settings. ACS Nano. 2018;12:9081–90.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  23. Koho S, Tortarolo G, Castello M, Deguchi T, Diaspro A, Vicidomini G. Fourier ring correlation simplifies image restoration in fluorescence microscopy. Nat Commun. 2019;10:3103.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Gao Huang ZL, Laurens van der Maaten, Kilian Q. Weinberger. Densely Connected Convolutional Networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2017:4700–8.

  25. Farsiu S, Chiu SJ, O’Connell RV, Folgar FA, Yuan E, Izatt JA, Toth CA. Quantitative classification of eyes with and without intermediate age-related macular degeneration using optical coherence tomography. Ophthalmology. 2014;121:162–72.

    Article  PubMed  Google Scholar 

  26. Wang Y, Zhang Y, Yao Z, Zhao R, Zhou F. Machine learning based detection of age-related macular degeneration (AMD) and diabetic macular edema (DME) from optical coherence tomography (OCT) images. Biomed Opt Express. 2016;7:4928–40.

    Article  PubMed  PubMed Central  Google Scholar 

  27. Rasti R, Rabbani H, Mehridehnavi A, Hajizadeh F. Macular OCT classification using a multi-scale convolutional neural network ensemble. IEEE Trans Med Imaging. 2018;37:1024–34.

    Article  PubMed  Google Scholar 

  28. Treder M, Lauermann JL, Eter N. Automated detection of exudative age-related macular degeneration in spectral domain optical coherence tomography using deep learning. Graefes Arch Clin Exp Ophthalmol. 2018;256:259–65.

    Article  CAS  PubMed  Google Scholar 

  29. Metz CE. Basic principles of ROC analysis. Semin Nucl Med. 1978;8:283–98.

    Article  CAS  PubMed  Google Scholar 

  30. Czajkowska J, Juszczyk J, Piejko L, Glenc-Ambroży M. High-frequency ultrasound dataset for deep learning-based image quality assessment. Sensors 2022;22.

  31. Faes L, Wagner SK, Fu DJ, Liu X, Korot E, Ledsam JR, Back T, Chopra R, Pontikos N, Kern C, Moraes G, Schmid MK, Sim D, Balaskas K, Bachmann LM, Denniston AK, Keane PA. Automated deep learning design for medical image classification by health-care professionals with no coding experience: a feasibility study. Lancet Digital Health. 2019;1:e232–42.

    Article  PubMed  Google Scholar 

  32. Li HH, Abraham JR, Sevgi DD, Srivastava SK, Hach JM, Whitney J, Vasanji A, Reese JL, Ehlers JP. Automated quality assessment and image selection of ultra-widefield fluorescein angiography images through deep learning. Trans Vision Sci Technol. 2020;9(2):52.

    Article  Google Scholar 

  33. Pfaehler E, Euba D, Rinscheid A, Hoekstra OS, Zijlstra J, Jv Sluis, Brouwers AH, Lapa C, Boellaard R. Convolutional neural networks for automatic image quality control and EARL compliance of PET images. Res Square. 2021;9(1):1–13.

    Google Scholar 

  34. Funck T, Larcher K, Toussaint PJ, Evans AC, Thiel A. APPIAN: Automated Pipeline for PET Image Analysis. Frontiers in Neuroinformatics 2018;12.

Download references

Acknowledgements

Not applicable.

Funding

This work was sponsored by CAMS fund for Rare Diseases Research (No. 2016ZX310174-4), Tsinghua University and PUMCH joint fund (No. PTQH201906006), National Key Research and Development Program of China (No. 2020YFC2002702). PTQH201906006, 2020YFC2002702 and 2016ZX310174-4 played no role in the design of the study and collection, analysis, interpretation of data, and in writing the manuscript.

Author information

Authors and Affiliations

Authors

Contributions

HQZ and LH were responsible for the research design. HQZ and YMW researched the literature and wrote the main manuscript text. HQZ, YRM, NN and HLJ completed data collection. HQZ, YL and YMW analyzed the data. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Li Huo.

Ethics declarations

Ethics approval and consent to participate

The authors are accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. The study was conducted in accordance with the Declaration of Helsinki (as revised in 2013), and all the experimental protocols were approved by the institutional review board of Peking Union Medical College Hospital as well as the requirement for informed consent was waived by the institutional review board of Peking Union Medical College Hospital since this was a retrospective study.

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1:

Distribution of image quality in training and validation sets for each fold.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, H., Liu, Y., Wang, Y. et al. Deep learning model for automatic image quality assessment in PET. BMC Med Imaging 23, 75 (2023). https://doi.org/10.1186/s12880-023-01017-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12880-023-01017-2

Keywords