Computed Tomography Image Quality Assessment Based on Deep Learning

Project Leaders

Siyi Xun

Partner Organisations

深圳人民医院

上海交通大学

In view of the problems existing in the non-homogeneity of CT image quality in clinical diagnosis and storage and management of medical data, scientific research on the following key issues is planned to focus on the development of effective automated CT-IQA datasets and models:

(1) Dataset

At present, there are few publicly available data sets for CT-IQA. Most of the existing data sets are simulated and synthesized by adding noise. The quality of simulation data is very different and easy to distinguish, but it cannot fully capture the physiological and anatomical differences between real patients, and it is difficult to accurately simulate the complex lesions and pathological conditions existing in the clinic. Therefore, we will make innovative use of real data collected from multiple medical institutions such as Shenzhen People's Hospital to build three datasets: 1) Chest CT image quality assessment dataset; 2) Chest CT body position and artifact assessment data set; 3) Cross-modal medical image quality assessment dataset. As far as we know, no such dataset currently exists, which would fill a gap in the field of medical data.

(2) A multi-task model for chest CT-IQA and classification is proposed for the first time.

The model is composed of backbone network, window and current classification module, assessment index regression module and quality assessment module modified by 3D VGG. This model can realize automatic classification and quality assessment of CT image window to a certain extent. To our knowledge, this is the first automated model to classify and evaluate the quality of CT images from different Windows.

(3) The pipeline of chest CT image position and artifact detection and quality assessment was innovatively proposed to effectively assist clinical diagnosis.

The deep learning model is first used to make regression prediction of the missing anatomical structure caused by the body dislocation, and then the presence and degree of artifacts are detected. The combination of the two effectively supplements the quality information of CT images, to automatically evaluate the quality of CT images.

(4) Innovative development of quality assessment models that can be used for different modes of medical images.

Based on the design model of global definition and local noise of medical images, the dual encoder structure of VGG improved by Hyper-conv and Swin-transformer is intended to be adopted for global and local feature extraction, and different modes are embedding as guidance information. It can improve network accuracy and generalization. The model enables automated quality assessment of medical images of multiple modes and anatomical sites. This will be the first universal quality assessment model for medical images of different modes.

Project Example


Deep Learning for Medical Image Quality Assessment

Project Leaders

Siyi Xun

This research is oriented towards the demands of medical image quality control. In response to the key scientific issues that urgently need to be addressed in the field of medical image quality control mentioned above, this study intends to conduct research at three levels: (1) Medical IQA databases, (2) Core technologies for medical IQA, and (3) Investigation of the correlation between image quality and clinical diagnosis. The specific aims and objectives of this research include:

(1) Addressing the critical gap in accessible and standardized datasets for medical image quality control, this research established a set of standardized, comprehensive, and well-interpretable quantitative criteria for chest CT image quality. We innovatively integrated real-world clinical data from Shenzhen People's Hospital with resources from multiple open-source public datasets. Leveraging this integration, we designed and constructed three novel datasets characterized by high practical utility: (i) a chest CT image classification and quality assessment dataset, (ii) a cross-modality 3D medical IQA dataset, and (iii) a cross-modal cross-organ medical IQA dataset. The establishment of these resources not only fills a significant void in the field of medical IQA data, but also provides a solid data foundation for the standardized and intelligent advancement of research in this area.

(2) To address the challenge of assessing and classifying chest CT image quality under complex conditions, we investigate a multi-task model capable of jointly performing CT image series type categorization, quality classification, and automatic regression of Signal-to-Noise Ratio and BRISQUE metrics. By integrating the results of these multiple tasks, the model achieves simultaneous classification and preliminary quality assessment of chest CT images.To further enhance the accuracy and clinical utility of the assessment results, the model incorporates auxiliary supervision from human annotations. The final output is an automated image quality score. This score provides clinicians with critical decision support when browsing and filtering vast volumes of medical images, particularly those of various types that are difficult to make intuitive and effective horizontal comparisons.

(3) To address the challenge of universal quality assessment for cross-modality and cross-organ medical images, we build a universal medical cross-modality IQA model. The model architecture adopts a two-stage progressive framework: Stage 1: Physical Parameter-Guided Mapping. Addressing the scarcity of quality annotation labels in medical imaging, this stage utilizes physical parameters as surrogate labels to establish a mapping relationship between these parameters and image quality. Building upon this, we propose a Hybrid-conv model with perceptual prompts. This model fuses modality-specific features with anatomical structure prior knowledge to achieve preliminary quality scoring for multi-modal, multi-organ images. Stage 2: Foundation Model with Expert Integration. Utilizing stage 1 IQA technology as a foundation, this stage constructs a physical parameter-based pretraining as an upstream task to learn common cross-domain quality features. Subsequently, it integrates expert experience as supervision to accomplish domain-specific downstream IQA tasks. Concurrently, the model incorporates a multi-scale saliency-aware slice selection module and an adaptive domain-aware prompt fusion module. This represents the first universal and scalable foundational model for medical IQA.

(4) Addressing the need to investigate the correlation between medical image quality and AI diagnostic performance in clinical environments, this research aims to perform a comprehensive systematic analysis to deep-explore the quantitative relationship between them. The study utilizes a stratified and balanced lung nodule CT image dataset constructed for this purpose. Nodule detection is performed using certified AI-assisted software. Additionally, average image quality scores are obtained through a standardized process involving multiple reviewers. Employing multidimensional statistical analysis methods the research precisely quantifies the impact of image quality on the lesion detection capability of the AI model while controlling for variables such as gender and age. The findings will provide empirical evidence for assessing the robustness of AI diagnostic models when applied to real-world clinical data.

(5) Computed tomography image quality is significantly impacted by various factors, including scanning parameters , patient factors, motion artifacts, and metal artifacts. While the medical imaging community has developed some tools for data quality validation, dedicated IQA methods for CT, especially for artifacts and positioning, remain challenging. This study proposes DEFCT-QA (Defective Computed Tomography Quality Assessment), a dedicated model for assessing medical image quality. DEFCT-QA simulates common clinical defect scenarios, including incomplete imaging, field-of-view errors, metal artifacts, and motion artifacts. Acknowledging the critical impact of noise and clarity, we integrated manually designed features extracted via traditional algorithms with semantic features derived from deep learning models to form a comprehensive geometric feature representation. Subsequent fusion yielded an image quality perception representation. Critically, rigorous multi-center external validation using the public multi-center Luna16 dataset further confirmed the model's effectiveness and generalizability in assessing diverse image defects.

Project Example