학술논문

Cytopathology image analysis method based on high-resolution medical representation learning in medical decision-making system
Document Type
article
Source
Complex & Intelligent Systems, Vol 10, Iss 3, Pp 4253-4274 (2024)
Subject
Cytopathology images
Artificial intelligence
Assisted analysis
Image segmentation
Knowledge sublimation
Electronic computers. Computer science
QA75.5-76.95
Information technology
T58.5-58.64
Language
English
ISSN
2199-4536
2198-6053
Abstract
Abstract Artificial intelligence has made substantial progress in many medical application scenarios. The quantity and complexity of pathology images are enormous, but conventional visual screening techniques are labor-intensive, time-consuming, and subject to some degree of subjectivity. Complex pathological data can be converted into mineable image features using artificial intelligence image analysis technology, enabling medical professionals to quickly and quantitatively identify regions of interest and extract information about cellular tissue. In this study, we designed a medical information assistance system for segmenting pathology images and quantifying statistical results, including data enhancement, cell nucleus segmentation, model tumor, and quantitative analysis. In cell nucleus segmentation, to address the problem of uneven healthcare resources, we designed a high-precision teacher model (HRMED_T) and a lightweight student model (HRMED_S). The HRMED_T model is based on visual Transformer and high-resolution representation learning. It achieves accurate segmentation by parallel low-resolution convolution and high-scaled image iterative fusion, while also maintaining the high-resolution representation. The HRMED_S model is based on the Channel-wise Knowledge Distillation approach to simplify the structure, achieve faster convergence, and refine the segmentation results by using conditional random fields instead of fully connected structures. The experimental results show that our system has better performance than other methods. The Intersection over the Union (IoU) of HRMED_T model reaches 0.756. The IoU of HRMED_S model also reaches 0.710 and params is only 3.99 M.