2.Full-size diffusion model for adaptive feature medical image fusion.
Jing DI ; Shuhui SHI ; Heran WANG ; Chan LIANG ; Yunlong ZHU
Journal of Biomedical Engineering 2025;42(5):871-882
To address issues such as loss of detailed information, blurred target boundaries, and unclear structural hierarchy in medical image fusion, this paper proposes an adaptive feature medical image fusion network based on a full-scale diffusion model. First, a region-level feature map is generated using a kernel-based saliency map to enhance local features and boundary details. Then, a full-scale diffusion feature extraction network is employed for global feature extraction, alongside a multi-scale denoising U-shaped network designed to fully capture cross-layer information. A multi-scale feature integration module is introduced to reinforce texture details and structural information extracted by the encoder. Finally, an adaptive fusion scheme is applied to progressively fuse region-level features, global features, and source images layer by layer, enhancing the preservation of detail information. To validate the effectiveness of the proposed method, this paper validates the proposed model on the publicly available Harvard dataset and an abdominal dataset. By comparing with nine other representative image fusion methods, the proposed approach achieved improvements across seven evaluation metrics. The results demonstrate that the proposed method effectively extracts both global and local features of medical images, enhances texture details and target boundary clarity, and generates fusion image with high contrast and rich information, providing more reliable support for subsequent clinical diagnosis.
Humans
;
Image Processing, Computer-Assisted/methods*
;
Algorithms
;
Neural Networks, Computer
;
Diagnostic Imaging/methods*
;
Image Interpretation, Computer-Assisted/methods*
3.A cephalometric landmark detection method using dual-encoder on X-ray image.
Chao DAI ; Chaolin HUANG ; Minpeng XU ; Yang WANG
Journal of Biomedical Engineering 2025;42(5):883-891
Accurate detection of cephalometric landmarks is crucial for orthodontic diagnosis and treatment planning. Current landmark detection methods are mainly divided into heatmap-based and regression-based approaches. However, these methods often rely on parallel computation of multiple models to improve accuracy, significantly increasing the complexity of training and deployment. This paper presented a novel regression method that can simultaneously detect all cephalometric landmarks in high-resolution X-ray images. By leveraging the encoder module of Transformer, a dual-encoder model was designed to achieve coarse-to-fine localization of cephalometric landmarks. The entire model consisted of three main components: a feature extraction module, a reference encoder module, and a fine-tuning encoder module, responsible for feature extraction and fusion of X-ray images, coarse localization of cephalometric landmarks, and fine localization of landmarks, respectively. The model was fully end-to-end differentiable and could learn the intercorrelation relationships between cephalometric landmarks. Experimental results showed that the successful detection rate (SDR) of our algorithm was superior to other existing methods. It attained the highest 2 mm SDR of 89.51% on test set 1 of the ISBI2015 dataset and 90.68% on the test set of the ISBI2023 dataset. Meanwhile, it reduces memory consumption and enhances the model's popularity and applicability, providing more reliable technical support for orthodontic diagnosis and treatment plan formulation.
Cephalometry/methods*
;
Humans
;
Algorithms
;
Anatomic Landmarks/diagnostic imaging*
;
Image Processing, Computer-Assisted/methods*
;
X-Rays
4.An adaptive multi-label classification model for diabetic retinopathy lesion recognition.
Xina LIU ; Jun XIE ; Junjun HOU ; Xinying XU ; Yan GUO
Journal of Biomedical Engineering 2025;42(5):892-900
Diabetic retinopathy is a common blinding complication in diabetic patients. Compared with conventional fundus color photography, fundus fluorescein angiography can dynamically display retinal vessel permeability changes, offering unique advantages in detecting early small lesions such as microaneurysms. However, existing intelligent diagnostic research on diabetic retinopathy images primarily focuses on fundus color photography, with relatively insufficient research on complex lesion recognition in fluorescein angiography images. This study proposed an adaptive multi-label classification model (D-LAM) to improve the recognition accuracy of small lesions by constructing a category-adaptive mapping module, a label-specific decoding module, and an innovative loss function. Experimental results on a self-built dataset demonstrated that the model achieved a mean average precision of 96.27%, a category F1-score of 91.21%, and an overall F1-score of 94.58%, with particularly outstanding performance in recognizing small lesions such as microaneurysms (AP = 1.00), significantly outperforming existing methods. The research provides reliable technical support for clinical diagnosis of diabetic retinopathy based on fluorescein angiography.
Diabetic Retinopathy/diagnostic imaging*
;
Humans
;
Fluorescein Angiography
;
Microaneurysm/diagnostic imaging*
;
Retinal Vessels
;
Algorithms
5.Research on attention-enhanced networks for subtype classification of age-related macular degeneration in optical coherence tomography.
Minghui CHEN ; Wenyi YANG ; Shiyi XU ; Yanqi LU ; Zhengqi YANG ; Fugang LI ; Zhensheng GU
Journal of Biomedical Engineering 2025;42(5):901-909
Subtype classification of age-related macular degeneration (AMD) based on optical coherence tomography (OCT) images serves as an effective auxiliary tool for clinicians in diagnosing disease progression and formulating treatment plans. To improve the classification accuracy of AMD subtypes, this study proposes a keypoint-based, attention-enhanced residual network (KPA-ResNet). The proposed architecture adopts a 50-layer residual network (ResNet-50) as the backbone, preceded by a keypoint localization module based on heatmap regression to outline critical lesion regions. A two-dimensional relative self-attention mechanism is incorporated into convolutional layers to enhance the representation of key lesion areas. Furthermore, the network depth is appropriately increased and an improved residual module, ConvNeXt, is introduced to enable comprehensive extraction of high-dimensional features and enrich the detail of lesion boundary contours, ultimately achieving higher classification accuracy of AMD subtypes. Experimental results demonstrate that KPA-ResNet achieves significant improvements in overall classification accuracy compared with conventional convolutional neural networks. Specifically, for the wet AMD subtypes, the classification accuracies for inactive choroidal neovascularization (CNV) and active CNV reach 92.8% and 95.2%, respectively, representing substantial improvement over ResNet-50. These findings validate the superior performance of KPA-ResNet in AMD subtype classification tasks. This work provides a high-accuracy, generalizable network architecture for OCT-based AMD subtype classification and offers new insights into integrating attention mechanisms with convolutional neural networks in ophthalmic image analysis.
Tomography, Optical Coherence/methods*
;
Humans
;
Macular Degeneration/diagnostic imaging*
;
Neural Networks, Computer
6.A multi-scale feature capturing and spatial position attention model for colorectal polyp image segmentation.
Wen GUO ; Xiangyang CHEN ; Jian WU ; Jiaqi LI ; Pengxue ZHU
Journal of Biomedical Engineering 2025;42(5):910-918
Colorectal polyps are important early markers of colorectal cancer, and their early detection is crucial for cancer prevention. Although existing polyp segmentation models have achieved certain results, they still face challenges such as diverse polyp morphology, blurred boundaries, and insufficient feature extraction. To address these issues, this study proposes a parallel coordinate fusion network (PCFNet), aiming to improve the accuracy and robustness of polyp segmentation. PCFNet integrates parallel convolutional modules and a coordinate attention mechanism, enabling the preservation of global feature information while precisely capturing detailed features, thereby effectively segmenting polyps with complex boundaries. Experimental results on Kvasir-SEG and CVC-ClinicDB demonstrate the outstanding performance of PCFNet across multiple metrics. Specifically, on the Kvasir-SEG dataset, PCFNet achieved an F1-score of 0.897 4 and a mean intersection over union (mIoU) of 0.835 8; on the CVC-ClinicDB dataset, it attained an F1-score of 0.939 8 and an mIoU of 0.892 3. Compared with other methods, PCFNet shows significant improvements across all performance metrics, particularly in multi-scale feature fusion and spatial information capture, demonstrating its innovativeness. The proposed method provides a more reliable AI-assisted diagnostic tool for early colorectal cancer screening.
Humans
;
Colonic Polyps/diagnostic imaging*
;
Colorectal Neoplasms/diagnostic imaging*
;
Neural Networks, Computer
;
Image Processing, Computer-Assisted/methods*
;
Algorithms
;
Early Detection of Cancer
7.Medical image segmentation method based on self-attention and multi-view attention.
Journal of Biomedical Engineering 2025;42(5):919-927
Most current medical image segmentation models are primarily built upon the U-shaped network (U-Net) architecture, which has certain limitations in capturing both global contextual information and fine-grained details. To address this issue, this paper proposes a novel U-shaped network model, termed the Multi-View U-Net (MUNet), which integrates self-attention and multi-view attention mechanisms. Specifically, a newly designed multi-view attention module is introduced to aggregate semantic features from different perspectives, thereby enhancing the representation of fine details in images. Additionally, the MUNet model leverages a self-attention encoding block to extract global image features, and by fusing global and local features, it improves segmentation performance. Experimental results demonstrate that the proposed model achieves superior segmentation performance in coronary artery image segmentation tasks, significantly outperforming existing models. By incorporating self-attention and multi-view attention mechanisms, this study provides a novel and efficient modeling approach for medical image segmentation, contributing to the advancement of intelligent medical image analysis.
Humans
;
Image Processing, Computer-Assisted/methods*
;
Neural Networks, Computer
;
Algorithms
;
Attention
;
Coronary Vessels/diagnostic imaging*
;
Diagnostic Imaging/methods*
8.Endometrial cancer lesion region segmentation based on large kernel convolution and combined attention.
Rushu PENG ; Qinghao ZENG ; Bin HE ; Junjie LIU ; Zhang XIAO
Journal of Biomedical Engineering 2025;42(5):928-935
Endometrial cancer (EC) is one of the most common gynecological malignancies, with an increasing incidence rate worldwide. Accurate segmentation of lesion areas in computed tomography (CT) images is a critical step in assisting clinical diagnosis. In this study, we propose a novel deep learning-based segmentation model, termed spatial choice and weight union network (SCWU-Net), which incorporates two newly designed modules: the spatial selection module (SSM) and the combination weight module (CWM). The SSM enhances the model's ability to capture contextual information through deep convolutional blocks, while the CWM, based on joint attention mechanisms, is employed within the skip connections to further boost segmentation performance. By integrating the strengths of both modules into a U-shaped multi-scale architecture, the model achieves precise segmentation of EC lesion regions. Experimental results on a public dataset demonstrate that SCWU-Net achieves a Dice similarity coefficient (DSC) of 82.98%, an intersection over union (IoU) of 78.63%, a precision of 92.36%, and a recall of 84.10%. Its overall performance is significantly outperforming other state-of-the-art models. This study enhances the accuracy of lesion segmentation in EC CT images and holds potential clinical value for the auxiliary diagnosis of endometrial cancer.
Humans
;
Endometrial Neoplasms/diagnostic imaging*
;
Female
;
Tomography, X-Ray Computed/methods*
;
Deep Learning
;
Algorithms
;
Image Processing, Computer-Assisted/methods*
;
Neural Networks, Computer
9.Deep overparameterized blood cell detection algorithm utilizing hybrid attention mechanisms.
Shuo ZHU ; Xukang ZHANG ; Zongyang WANG ; Rui JIANG ; Zhengda LIU
Journal of Biomedical Engineering 2025;42(5):936-944
To address the challenges in blood cell recognition caused by diverse morphology, dense distribution, and the abundance of small target information, this paper proposes a blood cell detection algorithm - the "You Only Look Once" model based on hybrid mixing attention and deep over-parameters (HADO-YOLO). First, a hybrid attention mechanism is introduced into the backbone network to enhance the model's sensitivity to detailed features. Second, the standard convolution layers with downsampling in the neck network are replaced with deep over-parameterized convolutions to expand the receptive field and improve feature representation. Finally, the detection head is decoupled to enhance the model's robustness for detecting abnormal cells. Experimental results on the Blood Cell Counting Dataset (BCCD) demonstrate that the HADO-YOLO algorithm achieves a mean average precision of 90.2% and a precision of 93.8%, outperforming the baseline YOLO model. Compared with existing blood cell detection methods, the proposed algorithm achieves state-of-the-art detection performance. In conclusion, HADO-YOLO offers a more efficient and accurate solution for identifying various types of blood cells, providing valuable technical support for future clinical diagnostic applications.
Algorithms
;
Humans
;
Blood Cells/cytology*
;
Blood Cell Count/methods*
;
Neural Networks, Computer
;
Deep Learning
;
Detection Algorithms
10.A head direction cell model based on a spiking neural network with landmark-free calibration.
Naigong YU ; Jingsen HUANG ; Ke LIN ; Zhiwen ZHANG
Journal of Biomedical Engineering 2025;42(5):970-976
In animal navigation, head direction is encoded by head direction cells within the olfactory-hippocampal structures of the brain. Even in darkness or unfamiliar environments, animals can estimate their head direction by integrating self-motion cues, though this process accumulates errors over time and undermines navigational accuracy. Traditional strategies rely on visual input to correct head direction, but visual scenes combined with self-motion information offer only partially accurate estimates. This study proposed an innovative calibration mechanism that dynamically adjusts the association between visual scenes and head direction based on the historical firing rates of head direction cells, without relying on specific landmarks. It also introduced a method to fine-tune error correction by modulating the strength of self-motion input to control the movement speed of the head direction cell activity bump. Experimental results showed that this approach effectively reduced the accumulation of self-motion-related errors and significantly enhanced the accuracy and robustness of the navigation system. These findings offer a new perspective for biologically inspired robotic navigation systems and underscore the potential of neural mechanisms in enabling efficient and reliable autonomous navigation.
Animals
;
Neural Networks, Computer
;
Calibration
;
Spatial Navigation/physiology*
;
Head Movements/physiology*
;
Neurons/physiology*
;
Models, Neurological
;
Head/physiology*
;
Action Potentials/physiology*


Result Analysis
Print
Save
E-mail