image segmentation

图像分割
  • 文章类型: Journal Article
    可变形图像配准(DIR)在许多临床任务中起着重要作用,在过去的几年中,深度学习在DIR方面取得了重大进展。
    提出一种用于单峰图像配准的快速多尺度无监督可变形图像配准(称为FMIRNet)方法。
    我们设计了一个多尺度融合模块,通过组合和细化三个尺度的变形场,来估计大位移场。在我们的融合模块中采用了空间注意力机制来逐个像素地加权位移场。除均方误差(MSE)外,我们还在训练阶段增加了结构相似性(ssim)度量,以增强变形图像和固定图像之间的结构一致性。
    我们的注册方法在EchoNet上进行了评估,混沌和SLIVER,在SSIM方面确实有了性能改进,NCC和NMI得分。此外,我们将FMIRNet集成到分段网络中(FCN,UNet)在我们的联合学习框架中使用很少的手动注释来增强数据集上的分段任务。实验结果表明,联合分割方法在Dice,HD和ASSD评分。
    我们提出的FMIRNet对于大变形估计是有效的,并且其注册能力在联合注册和分割框架中具有通用性和鲁棒性,可以为训练分割任务生成可靠的标签。
    UNASSIGNED: Deformable image registration (DIR) plays an important part in many clinical tasks, and deep learning has made significant progress in DIR over the past few years.
    UNASSIGNED: To propose a fast multiscale unsupervised deformable image registration (referred to as FMIRNet) method for monomodal image registration.
    UNASSIGNED: We designed a multiscale fusion module to estimate the large displacement field by combining and refining the deformation fields of three scales. The spatial attention mechanism was employed in our fusion module to weight the displacement field pixel by pixel. Except mean square error (MSE), we additionally added structural similarity (ssim) measure during the training phase to enhance the structural consistency between the deformed images and the fixed images.
    UNASSIGNED: Our registration method was evaluated on EchoNet, CHAOS and SLIVER, and had indeed performance improvement in terms of SSIM, NCC and NMI scores. Furthermore, we integrated the FMIRNet into the segmentation network (FCN, UNet) to boost the segmentation task on a dataset with few manual annotations in our joint leaning frameworks. The experimental results indicated that the joint segmentation methods had performance improvement in terms of Dice, HD and ASSD scores.
    UNASSIGNED: Our proposed FMIRNet is effective for large deformation estimation, and its registration capability is generalizable and robust in joint registration and segmentation frameworks to generate reliable labels for training segmentation tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在医疗临床场景中,由于患者隐私等原因,信息保护和数据迁移,当实际场景需要域自适应时,源域数据通常无法访问,并且只有源域上的预训练源模型可用。这类问题的现有解决方案在适应后往往会忘记以前在源域上学到的丰富的任务经验,这意味着该模型在适应时只是过度适应目标域数据,并且不会学习有助于实际任务决策的强大功能。我们通过探索无源域自适应在医学图像分割中的特殊应用来解决这个问题,并提出了一个两阶段加法无源域自适应框架。我们通过约束不同观点之间的核心病理结构和语义一致性来概括领域不变特征。并且我们减少了通过Monte-Carlo不确定性估计来定位和过滤可能有误差的元素所产生的分割。我们在跨设备息肉分割和跨模态脑肿瘤分割数据集上与其他方法进行了比较实验,目标域和源域的结果都验证了该方法能有效解决域偏移问题,并且在学习了目标域的新知识后,模型在源域上保持了优势。这项工作为在没有源数据的情况下实现目标域和源域的加性学习提供了有价值的探索,并为医学图像分割领域的适应性研究提供了新的思路和方法。
    In medical clinical scenarios for reasons such as patient privacy, information protection and data migration, when domain adaptation is needed for real scenarios, the source-domain data is often inaccessible and only the pre-trained source model on the source-domain is available. Existing solutions for this type of problem tend to forget the rich task experience previously learned on the source domain after adapting, which means that the model simply overfits the target-domain data when adapting and does not learn robust features that facilitate real task decisions. We address this problem by exploring the particular application of source-free domain adaptation in medical image segmentation and propose a two-stage additive source-free adaptation framework. We generalize the domain-invariant features by constraining the core pathological structure and semantic consistency between different perspectives. And we reduce the segmentation generated by locating and filtering elements that may have errors through Monte-Carlo uncertainty estimation. We conduct comparison experiments with some other methods on a cross-device polyp segmentation and a cross-modal brain tumor segmentation dataset, the results in both the target and source domains verify that the proposed method can effectively solve the domain offset problem and the model retains its dominance on the source domain after learning new knowledge of the target domain.This work provides valuable exploration for achieving additive learning on the target and source domains in the absence of source data and offers new ideas and methods for adaptation research in the field of medical image segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    肺部图像分割在计算机辅助肺部疾病诊断和治疗中起着重要作用。
    本文探讨了通过生成对抗网络进行肺部CT图像分割的方法。我们采用了各种生成对抗网络,并利用它们的图像平移功能来执行图像分割。采用生成对抗网络将原始肺部图像转换为分割图像。
    在真实的肺部图像数据集上测试了基于生成对抗网络的分割方法。实验结果表明,该方法优于最先进的方法。
    基于生成对抗网络的方法对肺部图像分割有效。
    UNASSIGNED: Lung image segmentation plays an important role in computer-aid pulmonary disease diagnosis and treatment.
    UNASSIGNED: This paper explores the lung CT image segmentation method by generative adversarial networks. We employ a variety of generative adversarial networks and used their capability of image translation to perform image segmentation. The generative adversarial network is employed to translate the original lung image into the segmented image.
    UNASSIGNED: The generative adversarial networks-based segmentation method is tested on real lung image data set. Experimental results show that the proposed method outperforms the state-of-the-art method.
    UNASSIGNED: The generative adversarial networks-based method is effective for lung image segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目标:与CNN相比,VisionTransformers最近取得了有竞争力的表现,因为它们具有出色的学习全球代表性的能力。然而,将它们应用于3D图像分割有两个主要挑战:i)由于3D医学图像的尺寸很大,由于巨大的计算成本,难以捕获全面的全球信息。ii)变压器中局部感应偏置不足会影响分割详细特征的能力,例如模糊和微妙定义的边界。因此,将VisionTransformer机制应用于医学图像分割领域,上述挑战需要充分克服。
    方法:我们提出了一种混合范式,称为可变形状混合变压器(VSmTrans),它整合了自我注意力和卷积,可以享受从自我注意力机制和卷积局部先验知识中自由学习复杂关系的好处。具体来说,我们设计了一种可变形状的自我注意机制,它可以快速扩展接受领域,而无需额外的计算成本,并在全球意识和本地细节之间实现良好的权衡。此外,并行卷积范式引入了强大的局部感应偏差,以促进挖掘细节的能力。同时,一对可学习的参数可以自动调整上述两个范例的重要性。对两个具有不同模态的公共医学图像数据集进行了广泛的实验:AMOSCT数据集和BraTS2021MRI数据集。
    结果:我们的方法在这些数据集上获得了88.3%和89.7%的最佳平均骰子得分,它优于以前最先进的基于SwinTransformer和基于CNN的架构。还进行了一系列消融实验,以验证所提出的混合机构和组件的效率,并探索VSmTrans中这些关键参数的有效性。
    结论:提出的用于3D医学图像分割的基于混合变压器的骨干网络可以紧密集成自注意和卷积,以利用这两种范例的优势。实验结果表明,与其他最先进的方法相比,我们的方法具有优越性。混合范例似乎最适合医学图像分割领域。消融实验还表明,所提出的混合机制可以有效地平衡具有局部感应偏差的大感受野,导致高度准确的分割结果,尤其是在捕捉细节方面。我们的代码可在https://github.com/qingze-bai/VSmTrans获得。
    OBJECTIVE: Vision Transformers recently achieved a competitive performance compared with CNNs due to their excellent capability of learning global representation. However, there are two major challenges when applying them to 3D image segmentation: i) Because of the large size of 3D medical images, comprehensive global information is hard to capture due to the enormous computational costs. ii) Insufficient local inductive bias in Transformers affects the ability to segment detailed features such as ambiguous and subtly defined boundaries. Hence, to apply the Vision Transformer mechanism in the medical image segmentation field, the above challenges need to be overcome adequately.
    METHODS: We propose a hybrid paradigm, called Variable-Shape Mixed Transformer (VSmTrans), that integrates self-attention and convolution and can enjoy the benefits of free learning of both complex relationships from the self-attention mechanism and the local prior knowledge from convolution. Specifically, we designed a Variable-Shape self-attention mechanism, which can rapidly expand the receptive field without extra computing cost and achieve a good trade-off between global awareness and local details. In addition, the parallel convolution paradigm introduces strong local inductive bias to facilitate the ability to excavate details. Meanwhile, a pair of learnable parameters can automatically adjust the importance of the above two paradigms. Extensive experiments were conducted on two public medical image datasets with different modalities: the AMOS CT dataset and the BraTS2021 MRI dataset.
    RESULTS: Our method achieves the best average Dice scores of 88.3 % and 89.7 % on these datasets, which are superior to the previous state-of-the-art Swin Transformer-based and CNN-based architectures. A series of ablation experiments were also conducted to verify the efficiency of the proposed hybrid mechanism and the components and explore the effectiveness of those key parameters in VSmTrans.
    CONCLUSIONS: The proposed hybrid Transformer-based backbone network for 3D medical image segmentation can tightly integrate self-attention and convolution to exploit the advantages of these two paradigms. The experimental results demonstrate our method\'s superiority compared to other state-of-the-art methods. The hybrid paradigm seems to be most appropriate to the medical image segmentation field. The ablation experiments also demonstrate that the proposed hybrid mechanism can effectively balance large receptive fields with local inductive biases, resulting in highly accurate segmentation results, especially in capturing details. Our code is available at https://github.com/qingze-bai/VSmTrans.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本研究旨在开发一种基于深度学习(DL)的上尿路(UUT)三维(3D)分割方法,包括输尿管和肾盂,非增强计算机断层扫描(NECT)扫描。本研究共选择150次NECT扫描,左侧UUT外观正常。将数据集分为训练集(n=130)和验证集(n=20)。测试集包含29例随机选择的病例,使用计算机断层扫描尿路造影(CTU)和NECT扫描,所有与左UUT的正常外观。有经验的放射科医师在每次扫描中标出左肾盂和输尿管。两种类型的框架(整体和部分),具有三种类型的DL模型(基本UNet,开发了UNet3+和ViT-UNet),并进行了评估。与所有其他测试方法相比,具有基本UNet模型的截面框架在测试集上实现了最高的平均精度(85.5%)和平均召回率(71.9%)。与CTU扫描相比,该方法的轴向UUT召回率高于CTU(82.5%对69.1%,P<0.01)。该方法在许多情况下实现了与CTU相似或更好的UUT可视化,然而,在某些情况下,它表现出不可忽视的假阳性率。所提出的DL方法在NECT扫描上的自动3DUUT分割中展示了有希望的潜力。所提出的DL模型可以显着提高UUT重建的效率,并有可能挽救许多患者免受CTU等侵入性检查的影响。DL模型也可以作为CTU的宝贵补充。
    This study aimed to develop a deep-learning (DL) based method for three-dimensional (3D) segmentation of the upper urinary tract (UUT), including ureter and renal pelvis, on non-enhanced computed tomography (NECT) scans. A total of 150 NECT scans with normal appearance of the left UUT were chosen for this study. The dataset was divided into training (n = 130) and validation sets (n = 20). The test set contained 29 randomly chosen cases with computed tomography urography (CTU) and NECT scans, all with normal appearance of the left UUT. An experienced radiologist marked out the left renal pelvis and ureter on each scan. Two types of frameworks (entire and sectional) with three types of DL models (basic UNet, UNet3 + and ViT-UNet) were developed, and evaluated. The sectional framework with basic UNet model achieved the highest mean precision (85.5%) and mean recall (71.9%) on the test set compared to all other tested methods. Compared with CTU scans, this method had higher axial UUT recall than CTU (82.5% vs 69.1%, P < 0.01). This method achieved similar or better visualization of UUT than CTU in many cases, however, in some cases, it exhibited a non-ignorable false-positive rate. The proposed DL method demonstrates promising potential in automated 3D UUT segmentation on NECT scans. The proposed DL models could remarkably improve the efficiency of UUT reconstruction, and have the potential to save many patients from invasive examinations such as CTU. DL models could also serve as a valuable complement to CTU.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    裂缝是混凝土表面的常见问题。随着基于机器视觉的检测系统的不断优化,有效的裂纹检测与识别是整个系统的核心。在这项研究中,支持向量机(SVM)用于区分裂缝与其他区域。为了完成SVM的识别系统,提出了一个由图像处理和识别模型组成的框架。提出了一种将Prewitt算子与Otsu阈值相结合的图像分割算法。新算法结合数学形态学处理的二值图像可以得到更完整的裂纹区域和更少的干涉区域。初始参数提取后,大多数杂质区域都是通过初步区分来筛选的,去除99%的杂质。该处理步骤确保了样品的平衡和有效性。建立基于径向基函数支持向量机的自动识别模型,紧密度,入住率,在将这三个特征与裂缝的所有六个特征进行比较后,选择了长宽比作为输入参数。该系统的识别准确率达到97.14%,证明了所提出的方法是有效的,满足了实际需求。
    Cracks are a common problem in concrete surfaces. With the continuous optimization of machine vision-based inspection systems, effective crack detection and recognition is the core of the entire system. In this study, support vector machine (SVM) was used to distinguish cracks from other regions. To complete the recognition system of the SVM, a framework consisting of an image processing and recognition model was proposed. An algorithm combining the Prewitt operator with the Otsu threshold was proposed for image segmentation. The binary image processed by the new algorithm combined with mathematical morphology can result in a more complete crack zone and fewer interference regions. After the initial parameter extraction, most of the impurity areas were screened by preliminary discrimination, removing 99% of the impurities. This processing step ensured the balance and effectiveness of the samples. To establish an automatic identification model based on SVM with a radial basis function, compactness, occupancy rate, and length-width ratio were selected as input parameters after comparing these three features with all six features of the crack. The recognition accuracy of this system reaches 97.14%, demonstrating that the proposed method is effective and satisfies practical requirements.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    足弓的形态特征和足底软组织厚度是评估足部健康的关键,这与各种足部和踝关节病变有关。通过将深度学习图像分割技术应用于侧向负重X射线图像,这项研究调查了足弓形态(FAM)与足底软组织厚度(PSTT)之间的相关性,检查年龄和性别的影响。具体来说,我们使用DeepLabV3+网络模型来准确描绘第一跖骨的边界,距骨,跟骨,舟骨,和整体的脚,实现FAM和PSTT的快速和自动测量。分析了包含1497张X射线图像的回顾性数据集,以探索FAM之间的关联,PSTT,和各种人口因素。我们的发现为足部形态提供了新的见解,为临床评估和干预提供强大的工具。精确数据支持提供的增强的检测和诊断能力促进了基于人群的研究和在临床环境中利用大数据。
    The morphological characteristics of the foot arch and the plantar soft tissue thickness are pivotal in assessing foot health, which is associated with various foot and ankle pathologies. By applying deep learning image segmentation techniques to lateral weight-bearing X-ray images, this study investigates the correlation between foot arch morphology (FAM) and plantar soft tissue thickness (PSTT), examining influences of age and sex. Specifically, we use the DeepLab V3+ network model to accurately delineate the boundaries of the first metatarsal, talus, calcaneus, navicular bones, and overall foot, enabling rapid and automated measurements of FAM and PSTT. A retrospective dataset containing 1497 X-ray images is analyzed to explore associations between FAM, PSTT, and various demographic factors. Our findings contribute novel insights into foot morphology, offering robust tools for clinical assessments and interventions. The enhanced detection and diagnostic capabilities provided by precise data support facilitate population-based studies and the leveraging of big data in clinical settings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:在磁共振成像(MRI)上可见的血管周围间隙(PVS)是与各种神经系统疾病相关的重要标志物。尽管PVS的定量分析可以提高敏感性并提高研究的一致性,该领域缺乏一种普遍验证的方法来分析来自多中心研究的图像。
    方法:我们在使用三大供应商(西门子,GeneralElectric,和飞利浦)。神经网络,MCPVS-Net(多中心PVS分割网络),使用来自40名受试者的数据进行训练,然后在15名受试者的单独队列中进行测试。我们根据为每个扫描仪供应商量身定制的地面实况掩模评估了分割准确性。此外,我们评估了每个扫描仪的分段PVS体积和视觉评分之间的一致性.我们还探讨了PVS体积与各种临床因素之间的相关性,例如年龄,高血压,和白质高强度(WMH)在1020名受试者的较大样本中。此外,mcPVS-Net被应用于包含来自联合成像扫描仪的T1w和T2加权(T2w)图像的新数据集以调查PVS体积是否可以区分具有不同视觉评分的受试者。我们还将mcPVS-Net与先前发布的从T1图像分割PVS的方法进行了比较。
    结果:在测试数据集中,mcPVS-Net的平均DICE系数为0.80,平均精度为0.81,Recall为0.79,表明具有良好的特异性和敏感性。分割的PVS体积与基底神经节(r=0.541,p<0.001)和白质区域(r=0.706,p<0.001)的视觉评分显着相关,和PVS体积在视觉评分不同的受试者之间存在显着差异。不同的扫描仪供应商之间的细分性能是一致的。PVS量与年龄显著相关,高血压,WMH。在联合成像扫描仪数据集中,PVS体积与在T1w或T2w图像上评估的PVS视觉评分显示出良好的关联。与以前发布的方法相比,mcPVS-Net显示出更高的准确性,并改善了基底神经节区域的PVS分割。
    结论:mcPVS-Net显示了从3DT1w图像中分割PVS的良好准确性。它可以作为未来PVS研究的有用工具。
    BACKGROUND: Perivascular spaces (PVS) visible on magnetic resonance imaging (MRI) are significant markers associated with various neurological diseases. Although quantitative analysis of PVS may enhance sensitivity and improve consistency across studies, the field lacks a universally validated method for analyzing images from multi-center studies.
    METHODS: We annotated PVS on multi-center 3D T1-weighted (T1w) images acquired using scanners from three major vendors (Siemens, General Electric, and Philips). A neural network, mcPVS-Net (multi-center PVS segmentation network), was trained using data from 40 subjects and then tested in a separate cohort of 15 subjects. We assessed segmentation accuracy against ground truth masks tailored for each scanner vendor. Additionally, we evaluated the agreement between segmented PVS volumes and visual scores for each scanner. We also explored correlations between PVS volumes and various clinical factors such as age, hypertension, and white matter hyperintensities (WMH) in a larger sample of 1020 subjects. Furthermore, mcPVS-Net was applied to a new dataset comprising both T1w and T2-weighted (T2w) images from a United Imaging scanner to investigate if PVS volumes could discriminate between subjects with differing visual scores. We also compared the mcPVS-Net with a previously published method that segments PVS from T1 images.
    RESULTS: In the test dataset, mcPVS-Net achieved a mean DICE coefficient of 0.80, with an average Precision of 0.81 and Recall of 0.79, indicating good specificity and sensitivity. The segmented PVS volumes were significantly associated with visual scores in both the basal ganglia (r = 0.541, p < 0.001) and white matter regions (r = 0.706, p < 0.001), and PVS volumes were significantly different among subjects with varying visual scores. Segmentation performance was consistent across different scanner vendors. PVS volumes exhibited significant associations with age, hypertension, and WMH. In the United Imaging scanner dataset, PVS volumes showed good associations with PVS visual scores evaluated on either T1w or T2w images. Compared to a previously published method, mcPVS-Net showed a higher accuracy and improved PVS segmentation in the basal ganglia region.
    CONCLUSIONS: The mcPVS-Net demonstrated good accuracy for segmenting PVS from 3D T1w images. It may serve as a useful tool for future PVS research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的:心肌声学造影(MCE)在诊断缺血中起着至关重要的作用。梗塞,肿块和其他心脏病。在MCE图像分析领域,准确和一致的心肌分割结果对于实现各种心脏疾病的自动分析至关重要。然而,当前MCE中的手动诊断方法的可重复性差,临床适用性有限。由于超声信号的不稳定性,MCE图像往往表现出低质量和高噪声,而干扰结构会进一步破坏分割的一致性。
    方法:为了克服这些挑战,我们提出了一个用于MCE分割的深度学习网络。这种架构利用扩张卷积来捕获大规模信息,而不牺牲位置准确性,并修改多头自我注意以增强全局上下文并确保一致性,有效地克服了与低图像质量和干扰相关的问题。此外,我们还调整了变压器与卷积神经网络的级联应用,以改善MCE中的分割。
    结果:在我们的实验中,与几种最先进的分割模型相比,我们的架构在标准MCE视图中获得了84.35%的最佳Dice评分.对于具有干扰结构(质量)的非标准视图和框架,我们的模型还获得了83.33%和83.97%的最佳骰子得分,分别。
    结论:这些研究证明我们的架构具有出色的形状一致性和坚固性,这使得它能够处理各种类型的MCE的分割。我们相对精确和一致的心肌分割结果为自动分析各种心脏病提供了基本条件,有可能发现潜在的病理特征并降低医疗保健成本。
    OBJECTIVE: Myocardial contrast echocardiography (MCE) plays a crucial role in diagnosing ischemia, infarction, masses and other cardiac conditions. In the realm of MCE image analysis, accurate and consistent myocardial segmentation results are essential for enabling automated analysis of various heart diseases. However, current manual diagnostic methods in MCE suffer from poor repeatability and limited clinical applicability. MCE images often exhibit low quality and high noise due to the instability of ultrasound signals, while interference structures can further disrupt segmentation consistency.
    METHODS: To overcome these challenges, we proposed a deep-learning network for the segmentation of MCE. This architecture leverages dilated convolutions to capture high-scale information without sacrificing positional accuracy and modifies multi-head self-attention to enhance global context and ensure consistency, effectively overcoming issues related to low image quality and interference. Furthermore, we also adapted the cascade application of transformers with convolutional neural networks for improved segmentation in MCE.
    RESULTS: In our experiments, our architecture achieved the best Dice score of 84.35% for standard MCE views compared with that of several state-of-the-art segmentation models. For non-standard views and frames with interfering structures (mass), our models also attained the best Dice scores of 83.33% and 83.97%, respectively.
    CONCLUSIONS: These studies proved that our architecture is of excellent shape consistency and robustness, which allows it to deal with segmentation of various types of MCE. Our relatively precise and consistent myocardial segmentation results provide fundamental conditions for the automated analysis of various heart diseases, with the potential to discover underlying pathological features and reduce healthcare costs.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    图像分割是图像处理领域中一个至关重要的过程。多级阈值分割是一种有效的图像分割方法,其中图像根据多级阈值被分割成不同的区域以进行信息分析。然而,多级阈值处理的复杂性随着阈值数量的增加而急剧增加。为了应对这一挑战,本文提出了一种新的混合算法,称为差分进化-黄金狼优化器(DEGJO),使用最小交叉熵(MCE)作为适应度函数进行多级阈值图像分割。将DE算法与GJO算法相结合,进行位置的迭代更新,这增强了GJO算法的搜索能力。在CEC2021基准函数上评估DEGJO算法的性能,并与最先进的优化算法进行比较。此外,通过对基准图像进行多级分割实验,评估了该算法的有效性。实验结果表明,与其他元启发式算法相比,DEGJO算法在适应度值方面具有优越的性能。此外,它还在定量性能指标方面产生良好的结果,如峰值信噪比(PSNR),结构相似性指数(SSIM),和特征相似性指数(FSIM)测量。
    Image segmentation is a crucial process in the field of image processing. Multilevel threshold segmentation is an effective image segmentation method, where an image is segmented into different regions based on multilevel thresholds for information analysis. However, the complexity of multilevel thresholding increases dramatically as the number of thresholds increases. To address this challenge, this article proposes a novel hybrid algorithm, termed differential evolution-golden jackal optimizer (DEGJO), for multilevel thresholding image segmentation using the minimum cross-entropy (MCE) as a fitness function. The DE algorithm is combined with the GJO algorithm for iterative updating of position, which enhances the search capacity of the GJO algorithm. The performance of the DEGJO algorithm is assessed on the CEC2021 benchmark function and compared with state-of-the-art optimization algorithms. Additionally, the efficacy of the proposed algorithm is evaluated by performing multilevel segmentation experiments on benchmark images. The experimental results demonstrate that the DEGJO algorithm achieves superior performance in terms of fitness values compared to other metaheuristic algorithms. Moreover, it also yields good results in quantitative performance metrics such as peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and feature similarity index (FSIM) measurements.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号