Conditional GAN

条件 GAN
  • 文章类型: Journal Article
    背景:基于图像的作物生长建模可以通过揭示空间作物随时间的发展来为精准农业做出重大贡献,它允许对相关的未来植物性状进行早期和特定位置的估计,如叶面积或生物量。生成现实而清晰的作物图像的前提是将多个生长影响条件集成到一个模型中,例如初始生长阶段的图像,相关的生长时间,以及有关现场治疗的更多信息。虽然基于图像的模型比基于过程的模型为作物生长建模提供了更大的灵活性,在各种影响增长的条件的综合整合方面仍然存在很大的研究差距。需要进一步的探索和调查来解决这一差距。
    方法:我们提出了一个两阶段框架,该框架由第一个图像生成模型和第二个增长估计模型组成,独立训练。图像生成模型是有条件的Wasserstein生成对抗网络(CWGAN)。在此模型的生成器中,条件批量归一化(CBN)用于集成不同类型的条件以及输入图像。这允许模型根据多个影响因素生成时变人工图像。框架的第二部分通过得出植物特定的性状并将其与非人工(真实)参考图像的性状进行比较来使用这些图像进行植物表型分析。此外,使用多尺度结构相似性(MS-SSIM)评估图像质量,学习感知图像块相似性(LPIPS),和Fréchet起始距离(FID)。在推理过程中,该框架允许为训练中使用的任何条件组合生成图像;我们称这种生成为数据驱动的作物生长模拟。
    结果:实验是在三个不同复杂度的数据集上进行的。这些数据集包括实验室植物拟南芥(拟南芥)和在实际田间条件下生长的作物,即花椰菜(GrowliFlower)和由蚕豆和春小麦(MixedCrop)组成的作物混合物。在所有情况下,该框架允许现实的,清晰的图像世代,从短期到长期预测的质量略有下降。对于在不同处理下生长的混合作物(不同品种,播种密度),结果表明,添加这些处理信息增加了一代质量和表型的准确性测量的估计生物量。用受过训练的框架对不同的生长影响条件进行模拟,为这些因素如何与作物外观相关提供了有价值的见解。这在复杂的情况下特别有用,探索较少的作物混合物系统。进一步的结果表明,添加基于过程的模拟生物量作为条件增加了来自预测图像的衍生表型性状的准确性。这证明了我们的框架作为数据驱动和基于过程的作物生长模型之间的接口的潜力。
    结论:通过多条件CWGAN,对未来植物外观的真实生成和模拟是充分可行的。提出的框架补充了基于流程的模型,克服了它们的局限性,例如对假设的依赖和低精确的现场定位特异性,通过对空间作物发育的逼真可视化,直接导致模型预测的高度可解释性。
    BACKGROUND: Image-based crop growth modeling can substantially contribute to precision agriculture by revealing spatial crop development over time, which allows an early and location-specific estimation of relevant future plant traits, such as leaf area or biomass. A prerequisite for realistic and sharp crop image generation is the integration of multiple growth-influencing conditions in a model, such as an image of an initial growth stage, the associated growth time, and further information about the field treatment. While image-based models provide more flexibility for crop growth modeling than process-based models, there is still a significant research gap in the comprehensive integration of various growth-influencing conditions. Further exploration and investigation are needed to address this gap.
    METHODS: We present a two-stage framework consisting first of an image generation model and second of a growth estimation model, independently trained. The image generation model is a conditional Wasserstein generative adversarial network (CWGAN). In the generator of this model, conditional batch normalization (CBN) is used to integrate conditions of different types along with the input image. This allows the model to generate time-varying artificial images dependent on multiple influencing factors. These images are used by the second part of the framework for plant phenotyping by deriving plant-specific traits and comparing them with those of non-artificial (real) reference images. In addition, image quality is evaluated using multi-scale structural similarity (MS-SSIM), learned perceptual image patch similarity (LPIPS), and Fréchet inception distance (FID). During inference, the framework allows image generation for any combination of conditions used in training; we call this generation data-driven crop growth simulation.
    RESULTS: Experiments are performed on three datasets of different complexity. These datasets include the laboratory plant Arabidopsis thaliana (Arabidopsis) and crops grown under real field conditions, namely cauliflower (GrowliFlower) and crop mixtures consisting of faba bean and spring wheat (MixedCrop). In all cases, the framework allows realistic, sharp image generations with a slight loss of quality from short-term to long-term predictions. For MixedCrop grown under varying treatments (different cultivars, sowing densities), the results show that adding these treatment information increases the generation quality and phenotyping accuracy measured by the estimated biomass. Simulations of varying growth-influencing conditions performed with the trained framework provide valuable insights into how such factors relate to crop appearances, which is particularly useful in complex, less explored crop mixture systems. Further results show that adding process-based simulated biomass as a condition increases the accuracy of the derived phenotypic traits from the predicted images. This demonstrates the potential of our framework to serve as an interface between a data-driven and a process-based crop growth model.
    CONCLUSIONS: The realistic generation and simulation  of future plant appearances is adequately feasible by multi-conditional CWGAN. The presented framework complements process-based models and overcomes their limitations, such as the reliance on assumptions and the low exact field-localization specificity, by realistic visualizations of the spatial crop development that directly lead to a high explainability of the model predictions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:心脏正电子发射断层扫描(PET)可以可视化和量化心脏功能的分子和生理途径。然而,心脏和呼吸运动会引入模糊,从而降低PET图像质量和定量精度。双心脏和呼吸门控PET重建可以减轻运动伪影,但增加噪声,因为仅数据的子集用于心动周期的每个时间帧。
    目的:本研究的目的是使用条件生成对抗网络(cGAN)创建零拍摄图像去噪框架,以提高非门控和双门控心脏PET图像的图像质量和定量准确性。
    方法:我们的研究包括40例接受18F-氟代脱氧葡萄糖(18F-FDG)心脏PET研究的患者的回顾性列表模式数据。我们最初训练并评估了3DcGAN,称为Pix2Pix-on模拟的非门控低计数PET数据与相应的全计数目标数据配对,然后将模型部署在同一PET/CT系统上采集的未知测试集上,包括非门控和双门控PET数据。
    结果:定量分析表明,3DPix2Pix网络架构在非门控和门控心脏PET图像中均实现了显着(p值<0.05)增强的图像质量和准确性。5%,10%,和15%保留的计数统计数据,该模型将峰值信噪比(PSNR)提高了33.7%,21.2%,和15.5%,结构相似性指数(SSIM)下降7.1%,3.3%,和2.2%,平均绝对误差(MAE)减少61.4%,54.3%,49.7%,分别。当在双门PET数据上测试时,该模型持续降低了噪音,不考虑心脏/呼吸运动阶段,同时保持图像分辨率和准确性。在所有大门上都观察到了显著的改善,包括PSNR增加34.7%,SSIM改善7.8%,MAE减少60.3%。
    结论:这项研究的结果表明,双门控心脏PET图像,通常具有可能影响诊断性能的重建后伪影,可以使用生成预训练去噪网络有效地改进。
    BACKGROUND: Cardiac positron emission tomography (PET) can visualize and quantify the molecular and physiological pathways of cardiac function. However, cardiac and respiratory motion can introduce blurring that reduces PET image quality and quantitative accuracy. Dual cardiac- and respiratory-gated PET reconstruction can mitigate motion artifacts but increases noise as only a subset of data are used for each time frame of the cardiac cycle.
    OBJECTIVE: The objective of this study is to create a zero-shot image denoising framework using a conditional generative adversarial networks (cGANs) for improving image quality and quantitative accuracy in non-gated and dual-gated cardiac PET images.
    METHODS: Our study included retrospective list-mode data from 40 patients who underwent an 18F-fluorodeoxyglucose (18F-FDG) cardiac PET study. We initially trained and evaluated a 3D cGAN-known as Pix2Pix-on simulated non-gated low-count PET data paired with corresponding full-count target data, and then deployed the model on an unseen test set acquired on the same PET/CT system including both non-gated and dual-gated PET data.
    RESULTS: Quantitative analysis demonstrated that the 3D Pix2Pix network architecture achieved significantly (p value<0.05) enhanced image quality and accuracy in both non-gated and gated cardiac PET images. At 5%, 10%, and 15% preserved count statistics, the model increased peak signal-to-noise ratio (PSNR) by 33.7%, 21.2%, and 15.5%, structural similarity index (SSIM) by 7.1%, 3.3%, and 2.2%, and reduced mean absolute error (MAE) by 61.4%, 54.3%, and 49.7%, respectively. When tested on dual-gated PET data, the model consistently reduced noise, irrespective of cardiac/respiratory motion phases, while maintaining image resolution and accuracy. Significant improvements were observed across all gates, including a 34.7% increase in PSNR, a 7.8% improvement in SSIM, and a 60.3% reduction in MAE.
    CONCLUSIONS: The findings of this study indicate that dual-gated cardiac PET images, which often have post-reconstruction artifacts potentially affecting diagnostic performance, can be effectively improved using a generative pre-trained denoising network.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在过去的十年里,深度学习(DL)算法已成为在超声(US)检查期间帮助临床医生识别胎头标准平面(FHSP)的有前景的工具.然而,这些算法在临床环境中的采用仍然受到缺乏大型注释数据集的阻碍.为了克服这个障碍,我们介绍FetalBrainAwareNet,一个创新的框架,旨在合成FHSP的解剖学精确图像。FetalBrainAwareNet引入了一种先进的方法,该方法在其条件对抗训练过程中利用类激活图作为先验。这种方法促进了合成图像中特定解剖标志的存在。此外,我们研究了对抗训练损失函数中的专门正则化术语,以控制胎儿颅骨的形态并促进标准平面之间的区分,确保合成图像在结构和整体外观上都忠实地代表真实的US扫描。我们的FetalBrainAwareNet框架的多功能性突出了它能够使用一个奇异的,生成三个主要的FHSP的高质量图像,集成框架。定量(Fréchet起始距离为88.52)和定性(t-SNE)结果表明,与最先进的方法相比,我们的框架生成的US图像具有更大的可变性。通过使用我们的框架生成的合成图像,我们将FHSP分类器的准确率提高了3.2%,与仅使用实际采集来训练相同的分类器相比.这些成就表明,使用我们的合成图像来增加训练集可以为增强可以集成在实际临床场景中的FHSP分类的DL算法的性能提供好处。
    Over the past decade, deep-learning (DL) algorithms have become a promising tool to aid clinicians in identifying fetal head standard planes (FHSPs) during ultrasound (US) examination. However, the adoption of these algorithms in clinical settings is still hindered by the lack of large annotated datasets. To overcome this barrier, we introduce FetalBrainAwareNet, an innovative framework designed to synthesize anatomically accurate images of FHSPs. FetalBrainAwareNet introduces a cutting-edge approach that utilizes class activation maps as a prior in its conditional adversarial training process. This approach fosters the presence of the specific anatomical landmarks in the synthesized images. Additionally, we investigate specialized regularization terms within the adversarial training loss function to control the morphology of the fetal skull and foster the differentiation between the standard planes, ensuring that the synthetic images faithfully represent real US scans in both structure and overall appearance. The versatility of our FetalBrainAwareNet framework is highlighted by its ability to generate high-quality images of three predominant FHSPs using a singular, integrated framework. Quantitative (Fréchet inception distance of 88.52) and qualitative (t-SNE) results suggest that our framework generates US images with greater variability compared to state-of-the-art methods. By using the synthetic images generated with our framework, we increase the accuracy of FHSP classifiers by 3.2% compared to training the same classifiers solely with real acquisitions. These achievements suggest that using our synthetic images to increase the training set could provide benefits to enhance the performance of DL algorithms for FHSPs classification that could be integrated in real clinical scenarios.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    三维(3D)点云重建的性能受到诸如植被的动态特征的影响。植被可以通过基于近红外(NIR)的指数来检测;但是,提供多光谱数据的传感器是资源密集型的。为了解决这个问题,本研究提出了一个两阶段的框架,首先提高性能的三维点云生成的建筑物与两视图SfM算法,其次,减少由植被引起的噪音。拟议的框架还可以克服在识别植被区域以减少SfM过程中的干扰时缺乏近红外数据的问题。第一阶段包括跨传感器训练,模型选择以及使用生成对抗网络(GAN)对图像到图像的RGB到彩色红外(CIR)转换的评估。第二阶段包括具有多个特征检测器运算符的特征检测,关于基于NDVI的植被分类的特征去除,掩蔽,匹配,姿态估计和三角测量生成稀疏三维点云。在这两个阶段中使用的材料是公开可用的RGB-NIR数据集,以及卫星和无人机图像。实验结果表明,交叉传感器和分类验证的精度分别为0.9466和0.9024,kappa系数分别为0.8932和0.9110。基于直方图的评估表明,预测的NIR波段与卫星测试数据集的原始NIR数据一致。最后,通过对无人机RGB和人工生成的NIR进行分段驱动的两视图SfM测试,证明了所提出的框架可以有效地将RGB转换为CIR进行NDVI计算。Further,人工生成的NDVI能够对植被进行分割和分类。因此,生成的点云噪音较小,并且增强了3D模型。
    The performance of three-dimensional (3D) point cloud reconstruction is affected by dynamic features such as vegetation. Vegetation can be detected by near-infrared (NIR)-based indices; however, the sensors providing multispectral data are resource intensive. To address this issue, this study proposes a two-stage framework to firstly improve the performance of the 3D point cloud generation of buildings with a two-view SfM algorithm, and secondly, reduce noise caused by vegetation. The proposed framework can also overcome the lack of near-infrared data when identifying vegetation areas for reducing interferences in the SfM process. The first stage includes cross-sensor training, model selection and the evaluation of image-to-image RGB to color infrared (CIR) translation with Generative Adversarial Networks (GANs). The second stage includes feature detection with multiple feature detector operators, feature removal with respect to the NDVI-based vegetation classification, masking, matching, pose estimation and triangulation to generate sparse 3D point clouds. The materials utilized in both stages are a publicly available RGB-NIR dataset, and satellite and UAV imagery. The experimental results indicate that the cross-sensor and category-wise validation achieves an accuracy of 0.9466 and 0.9024, with a kappa coefficient of 0.8932 and 0.9110, respectively. The histogram-based evaluation demonstrates that the predicted NIR band is consistent with the original NIR data of the satellite test dataset. Finally, the test on the UAV RGB and artificially generated NIR with a segmentation-driven two-view SfM proves that the proposed framework can effectively translate RGB to CIR for NDVI calculation. Further, the artificially generated NDVI is able to segment and classify vegetation. As a result, the generated point cloud is less noisy, and the 3D model is enhanced.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    准确的脑肿瘤分割对于手术计划等任务至关重要,诊断,和分析,磁共振成像(MRI)由于其对脑组织的出色可视化而成为首选方式。然而,MR扫描中宽强度范围的体素值通常导致不同肿瘤组织的密度分布之间的显著重叠,导致对比度和分割精度降低。本文介绍了一种基于条件生成对抗网络(cGAN)的新颖框架,旨在增强逐体素和逐区域分割方法的肿瘤子区域的对比度。我们提出了两个模型:增强和分割GAN(ESGAN),它将分类器损失与对抗性损失相结合,以预测输入补丁的中心标签,和增强GAN(增强GAN),生成高对比度合成图像,减少类间重叠。然后将这些合成图像与相应的模式融合,以强调有意义的组织,同时抑制较弱的组织。我们还介绍了一种新颖的生成器,该生成器自适应地校准输入面片内的体素值,利用完全卷积网络。两种模型都采用多尺度马尔可夫网络作为GAN鉴别器,以捕获局部斑块统计数据并估计复杂上下文中MR图像的分布。公开的MR脑肿瘤数据集的实验结果表明,与当前的脑肿瘤分割技术相比,我们的模型具有竞争力。
    Accurate brain tumour segmentation is critical for tasks such as surgical planning, diagnosis, and analysis, with magnetic resonance imaging (MRI) being the preferred modality due to its excellent visualisation of brain tissues. However, the wide intensity range of voxel values in MR scans often results in significant overlap between the density distributions of different tumour tissues, leading to reduced contrast and segmentation accuracy. This paper introduces a novel framework based on conditional generative adversarial networks (cGANs) aimed at enhancing the contrast of tumour subregions for both voxel-wise and region-wise segmentation approaches. We present two models: Enhancement and Segmentation GAN (ESGAN), which combines classifier loss with adversarial loss to predict central labels of input patches, and Enhancement GAN (EnhGAN), which generates high-contrast synthetic images with reduced inter-class overlap. These synthetic images are then fused with corresponding modalities to emphasise meaningful tissues while suppressing weaker ones. We also introduce a novel generator that adaptively calibrates voxel values within input patches, leveraging fully convolutional networks. Both models employ a multi-scale Markovian network as a GAN discriminator to capture local patch statistics and estimate the distribution of MR images in complex contexts. Experimental results on publicly available MR brain tumour datasets demonstrate the competitive accuracy of our models compared to current brain tumour segmentation techniques.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    目的:前列腺癌是影响男性的最常见疾病之一。主要的诊断和预后参考工具是格里森评分系统。专家病理学家将Gleason等级分配给前列腺组织样本。由于这个过程非常耗时,一些人工智能应用程序被开发来自动化它。训练过程经常面临不足和不平衡的数据库,这影响了模型的可泛化性。因此,这项工作的目的是开发一种生成深度学习模型,能够合成任何选定的格里森等级的补丁,以对不平衡数据进行数据增强,并测试分类模型的改进。
    方法:这项工作中提出的方法包括条件渐进式生长GAN(ProGleason-GAN),该方法能够通过在合成样品中选择所需的GleasonGrade癌症模式来合成前列腺组织病理学组织斑块。通过嵌入层将有条件的格里森等级信息引入到模型中,因此不需要在Wasserstein损失函数中添加项。我们使用小批量标准偏差和像素归一化来提高训练过程的性能和稳定性。
    结果:使用FrechetInceptionDistance(FID)对合成样品进行了现实评估。我们获得了非癌模式的88.85的FID度量,处理后污点归一化后,GG3为81.86,GG4为49.32,GG5为108.69。此外,我们选择了一组专家病理学家对所提出的框架进行外部验证.最后,我们提出的框架的应用改进了SICAPv2数据集中的分类结果,证明其作为数据增强方法的有效性。
    结论:ProGleason-GAN方法结合染色标准化后处理可提供有关Frechet的初始距离的最新结果。这个模型可以合成非癌症模式的样本,GG3、GG4或GG5。在训练过程中包含关于格里森等级的条件信息允许模型选择合成样本中的癌变模式。所提出的框架可以用作数据增强方法。
    OBJECTIVE: Prostate cancer is one of the most common diseases affecting men. The main diagnostic and prognostic reference tool is the Gleason scoring system. An expert pathologist assigns a Gleason grade to a sample of prostate tissue. As this process is very time-consuming, some artificial intelligence applications were developed to automatize it. The training process is often confronted with insufficient and unbalanced databases which affect the generalisability of the models. Therefore, the aim of this work is to develop a generative deep learning model capable of synthesising patches of any selected Gleason grade to perform data augmentation on unbalanced data and test the improvement of classification models.
    METHODS: The methodology proposed in this work consists of a conditional Progressive Growing GAN (ProGleason-GAN) capable of synthesising prostate histopathological tissue patches by selecting the desired Gleason Grade cancer pattern in the synthetic sample. The conditional Gleason Grade information is introduced into the model through the embedding layers, so there is no need to add a term to the Wasserstein loss function. We used minibatch standard deviation and pixel normalisation to improve the performance and stability of the training process.
    RESULTS: The reality assessment of the synthetic samples was performed with the Frechet Inception Distance (FID). We obtained an FID metric of 88.85 for non-cancerous patterns, 81.86 for GG3, 49.32 for GG4 and 108.69 for GG5 after post-processing stain normalisation. In addition, a group of expert pathologists was selected to perform an external validation of the proposed framework. Finally, the application of our proposed framework improved the classification results in SICAPv2 dataset, proving its effectiveness as a data augmentation method.
    CONCLUSIONS: ProGleason-GAN approach combined with a stain normalisation post-processing provides state-of-the-art results regarding Frechet\'s Inception Distance. This model can synthesise samples of non-cancerous patterns, GG3, GG4 or GG5. The inclusion of conditional information about the Gleason grade during the training process allows the model to select the cancerous pattern in a synthetic sample. The proposed framework can be used as a data augmentation method.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Background: Ultra-Wide-Field (UWF) fundus imaging is an essential diagnostic tool for identifying ophthalmologic diseases, as it captures detailed retinal structures within a wider field of view (FOV). However, the presence of eyelashes along the edge of the eyelids can cast shadows and obscure the view of fundus imaging, which hinders reliable interpretation and subsequent screening of fundus diseases. Despite its limitations, there are currently no effective methods or datasets available for removing eyelash artifacts from UWF fundus images. This research aims to develop an effective approach for eyelash artifact removal and thus improve the visual quality of UWF fundus images for accurate analysis and diagnosis. Methods: To address this issue, we first constructed two UWF fundus datasets: the paired synthetic eyelashes (PSE) dataset and the unpaired real eyelashes (uPRE) dataset. Then we proposed a deep learning architecture called Joint Conditional Generative Adversarial Networks (JcGAN) to remove eyelash artifacts from UWF fundus images. JcGAN employs a shared generator with two discriminators for joint learning of both real and synthetic eyelash artifacts. Furthermore, we designed a background refinement module that refines background information and is trained with the generator in an end-to-end manner. Results: Experimental results on both PSE and uPRE datasets demonstrate the superiority of the proposed JcGAN over several state-of-the-art deep learning approaches. Compared with the best existing method, JcGAN improves PSNR and SSIM by 4.82% and 0.23%, respectively. In addition, we also verified that eyelash artifact removal via JcGAN could significantly improve vessel segmentation performance in UWF fundus images. Assessment via vessel segmentation illustrates that the sensitivity, Dice coefficient and area under curve (AUC) of ResU-Net have respectively increased by 3.64%, 1.54%, and 1.43% after eyelash artifact removal using JcGAN. Conclusion: The proposed JcGAN effectively removes eyelash artifacts in UWF images, resulting in improved visibility of retinal vessels. Our method can facilitate better processing and analysis of retinal vessels and has the potential to improve diagnostic outcomes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    生物医学数据采集,达到足够的参与者样本是困难的,时间和精力消耗的过程。另一方面,计算机辅助诊断(CAD)算法的成功率取决于样本和特征空间。在本文中,提出了基于条件生成对抗网络(CGAN)的增强特征生成来合成具有较高类可分性的大样本数据集。五个医疗数据集中的25%用于训练CGAN,并对具有任何样本量的合成数据集进行评估,并与原件进行比较。因此,可以在CGAN模型和较低样本收集的帮助下生成新的数据集。它帮助医生减少样本收集过程,并且它使用具有增强特征向量的生成的增强数据来提高CAD系统的准确率。合成的数据集使用最近邻进行分类,径向基函数支持向量机和人工神经网络来分析所提出CGAN模型的有效性。
    Biomedical data acquisition, and reaching sufficient samples of participants are difficult and time ans effort consuming processes. On the other hand, the success rates of computer aided diagnosis (CAD) algorithms are sample and feature space depended. In this paper, conditional generative adversarial network (CGAN) based enhanced feature generation is proposed to synthesize large sample datasets having higher class separability. Twenty five percent of five medical datasets are used to train CGAN, and the synthetic datasets with any sample size are evaluated and compared to originals. Thus, new datasets can be generated with the help of the CGAN model and lower sample collection. It helps physicians decreasing sample collection processes, and it increases accuracy rates of the CAD systems using generated enhanced data with enhanced feature vectors. The synthesized datasets are classified using nearest neighbor, radial basis function support vector machine and artificial neural network to analyze the effectiveness of the proposed CGAN model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,基于深度学习技术和超宽视野(UWF)图像的视网膜疾病自动检测在临床实践中发挥着重要作用。然而,由于小的病变和有限的数据样本,训练具有强大泛化能力的检测精确模型并不容易。在本文中,我们提出了一种病变注意条件生成对抗网络(LAC-GAN)来合成具有真实病变细节的视网膜图像,以改进疾病检测模型的训练.具体来说,生成器将血管掩码和类标签作为条件输入,并通过一系列残差块处理随机高斯噪声以生成合成图像。专注于病理信息,提出了一种基于随机森林(RF)方法的病变特征注意机制,构建其反向激活网络以激活病变特征。对于鉴别器,设计了一个权重共享的多鉴别器,通过仿射变换来提高模型的性能。在多中心UWF图像数据集上的实验结果表明,该方法可以生成具有合理细节的视网膜图像,这有助于提高疾病检测模型的性能。
    Automatic detection of retinal diseases based on deep learning technology and Ultra-widefield (UWF) images plays an important role in clinical practices in recent years. However, due to small lesions and limited data samples, it is not easy to train a detection-accurate model with strong generalization ability. In this paper, we propose a lesion attention conditional generative adversarial network (LAC-GAN) to synthesize retinal images with realistic lesion details to improve the training of the disease detection model. Specifically, the generator takes the vessel mask and class label as the conditional inputs, and processes the random Gaussian noise by a series of residual block to generate the synthetic images. To focus on pathological information, we propose a lesion feature attention mechanism based on random forest (RF) method, which constructs its reverse activation network to activate the lesion features. For discriminator, a weight-sharing multi-discriminator is designed to improve the performance of model by affine transformations. Experimental results on multi-center UWF image datasets demonstrate that the proposed method can generate retinal images with reasonable details, which helps to enhance the performance of the disease detection model.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Objective.我们提出了一种方法,用于在单光子发射计算机断层扫描成像设备的蒙特卡罗模拟过程中,使用条件生成对抗网络(condGAN)对退出体模的粒子分布族进行建模。方法。建议的condGAN在包含能量的低统计数据集上进行训练,时间,离开粒子的位置和方向。此外,它还包含由四个维度组成的条件向量:初始能量和体模内发射粒子的位置(总共12个维度)。与体模内吸收的gamma相关的信息也被添加到数据集中。在培训过程结束时,秃鹰的一个组成部分,发电机(G),已获得。主要结果。然后,可以用G生成具有特定能量和体模内发射位置的粒子,以代替体模内粒子的跟踪,与传统的蒙特卡罗模拟相比,可以减少计算时间。意义。对于给定的体模,condGAN生成器仅训练一次,但可以从各种活动源分布生成粒子。
    Objective.We propose a method to model families of distributions of particles exiting a phantom with a conditional generative adversarial network (condGAN) during Monte Carlo simulation of single photon emission computed tomography imaging devices.Approach.The proposed condGAN is trained on a low statistics dataset containing the energy, the time, the position and the direction of exiting particles. In addition, it also contains a vector of conditions composed of four dimensions: the initial energy and the position of emitted particles within the phantom (a total of 12 dimensions). The information related to the gammas absorbed within the phantom is also added in the dataset. At the end of the training process, one component of the condGAN, the generator (G), is obtained.Main results.Particles with specific energies and positions of emission within the phantom can then be generated withGto replace the tracking of particle within the phantom, allowing reduced computation time compared to conventional Monte Carlo simulation.Significance.The condGAN generator is trained only once for a given phantom but can generate particles from various activity source distributions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号