convolutional network

卷积网络
  • 文章类型: Journal Article
    大体肿瘤体积(GTV)的准确描绘对于放射治疗至关重要。深度学习驱动的GTV分割技术在快速准确地描绘GTV、为放射科医师制定放射计划提供依据。现有的基于深度学习的GTV二维和三维分割模型分别受到空间特征损失和各向异性,并且都受到肿瘤特征变异性的影响,模糊的边界,背景干扰。所有这些因素都严重影响分割性能。为了解决上述问题,本研究提出了一种基于2D-3D架构的层-体并行注意力(LVPA)-UNet模型,其中介绍了三种策略。首先,在LVPA-UNet中引入了2D和3D工作流程。它们并行工作,可以相互引导。通过它们可以提取2DMRI的每个切片的精细特征以及肿瘤的3D解剖结构和空间特征。其次,平行多分支深度条带卷积使模型适应切片和体积空间内不同形状和大小的肿瘤,并实现模糊边界的精细处理。最后,提出了一种层-通道注意力机制,根据切片和通道的不同肿瘤信息自适应调整其权重,然后突出切片和肿瘤通道。LVPA-UNet对来自三个中心的1010个鼻咽癌(NPC)MRI数据集的实验显示,DSC为0.7907,精度为0.7929,召回率为0.8025,HD95为1.8702mm,优于八种典型型号。与基线模型相比,它使DSC提高了2.14%,精度为2.96%,召回率为1.01%,而减少HD950.5434毫米。因此,在通过深度学习确保分割效率的同时,LVPA-UNet能够为放射治疗提供优越的GTV勾画结果,为精准医学提供技术支持。
    Accurate delineation of Gross Tumor Volume (GTV) is crucial for radiotherapy. Deep learning-driven GTV segmentation technologies excel in rapidly and accurately delineating GTV, providing a basis for radiologists in formulating radiation plans. The existing 2D and 3D segmentation models of GTV based on deep learning are limited by the loss of spatial features and anisotropy respectively, and are both affected by the variability of tumor characteristics, blurred boundaries, and background interference. All these factors seriously affect the segmentation performance. To address the above issues, a Layer-Volume Parallel Attention (LVPA)-UNet model based on 2D-3D architecture has been proposed in this study, in which three strategies are introduced. Firstly, 2D and 3D workflows are introduced in the LVPA-UNet. They work in parallel and can guide each other. Both the fine features of each slice of 2D MRI and the 3D anatomical structure and spatial features of the tumor can be extracted by them. Secondly, parallel multi-branch depth-wise strip convolutions adapt the model to tumors of varying shapes and sizes within slices and volumetric spaces, and achieve refined processing of blurred boundaries. Lastly, a Layer-Channel Attention mechanism is proposed to adaptively adjust the weights of slices and channels according to their different tumor information, and then to highlight slices and channels with tumor. The experiments by LVPA-UNet on 1010 nasopharyngeal carcinoma (NPC) MRI datasets from three centers show a DSC of 0.7907, precision of 0.7929, recall of 0.8025, and HD95 of 1.8702 mm, outperforming eight typical models. Compared to the baseline model, it improves DSC by 2.14 %, precision by 2.96 %, and recall by 1.01 %, while reducing HD95 by 0.5434 mm. Consequently, while ensuring the efficiency of segmentation through deep learning, LVPA-UNet is able to provide superior GTV delineation results for radiotherapy and offer technical support for precision medicine.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    创伤性脑损伤导致的颅内出血(ICH)是一个严重的问题,如果不及时诊断,通常会导致死亡或长期残疾。目前,医生主要使用计算机断层扫描(CT)扫描来检测和精确定位出血,通常由放射科医生解释。然而,这种诊断过程在很大程度上依赖于医疗专业人员的专业知识。为了解决潜在的错误,计算机辅助诊断系统已经开发。在这项研究中,我们提出了一种新方法,通过使用通过不同数据增强技术创建的多张图像,增强CT扫描中ICH病变的定位和分割.我们将残差连接集成到基于U-Net的分段网络中,以提高训练效率。我们的实验,根据82例颅脑外伤患者的CT扫描,验证我们方法的有效性,使用10倍交叉验证,ICH分割的IOU评分为0.807±0.03。
    Intracranial hemorrhage (ICH) resulting from traumatic brain injury is a serious issue, often leading to death or long-term disability if not promptly diagnosed. Currently, doctors primarily use Computerized Tomography (CT) scans to detect and precisely locate a hemorrhage, typically interpreted by radiologists. However, this diagnostic process heavily relies on the expertise of medical professionals. To address potential errors, computer-aided diagnosis systems have been developed. In this study, we propose a new method that enhances the localization and segmentation of ICH lesions in CT scans by using multiple images created through different data augmentation techniques. We integrate residual connections into a U-Net-based segmentation network to improve the training efficiency. Our experiments, based on 82 CT scans from traumatic brain injury patients, validate the effectiveness of our approach, achieving an IOU score of 0.807 ± 0.03 for ICH segmentation using 10-fold cross-validation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    大豆荚数是大豆产量的关键指标之一。然而,由于与计算豆荚相关的挑战,如拥挤和不均匀的豆荚分布,现有的pod计数模型将准确性优先于效率,这不满足轻量级和实时任务的要求。
    为了实现这一目标,我们设计了一个名为PodNet的深度卷积网络。它采用了轻量级编码器和有效的解码器,可以有效地解码浅层和深层信息,减轻非相邻层之间信息丢失和退化引起的间接相互作用。
    我们利用田间收获的大豆豆荚的高分辨率数据集来评估模型的泛化能力。通过人工计数和模型收益率估计之间的实验比较,我们证实了PodNet模型的有效性。实验结果表明,与地面实况相比,PodNet的大豆豆荚数量预测的R2为0.95,只有2.48M参数,比目前的SOTA模型YOLOPOD低一个数量级,FPS远高于YOLOPOD。
    与先进的计算机视觉方法相比,PodNet显着提高效率,几乎不牺牲精度。其轻量级架构和高FPS使其适合实时应用,为计算和定位密集物体提供了一种新的解决方案。
    UNASSIGNED: Soybean pod count is one of the crucial indicators of soybean yield. Nevertheless, due to the challenges associated with counting pods, such as crowded and uneven pod distribution, existing pod counting models prioritize accuracy over efficiency, which does not meet the requirements for lightweight and real-time tasks.
    UNASSIGNED: To address this goal, we have designed a deep convolutional network called PodNet. It employs a lightweight encoder and an efficient decoder that effectively decodes both shallow and deep information, alleviating the indirect interactions caused by information loss and degradation between non-adjacent levels.
    UNASSIGNED: We utilized a high-resolution dataset of soybean pods from field harvesting to evaluate the model\'s generalization ability. Through experimental comparisons between manual counting and model yield estimation, we confirmed the effectiveness of the PodNet model. The experimental results indicate that PodNet achieves an R2 of 0.95 for the prediction of soybean pod quantities compared to ground truth, with only 2.48M parameters, which is an order of magnitude lower than the current SOTA model YOLO POD, and the FPS is much higher than YOLO POD.
    UNASSIGNED: Compared to advanced computer vision methods, PodNet significantly enhances efficiency with almost no sacrifice in accuracy. Its lightweight architecture and high FPS make it suitable for real-time applications, providing a new solution for counting and locating dense objects.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:准确、及时地评估儿童的发育状况对于早期诊断和干预至关重要。由于缺乏训练有素的医疗保健提供者和不精确的父母报告,更准确和自动化的发展评估至关重要。在发展的各个领域,众所周知,幼儿的粗大运动发育可以预测随后的童年发展。
    目的:这项研究的目的是开发一种模型来评估粗大运动行为,并将结果整合以确定幼儿的整体粗大运动状态。这项研究还旨在确定在评估总体总体运动技能方面很重要的行为,并检测关键时刻和重要的身体部位,以评估每种行为。
    方法:我们使用了18-35个月幼儿的行为视频。为了评估电机总体发展,我们选择了4种行为(爬楼梯,走下楼梯,扔球,并站在1英尺上),已通过韩国婴儿和儿童发育筛查测试进行了验证。在儿童行为视频中,我们将每个孩子的位置估计为边界框,并在框内提取人类关键点。在第一阶段,使用基于图形卷积网络(GCN)的算法分别评估具有每种行为的提取的人类关键点的视频。在第一阶段模型中获得的每个标签的概率值用作第二阶段模型的输入,极端梯度提升(XGBoost)算法,预测总体运动状态。为了可解释性,我们使用梯度加权类激活映射(Grad-CAM)来识别运动过程中的重要时刻和相关身体部位。Shapley加性解释方法用于评估变量重要性,以确定对整体发展评估贡献最大的运动。
    结果:从147名儿童中收集了4种粗大运动技能的行为视频,共产生2395个视频。评估每种行为的阶段1GCN模型的接受者工作特征曲线下面积(AUROC)为0.79至0.90。关键点映射Grad-CAM可视化识别了每个行为中的重要时刻以及重要身体部位的差异。评估总体粗大运动状态的阶段2XGBoost模型的AUROC为0.90。在这四种行为中,“下楼梯”对整体发展评估的贡献最大。
    结论:使用18-35个月幼儿的运动视频,我们开发了客观和自动化的模型来评估每个行为和评估每个孩子的整体粗大运动表现。我们确定了评估总体电机性能的重要行为,并开发了在评估总体电机性能时识别重要力矩和身体部位的方法。
    BACKGROUND: Accurate and timely assessment of children\'s developmental status is crucial for early diagnosis and intervention. More accurate and automated developmental assessments are essential due to the lack of trained health care providers and imprecise parental reporting. In various areas of development, gross motor development in toddlers is known to be predictive of subsequent childhood developments.
    OBJECTIVE: The purpose of this study was to develop a model to assess gross motor behavior and integrate the results to determine the overall gross motor status of toddlers. This study also aimed to identify behaviors that are important in the assessment of overall gross motor skills and detect critical moments and important body parts for the assessment of each behavior.
    METHODS: We used behavioral videos of toddlers aged 18-35 months. To assess gross motor development, we selected 4 behaviors (climb up the stairs, go down the stairs, throw the ball, and stand on 1 foot) that have been validated with the Korean Developmental Screening Test for Infants and Children. In the child behavior videos, we estimated each child\'s position as a bounding box and extracted human keypoints within the box. In the first stage, the videos with the extracted human keypoints of each behavior were evaluated separately using a graph convolutional networks (GCN)-based algorithm. The probability values obtained for each label in the first-stage model were used as input for the second-stage model, the extreme gradient boosting (XGBoost) algorithm, to predict the overall gross motor status. For interpretability, we used gradient-weighted class activation mapping (Grad-CAM) to identify important moments and relevant body parts during the movements. The Shapley additive explanations method was used for the assessment of variable importance, to determine the movements that contributed the most to the overall developmental assessment.
    RESULTS: Behavioral videos of 4 gross motor skills were collected from 147 children, resulting in a total of 2395 videos. The stage-1 GCN model to evaluate each behavior had an area under the receiver operating characteristic curve (AUROC) of 0.79 to 0.90. Keypoint-mapping Grad-CAM visualization identified important moments in each behavior and differences in important body parts. The stage-2 XGBoost model to assess the overall gross motor status had an AUROC of 0.90. Among the 4 behaviors, \"go down the stairs\" contributed the most to the overall developmental assessment.
    CONCLUSIONS: Using movement videos of toddlers aged 18-35 months, we developed objective and automated models to evaluate each behavior and assess each child\'s overall gross motor performance. We identified the important behaviors for assessing gross motor performance and developed methods to recognize important moments and body parts while evaluating gross motor performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:在MRI图像上准确分割中风病变对于神经科医生在中风后护理计划中非常重要。分割有助于临床医生更好地诊断和评估任何治疗风险。然而,脑部病变的手动分割依赖于神经科医生的经验,也是一个非常繁琐和耗时的过程。所以,在这项研究中,我们提出了一种新的深度卷积神经网络(CNN-Res),该网络可自动从多模态MRI中分割缺血性卒中病变.
    方法:CNN-Res使用U形结构,所以网络有加密和解密路径。残差单元嵌入在编码器路径中。在这个模型中,为了减少梯度下降,使用了剩余单位,并在图像中提取更复杂的信息,应用多模态MRI数据。在加密和解密子网之间的链接中,使用了瓶颈策略,与同类研究相比,减少了参数数量和训练时间。
    结果:CNN-Res在两个不同的数据集上进行了评估。首先,它是在大不里士医学院神经科学中心收集的数据集上检查的,其中平均骰子系数等于85.43%。然后,为了将模型的效率和性能与其他类似作品进行比较,CNN-Res在流行的SPES2015比赛数据集上进行了评估,其中平均骰子系数为79.23%。
    结论:这项研究提出了一种使用称为CNN-Res的深度卷积神经网络对MRI医学图像进行分割的新的准确方法。它直接从原始输入像素预测分段图。
    Accurate segmentation of stroke lesions on MRI images is very important for neurologists in the planning of post-stroke care. Segmentation helps clinicians to better diagnose and evaluation of any treatment risks. However, manual segmentation of brain lesions relies on the experience of neurologists and is also a very tedious and time-consuming process. So, in this study, we proposed a novel deep convolutional neural network (CNN-Res) that automatically performs the segmentation of ischemic stroke lesions from multimodal MRIs.
    CNN-Res used a U-shaped structure, so the network has encryption and decryption paths. The residual units are embedded in the encoder path. In this model, to reduce gradient descent, the residual units were used, and to extract more complex information in images, multimodal MRI data were applied. In the link between the encryption and decryption subnets, the bottleneck strategy was used, which reduced the number of parameters and training time compared to similar research.
    CNN-Res was evaluated on two distinct datasets. First, it was examined on a dataset collected from the Neuroscience Center of Tabriz University of Medical Sciences, where the average Dice coefficient was equal to 85.43%. Then, to compare the efficiency and performance of the model with other similar works, CNN-Res was evaluated on the popular SPES 2015 competition dataset where the average Dice coefficient was 79.23%.
    This study presented a new and accurate method for the segmentation of MRI medical images using a deep convolutional neural network called CNN-Res, which directly predicts segment maps from raw input pixels.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    我们使用机器学习来评估机器人辅助肾部分切除术(RAPN)的肿瘤切除和肾图步骤过程中视频的手术技巧。这扩展了使用合成组织的先前工作,以包括实际手术。我们研究了级联神经网络,用于从达芬奇系统记录的RPN视频中预测手术熟练程度得分(OSATS和GEARS)。语义分割任务生成掩模并跟踪各种手术器械。通过语义分割发现的仪器的运动由评分网络处理,该评分网络对每个子类别的齿轮和OSATS评分进行回归(预测)。总的来说,该模型对许多子类别表现良好,如力灵敏度和齿轮和OSATS评分的仪器知识,但可能遭受假阳性和假阴性,这在人类评估者中是不会想到的。这主要归因于有限的训练数据可变性和稀疏性。
    We use machine learning to evaluate surgical skill from videos during the tumor resection and renography steps of a robotic assisted partial nephrectomy (RAPN). This expands previous work using synthetic tissue to include actual surgeries. We investigate cascaded neural networks for predicting surgical proficiency scores (OSATS and GEARS) from RAPN videos recorded from the DaVinci system. The semantic segmentation task generates a mask and tracks the various surgical instruments. The movements from the instruments found via semantic segmentation are processed by a scoring network that regresses (predicts) GEARS and OSATS scoring for each subcategory. Overall, the model performs well for many subcategories such as force sensitivity and knowledge of instruments of GEARS and OSATS scoring, but can suffer from false positives and negatives that would not be expected of human raters. This is mainly attributed to limited training data variability and sparsity.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在癫痫发作发生之前预测癫痫发作可以帮助通过药物预防癫痫发作。在这项研究中,首先,从5-s分割的脑电信号中总共提取了22个特征。第二,张量被开发为不同深度迁移学习模型的输入,以找到预测癫痫发作的最佳模型。还通过选择10、20、30和40分钟的四个不同间隔来研究发作前状态持续时间的影响。然后,通过将3种ImageNet卷积网络与3种分类器相结合,创建了9种模型,并对患者依赖的癫痫发作进行了预测.具有完全连接(FC)分类器的Xception卷积网络在40分钟的发病前状态下从欧洲数据库中获得了98.47%的平均灵敏度和0.031h-1的错误预测率(FPR)。这项研究的最有希望的结果是癫痫发作的独立于患者的预测;具有FC分类器的MobileNet-V2模型用一名患者的数据进行了训练,并在其他六名患者上进行了测试。对于40分钟的预发作方案,灵敏度为98.39%,FPR为0.029h-1。
    Predicting seizures before they happen can help prevent them through medication. In this research, first, a total of 22 features were extracted from 5-s segmented EEG signals. Second, tensors were developed as inputs for different deep transfer learning models to find the best model for predicting epileptic seizures. The effect of Pre-ictal state duration was also investigated by selecting four different intervals of 10, 20, 30, and 40 min. Then, nine models were created by combining three ImageNet convolutional networks with three classifiers and were examined for predicting seizures patient-dependently. The Xception convolutional network with a Fully Connected (FC) classifier achieved an average sensitivity of 98.47% and a False Prediction Rate (FPR) of 0.031 h-1 in a 40-min Pre-ictal state for ten patients from the European database. The most promising result of this study was the patient-independent prediction of epileptic seizures; the MobileNet-V2 model with an FC classifier was trained with one patient\'s data and tested on six other patients, achieving a sensitivity rate of 98.39% and an FPR of 0.029 h-1 for a 40-min Pre-ictal scheme.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    帕金森病(PD)是最常见的帕金森病,这是一组具有PD样运动障碍的神经系统疾病。该疾病影响全球600多万人,其特征是运动和非运动症状。受影响的人在控制动作方面有困难,这可能会影响简单的日常生活任务,比如在电脑上打字。我们提出了一种改进的SqueezeNet卷积神经网络(CNN)的应用,用于基于受试者的关键类型模式检测PD。首先,使用数据标准化和合成少数过采样技术(SMOTE)对数据进行预处理,然后应用连续小波变换来生成用于训练和测试改进的SqueezeNet模型的频谱图。修改后的SqueezeNet模型达到了90%的准确率,与其他方法相比,有了显著的改进。
    Parkinson\'s disease (PD) is the most common form of Parkinsonism, which is a group of neurological disorders with PD-like motor impairments. The disease affects over 6 million people worldwide and is characterized by motor and non-motor symptoms. The affected person has trouble in controlling movements, which may affect simple daily-life tasks, such as typing on a computer. We propose the application of a modified SqueezeNet convolutional neural network (CNN) for detecting PD based on the subject\'s key-typing patterns. First, the data are pre-processed using data standardization and the Synthetic Minority Oversampling Technique (SMOTE), and then a Continuous Wavelet Transformation is applied to generate spectrograms used for training and testing a modified SqueezeNet model. The modified SqueezeNet model achieved an accuracy of 90%, representing a noticeable improvement in comparison to other approaches.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    持续的COVID-19大流行给全球供应链(SC)造成了前所未有的困境。基本和救生产品的装运,从制药,农业,和医疗保健,制造业,受到重大影响或延误,使全球SC脆弱。更好地了解装运风险可以大大减少这种紧张情绪。此后,本文提出了一些深度学习(DL)方法,通过预测“货物是否可以从一个来源出口到另一个来源”来减轻装运风险,尽管COVID-19大流行施加了限制。拟议的DL方法有四个主要阶段:数据捕获,去噪或预处理,特征提取,和分类。特征提取阶段取决于DL模型的两个主要变体。第一个变体涉及三个递归神经网络(RNN)结构(即,长短期记忆(LSTM),双向长短期记忆(BiLSTM),和门控经常性单位(GRU)),第二个变体是时间卷积网络(TCN)。就分类阶段而言,六个不同的分类器被用来测试整个方法。这些分类器是SoftMax,随机树(RT),随机森林(RF),k-最近邻(KNN),人工神经网络(ANN),和支持向量机(SVM)。基于在线数据集(作为案例研究)评估了所提出的DL模型的性能。数值结果表明,提出的模型之一(即,TCN)在预测COVID-19限制下运往特定目的地的风险方面约100%准确。毫无疑问,这项工作的后果将有助于决策者主动预测供应链风险,以提高SCs的弹性.
    The ongoing COVID-19 pandemic has created an unprecedented predicament for global supply chains (SCs). Shipments of essential and life-saving products, ranging from pharmaceuticals, agriculture, and healthcare, to manufacturing, have been significantly impacted or delayed, making the global SCs vulnerable. A better understanding of the shipment risks can substantially reduce that nervousness. Thenceforth, this paper proposes a few Deep Learning (DL) approaches to mitigate shipment risks by predicting \"if a shipment can be exported from one source to another\", despite the restrictions imposed by the COVID-19 pandemic. The proposed DL methodologies have four main stages: data capturing, de-noising or pre-processing, feature extraction, and classification. The feature extraction stage depends on two main variants of DL models. The first variant involves three recurrent neural networks (RNN) structures (i.e., long short-term memory (LSTM), Bidirectional long short-term memory (BiLSTM), and gated recurrent unit (GRU)), and the second variant is the temporal convolutional network (TCN). In terms of the classification stage, six different classifiers are applied to test the entire methodology. These classifiers are SoftMax, random trees (RT), random forest (RF), k-nearest neighbor (KNN), artificial neural network (ANN), and support vector machine (SVM). The performance of the proposed DL models is evaluated based on an online dataset (taken as a case study). The numerical results show that one of the proposed models (i.e., TCN) is about 100% accurate in predicting the risk of shipment to a particular destination under COVID-19 restrictions. Unarguably, the aftermath of this work will help the decision-makers to predict supply chain risks proactively to increase the resiliency of the SCs.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:图像分割中反复出现的问题是缺少标记数据。在2019年冠状病毒病(COVID-19)患者的肺部计算机断层扫描(CT)分割中,这个问题尤其严重。原因很简单:这种疾病的流行时间不足以产生大量的标签。半监督学习提供了一种从未标记的数据中学习的方法,并且近年来取得了巨大的进步。然而,由于标签空间的复杂性,这些进步不能应用于图像分割。话虽如此,正是这种复杂性使得获得像素级标签极其昂贵,使半监督学习更具吸引力。本研究旨在通过提出一种新颖的模型来弥合这一差距,该模型利用深度卷积网络的图像分割能力和生成模型的半监督学习能力,用于COVID-19患者的胸部CT图像。
    结果:我们提出了一种称为共享变分自动编码器(SVAE)的新型生成模型。SVAE利用潜在变量的五层深层结构以及它们之间的深度卷积映射,产生了一个非常适用于肺部CT图像的生成模型。然后,我们在SVAE的最后一层添加了一个新的组件,该组件迫使模型使用必须匹配地面真实分割的分割来重建输入图像。我们将这个最终模型命名为StitchNet。
    结论:我们在COVID-19患者的高质量CT图像数据集上比较了StitchNet和其他图像分割模型。我们证明了我们的模型与其他分割模型具有相当的性能。我们还探讨了我们提出的算法的潜在局限性和优势,并为这个具有挑战性的问题提出了一些潜在的未来研究方向。
    BACKGROUND: A recurring problem in image segmentation is a lack of labelled data. This problem is especially acute in the segmentation of lung computed tomography (CT) of patients with Coronavirus Disease 2019 (COVID-19). The reason for this is simple: the disease has not been prevalent long enough to generate a great number of labels. Semi-supervised learning promises a way to learn from data that is unlabelled and has seen tremendous advancements in recent years. However, due to the complexity of its label space, those advancements cannot be applied to image segmentation. That being said, it is this same complexity that makes it extremely expensive to obtain pixel-level labels, making semi-supervised learning all the more appealing. This study seeks to bridge this gap by proposing a novel model that utilizes the image segmentation abilities of deep convolution networks and the semi-supervised learning abilities of generative models for chest CT images of patients with the COVID-19.
    RESULTS: We propose a novel generative model called the shared variational autoencoder (SVAE). The SVAE utilizes a five-layer deep hierarchy of latent variables and deep convolutional mappings between them, resulting in a generative model that is well suited for lung CT images. Then, we add a novel component to the final layer of the SVAE which forces the model to reconstruct the input image using a segmentation that must match the ground truth segmentation whenever it is present. We name this final model StitchNet.
    CONCLUSIONS: We compare StitchNet to other image segmentation models on a high-quality dataset of CT images from COVID-19 patients. We show that our model has comparable performance to the other segmentation models. We also explore the potential limitations and advantages in our proposed algorithm and propose some potential future research directions for this challenging issue.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号