Post-filtering

  • 文章类型: Journal Article
    背景:在正电子发射断层扫描(PET)中,残余图像噪声是大量的,是限制病变检测的因素之一,量化,和整体图像质量。因此,改善降噪仍然相当感兴趣。对于呼吸门控PET研究尤其如此。PET成像中唯一广泛使用的降噪方法是应用低通滤波器,通常是高斯,然而,这会导致空间分辨率的损失和部分体积效应的增加,从而影响小病变的可检测性和定量数据评估。双边滤波器(BF)-一种局部自适应图像滤波器-允许减少图像噪声,同时保留定义明确的对象边缘,但手动优化给定PET扫描的滤波器参数可能是繁琐且耗时的。妨碍其临床使用。在这项工作中,我们研究了一种合适的基于深度学习的方法在多大程度上可以通过训练一个合适的网络来解决这个问题,该网络的目标是再现手动调整的特定案例双边过滤的结果。
    方法:总之,使用三种不同的示踪剂进行69次呼吸门控临床PET/CT扫描([18F]FDG,[18F]L-DOPA,[68Ga]DOTATATE)用于本研究。在数据处理之前,门控数据集被拆分,导致总共552个单门图像卷。对于这些图像体积中的每一个,描绘了四个3DROI:一个ROI用于图像噪声评估,三个ROI用于不同目标/背景对比水平下的局灶性摄取(例如肿瘤病变)测量。使用自动程序对每个数据集的二维BF参数空间进行强力搜索,以识别“最佳”滤波器参数,以生成用户批准的由原始和最佳BF滤波图像对组成的地面实况输入数据。为了再现最佳BF滤波,我们采用了一种结合残差学习原理的改进的3DU-NetCNN。使用5倍交叉验证方案进行网络训练和评估。通过计算CNN之间的绝对和分数差异来评估滤波对病变SUV量化和图像噪声水平的影响,手动BF,或先前定义的ROI中的原始(STD)数据集。
    结果:用于过滤器参数确定的自动化程序为大多数数据集选择了足够的过滤器参数,只有19个患者数据集需要手动调整。对聚焦吸收ROI的评估表明,CNN以及基于BF的滤波基本上保持了未滤波图像的聚焦SUV最大值,δSUVmaxCNN的平均值较低,STD=(-3.9±5.2)%,δSUVmaxBF,STD=(-4.4±5.3)%。关于CNN与BF的相对性能,在绝大多数情况下,这两种方法都导致了非常相似的SUV最大值,总平均差为δSUVmaxCNN,BF=(0.5±4.8)%。对噪声特性的评估表明,CNN滤波可以很好地再现具有δNoiseCNN的BF的噪声水平和特性,BF=(5.6±10.5)%。在CNN和BF之间没有观察到显著的示踪剂依赖性差异。
    结论:我们的结果表明,基于神经网络的去噪可以完全自动化的方式再现逐例优化BF的结果。除了罕见的情况下,它导致图像的噪声水平几乎相同的质量,边缘保护,和信号恢复。我们相信这样的网络可能证明在改进呼吸门控PET研究的运动校正的背景下特别有用,但也可以帮助在临床PET中建立BF等效边缘保留CNN滤波,因为它避免了耗时的手动BF参数调整。
    BACKGROUND: Residual image noise is substantial in positron emission tomography (PET) and one of the factors limiting lesion detection, quantification, and overall image quality. Thus, improving noise reduction remains of considerable interest. This is especially true for respiratory-gated PET investigations. The only broadly used approach for noise reduction in PET imaging has been the application of low-pass filters, usually Gaussians, which however leads to loss of spatial resolution and increased partial volume effects affecting detectability of small lesions and quantitative data evaluation. The bilateral filter (BF) - a locally adaptive image filter - allows to reduce image noise while preserving well defined object edges but manual optimization of the filter parameters for a given PET scan can be tedious and time-consuming, hampering its clinical use. In this work we have investigated to what extent a suitable deep learning based approach can resolve this issue by training a suitable network with the target of reproducing the results of manually adjusted case-specific bilateral filtering.
    METHODS: Altogether, 69 respiratory-gated clinical PET/CT scans with three different tracers ( [ 18 F ] FDG, [ 18 F ] L-DOPA, [ 68 Ga ] DOTATATE) were used for the present investigation. Prior to data processing, the gated data sets were split, resulting in a total of 552 single-gate image volumes. For each of these image volumes, four 3D ROIs were delineated: one ROI for image noise assessment and three ROIs for focal uptake (e.g. tumor lesions) measurements at different target/background contrast levels. An automated procedure was used to perform a brute force search of the two-dimensional BF parameter space for each data set to identify the \"optimal\" filter parameters to generate user-approved ground truth input data consisting of pairs of original and optimally BF filtered images. For reproducing the optimal BF filtering, we employed a modified 3D U-Net CNN incorporating residual learning principle. The network training and evaluation was performed using a 5-fold cross-validation scheme. The influence of filtering on lesion SUV quantification and image noise level was assessed by calculating absolute and fractional differences between the CNN, manual BF, or original (STD) data sets in the previously defined ROIs.
    RESULTS: The automated procedure used for filter parameter determination chose adequate filter parameters for the majority of the data sets with only 19 patient data sets requiring manual tuning. Evaluation of the focal uptake ROIs revealed that CNN as well as BF based filtering essentially maintain the focal SUV max values of the unfiltered images with a low mean ± SD difference of δ SUV max CNN , STD = (-3.9 ± 5.2)% and δ SUV max BF , STD = (-4.4 ± 5.3)%. Regarding relative performance of CNN versus BF, both methods lead to very similar SUV max values in the vast majority of cases with an overall average difference of δ SUV max CNN , BF = (0.5 ± 4.8)%. Evaluation of the noise properties showed that CNN filtering mostly satisfactorily reproduces the noise level and characteristics of BF with δ Noise CNN , BF = (5.6 ± 10.5)%. No significant tracer dependent differences between CNN and BF were observed.
    CONCLUSIONS: Our results show that a neural network based denoising can reproduce the results of a case by case optimized BF in a fully automated way. Apart from rare cases it led to images of practically identical quality regarding noise level, edge preservation, and signal recovery. We believe such a network might proof especially useful in the context of improved motion correction of respiratory-gated PET studies but could also help to establish BF-equivalent edge-preserving CNN filtering in clinical PET since it obviates time consuming manual BF parameter tuning.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在全球范围内视频应用不断增加的当前场景中,对高效视频编码技术的需求比以往任何时候都更加重要,和物联网(IoT)设备正在变得普遍。在这种情况下,有必要仔细审查最近完成的MPEG-5基本视频编码(EVC)标准,因为EVC基线配置文件是定制的,以满足在低复杂度方面处理物联网视频数据所需的特定要求。然而,EVC基线配置文件具有显著的缺点。由于它是仅由20多年开发的简单工具组成的编解码器,它往往代表许多编码工件。特别是,块边界处的块效应的存在被认为是必须解决的关键问题。为了解决这个问题,本文提出了一种基于块分割信息的卷积神经网络(CNN)的后置滤波器。所提出的方法在实验结果中客观地显示了大约0.57dB的全帧(AI)和0.37dB的低延迟(LD)的改进,在每个配置与前置后滤波视频相比,增强的PSNR导致Luma和Chroma组件中AI的整体比特率降低11.62%,LD的整体比特率降低10.91%,分别。由于PSNR的巨大改进,该方法在主观上显著提高了视觉质量,特别是在编码块边界处的块伪影中。
    The need for efficient video coding technology is more important than ever in the current scenario where video applications are increasing worldwide, and Internet of Things (IoT) devices are becoming widespread. In this context, it is necessary to carefully review the recently completed MPEG-5 Essential Video Coding (EVC) standard because the EVC Baseline profile is customized to meet the specific requirements needed to process IoT video data in terms of low complexity. Nevertheless, the EVC Baseline profile has a notable disadvantage. Since it is a codec composed only of simple tools developed over 20 years, it tends to represent numerous coding artifacts. In particular, the presence of blocking artifacts at the block boundary is regarded as a critical issue that must be addressed. To address this, this paper proposes a post-filter using a block partitioning information-based Convolutional Neural Network (CNN). The proposed method in the experimental results objectively shows an approximately 0.57 dB for All-Intra (AI) and 0.37 dB for Low-Delay (LD) improvements in each configuration by the proposed method when compared to the pre-post-filter video, and the enhanced PSNR results in an overall bitrate reduction of 11.62% for AI and 10.91% for LD in the Luma and Chroma components, respectively. Due to the huge improvement in the PSNR, the proposed method significantly improved the visual quality subjectively, particularly in blocking artifacts at the coding block boundary.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文提出了一种基于卷积神经网络(CNN)的新型后滤波方法,用于增强RGB/灰度图像和视频序列的质量。使用常见的图像编解码器对有损图像进行编码,例如JPEG和JPEG2000。使用先前和正在进行的视频编码标准对视频序列进行编码,高效视频编码(HEVC)和多功能视频编码(VVC),分别。提出了一种新颖的深度神经网络架构来估计完整的精细细化细节,half-,和季度补丁分辨率。所提出的体系结构是使用基于以下概念设计的一组高效处理块构建的:(i)用于细化特征图的多头注意力机制,(ii)降低网络复杂性的权重共享概念,和(iii)用于多分辨率特征融合的层结构的新颖块设计。与常见的图像编解码器和视频编码标准相比,所提出的方法提供了实质性的性能改进。对高分辨率图像和标准视频序列的实验结果表明,所提出的后滤波方法为RGB图像提供了比JPEG平均31.44%和比HEVC(x265)平均54.61%的BD率节省。对于灰度图像,Y-BD速率比JPEG节省26.21%,比VVC(VTM)节省15.28%,视频序列比HEVC高15.47%,比VVC高14.66%。
    The paper proposes a novel post-filtering method based on convolutional neural networks (CNNs) for quality enhancement of RGB/grayscale images and video sequences. The lossy images are encoded using common image codecs, such as JPEG and JPEG2000. The video sequences are encoded using previous and ongoing video coding standards, high-efficiency video coding (HEVC) and versatile video coding (VVC), respectively. A novel deep neural network architecture is proposed to estimate fine refinement details for full-, half-, and quarter-patch resolutions. The proposed architecture is built using a set of efficient processing blocks designed based on the following concepts: (i) the multi-head attention mechanism for refining the feature maps, (ii) the weight sharing concept for reducing the network complexity, and (iii) novel block designs of layer structures for multiresolution feature fusion. The proposed method provides substantial performance improvements compared with both common image codecs and video coding standards. Experimental results on high-resolution images and standard video sequences show that the proposed post-filtering method provides average BD-rate savings of 31.44% over JPEG and 54.61% over HEVC (x265) for RGB images, Y-BD-rate savings of 26.21% over JPEG and 15.28% over VVC (VTM) for grayscale images, and 15.47% over HEVC and 14.66% over VVC for video sequences.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    为了简化麦克风阵列的复杂度和降低成本,本文提出了一种基于双麦克风的声音定位和语音增强算法。基于双麦克风接收的信号的时间延迟估计,本文将能量差估计和可控波束响应功率相结合,实现了声源的三维坐标计算和双麦克风声音定位。通过对声源方位角和语音信号独立量的分析,实现了声源扬声器信号的分离。在此基础上,后维纳滤波用于放大和抑制扬声器的语音信号,这可以帮助实现语音增强。实验结果表明,本文提出的双麦克风声音定位算法能够准确识别声音位置,语音增强算法比原算法具有更强的鲁棒性和适应性。
    In order to simplify the complexity and reduce the cost of the microphone array, this paper proposes a dual-microphone based sound localization and speech enhancement algorithm. Based on the time delay estimation of the signal received by the dual microphones, this paper combines energy difference estimation and controllable beam response power to realize the 3D coordinate calculation of the acoustic source and dual-microphone sound localization. Based on the azimuth angle of the acoustic source and the analysis of the independent quantity of the speech signal, the separation of the speaker signal of the acoustic source is realized. On this basis, post-wiener filtering is used to amplify and suppress the voice signal of the speaker, which can help to achieve speech enhancement. Experimental results show that the dual-microphone sound localization algorithm proposed in this paper can accurately identify the sound location, and the speech enhancement algorithm is more robust and adaptable than the original algorithm.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Several researchers have contemplated deep learning-based post-filters to increase the quality of statistical parametric speech synthesis, which perform a mapping of the synthetic speech to the natural speech, considering the different parameters separately and trying to reduce the gap between them. The Long Short-term Memory (LSTM) Neural Networks have been applied successfully in this purpose, but there are still many aspects to improve in the results and in the process itself. In this paper, we introduce a new pre-training approach for the LSTM, with the objective of enhancing the quality of the synthesized speech, particularly in the spectrum, in a more efficient manner. Our approach begins with an auto-associative training of one LSTM network, which is used as an initialization for the post-filters. We show the advantages of this initialization for the enhancing of the Mel-Frequency Cepstral parameters of synthetic speech. Results show that the initialization succeeds in achieving better results in enhancing the statistical parametric speech spectrum in most cases when compared to the common random initialization approach of the networks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    BACKGROUND: Positron emission tomography (PET) imaging has a wide applicability in oncology, cardiology and neurology. However, a major drawback when imaging very active regions such as the bladder is the spill-in effect, leading to inaccurate quantification and obscured visualisation of nearby lesions. Therefore, this study aims at investigating and correcting for the spill-in effect from high-activity regions to the surroundings as a function of activity in the hot region, lesion size and location, system resolution and application of post-filtering using a recently proposed background correction technique. This study involves analytical simulations for the digital XCAT2 phantom and validation acquiring NEMA phantom and patient data with the GE Signa PET/MR scanner. Reconstructions were done using the ordered subset expectation maximisation (OSEM) algorithm. Dedicated point-spread function (OSEM+PSF) and a recently proposed background correction (OSEM+PSF+BC) were incorporated into the reconstruction for spill-in correction. The standardised uptake values (SUV) were compared for all reconstruction algorithms.
    RESULTS: The simulation study revealed that lesions within 15-20 mm from the hot region were predominantly affected by the spill-in effect, leading to an increased bias and impaired lesion visualisation within the region. For OSEM, lesion SUVmax converged to the true value at low bladder activity, but as activity increased, there was an overestimation as much as 19% for proximal lesions (distance around 15-20 mm from the bladder edge) and 2-4% for distant lesions (distance larger than 20 mm from the bladder edge). As bladder SUV increases, the % SUV change for proximal lesions is about 31% and 6% for SUVmax and SUVmean, respectively, showing that the spill-in effect is more evident for the SUVmax than the SUVmean. Also, the application of post-filtering resulted in up to 65% increment in the spill-in effect around the bladder edges. For proximal lesions, PSF has no major improvement over OSEM because of the spill-in effect, coupled with the blurring effect by post-filtering. Within two voxels around the bladder, the spill-in effect in OSEM is 42% (32%), while for OSEM+PSF, it is 31% (19%), with (and without) post-filtering, respectively. But with OSEM+PSF+BC, the spill-in contribution from the bladder was relatively low (below 5%, either with or without post-filtering). These results were further validated using the NEMA phantom and patient data for which OSEM+PSF+BC showed about 70-80% spill-in reduction around the bladder edges and increased contrast-to-noise ratio up to 36% compared to OSEM and OSEM+PSF reconstructions without post-filtering.
    CONCLUSIONS: The spill-in effect is dependent on the activity in the hot region, lesion size and location, as well as post-filtering; and this is more evident in SUVmax than SUVmean. However, the recently proposed background correction method facilitates stability in quantification and enhances the contrast in lesions with low uptake.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    准确预测蛋白质-蛋白质相互作用位点(PPI)是当前的热门话题,因为它已被证明对理解疾病机制和设计药物非常有用。基于机器学习的计算方法已被广泛使用,并被证明对PPI预测有用。然而,直接应用传统的机器学习算法,通常假设不同类别的样本是平衡的,由于PPI预测问题中存在严重的类不平衡,因此通常会导致性能不佳。在这项研究中,我们提出了一种新的方法,通过使用数据清理程序减轻类不平衡的严重程度,并通过后过滤程序减少预测的误报来提高PPI预测性能:首先,应用基于机器学习的数据清理过程来删除这些边缘目标,这可能会对训练具有清晰分类边界的模型产生负面影响,从大多数样本中提取,以减轻原始训练数据集中类别不平衡的严重程度;然后,在清理后的数据集上训练预测模型;最后,有效的后过滤程序进一步用于减少潜在的假阳性预测。在基准数据集上进行严格的交叉验证和独立验证测试,证明了所提出方法的有效性。与现有的最先进的基于序列的PPI预测因子相比,它表现出高度的竞争力,并且应该补充现有的PPI预测方法。
    Accurately predicting protein-protein interaction sites (PPIs) is currently a hot topic because it has been demonstrated to be very useful for understanding disease mechanisms and designing drugs. Machine-learning-based computational approaches have been broadly utilized and demonstrated to be useful for PPI prediction. However, directly applying traditional machine learning algorithms, which often assume that samples in different classes are balanced, often leads to poor performance because of the severe class imbalance that exists in the PPI prediction problem. In this study, we propose a novel method for improving PPI prediction performance by relieving the severity of class imbalance using a data-cleaning procedure and reducing predicted false positives with a post-filtering procedure: First, a machine-learning-based data-cleaning procedure is applied to remove those marginal targets, which may potentially have a negative effect on training a model with a clear classification boundary, from the majority samples to relieve the severity of class imbalance in the original training dataset; then, a prediction model is trained on the cleaned dataset; finally, an effective post-filtering procedure is further used to reduce potential false positive predictions. Stringent cross-validation and independent validation tests on benchmark datasets demonstrated the efficacy of the proposed method, which exhibits highly competitive performance compared with existing state-of-the-art sequence-based PPIs predictors and should supplement existing PPI prediction methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号