super-resolution

超分辨率
  • 文章类型: Journal Article
    磁共振成像(MRI)通常用于研究婴儿的大脑发育。然而,由于图像采集时间长,受试者依从性有限,高质量的婴儿MRI可能具有挑战性。在不给图像采集带来额外负担的情况下,图像超分辨率(SR)可用于增强采集后的图像质量。大多数SR技术在多个对齐的低分辨率(LR)和高分辨率(HR)图像对上进行监督和训练,这在实践中通常是不可用的。与监督方法不同,深度图像先验(DIP)可以用于无监督的单图像SR,仅利用输入LR图像进行从头优化以产生HR图像。然而,确定何时在DIP训练早期停止是不平凡的,并且提出了完全自动化SR过程的挑战。为了解决这个问题,我们将SR图像的低频k空间限制为与LR图像相似。我们通过设计一个双模态框架来进一步提高性能,该框架利用T1加权和T2加权图像之间的共享解剖信息。我们评估了我们的模型,双模态DIP(dmDIP),从出生到一岁的婴儿MRI数据,这表明,增强的图像质量可以获得显著降低的敏感性提前停止。
    Magnetic resonance imaging (MRI) is commonly used for studying infant brain development. However, due to the lengthy image acquisition time and limited subject compliance, high-quality infant MRI can be challenging. Without imposing additional burden on image acquisition, image super-resolution (SR) can be used to enhance image quality post-acquisition. Most SR techniques are supervised and trained on multiple aligned low-resolution (LR) and high-resolution (HR) image pairs, which in practice are not usually available. Unlike supervised approaches, Deep Image Prior (DIP) can be employed for unsupervised single-image SR, utilizing solely the input LR image for de novo optimization to produce an HR image. However, determining when to stop early in DIP training is non-trivial and presents a challenge to fully automating the SR process. To address this issue, we constrain the low-frequency k-space of the SR image to be similar to that of the LR image. We further improve performance by designing a dual-modal framework that leverages shared anatomical information between T1-weighted and T2-weighted images. We evaluated our model, dual-modal DIP (dmDIP), on infant MRI data acquired from birth to one year of age, demonstrating that enhanced image quality can be obtained with substantially reduced sensitivity to early stopping.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:与磁共振成像(MRI)兼容的手术机器人的技术进步对术前和术中MRI的实时可变形图像配准(DIR)产生了不可或缺的需求,但是缺乏相关方法。挑战来自维度不匹配,分辨率差异,非刚性变形和实时配准的要求。
    方法:在本文中,我们提出了一个叫做MatchMorph的实时DIR框架,专为低分辨率局部术中MRI和高分辨率全局术前MRI的配准而设计。首先,开发了一种基于全局推理的超分辨率网络,以将术中MRI的分辨率提高到术前MRI的分辨率。从而解决了分辨率差异。其次,设计了一种快速匹配算法,用于确定术中MRI在相应术前MRI中的最佳位置,以解决维度不匹配问题.Further,构建了一个基于交叉注意力的双流DIR网络来操纵术前和术中MRI之间的变形,真正及时。
    结果:我们对公开可用的数据集IXI和OASIS进行了全面的实验,以评估所提出的MatchMorph框架的性能。与最先进的(SOTA)网络TransMorph相比,MatchMorph设计的双流DIR网络在IXI数据集上具有1.306mm小的HD和0.07mm小的ASD评分,实现了卓越的性能.此外,MatchMorph框架演示了大约280毫秒的推理速度。
    结论:从高分辨率全局术前MRI和模拟的低分辨率局部术中MRI获得的定性和定量配准结果验证了所提出的MatchMorph框架的有效性和效率。
    OBJECTIVE: The technological advancements in surgical robots compatible with magnetic resonance imaging (MRI) have created an indispensable demand for real-time deformable image registration (DIR) of pre- and intra-operative MRI, but there is a lack of relevant methods. Challenges arise from dimensionality mismatch, resolution discrepancy, non-rigid deformation and requirement for real-time registration.
    METHODS: In this paper, we propose a real-time DIR framework called MatchMorph, specifically designed for the registration of low-resolution local intraoperative MRI and high-resolution global preoperative MRI. Firstly, a super-resolution network based on global inference is developed to enhance the resolution of intraoperative MRI to the same as preoperative MRI, thus resolving the resolution discrepancy. Secondly, a fast-matching algorithm is designed to identify the optimal position of the intraoperative MRI within the corresponding preoperative MRI to address the dimensionality mismatch. Further, a cross-attention-based dual-stream DIR network is constructed to manipulate the deformation between pre- and intra-operative MRI, real-timely.
    RESULTS: We conducted comprehensive experiments on publicly available datasets IXI and OASIS to evaluate the performance of the proposed MatchMorph framework. Compared to the state-of-the-art (SOTA) network TransMorph, the designed dual-stream DIR network of MatchMorph achieved superior performance with a 1.306 mm smaller HD and a 0.07 mm smaller ASD score on the IXI dataset. Furthermore, the MatchMorph framework demonstrates an inference speed of approximately 280 ms.
    CONCLUSIONS: The qualitative and quantitative registration results obtained from high-resolution global preoperative MRI and simulated low-resolution local intraoperative MRI validated the effectiveness and efficiency of the proposed MatchMorph framework.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在单幅图像超分辨率背景下,特征提取起着举足轻重的作用。尽管如此,依靠单一的特征提取方法往往会破坏特征表示的全部潜力,妨碍模型的整体性能。为了解决这个问题,这项研究介绍了广泛的活化特征蒸馏网络(WFDN),通过双路径学习实现单幅图像的超分辨率。最初,采用双路径并行网络结构,利用剩余网络作为骨干,并结合全局剩余连接,以增强功能开发并加快网络融合。随后,采用了特征蒸馏块,其特点是训练速度快,参数计数低。同时,整合了广泛的激活机制,以进一步提高高频特征的表示能力。最后,引入门控融合机制对双分支提取的特征信息进行加权融合。该机制增强了重建性能,同时减轻了信息冗余。大量的实验表明,与最先进的方法相比,该算法获得了稳定和优越的结果,对四个基准数据集进行的定量评估指标测试证明了这一点。此外,我们的WFDN擅长重建具有更丰富详细纹理的图像,更现实的线条,更清晰的结构,肯定了其非凡的优越性和稳健性。
    Feature extraction plays a pivotal role in the context of single image super-resolution. Nonetheless, relying on a single feature extraction method often undermines the full potential of feature representation, hampering the model\'s overall performance. To tackle this issue, this study introduces the wide-activation feature distillation network (WFDN), which realizes single image super-resolution through dual-path learning. Initially, a dual-path parallel network structure is employed, utilizing a residual network as the backbone and incorporating global residual connections to enhance feature exploitation and expedite network convergence. Subsequently, a feature distillation block is adopted, characterized by fast training speed and a low parameter count. Simultaneously, a wide-activation mechanism is integrated to further enhance the representational capacity of high-frequency features. Lastly, a gated fusion mechanism is introduced to weight the fusion of feature information extracted from the dual branches. This mechanism enhances reconstruction performance while mitigating information redundancy. Extensive experiments demonstrate that the proposed algorithm achieves stable and superior results compared to the state-of-the-art methods, as evidenced by quantitative evaluation metrics tests conducted on four benchmark datasets. Furthermore, our WFDN excels in reconstructing images with richer detailed textures, more realistic lines, and clearer structures, affirming its exceptional superiority and robustness.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    自从它的发明,超分辨率显微镜已成为生物结构高级成像的流行工具,允许在低于衍射极限的空间尺度下可视化亚细胞结构。因此,最近,这并不奇怪,不同的超分辨率技术正在应用于神经科学,例如,解决神经递质受体和蛋白质复合物组成在突触前终末的聚集。尽管如此,这些实验绝大多数都是在细胞培养或非常薄的组织切片中进行的,而在生物样品的较深层(30-50μm)中只有少数超分辨率成像的例子。在这种情况下,哺乳动物的全视网膜已很少被研究与超分辨率显微镜。这里,我们的目标是建立一个受激发射损耗(STED)显微镜对整个视网膜成像方案.为此,我们开发了样品制备,包括视网膜组织的水平切片,与STED兼容的荧光团的免疫标记方案,并优化了图像采集设置。我们标记了躯体中的亚细胞结构,树突,和小鼠内部视网膜中的视网膜神经节细胞轴突。通过测量我们制备中最薄的丝状结构的半峰全宽,与传统的共焦图像相比,我们实现了两个或更高的分辨率增强。当与视网膜的水平切片相结合时,这些设置允许可视化外视网膜中推定的GABA能水平细胞突触。一起来看,我们成功地建立了一个STED协议,用于在30到50µm深度的全装鼠标视网膜中进行可靠的超分辨率成像,这使得调查,例如,健康和疾病中视网膜突触的蛋白质复合物组成和细胞骨架超微结构。
    Since its invention, super-resolution microscopy has become a popular tool for advanced imaging of biological structures, allowing visualisation of subcellular structures at a spatial scale below the diffraction limit. Thus, it is not surprising that recently, different super-resolution techniques are being applied in neuroscience, e.g. to resolve the clustering of neurotransmitter receptors and protein complex composition in presynaptic terminals. Still, the vast majority of these experiments were carried out either in cell cultures or very thin tissue sections, while there are only a few examples of super-resolution imaging in deeper layers (30 - 50 µm) of biological samples. In that context, the mammalian whole-mount retina has rarely been studied with super-resolution microscopy. Here, we aimed at establishing a stimulated-emission-depletion (STED) microscopy protocol for imaging whole-mount retina. To this end, we developed sample preparation including horizontal slicing of retinal tissue, an immunolabeling protocol with STED-compatible fluorophores and optimised the image acquisition settings. We labelled subcellular structures in somata, dendrites, and axons of retinal ganglion cells in the inner mouse retina. By measuring the full width at half maximum of the thinnest filamentous structures in our preparation, we achieved a resolution enhancement of two or higher compared to conventional confocal images. When combined with horizontal slicing of the retina, these settings allowed visualisation of putative GABAergic horizontal cell synapses in the outer retina. Taken together, we successfully established a STED protocol for reliable super-resolution imaging in the whole-mount mouse retina at depths between 30 and 50 µm, which enables investigating, for instance, protein complex composition and cytoskeletal ultrastructure at retinal synapses in health and disease.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    目的:评估基于深度学习的三维(3D)超分辨率扩散加权成像(DWI)影像组学模型预测高强度聚焦超声(HIFU)消融子宫肌瘤预后的可行性和有效性。
    方法:这项回顾性研究包括360例接受HIFU治疗的子宫肌瘤患者,包括中心A(训练集:N=240;内部测试集:N=60)和中心B(外部测试集:N=60),并根据术后非灌注体积比分类为预后良好或不良。在传统高分辨率DWI(HR-DWI)的基础上,采用深度迁移学习方法构建超分辨率DWI(SR-DWI),从两种图像类型的手动分割的感兴趣区域中提取1198个影像组学特征。在数据预处理和特征选择之后,使用支持向量机(SVM)构建HR-DWI和SR-DWI的影像组学模型,随机森林(RF),和光梯度提升机(LightGBM)算法,使用曲线下面积(AUC)和决策曲线评估性能。
    结果:与放射科专家相比,所有DWI影像组学模型在预测HIFU消融子宫肌瘤预后方面均表现出优异的AUC(AUC:0.706,95%CI:0.647-0.748)。当使用不同的机器学习算法时,支持向量机的HR-DWI模型的AUC值为0.805(95%CI:0.679-0.931),0.797(95%CI:0.672-0.921)射频,和0.770(95%CI:0.631-0.908)与LightGBM。同时,在所有算法中,SR-DWI模型优于HR-DWI模型(P<0.05),SVM的AUC值为0.868(95%CI:0.775-0.960),0.824(95%CI:0.715-0.934)与RF,和0.821(95%CI:0.709-0.933)与LightGBM。而决策曲线分析进一步证实了该模型良好的临床应用价值。
    结论:基于深度学习的3DSR-DWI影像组学模型在预测HIFU消融子宫肌瘤预后方面具有良好的可行性和有效性。优于HR-DWI模型和放射科专家的评估。
    OBJECTIVE: To assess the feasibility and efficacy of a deep learning-based three-dimensional (3D) super-resolution diffusion-weighted imaging (DWI) radiomics model in predicting the prognosis of high-intensity focused ultrasound (HIFU) ablation of uterine fibroids.
    METHODS: This retrospective study included 360 patients with uterine fibroids who received HIFU treatment, including Center A (training set: N = 240; internal testing set: N = 60) and Center B (external testing set: N = 60) and were classified as having a favorable or unfavorable prognosis based on the postoperative non-perfusion volume ratio. A deep transfer learning approach was used to construct super-resolution DWI (SR-DWI) based on conventional high-resolution DWI (HR-DWI), and 1198 radiomics features were extracted from manually segmented regions of interest in both image types. Following data preprocessing and feature selection, radiomics models were constructed for HR-DWI and SR-DWI using Support Vector Machine (SVM), Random Forest (RF), and Light Gradient Boosting Machine (LightGBM) algorithms, with performance evaluated using area under the curve (AUC) and decision curves.
    RESULTS: All DWI radiomics models demonstrated superior AUC in predicting HIFU ablated uterine fibroids prognosis compared to expert radiologists (AUC: 0.706, 95% CI: 0.647-0.748). When utilizing different machine learning algorithms, the HR-DWI model achieved AUC values of 0.805 (95% CI: 0.679-0.931) with SVM, 0.797 (95% CI: 0.672-0.921) with RF, and 0.770 (95% CI: 0.631-0.908) with LightGBM. Meanwhile, the SR-DWI model outperformed the HR-DWI model (P < 0.05) across all algorithms, with AUC values of 0.868 (95% CI: 0.775-0.960) with SVM, 0.824 (95% CI: 0.715-0.934) with RF, and 0.821 (95% CI: 0.709-0.933) with LightGBM. And decision curve analysis further confirmed the good clinical value of the models.
    CONCLUSIONS: Deep learning-based 3D SR-DWI radiomics model demonstrated favorable feasibility and effectiveness in predicting the prognosis of HIFU ablated uterine fibroids, which was superior to HR-DWI model and assessment by expert radiologists.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    为了方便传输,全向图像(ODI)通常遵循等矩形投影(ERP)格式,并且分辨率较低。为了提供更好的身临其境的体验,全向图像超分辨率(ODISR)至关重要。然而,ERPODI遭受严重的几何失真和跨纬度的像素拉伸,在高纬度产生大量冗余信息。这一特点对传统的SR方法提出了巨大的挑战,这只能获得次优的ODISR性能。为了解决这个问题,本文提出了一种新的ODISR位置注意网络(PAN)。具体来说,引入了两分支结构,其中基本增强分支(BE)用于实现提取的浅层特征的粗深特征增强。同时,位置注意力增强分支(PAE)构建位置注意力机制,根据不同纬度特征在ERP表示中的位置和拉伸程度动态调整贡献,实现了对差异化信息的增强,抑制冗余信息,并以空间失真调制深层特征。随后,有效地融合了两个分支的特征,以实现进一步的细化和适应ODI的失真特性。之后,我们利用长期记忆模块(LM),促进分支之间的信息交互和融合,以增强对失真的感知,聚合先前的分层功能以保留长期内存并提高ODISR性能。广泛的结果证明了我们的PAN在ODISR中的最先进性能和高效率。
    For convenient transmission, omnidirectional images (ODIs) usually follow the equirectangular projection (ERP) format and are low-resolution. To provide better immersive experience, omnidirectional image super resolution (ODISR) is essential. However, ERP ODIs suffer from serious geometric distortion and pixel stretching across latitudes, generating massive redundant information at high latitudes. This characteristic poses a huge challenge for the traditional SR methods, which can only obtain the suboptimal ODISR performance. To address this issue, we propose a novel position attention network (PAN) for ODISR in this paper. Specifically, a two-branch structure is introduced, in which the basic enhancement branch (BE) serves to achieve coarse deep feature enhancement for extracted shallow features. Meanwhile, the position attention enhancement branch (PAE) builds a positional attention mechanism to dynamically adjust the contribution of features at different latitudes in the ERP representation according to their positions and stretching degrees, which achieves the enhancement for the differentiated information, suppresses the redundant information, and modulate the deep features with spatial distortion. Subsequently, the features of two branches are fused effectively to achieve the further refinement and adapt the distortion characteristic of ODIs. After that, we exploit a long-term memory module (LM), promoting information interactions and fusions between the branches to enhance the perception of the distortion, aggregating the prior hierarchical features to keep the long-term memory and boosting the ODISR performance. Extensive results demonstrate the state-of-the-art performance and the high efficiency of our PAN in ODISR.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    在医学诊断中,仅依靠一种类型的生物标志物不足以准确识别癌症.基于血液的多癌症早期检测可以帮助从单个血液样本中识别多种类型的癌症。在这项研究中,开发了一种超分辨率多光谱成像纳米免疫传感器(srMINI),该传感器基于与链霉亲和素偶联的三种不同颜色的量子点(QD),用于在单分子水平同时筛查血液中的各种癌症生物标志物.在实验中,srMINI芯片用于同时检测三种关键的癌症生物标志物:癌胚抗原(CEA),C反应蛋白(CRP),甲胎蛋白(AFP)。srMINI芯片对这些癌症生物标志物的检测灵敏度为0.18-0.5ag/mL(1.1-2.6zM),比商业酶联免疫吸附测定试剂盒高108倍,因为不存在来自底物的干扰信号。为血液中癌症生物标志物的多重检测建立了相当大的潜力。因此,使用开发的srMINI芯片同时检测各种癌症生物标志物,具有较高的诊断精度和准确性,有望作为单分子生物传感器在早期诊断或社区筛查中发挥决定性作用.
    In medical diagnosis, relying on only one type of biomarker is insufficient to accurately identify cancer. Blood-based multicancer early detection can help identify more than one type of cancer from a single blood sample. In this study, a super-resolution multispectral imaging nanoimmunosensor (srMINI) based on three quantum dots (QDs) of different color conjugated with streptavidin was developed for the simultaneous screening of various cancer biomarkers in blood at the single-molecule level. In the experiment, the srMINI chip was used to simultaneously detect three key cancer biomarkers: carcinoembryonic antigen (CEA), C-reactive protein (CRP), and alpha-fetoprotein (AFP). The srMINI chip exhibited 108 times higher detection sensitivity of 0.18-0.5 ag/mL (1.1-2.6 zM) for these cancer biomarkers than commercial enzyme-linked immunosorbent assay kits because of the absence of interfering signals from the substrate, establishing considerable potential for multiplex detection of cancer biomarkers in blood. Therefore, the simultaneous detection of various cancer biomarkers using the developed srMINI chip with high diagnostic precision and accuracy is expected to play a decisive role in early diagnosis or community screening as a single-molecule biosensor.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    最近,基于金字塔的多分辨率技术已经成为图像超分辨率的主要研究方法。然而,这些方法通常依赖于级别之间的信息传输的单一模式。在我们的方法中,提出了一种基于小波能量熵(WEE)约束的小波金字塔递归神经网络(WPRNN)。该网络传输先前级别的小波系数和附加的浅系数特征以捕获局部细节。此外,每个金字塔级别和跨金字塔级别的低频和高频小波系数的参数是共享的。设计了多分辨率小波金字塔融合(WPF)模块,以促进跨网络金字塔级别的信息传递。此外,从信号能量分布的角度出发,提出了一种小波能量熵损失来约束小波系数的重构。最后,我们的方法通过在公开可用的数据集上进行的一系列广泛的实验,以最小的参数实现了具有竞争力的重建性能,这证明了它的实际效用。
    Recently, multi-resolution pyramid-based techniques have emerged as the prevailing research approach for image super-resolution. However, these methods typically rely on a single mode of information transmission between levels. In our approach, a wavelet pyramid recursive neural network (WPRNN) based on wavelet energy entropy (WEE) constraint is proposed. This network transmits previous-level wavelet coefficients and additional shallow coefficient features to capture local details. Besides, the parameter of low- and high-frequency wavelet coefficients within each pyramid level and across pyramid levels is shared. A multi-resolution wavelet pyramid fusion (WPF) module is devised to facilitate information transfer across network pyramid levels. Additionally, a wavelet energy entropy loss is proposed to constrain the reconstruction of wavelet coefficients from the perspective of signal energy distribution. Finally, our method achieves the competitive reconstruction performance with the minimal parameters through an extensive series of experiments conducted on publicly available datasets, which demonstrates its practical utility.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    溶酶体是动态的细胞结构,可以响应刺激自适应地重塑其膜,包括膜损坏。我们之前发现了一个过程,我们称之为LYTL(由富含亮氨酸的重复激酶2[LRRK2]驱动的溶酶体导管/分选),其中受损的溶酶体产生分选成移动囊泡的小管。LYTL由帕金森病相关激酶LRRK2协调,该激酶通过磷酸化RAB蛋白将运动衔接蛋白和RHD家族成员JIP4募集到溶酶体。为了确定参与LYTL的新玩家,我们对LRRK2激酶抑制后分离的溶酶体进行了无偏倚的蛋白质组学。我们的结果表明,RILPL1通过LRRK2活性募集到破裂的溶酶体中,以促进溶酶体表面RAB蛋白的磷酸化。RILPL1,也是RHD家族的成员,增强了LRRK2阳性溶酶体在核周区域的聚集,并导致LYTL小管的收缩,与促进LYTL小管延伸的JIP4相反。机械上,RILPL1结合p150胶合,一个动态肌动蛋白亚基,促进溶酶体和小管运输到微管的负端。对插管过程的进一步表征表明,LYTL小管沿着酪氨酸微管移动,微管蛋白酪氨酸化被证明是小管伸长所必需的。总之,我们的发现强调了两种不同的RHD蛋白和pRAB效应子对LYTL小管的动态调节,作为相反的运动衔接蛋白:JIP4,通过驱动蛋白促进输卵管,和RILPL1,通过动力蛋白/动力蛋白促进小管收缩。我们推断,这两个相反的过程会产生亚稳态的溶酶体膜变形,从而促进动态插管事件。
    Lysosomes are dynamic cellular structures that adaptively remodel their membrane in response to stimuli, including membrane damage. We previously uncovered a process we term LYTL (LYsosomal Tubulation/sorting driven by Leucine-Rich Repeat Kinase 2 [LRRK2]), wherein damaged lysosomes generate tubules sorted into mobile vesicles. LYTL is orchestrated by the Parkinson\'s disease-associated kinase LRRK2 that recruits the motor adaptor protein and RHD family member JIP4 to lysosomes via phosphorylated RAB proteins. To identify new players involved in LYTL, we performed unbiased proteomics on isolated lysosomes after LRRK2 kinase inhibition. Our results demonstrate that there is recruitment of RILPL1 to ruptured lysosomes via LRRK2 activity to promote phosphorylation of RAB proteins at the lysosomal surface. RILPL1, which is also a member of the RHD family, enhances the clustering of LRRK2-positive lysosomes in the perinuclear area and causes retraction of LYTL tubules, in contrast to JIP4 which promotes LYTL tubule extension. Mechanistically, RILPL1 binds to p150Glued, a dynactin subunit, facilitating the transport of lysosomes and tubules to the minus end of microtubules. Further characterization of the tubulation process revealed that LYTL tubules move along tyrosinated microtubules, with tubulin tyrosination proving essential for tubule elongation. In summary, our findings emphasize the dynamic regulation of LYTL tubules by two distinct RHD proteins and pRAB effectors, serving as opposing motor adaptor proteins: JIP4, promoting tubulation via kinesin, and RILPL1, facilitating tubule retraction through dynein/dynactin. We infer that the two opposing processes generate a metastable lysosomal membrane deformation that facilitates dynamic tubulation events.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着深度学习的引入,在过去的十年中,在计算机视觉领域进行了大量的研究。特别是,物体检测(OD)的研究持续快速发展。然而,尽管取得了这些进步,需要克服一些限制,以实现基于深度学习的OD模型的实际应用。一个这样的限制是当图像质量差或目标对象小时不准确的OD。小物体的性能退化现象类似于OD模型的基本限制,例如接受野的约束,这是仅使用OD模型难以解决的问题。因此,OD性能可能会受到低图像质量或小目标物体的阻碍。为了解决这个问题,这项研究调查了超分辨率(SR)和OD技术的兼容性,以提高检测,特别是小物件。我们分析了SR和OD模型的组合,根据建筑特征对它们进行分类。实验结果表明,将OD检测器与SR模型集成在一起时会有很大的改善。总的来说,事实证明,当评估指标(PSNR,SSIM)的SR模型很高,OD的性能也相应较高。尤其是,对MSCOCO数据集的评估显示,与所有对象相比,小对象的增强率高出9.4%。这项工作提供了SR和OD模型兼容性的分析,证明了它们协同组合的潜在好处。实验代码可以在我们的GitHub存储库中找到。
    With the introduction of deep learning, a significant amount of research has been conducted in the field of computer vision in the past decade. In particular, research on object detection (OD) continues to progress rapidly. However, despite these advances, some limitations need to be overcome to enable real-world applications of deep learning-based OD models. One such limitation is inaccurate OD when image quality is poor or a target object is small. The performance degradation phenomenon for small objects is similar to the fundamental limitations of an OD model, such as the constraint of the receptive field, which is a difficult problem to solve using only an OD model. Therefore, OD performance can be hindered by low image quality or small target objects. To address this issue, this study investigates the compatibility of super-resolution (SR) and OD techniques to improve detection, particularly for small objects. We analyze the combination of SR and OD models, classifying them based on architectural characteristics. The experimental results show a substantial improvement when integrating OD detectors with SR models. Overall, it was demonstrated that, when the evaluation metrics (PSNR, SSIM) of the SR models are high, the performance in OD is correspondingly high as well. Especially, evaluations on the MS COCO dataset reveal that the enhancement rate for small objects is 9.4% higher compared to all objects. This work provides an analysis of SR and OD model compatibility, demonstrating the potential benefits of their synergistic combination. The experimental code can be found on our GitHub repository.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号