Convolutional neural network

卷积神经网络
  • 文章类型: Journal Article
    对图像进行分类是计算机视觉中最重要的任务之一。最近,图像分类任务的最佳性能已由深度和良好连接的网络显示。这些天,大多数数据集由固定数量的彩色图像组成。输入图像采用红绿蓝(RGB)格式,并进行分类,而不对原始图像进行任何更改。观察到颜色空间(基本上改变原始RGB图像)对分类精度有重大影响,我们深入研究颜色空间的意义。此外,具有高度可变数量的类的数据集,例如PlantVillage数据集利用在同一模型中包含大量颜色空间的模型,达到很高的精度,和不同类别的图像更好地表示在不同的颜色空间。此外,我们证明了这种类型的模型,其中输入被同时预处理到许多颜色空间中,需要更少的参数来实现分类的高精度。所提出的模型基本上以RGB图像作为输入,一次把它变成七个独立的颜色空间,然后将这些颜色空间中的每一个都输入到自己的卷积神经网络(CNN)模型中。为了减轻计算机的负载和所需的超参数数量,我们在提出的CNN模型中使用组卷积层。与目前最先进的作物病害分类方法相比,我们取得了实质性的进展。
    Classifying images is one of the most important tasks in computer vision. Recently, the best performance for image classification tasks has been shown by networks that are both deep and well-connected. These days, most datasets are made up of a fixed number of color images. The input images are taken in red green blue (RGB) format and classified without any changes being made to the original. It is observed that color spaces (basically changing original RGB images) have a major impact on classification accuracy, and we delve into the significance of color spaces. Moreover, datasets with a highly variable number of classes, such as the PlantVillage dataset utilizing a model that incorporates numerous color spaces inside the same model, achieve great levels of accuracy, and different classes of images are better represented in different color spaces. Furthermore, we demonstrate that this type of model, in which the input is preprocessed into many color spaces simultaneously, requires significantly fewer parameters to achieve high accuracy for classification. The proposed model basically takes an RGB image as input, turns it into seven separate color spaces at once, and then feeds each of those color spaces into its own Convolutional Neural Network (CNN) model. To lessen the load on the computer and the number of hyperparameters needed, we employ group convolutional layers in the proposed CNN model. We achieve substantial gains over the present state-of-the-art methods for the classification of crop disease.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    步态识别,生物识别方法,由于其独特的属性而引起了极大的关注,包括非侵入性,远距离捕获,以及对模仿的抵制。在深度学习从数据中提取复杂特征的卓越能力的推动下,步态识别经历了一场革命。本文概述了基于深度学习的步态识别方法的当前发展。我们探索和分析步态识别的发展,并强调其在取证中的用途,安全,和刑事调查。这篇文章深入探讨了与步态识别相关的挑战,例如步行条件的变化,视角,衣服也是。我们通过对最先进的架构进行全面分析,讨论深度神经网络在解决这些挑战方面的有效性。包括卷积神经网络(CNN),递归神经网络(RNN),注意机制。多种基于神经网络的步态识别模型,如门控和共享注意力ICDNet(GA-ICDNet),多尺度时间特征提取器(MSTFE),GaitNet,和各种基于CNN的方法,在不同的步行条件下表现出令人印象深刻的准确性,展示这些模型在捕捉独特步态模式方面的有效性。GaitNet实现了99.7%的卓越识别精度,而GA-ICDNet在验证任务中表现出高精度,误差率为0.67%。GaitGraph(ResGCN+2DCNN)实现了从66.3%到87.7%的秩1精度,而与Koopman运营商的完全连接网络在各种条件下实现了OU-MVLP的平均秩1精度74.7%。然而,利用图卷积网络(GCN)和GaitSet的GCPFP(具有基于图卷积的零件特征轮询的GCN)对于CASIA-B实现了62.4%的最低平均秩1精度,而MFINet(多因素推理网络)在CASIA-B上的服装变化条件下表现出11.72%至19.32%的最低精度范围。除了全面分析最近在步态识别方面的突破,还评估了未来潜在研究方向的范围。
    Gait recognition, a biometric identification method, has garnered significant attention due to its unique attributes, including non-invasiveness, long-distance capture, and resistance to impersonation. Gait recognition has undergone a revolution driven by the remarkable capacity of deep learning to extract complicated features from data. An overview of the current developments in deep learning-based gait identification methods is provided in this work. We explore and analyze the development of gait recognition and highlight its uses in forensics, security, and criminal investigations. The article delves into the challenges associated with gait recognition, such as variations in walking conditions, viewing angles, and clothing as well. We discuss about the effectiveness of deep neural networks in addressing these challenges by providing a comprehensive analysis of state-of-the-art architectures, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), and attention mechanisms. Diverse neural network-based gait recognition models, such as Gate Controlled and Shared Attention ICDNet (GA-ICDNet), Multi-Scale Temporal Feature Extractor (MSTFE), GaitNet, and various CNN-based approaches, demonstrate impressive accuracy across different walking conditions, showcasing the effectiveness of these models in capturing unique gait patterns. GaitNet achieved an exceptional identification accuracy of 99.7%, whereas GA-ICDNet showed high precision with an equal error rate of 0.67% in verification tasks. GaitGraph (ResGCN+2D CNN) achieved rank-1 accuracies ranging from 66.3% to 87.7%, whereas a Fully Connected Network with Koopman Operator achieved an average rank-1 accuracy of 74.7% for OU-MVLP across various conditions. However, GCPFP (GCN with Graph Convolution-Based Part Feature Polling) utilizing graph convolutional network (GCN) and GaitSet achieves the lowest average rank-1 accuracy of 62.4% for CASIA-B, while MFINet (Multiple Factor Inference Network) exhibits the lowest accuracy range of 11.72% to 19.32% under clothing variation conditions on CASIA-B. In addition to an across-the-board analysis of recent breakthroughs in gait recognition, the scope for potential future research direction is also assessed.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    手势是一种有效的沟通工具,可以在各个领域传达丰富的信息,包括医疗和教育。电子学习在过去几年中取得了显着增长,现在已成为许多企业的重要资源。尽管如此,关于在电子学习中使用手势的研究并不多。与此类似,医疗专业人员经常使用手势来帮助诊断和治疗。
    我们的目标是改进教师的方式,学生,和医疗专业人员通过引入动态的手势监测和识别方法来接收信息。六个模块组成了我们的方法:视频到帧转换,质量增强的预处理,手骨架映射与单发多盒检测器(SSMD)跟踪,使用背景建模和卷积神经网络(CNN)边界框技术的手检测,使用基于点的和全手覆盖技术的特征提取,并使用基于种群的增量学习算法进行优化。接下来,1DCNN分类器用于识别手部运动。
    经过大量的试验和错误,我们能够在印度手语和WLASL数据集上获得83.71%和85.71%的手跟踪精度,分别。我们的发现表明了我们的方法识别手部动作的效果。
    教师,学生,和医疗专业人员都可以通过利用我们建议的系统有效地传输和理解信息。获得的准确率凸显了我们的方法如何改善通信并使各个领域的信息交换更加容易。
    UNASSIGNED: Hand gestures are an effective communication tool that may convey a wealth of information in a variety of sectors, including medical and education. E-learning has grown significantly in the last several years and is now an essential resource for many businesses. Still, there has not been much research conducted on the use of hand gestures in e-learning. Similar to this, gestures are frequently used by medical professionals to help with diagnosis and treatment.
    UNASSIGNED: We aim to improve the way instructors, students, and medical professionals receive information by introducing a dynamic method for hand gesture monitoring and recognition. Six modules make up our approach: video-to-frame conversion, preprocessing for quality enhancement, hand skeleton mapping with single shot multibox detector (SSMD) tracking, hand detection using background modeling and convolutional neural network (CNN) bounding box technique, feature extraction using point-based and full-hand coverage techniques, and optimization using a population-based incremental learning algorithm. Next, a 1D CNN classifier is used to identify hand motions.
    UNASSIGNED: After a lot of trial and error, we were able to obtain a hand tracking accuracy of 83.71% and 85.71% over the Indian Sign Language and WLASL datasets, respectively. Our findings show how well our method works to recognize hand motions.
    UNASSIGNED: Teachers, students, and medical professionals can all efficiently transmit and comprehend information by utilizing our suggested system. The obtained accuracy rates highlight how our method might improve communication and make information exchange easier in various domains.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    癫痫,这与神经元损伤和功能衰退有关,通常会给患者带来日常生活中的许多挑战。早期诊断在控制病情和减轻患者痛苦中起着至关重要的作用。基于脑电图(EEG)的方法由于其有效性和非侵入性而通常用于诊断癫痫。在这项研究中,提出了一种分类方法,该方法使用快速傅里叶变换(FFT)提取结合卷积神经网络(CNN)和长短期记忆(LSTM)模型。
    大多数方法使用传统框架对癫痫进行分类,我们提出了一种新的方法来解决这个问题,即从源数据中提取特征,然后将它们输入到网络中进行训练和识别。它将源数据预处理为训练和验证数据,然后使用CNN和LSTM对数据的样式进行分类。
    在分析公共测试数据集时,用于癫痫分类的全CNN嵌套LSTM模型中表现最好的特征是3种特征中的FFT特征.值得注意的是,所有进行的实验都有很高的准确率,准确度超过96%的值,93%的灵敏度,和96%的特异性。这些结果进一步以当前的方法为基准,在所有试验中展示一致和强大的性能。我们的方法始终如一地实现了超过97.00%的准确率,在单个实验中的值范围为97.95%至99.83%。特别值得注意的是,我们的方法在AB与(与)CDE比较,注册为99.06%。
    我们的方法具有区分癫痫和非癫痫个体的精确分类能力,无论参与者的眼睛是闭上还是睁开。此外,我们的技术在有效地对癫痫类型进行分类方面显示出显著的性能,区分癫痫发作和发作间状态与非癫痫状态。我们的自动分类方法的固有优点是其能够忽略在闭眼或睁眼状态期间获取的EEG数据。这种创新为现实世界的应用带来了希望,可能帮助医疗专业人员更有效地诊断癫痫。
    UNASSIGNED: Epilepsy, which is associated with neuronal damage and functional decline, typically presents patients with numerous challenges in their daily lives. An early diagnosis plays a crucial role in managing the condition and alleviating the patients\' suffering. Electroencephalogram (EEG)-based approaches are commonly employed for diagnosing epilepsy due to their effectiveness and non-invasiveness. In this study, a classification method is proposed that use fast Fourier Transform (FFT) extraction in conjunction with convolutional neural networks (CNN) and long short-term memory (LSTM) models.
    UNASSIGNED: Most methods use traditional frameworks to classify epilepsy, we propose a new approach to this problem by extracting features from the source data and then feeding them into a network for training and recognition. It preprocesses the source data into training and validation data and then uses CNN and LSTM to classify the style of the data.
    UNASSIGNED: Upon analyzing a public test dataset, the top-performing features in the fully CNN nested LSTM model for epilepsy classification are FFT features among three types of features. Notably, all conducted experiments yielded high accuracy rates, with values exceeding 96% for accuracy, 93% for sensitivity, and 96% for specificity. These results are further benchmarked against current methodologies, showcasing consistent and robust performance across all trials. Our approach consistently achieves an accuracy rate surpassing 97.00%, with values ranging from 97.95 to 99.83% in individual experiments. Particularly noteworthy is the superior accuracy of our method in the AB versus (vs.) CDE comparison, registering at 99.06%.
    UNASSIGNED: Our method exhibits precise classification abilities distinguishing between epileptic and non-epileptic individuals, irrespective of whether the participant\'s eyes are closed or open. Furthermore, our technique shows remarkable performance in effectively categorizing epilepsy type, distinguishing between epileptic ictal and interictal states versus non-epileptic conditions. An inherent advantage of our automated classification approach is its capability to disregard EEG data acquired during states of eye closure or eye-opening. Such innovation holds promise for real-world applications, potentially aiding medical professionals in diagnosing epilepsy more efficiently.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本研究旨在(1)在TOF-MRA中复制一种基于深度学习的脑动脉瘤分割模型,(2)通过测试各种全自动预处理管道来改进方法,(3)严格验证模型的独立性,外部测试数据集。在单个供应商的本地扫描仪上获取的235个TOF-MRA上训练了卷积神经网络,以分割颅内动脉瘤。不同的预处理管道,包括偏置场校正,重新采样,比较了种植和强度归一化对模型性能的影响。这些模型在独立的环境下进行了测试,外部同一供应商和其他供应商测试数据集,每个由70个TOF-MRA组成,包括有和没有动脉瘤的患者。性能最佳的模型在外部同一供应商测试数据集上取得了出色的结果,超越以前出版物的结果,提高了灵敏度(0.97vs.~0.86),较高的骰子得分系数(DSC,0.60±0.25vs.0.53±0.31),和提高的假阳性率(0.87±1.35vs.~2.7FPs/case)。该模型在外部其他供应商测试数据集中进一步显示出优异的性能(DSC0.65±0.26;灵敏度0.92,0.96±2.38FPs/case)。特异性分别为0.38和0.53。将体素尺寸从0.5×0.5×0.5mm提高到1×1×1mm可将假阳性率降低七倍。这项研究成功地复制了以前的方法的核心原理,用于在TOF-MRA中检测和分割脑动脉瘤,完全自动化的预处理管道。该模型在两个独立的外部数据集上展示了强大的可转移性,使用来自与训练数据集相同的扫描仪供应商和其他供应商的TOF-MRA。这些发现对于这种方法的临床应用非常令人鼓舞。
    This study aimed to (1) replicate a deep-learning-based model for cerebral aneurysm segmentation in TOF-MRAs, (2) improve the approach by testing various fully automatic pre-processing pipelines, and (3) rigorously validate the model\'s transferability on independent, external test-datasets. A convolutional neural network was trained on 235 TOF-MRAs acquired on local scanners from a single vendor to segment intracranial aneurysms. Different pre-processing pipelines including bias field correction, resampling, cropping and intensity-normalization were compared regarding their effect on model performance. The models were tested on independent, external same-vendor and other-vendor test-datasets, each comprised of 70 TOF-MRAs, including patients with and without aneurysms. The best-performing model achieved excellent results on the external same-vendor test-dataset, surpassing the results of the previous publication with an improved sensitivity (0.97 vs. ~ 0.86), a higher Dice score coefficient (DSC, 0.60 ± 0.25 vs. 0.53 ± 0.31), and an improved false-positive rate (0.87 ± 1.35 vs. ~ 2.7 FPs/case). The model further showed excellent performance in the external other-vendor test-datasets (DSC 0.65 ± 0.26; sensitivity 0.92, 0.96 ± 2.38 FPs/case). Specificity was 0.38 and 0.53, respectively. Raising the voxel-size from 0.5 × 0.5×0.5 mm to 1 × 1×1 mm reduced the false-positive rate seven-fold. This study successfully replicated core principles of a previous approach for detecting and segmenting cerebral aneurysms in TOF-MRAs with a robust, fully automatable pre-processing pipeline. The model demonstrated robust transferability on two independent external datasets using TOF-MRAs from the same scanner vendor as the training dataset and from other vendors. These findings are very encouraging regarding the clinical application of such an approach.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    快速准确的诊断测试是改善患者预后和对抗传染病的基础。聚集的定期间隔短回文重复(CRISPR)基于Cas12a的检测系统已成为现场核酸测试的有希望的解决方案。尽管如此,用于基于Cas12a的检测的CRISPRRNA(crRNA)的有效设计仍然具有挑战性且耗时。在这项研究中,我们提出了一个增强的crRNA设计系统,用于Cas12a介导的诊断,被称为EasyDesign。该系统采用优化的卷积神经网络(CNN)预测模型,在包括11,496个实验验证的基于Cas12a的检测案例的综合数据集上进行训练,涵盖了广泛的流行病原体,达到斯皮尔曼的ρ=0.812。我们进一步评估了crRNA设计中四种未包含在训练数据中的病原体的模型性能:猴痘病毒,肠道病毒71型、柯萨奇病毒A16型和单核细胞增生李斯特菌。结果表明,与传统的实验筛选相比,预测性能更好。此外,我们开发了一个交互式网络服务器(https://crispr。zhejianglab.com/),将EasyDesign与重组酶聚合酶扩增(RPA)引物设计集成在一起,增强用户可访问性。通过这个基于Web的平台,我们成功设计了6种人乳头瘤病毒(HPV)亚型的最佳Cas12acrRNA。值得注意的是,每个HPV亚型的所有前5个预测的crRNA在CRISPR分析中都表现出强大的荧光信号,从而表明该平台可以有效地促进临床样本检测。总之,EasyDesign为基于Cas12a的检测中的crRNA设计提供了快速可靠的解决方案,它可以作为临床诊断和研究应用的有价值的工具。
    Rapid and accurate diagnostic tests are fundamental for improving patient outcomes and combating infectious diseases. The Clustered Regularly Interspaced Short Palindromic Repeats (CRISPR) Cas12a-based detection system has emerged as a promising solution for on-site nucleic acid testing. Nonetheless, the effective design of CRISPR RNA (crRNA) for Cas12a-based detection remains challenging and time-consuming. In this study, we propose an enhanced crRNA design system with deep learning for Cas12a-mediated diagnostics, referred to as EasyDesign. This system employs an optimized convolutional neural network (CNN) prediction model, trained on a comprehensive data set comprising 11,496 experimentally validated Cas12a-based detection cases, encompassing a wide spectrum of prevalent pathogens, achieving Spearman\'s ρ = 0.812. We further assessed the model performance in crRNA design for four pathogens not included in the training data: Monkeypox Virus, Enterovirus 71, Coxsackievirus A16, and Listeria monocytogenes. The results demonstrated superior prediction performance compared to the traditional experiment screening. Furthermore, we have developed an interactive web server (https://crispr.zhejianglab.com/) that integrates EasyDesign with recombinase polymerase amplification (RPA) primer design, enhancing user accessibility. Through this web-based platform, we successfully designed optimal Cas12a crRNAs for six human papillomavirus (HPV) subtypes. Remarkably, all the top five predicted crRNAs for each HPV subtype exhibited robust fluorescent signals in CRISPR assays, thereby suggesting that the platform could effectively facilitate clinical sample testing. In conclusion, EasyDesign offers a rapid and reliable solution for crRNA design in Cas12a-based detection, which could serve as a valuable tool for clinical diagnostics and research applications.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在经食管超声心动图(TEE)上从经胃短轴视图(TSV)分割左心室是围手术期心血管评估的基础。即使是经验丰富的专业人士,该过程仍然耗时且依赖于经验。当前的研究旨在通过评估不同U-Net算法的有效性来评估深度学习用于自动分割的可行性。回顾性收集了一个包含1388例TSV采集的大型数据集,该数据集来自451例患者(32%的女性,平均年龄53.42岁),在2015年7月至2023年10月期间接受围手术期TEE。通过图像预处理和数据增强,训练集中包含3336张图像,验证集中的138个图像,和测试集中的138个图像。四个深度神经网络(U-Net,注意U-Net,UNet++,和UNeXt)用于左心室分割,并根据测试集上的Jaccard相似系数(JSC)和Dice相似系数(DSC)进行比较,以及网络参数的数量,培训时间,和推理时间。注意U-Net和U-Net++模型在JSC(最高平均JSC:86.02%)和DSC(最高平均DSC:92.00%)方面表现更好,UNeXt模型的网络参数最小(147万),U-Net模型的训练时间(6428.65s)和推断时间(101.75ms)最少。注意力U-Net模型在挑战性案例中优于其他三个模型,包括左心室边界受损和乳头状肌伪影。这一开创性的探索证明了深度学习在TEE上从TSV分割左心室的可行性,这将促进心血管评估的加速和客观替代围手术期管理。
    Segmenting the left ventricle from the transgastric short-axis views (TSVs) on transesophageal echocardiography (TEE) is the cornerstone for cardiovascular assessment during perioperative management. Even for seasoned professionals, the procedure remains time-consuming and experience-dependent. The current study aims to evaluate the feasibility of deep learning for automatic segmentation by assessing the validity of different U-Net algorithms. A large dataset containing 1388 TSV acquisitions was retrospectively collected from 451 patients (32% women, average age 53.42 years) who underwent perioperative TEE between July 2015 and October 2023. With image preprocessing and data augmentation, 3336 images were included in the training set, 138 images in the validation set, and 138 images in the test set. Four deep neural networks (U-Net, Attention U-Net, UNet++, and UNeXt) were employed for left ventricle segmentation and compared in terms of the Jaccard similarity coefficient (JSC) and Dice similarity coefficient (DSC) on the test set, as well as the number of network parameters, training time, and inference time. The Attention U-Net and U-Net++ models performed better in terms of JSC (the highest average JSC: 86.02%) and DSC (the highest average DSC: 92.00%), the UNeXt model had the smallest network parameters (1.47 million), and the U-Net model had the least training time (6428.65 s) and inference time for a single image (101.75 ms). The Attention U-Net model outperformed the other three models in challenging cases, including the impaired boundary of left ventricle and the artifact of the papillary muscle. This pioneering exploration demonstrated the feasibility of deep learning for the segmentation of the left ventricle from TSV on TEE, which will facilitate an accelerated and objective alternative of cardiovascular assessment for perioperative management.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    (1)背景:在繁忙的急诊科(ED)中识别急性主动脉综合征(AAS)和胸主动脉瘤(TAA)至关重要,因为它们具有危及生命的性质,需要及时准确的诊断。(2)方法:采用回顾性病例对照研究方法,对3家医院的ED进行分析。在2010年1月1日至2020年1月1日期间主诉胸痛或背痛的成年患者被纳入研究。收集的胸部X线摄影(CXR)数据分为训练(80%)和测试(20%)数据集。训练数据集由四个不同的卷积神经网络(CNN)模型训练。(3)结果:本研究共纳入1625例患者。InceptionV3模型获得了最高的F1评分0.76。(4)结论:使用基于CNN的模型分析CXR为临床医生提供了一种新的工具来解释患有胸痛和可疑AAS和TAA的ED患者。将来可以考虑将这种成像工具集成到ED中以增强临床致命疾病的诊断工作流程。
    (1) Background: Identifying acute aortic syndrome (AAS) and thoracic aortic aneurysm (TAA) in busy emergency departments (EDs) is crucial due to their life-threatening nature, necessitating timely and accurate diagnosis. (2) Methods: This retrospective case-control study was conducted in the ED of three hospitals. Adult patients visiting the ED between 1 January 2010 and 1 January 2020 with a chief complaint of chest or back pain were enrolled in the study. The collected chest radiography (CXRs) data were divided into training (80%) and testing (20%) datasets. The training dataset was trained by four different convolutional neural network (CNN) models. (3) Results: A total of 1625 patients were enrolled in this study. The InceptionV3 model achieved the highest F1 score of 0.76. (4) Conclusions: Analysis of CXRs using a CNN-based model provides a novel tool for clinicians to interpret ED patients with chest pain and suspected AAS and TAA. The integration of such imaging tools into ED could be considered in the future to enhance the diagnostic workflow for clinically fatal diseases.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    微光图像在智能监控和许多其他应用中普遍存在,低亮度阻碍进一步加工。尽管弱光图像增强可以减少此类问题的影响,当前的方法通常涉及复杂的网络结构或许多迭代,这不利于他们的效率。本文提出了一种使用相机响应模型的零参考相机响应网络,以实现对任意弱光图像的有效增强。建立具有流线型结构的双层参数生成网络,从辐射图中提取曝光率K,这是通过相机响应函数反转输入获得的。然后,K被用作用于对弱光图像进行一次变换以实现增强的亮度变换函数的参数。此外,设计了对比度保持亮度损失和边缘保持平滑度损失,而无需从数据集中引用。两者都可以进一步在输入中保留一些关键信息以提高精度。该增强被简化并且可以达到类似方法的两倍以上的速度。在多个LLIE数据集和DARKFACE人脸检测数据集上的大量实验充分证明了我们方法的优势,主观和客观。
    Low-light images are prevalent in intelligent monitoring and many other applications, with low brightness hindering further processing. Although low-light image enhancement can reduce the influence of such problems, current methods often involve a complex network structure or many iterations, which are not conducive to their efficiency. This paper proposes a Zero-Reference Camera Response Network using a camera response model to achieve efficient enhancement for arbitrary low-light images. A double-layer parameter-generating network with a streamlined structure is established to extract the exposure ratio K from the radiation map, which is obtained by inverting the input through a camera response function. Then, K is used as the parameter of a brightness transformation function for one transformation on the low-light image to realize enhancement. In addition, a contrast-preserving brightness loss and an edge-preserving smoothness loss are designed without the requirement for references from the dataset. Both can further retain some key information in the inputs to improve precision. The enhancement is simplified and can reach more than twice the speed of similar methods. Extensive experiments on several LLIE datasets and the DARK FACE face detection dataset fully demonstrate our method\'s advantages, both subjectively and objectively.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在本文中,人工智能(AI)技术应用于各向异性物体的电磁成像。磁异常传感系统和电磁成像的进展使用电磁原理来检测和表征地下或隐藏物体。我们使用测量的多频散射场通过反向传播方案(BPS)计算各向异性物体的初始介电常数分布。稍后,将估计的多频介电常数分布输入到卷积神经网络(CNN),用于自适应矩估计(ADAM)方法,以重建更准确的图像。同时,我们还改进了CNN中损失函数的定义。数值结果表明,将结构相似指数度量(SSIM)和均方根误差(RMSE)统一的改进损失函数可以有效地提高图像质量。在我们的模拟环境中,TE(横向电)和TM(横向磁)波都考虑了噪声干扰,以重建各向异性散射体。最后,我们得出的结论是,多频重建比单频重建更稳定和精确。
    In this paper, artificial intelligence (AI) technology is applied to the electromagnetic imaging of anisotropic objects. Advances in magnetic anomaly sensing systems and electromagnetic imaging use electromagnetic principles to detect and characterize subsurface or hidden objects. We use measured multifrequency scattered fields to calculate the initial dielectric constant distribution of anisotropic objects through the backpropagation scheme (BPS). Later, the estimated multifrequency permittivity distribution is input to a convolutional neural network (CNN) for the adaptive moment estimation (ADAM) method to reconstruct a more accurate image. In the meantime, we also improve the definition of loss function in the CNN. Numerical results show that the improved loss function unifying the structural similarity index measure (SSIM) and root mean square error (RMSE) can effectively enhance image quality. In our simulation environment, noise interference is considered for both TE (transverse electric) and TM (transverse magnetic) waves to reconstruct anisotropic scatterers. Lastly, we conclude that multifrequency reconstructions are more stable and precise than single-frequency reconstructions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号