MobileNet

MobileNet
  • 文章类型: Journal Article
    对微藻物种进行准确分类对于监测海洋生态系统和管理海洋粘液的出现至关重要,这对于监测海洋环境中的粘液现象至关重要。由于耗时的过程和对专家知识的需求,传统方法已经不足。本文的目的是采用卷积神经网络(CNN)和支持向量机(SVM)来提高分类精度和效率。通过采用先进的计算技术,包括MobileNet和GoogleNet模型,与SVM分类一起,这项研究表明,与传统的识别方法相比,有了显著的进步。在使用四种不同的SVM核函数对由7820图像组成的数据集进行分类时,线性内核的成功率最高,为98.79%。其次是RBF内核,占98.73%,多项式内核为97.84%,乙状核为97.20%。这项研究不仅为海洋生物多样性监测的未来研究提供了方法论框架,而且还强调了在生态保护和了解气候变化和环境污染中的粘液动态方面实时应用的潜力。
    Accurately classifying microalgae species is vital for monitoring marine ecosystems and managing the emergence of marine mucilage, which is crucial for monitoring mucilage phenomena in marine environments. Traditional methods have been inadequate due to time-consuming processes and the need for expert knowledge. The purpose of this article is to employ convolutional neural networks (CNNs) and support vector machines (SVMs) to improve classification accuracy and efficiency. By employing advanced computational techniques, including MobileNet and GoogleNet models, alongside SVM classification, the study demonstrates significant advancements over conventional identification methods. In the classification of a dataset consisting of 7820 images using four different SVM kernel functions, the linear kernel achieved the highest success rate at 98.79 %. It is followed by the RBF kernel at 98.73 %, the polynomial kernel at 97.84 %, and the sigmoid kernel at 97.20 %. This research not only provides a methodological framework for future studies in marine biodiversity monitoring but also highlights the potential for real-time applications in ecological conservation and understanding mucilage dynamics amidst climate change and environmental pollution.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    蓝藻是水生环境中的主要微生物,由于饮用水水库中的毒素产生,对公众健康构成重大风险。传统的水质评估水样中产毒属的丰度既费时又容易出错,强调迫切需要快速准确的自动化方法。这项研究通过引入一个新的公共数据集来解决这一差距,TCB-DS(产毒性蓝细菌数据集),包括2593张10个产毒蓝细菌属的显微图像,随后,识别这些属的自动化系统,可分为两部分。最初,采用特征提取器卷积神经网络(CNN)模型,与MobileNet在与各种其他流行的架构(如MobileNetV2、VGG、等。其次,要对第一部分的提取特征执行分类算法,测试了多种方法,实验结果表明,全连接神经网络(FCNN)具有最佳性能,加权精度和f1得分分别为94.79%和94.91%,分别。使用MobileNetV2作为特征提取器和FCNN作为分类器获得的最高宏观精度和f1得分分别为90.17%和87.64%。这些结果表明,所提出的方法可以用作自动筛选工具,用于识别产毒蓝藻,对水质控制具有实际意义,取代了实验室操作员在显微镜观察后给出的传统估计。本文的数据集和代码可在https://github.com/iman2693/CTCB上公开获得。
    Cyanobacteria are the dominating microorganisms in aquatic environments, posing significant risks to public health due to toxin production in drinking water reservoirs. Traditional water quality assessments for abundance of the toxigenic genera in water samples are both time-consuming and error-prone, highlighting the urgent need for a fast and accurate automated approach. This study addresses this gap by introducing a novel public dataset, TCB-DS (Toxigenic Cyanobacteria Dataset), comprising 2593 microscopic images of 10 toxigenic cyanobacterial genera and subsequently, an automated system to identify these genera which can be divided into two parts. Initially, a feature extractor Convolutional Neural Network (CNN) model was employed, with MobileNet emerging as the optimal choice after comparing it with various other popular architectures such as MobileNetV2, VGG, etc. Secondly, to perform classification algorithms on the extracted features of the first section, multiple approaches were tested and the experimental results indicate that a Fully Connected Neural Network (FCNN) had the optimal performance with weighted accuracy and f1-score of 94.79% and 94.91%, respectively. The highest macro accuracy and f1-score were 90.17% and 87.64% which were acquired using MobileNetV2 as the feature extractor and FCNN as the classifier. These results demonstrate that the proposed approach can be employed as an automated screening tool for identifying toxigenic Cyanobacteria with practical implications for water quality control replacing the traditional estimation given by the lab operator following microscopic observations. The dataset and code of this paper are publicly available at https://github.com/iman2693/CTCB.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    聋人和听力困难的人主要使用手语进行交流,这是一组符号,使用手势与面部表情相结合来制作有意义和完整的句子。聋人和听力障碍人士面临的问题是缺乏将手语翻译成书面或口头文本的自动工具,这导致了他们和社区之间的沟通差距。最先进的基于视觉的手语识别方法侧重于翻译非阿拉伯手语,很少有针对阿拉伯手语(ArSL)的,甚至更少的针对沙特手语(SSL)的。本文提出了一种移动应用程序,可以帮助沙特阿拉伯的聋人和听力障碍人士与他们的社区进行有效的沟通。该原型是一个基于Android的移动应用程序,应用深度学习技术将隔离的SSL转换为文本和音频,并包含其他针对ArSL的相关应用程序所没有的独特功能。拟议的方法,当在一个全面的数据集上评估时,通过超越几种最先进的方法并产生与这些方法相当的结果,证明了其有效性。此外,在几个聋哑和听力障碍用户身上测试原型,除了听力用户,证明了它的有用性。在未来,我们的目标是提高模型的准确性,并以更多的功能丰富应用。
    Deaf and hard-of-hearing people mainly communicate using sign language, which is a set of signs made using hand gestures combined with facial expressions to make meaningful and complete sentences. The problem that faces deaf and hard-of-hearing people is the lack of automatic tools that translate sign languages into written or spoken text, which has led to a communication gap between them and their communities. Most state-of-the-art vision-based sign language recognition approaches focus on translating non-Arabic sign languages, with few targeting the Arabic Sign Language (ArSL) and even fewer targeting the Saudi Sign Language (SSL). This paper proposes a mobile application that helps deaf and hard-of-hearing people in Saudi Arabia to communicate efficiently with their communities. The prototype is an Android-based mobile application that applies deep learning techniques to translate isolated SSL to text and audio and includes unique features that are not available in other related applications targeting ArSL. The proposed approach, when evaluated on a comprehensive dataset, has demonstrated its effectiveness by outperforming several state-of-the-art approaches and producing results that are comparable to these approaches. Moreover, testing the prototype on several deaf and hard-of-hearing users, in addition to hearing users, proved its usefulness. In the future, we aim to improve the accuracy of the model and enrich the application with more features.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    已经提出了几种深度学习辅助疾病评估方案(DAS),以增强对严重医疗紧急情况COVID-19的准确检测,通过对临床数据的分析。肺部成像,尤其是CT扫描,在确定和评估COVID-19感染的严重程度中起着关键作用。利用深度学习的现有自动化方法显著有助于减少与该过程相关的诊断负担。这项研究旨在开发一种用于COVID-19检测的简单DAS,使用应用于肺部CT切片的预训练轻量级深度学习方法(LDMs)。LDM的使用有助于不太复杂但高度准确的检测系统。开发的DAS的关键阶段包括图像收集和使用香农阈值的初始处理,LDM支持的深度特征挖掘,利用布朗蝴蝶算法(BBA)的特征优化,并通过三次交叉验证进行二元分类。拟议方案的绩效评估包括评估个人,融合,和合奏功能。调查显示,开发的DAS在具有单个特征的情况下实现了93.80%的检测精度,96%的精度与融合的特征,和一个令人印象深刻的99.10%的精度与合奏功能。这些结果肯定了所提出的方案在所选肺部CT数据库中显着提高COVID-19检测准确性的有效性。
    Several deep-learning assisted disease assessment schemes (DAS) have been proposed to enhance accurate detection of COVID-19, a critical medical emergency, through the analysis of clinical data. Lung imaging, particularly from CT scans, plays a pivotal role in identifying and assessing the severity of COVID-19 infections. Existing automated methods leveraging deep learning contribute significantly to reducing the diagnostic burden associated with this process. This research aims in developing a simple DAS for COVID-19 detection using the pre-trained lightweight deep learning methods (LDMs) applied to lung CT slices. The use of LDMs contributes to a less complex yet highly accurate detection system. The key stages of the developed DAS include image collection and initial processing using Shannon\'s thresholding, deep-feature mining supported by LDMs, feature optimization utilizing the Brownian Butterfly Algorithm (BBA), and binary classification through three-fold cross-validation. The performance evaluation of the proposed scheme involves assessing individual, fused, and ensemble features. The investigation reveals that the developed DAS achieves a detection accuracy of 93.80% with individual features, 96% accuracy with fused features, and an impressive 99.10% accuracy with ensemble features. These outcomes affirm the effectiveness of the proposed scheme in significantly enhancing COVID-19 detection accuracy in the chosen lung CT database.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在本文中,对轮盘异常检测进行了研究。在铁路领域,由于乘客的安全问题和维护问题,对实际铁路车辆进行测试是一项挑战,因为这是一个公共行业。因此,利用动力学软件。接下来,进行STFT(短时傅里叶变换)以产生谱图图像。在铁路车辆的情况下,control,监测,通过TCMS进行通信,但是复杂的分析和数据处理是困难的,因为没有像GPU这样的设备。此外,有记忆限制。因此,在本文中,选择了相对轻量级的LeNet-5、ResNet-20和MobileNet-V3模型进行深度学习实验。此时,LeNet-5和MobileNet-V3模型从基本架构进行了修改。由于对铁路车辆进行了预防性维护,很难获得故障数据。因此,还进行了半监督学习。此时,引用了DeepOneClass分类论文。评估结果表明,改进的LeNet-5和MobileNet-V3模型获得了大约97%和96%的准确率,分别。在这一点上,LeNet-5模型的训练时间比MobileNet-V3模型快12分钟。此外,半监督学习结果显示,当考虑铁路维护环境时,准确率约为94%。总之,考虑到铁路车辆维修环境和设备规格,推断,相对简单和轻量级的LeNet-5模型可以在使用小图像时有效地利用。
    In this paper, research was conducted on anomaly detection of wheel flats. In the railway sector, conducting tests with actual railway vehicles is challenging due to safety concerns for passengers and maintenance issues as it is a public industry. Therefore, dynamics software was utilized. Next, STFT (short-time Fourier transform) was performed to create spectrogram images. In the case of railway vehicles, control, monitoring, and communication are performed through TCMS, but complex analysis and data processing are difficult because there are no devices such as GPUs. Furthermore, there are memory limitations. Therefore, in this paper, the relatively lightweight models LeNet-5, ResNet-20, and MobileNet-V3 were selected for deep learning experiments. At this time, the LeNet-5 and MobileNet-V3 models were modified from the basic architecture. Since railway vehicles are given preventive maintenance, it is difficult to obtain fault data. Therefore, semi-supervised learning was also performed. At this time, the Deep One Class Classification paper was referenced. The evaluation results indicated that the modified LeNet-5 and MobileNet-V3 models achieved approximately 97% and 96% accuracy, respectively. At this point, the LeNet-5 model showed a training time of 12 min faster than the MobileNet-V3 model. In addition, the semi-supervised learning results showed a significant outcome of approximately 94% accuracy when considering the railway maintenance environment. In conclusion, considering the railway vehicle maintenance environment and device specifications, it was inferred that the relatively simple and lightweight LeNet-5 model can be effectively utilized while using small images.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    研究人员已经探索了ASD的各种潜在指标,包括大脑结构和活动的变化,遗传学,免疫系统异常,但是还没有找到确定的指标。因此,本研究旨在使用两种类型的磁共振图像(MRI)来调查ASD指标,结构(sMRI)和功能(fMRI),并解决数据可用性有限的问题。当处理有限的数据时,迁移学习是一种有价值的技术,因为它利用了从具有丰富数据的领域中的预训练模型获得的知识。这项研究提出了使用四个视觉变压器,即ConvNeXT,MobileNet,斯温,和ViT使用sMRI模式。该研究还调查了具有sMRI和fMRI模式的3D-CNN模型的使用。我们的实验涉及从沿轴向的原始3DsMRI和4DfMRI扫描中生成数据和提取切片的不同方法,日冕,和矢状大脑平面。为了评估我们的方法,我们利用来自ABIDE存储库的名为NYU的标准神经成像数据集对ASD受试者和典型对照受试者进行分类.我们的模型的性能是根据几个基线进行评估的,包括实施VGG和ResNet迁移学习模型的研究。我们的实验结果验证了使用3D-CNN和迁移学习方法进行多切片生成的有效性,因为它们实现了最先进的结果。特别是,来自fMRI和3D-CNN的50个中间切片的结果在ASD可分类性方面显示出深远的希望,因为当使用跨轴向4D图像的平均值时,它获得了0.8710的最大准确性和0.8261的F1评分。日冕,和矢状.此外,除了大脑视图的开始和结束之外,在fMRI中使用整个切片有助于减少无关信息,并显示出0.8387准确性和0.7727F1评分的良好表现.最后,使用ConvNeXt模型的迁移学习在沿轴向使用50个中间切片sMRI时获得的结果高于其他变压器,日冕,和矢状平面。
    Researchers have explored various potential indicators of ASD, including changes in brain structure and activity, genetics, and immune system abnormalities, but no definitive indicator has been found yet. Therefore, this study aims to investigate ASD indicators using two types of magnetic resonance images (MRI), structural (sMRI) and functional (fMRI), and to address the issue of limited data availability. Transfer learning is a valuable technique when working with limited data, as it utilizes knowledge gained from a pre-trained model in a domain with abundant data. This study proposed the use of four vision transformers namely ConvNeXT, MobileNet, Swin, and ViT using sMRI modalities. The study also investigated the use of a 3D-CNN model with sMRI and fMRI modalities. Our experiments involved different methods of generating data and extracting slices from raw 3D sMRI and 4D fMRI scans along the axial, coronal, and sagittal brain planes. To evaluate our methods, we utilized a standard neuroimaging dataset called NYU from the ABIDE repository to classify ASD subjects from typical control subjects. The performance of our models was evaluated against several baselines including studies that implemented VGG and ResNet transfer learning models. Our experimental results validate the effectiveness of the proposed multi-slice generation with the 3D-CNN and transfer learning methods as they achieved state-of-the-art results. In particular, results from 50-middle slices from the fMRI and 3D-CNN showed a profound promise in ASD classifiability as it obtained a maximum accuracy of 0.8710 and F1-score of 0.8261 when using the mean of 4D images across the axial, coronal, and sagittal. Additionally, the use of the whole slices in fMRI except the beginnings and the ends of brain views helped to reduce irrelevant information and showed good performance of 0.8387 accuracy and 0.7727 F1-score. Lastly, the transfer learning with the ConvNeXt model achieved results higher than other transformers when using 50-middle slices sMRI along the axial, coronal, and sagittal planes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    口腔病变是口腔疾病的普遍表现,及时识别口腔病变对有效干预势在必行。幸运的是,深度学习算法已显示出自动化病变检测的巨大潜力。这项研究的主要目的是采用基于深度学习的图像分类算法来识别口腔病变。我们使用了三种深度学习模型,即VGG19,DeIT,和MobileNet,评估各种分类方法的有效性。为了评估模型的准确性和可靠性,我们使用了一个由口腔图片组成的数据集,其中包含两个不同的类别:良性和恶性病变。实验结果表明,VGG19和MobileNet几乎达到了100%的完美准确率,而DeIT的准确率略低,为98.73%。这项研究的结果表明,用于图片分类的深度学习算法在检测口腔病变方面表现出很高的有效性,VGG19和MobileNet达到100%,DeIT达到98.73%。具体来说,VGG19和MobileNet模型对这一特定任务表现出显著的适用性。
    Oral lesions are a prevalent manifestation of oral disease, and the timely identification of oral lesions is imperative for effective intervention. Fortunately, deep learning algorithms have shown great potential for automated lesion detection. The primary aim of this study was to employ deep learning-based image classification algorithms to identify oral lesions. We used three deep learning models, namely VGG19, DeIT, and MobileNet, to assess the efficacy of various categorization methods. To evaluate the accuracy and reliability of the models, we employed a dataset consisting of oral pictures encompassing two distinct categories: benign and malignant lesions. The experimental findings indicate that VGG19 and MobileNet attained an almost perfect accuracy rate of 100%, while DeIT achieved a slightly lower accuracy rate of 98.73%. The results of this study indicate that deep learning algorithms for picture classification demonstrate a high level of effectiveness in detecting oral lesions by achieving 100% for VGG19 and MobileNet and 98.73% for DeIT. Specifically, the VGG19 and MobileNet models exhibit notable suitability for this particular task.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:在当代食品工业中,乌龙茶品种的准确和快速分化对于可追溯性和质量控制至关重要。然而,实现这一目标仍然是一个巨大的挑战。本研究通过采用机器学习算法-即支持向量机(SVM)和卷积神经网络(CNN)-以及基于视觉属性的乌龙茶叶自动分类的计算机视觉技术来解决这一问题。
    结果:13个不同的特征,包括颜色和纹理,从五个独特的乌龙茶品种中鉴定出。为了增强预测模型的鲁棒性,采用数据增强和图像裁剪方法。对基于SVM和CNN的模型的比较分析表明,ResNet50模型实现了超过93%的高Top1准确率。这种强大的性能证实了所实施的快速和精确的乌龙茶分类方法的有效性。
    结论:该研究阐明了计算机视觉与机器学习算法的集成,乌龙茶品种快速准确分类的非侵入性方法。这些发现对过程监控有重大影响,质量保证,真实性验证,和茶叶行业内的掺假检测。本文受版权保护。保留所有权利。
    BACKGROUND: In the contemporary food industry, accurate and rapid differentiation of oolong tea varieties holds paramount importance for traceability and quality control. However, achieving this remains a formidable challenge. This study addresses this lacuna by employing machine learning algorithms - namely support vector machines (SVMs) and convolutional neural networks (CNNs) - alongside computer vision techniques for the automated classification of oolong tea leaves based on visual attributes.
    RESULTS: An array of 13 distinct characteristics, encompassing color and texture, were identified from five unique oolong tea varieties. To fortify the robustness of the predictive models, data augmentation and image cropping methods were employed. A comparative analysis of SVM- and CNN-based models revealed that the ResNet50 model achieved a high Top-1 accuracy rate exceeding 93%. This robust performance substantiates the efficacy of the implemented methodology for rapid and precise oolong tea classification.
    CONCLUSIONS: The study elucidates that the integration of computer vision with machine learning algorithms constitutes a promising, non-invasive approach for the quick and accurate categorization of oolong tea varieties. The findings have significant ramifications for process monitoring, quality assurance, authenticity validation and adulteration detection within the tea industry. © 2023 Society of Chemical Industry.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    红外热成像(IRT)是一种用于诊断光伏(PV)装置以检测次优条件的技术。智能城市中光伏安装的增加引发了对改善IRT使用的技术的搜索,这要求辐照度条件大于700W/m2,使得在辐照度低于该值时无法使用。该项目提出了一个基于人工智能(AI)的物联网平台,该平台通过分析暴露于大于300W/m2的辐照度的模块之间的温差来自动检测光伏模块中的热点。为此,两个AI(深度学习和机器学习)在一个真实的光伏安装中进行了训练和测试,其中热点被诱导。该系统能够以0.995的灵敏度和0.923的精度在脏的情况下检测热点,短路,和部分阴影条件。该项目与其他项目不同,因为它提出了一种替代方案,以促进IRT诊断的实施并评估光伏组件的实际温度,这代表了光伏安装经理和检查员的潜在经济节约。
    Infrared thermography (IRT) is a technique used to diagnose Photovoltaic (PV) installations to detect sub-optimal conditions. The increase of PV installations in smart cities has generated the search for technology that improves the use of IRT, which requires irradiance conditions to be greater than 700 W/m2, making it impossible to use at times when irradiance goes under that value. This project presents an IoT platform working on artificial intelligence (AI) which automatically detects hot spots in PV modules by analyzing the temperature differentials between modules exposed to irradiances greater than 300 W/m2. For this purpose, two AI (Deep learning and machine learning) were trained and tested in a real PV installation where hot spots were induced. The system was able to detect hot spots with a sensitivity of 0.995 and an accuracy of 0.923 under dirty, short-circuited, and partially shaded conditions. This project differs from others because it proposes an alternative to facilitate the implementation of diagnostics with IRT and evaluates the real temperatures of PV modules, which represents a potential economic saving for PV installation managers and inspectors.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    脑肿瘤,以及其他伤害神经系统的疾病,是全球死亡率的重要贡献者。早期诊断在有效治疗脑肿瘤中起着至关重要的作用。为了区分有肿瘤的人和没有肿瘤的人,这项研究采用了图像和基于数据的特征的组合。在初始阶段,图像数据集得到增强,然后应用基于UNet迁移学习的模型将患者准确分类为患有肿瘤或正常。在第二阶段,这项研究利用13个特征结合投票分类器。投票分类器融合了从深度卷积层中提取的特征,并将随机梯度下降与逻辑回归相结合,以获得更好的分类结果。两种提出的模型所达到的0.99的准确度分数表明其优越的性能。此外,将结果与其他监督学习算法和最先进的模型进行比较,可以验证其性能。
    Brain tumors, along with other diseases that harm the neurological system, are a significant contributor to global mortality. Early diagnosis plays a crucial role in effectively treating brain tumors. To distinguish individuals with tumors from those without, this study employs a combination of images and data-based features. In the initial phase, the image dataset is enhanced, followed by the application of a UNet transfer-learning-based model to accurately classify patients as either having tumors or being normal. In the second phase, this research utilizes 13 features in conjunction with a voting classifier. The voting classifier incorporates features extracted from deep convolutional layers and combines stochastic gradient descent with logistic regression to achieve better classification results. The reported accuracy score of 0.99 achieved by both proposed models shows its superior performance. Also, comparing results with other supervised learning algorithms and state-of-the-art models validates its performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号