Feature Extraction

特征提取
  • 文章类型: Journal Article
    这项研究描述了一种对胶质瘤病理切片进行分级的新方法。我们自己的集成高光谱成像系统用于表征来自神经胶质瘤微阵列载玻片的270条带癌组织样本。然后根据世界卫生组织制定的指南对这些样本进行分类,定义了弥漫性神经胶质瘤的亚型和等级。我们使用不同恶性等级的脑胶质瘤的显微高光谱图像探索了一种称为SMLMER-ResNet的高光谱特征提取模型。该模型结合通道注意机制和多尺度图像特征,自动学习胶质瘤的病理组织,获得分层特征表示,有效去除冗余信息的干扰。它还完成了多模态,多尺度空间谱特征提取提高胶质瘤亚型的自动分类。所提出的分类方法具有较高的平均分类精度(>97.3%)和Kappa系数(0.954),表明其在提高高光谱胶质瘤自动分类方面的有效性。该方法很容易适用于广泛的临床环境。为减轻临床病理学家的工作量提供宝贵的帮助。此外,这项研究有助于制定更个性化和更精细的治疗计划,以及随后的随访和治疗调整,通过为医生提供对神经胶质瘤潜在病理组织的见解。
    This study describes a novel method for grading pathological sections of gliomas. Our own integrated hyperspectral imaging system was employed to characterize 270 bands of cancerous tissue samples from microarray slides of gliomas. These samples were then classified according to the guidelines developed by the World Health Organization, which define the subtypes and grades of diffuse gliomas. We explored a hyperspectral feature extraction model called SMLMER-ResNet using microscopic hyperspectral images of brain gliomas of different malignancy grades. The model combines the channel attention mechanism and multi-scale image features to automatically learn the pathological organization of gliomas and obtain hierarchical feature representations, effectively removing the interference of redundant information. It also completes multi-modal, multi-scale spatial-spectral feature extraction to improve the automatic classification of glioma subtypes. The proposed classification method demonstrated high average classification accuracy (>97.3%) and a Kappa coefficient (0.954), indicating its effectiveness in improving the automatic classification of hyperspectral gliomas. The method is readily applicable in a wide range of clinical settings, offering valuable assistance in alleviating the workload of clinical pathologists. Furthermore, the study contributes to the development of more personalized and refined treatment plans, as well as subsequent follow-up and treatment adjustment, by providing physicians with insights into the underlying pathological organization of gliomas.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    有效的特征提取和选择对于基于肌电信号的手势的准确分类和预测至关重要。在本文中,我们系统地比较了六种过滤器和包装器特征评估方法,并研究了它们各自对手势识别准确性的影响。调查基于几个基准数据集和一个真实的手势数据集,包括使用八个商业sEMG传感器从14名健康受试者收集的15个手部力量练习。从每个sEMG通道中提取了总共37个时域和频域特征。基准数据集显示,最小冗余最大相关性(mRMR)特征评估方法的性能最差,导致分类精度下降。然而,RFE方法证明了在大多数数据集中提高分类准确性的潜力。它选择了一个包含65个特征的特征子集,这导致了97.14%的准确率。互信息(MI)方法选择了200个特征,精度达到97.38%。特征重要性(FI)方法达到了97.62%的较高准确性,但选择了140个特征。进一步的研究表明,使用RFE方法选择65和75个特征导致97.14%的相同精度。对所选特征的彻底检查揭示了来自三个特定传感器的三个附加特征的潜力,以将分类准确性提高到97.38%。这些结果强调了采用适当的特征选择方法在保持分类准确性的同时显着减少必要特征数量的重要性。他们还强调了进一步分析和完善以实现最佳解决方案的必要性。
    Effective feature extraction and selection are crucial for the accurate classification and prediction of hand gestures based on electromyographic signals. In this paper, we systematically compare six filter and wrapper feature evaluation methods and investigate their respective impacts on the accuracy of gesture recognition. The investigation is based on several benchmark datasets and one real hand gesture dataset, including 15 hand force exercises collected from 14 healthy subjects using eight commercial sEMG sensors. A total of 37 time- and frequency-domain features were extracted from each sEMG channel. The benchmark dataset revealed that the minimum Redundancy Maximum Relevance (mRMR) feature evaluation method had the poorest performance, resulting in a decrease in classification accuracy. However, the RFE method demonstrated the potential to enhance classification accuracy across most of the datasets. It selected a feature subset comprising 65 features, which led to an accuracy of 97.14%. The Mutual Information (MI) method selected 200 features to reach an accuracy of 97.38%. The Feature Importance (FI) method reached a higher accuracy of 97.62% but selected 140 features. Further investigations have shown that selecting 65 and 75 features with the RFE methods led to an identical accuracy of 97.14%. A thorough examination of the selected features revealed the potential for three additional features from three specific sensors to enhance the classification accuracy to 97.38%. These results highlight the significance of employing an appropriate feature selection method to significantly reduce the number of necessary features while maintaining classification accuracy. They also underscore the necessity for further analysis and refinement to achieve optimal solutions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:声乐生物标志物,从声音特征的声学分析中得出,提供非侵入性的医疗筛查途径,诊断,和监测。先前的研究证明了通过智能手机记录语音的声学分析来预测2型糖尿病的可行性。在这项工作的基础上,这项研究探讨了音频数据压缩对声学声乐生物标志物开发的影响,这对于在医疗保健中更广泛的适用性至关重要。
    目的:本研究的目的是分析常见的音频压缩算法(MP3,M4A,和WMA)由3种不同的转换工具以2种比特率应用,影响对声音生物标志物检测至关重要的特征。
    方法:使用转换为MP3,M4A的未压缩语音样本,研究了音频数据压缩对声学声乐生物标志物开发的影响。和WMA格式在2比特率(320和128kbps)与MediaHuman(MH)音频转换器,WonderShare(WS)UniConverter,和快进运动图像专家组(FFmpeg)。数据集包括来自505名参与者的记录,总共17298个音频文件,使用智能手机收集。参与者每天记录一个固定的英语句子,最多6次,最长14天。特征提取,包括音高,抖动,强度,和梅尔频率倒谱系数(MFCC),是使用Python和Parselmouth进行的。使用Wilcoxon符号秩检验和Bonferroni校正进行多重比较用于统计分析。
    结果:在这项研究中,最初从505名参与者那里录制了36,970个音频文件,筛选后,有17298张录音符合固定的句子标准。音频转换软件之间的差异,MH,WS,和FFmpeg,值得注意的是,影响压缩结果,如恒定或可变比特率。分析包括不同的数据压缩格式和广泛的语音特征和MFCC。Wilcoxon符号秩检验得出P值,低于Bonferroni校正的显著性水平的那些表明由于压缩引起的显著改变。结果表明了跨格式和比特率的压缩的特定特征影响。与WS转换的文件相比,MH转换的文件表现出更大的弹性。比特率也影响了功能稳定性,38例唯一受单一比特率影响。值得注意的是,语音特征在各种转换方法中显示出比MFCC更高的稳定性。
    结论:发现压缩效果具有特定特征,MH和FFmpeg表现出更大的弹性。某些功能一直受到影响,强调理解特征弹性对诊断应用的重要性。考虑到声乐生物标志物在医疗保健中的实施,为数据存储或传输目的找到通过压缩保持一致的功能是很有价值的。专注于特定的功能和格式,未来的研究可以拓宽范围,包括不同的特征,实时压缩算法,和各种记录方法。这项研究增强了我们对音频压缩对语音特征和MFCC的影响的理解,为跨领域开发应用程序提供见解。该研究强调了特征稳定性在处理压缩音频数据中的重要性,为在不断发展的技术环境中使用明智的语音数据奠定基础。
    BACKGROUND: Vocal biomarkers, derived from acoustic analysis of vocal characteristics, offer noninvasive avenues for medical screening, diagnostics, and monitoring. Previous research demonstrated the feasibility of predicting type 2 diabetes mellitus through acoustic analysis of smartphone-recorded speech. Building upon this work, this study explores the impact of audio data compression on acoustic vocal biomarker development, which is critical for broader applicability in health care.
    OBJECTIVE: The objective of this research is to analyze how common audio compression algorithms (MP3, M4A, and WMA) applied by 3 different conversion tools at 2 bitrates affect features crucial for vocal biomarker detection.
    METHODS: The impact of audio data compression on acoustic vocal biomarker development was investigated using uncompressed voice samples converted into MP3, M4A, and WMA formats at 2 bitrates (320 and 128 kbps) with MediaHuman (MH) Audio Converter, WonderShare (WS) UniConverter, and Fast Forward Moving Picture Experts Group (FFmpeg). The data set comprised recordings from 505 participants, totaling 17,298 audio files, collected using a smartphone. Participants recorded a fixed English sentence up to 6 times daily for up to 14 days. Feature extraction, including pitch, jitter, intensity, and Mel-frequency cepstral coefficients (MFCCs), was conducted using Python and Parselmouth. The Wilcoxon signed rank test and the Bonferroni correction for multiple comparisons were used for statistical analysis.
    RESULTS: In this study, 36,970 audio files were initially recorded from 505 participants, with 17,298 recordings meeting the fixed sentence criteria after screening. Differences between the audio conversion software, MH, WS, and FFmpeg, were notable, impacting compression outcomes such as constant or variable bitrates. Analysis encompassed diverse data compression formats and a wide array of voice features and MFCCs. Wilcoxon signed rank tests yielded P values, with those below the Bonferroni-corrected significance level indicating significant alterations due to compression. The results indicated feature-specific impacts of compression across formats and bitrates. MH-converted files exhibited greater resilience compared to WS-converted files. Bitrate also influenced feature stability, with 38 cases affected uniquely by a single bitrate. Notably, voice features showed greater stability than MFCCs across conversion methods.
    CONCLUSIONS: Compression effects were found to be feature specific, with MH and FFmpeg showing greater resilience. Some features were consistently affected, emphasizing the importance of understanding feature resilience for diagnostic applications. Considering the implementation of vocal biomarkers in health care, finding features that remain consistent through compression for data storage or transmission purposes is valuable. Focused on specific features and formats, future research could broaden the scope to include diverse features, real-time compression algorithms, and various recording methods. This study enhances our understanding of audio compression\'s influence on voice features and MFCCs, providing insights for developing applications across fields. The research underscores the significance of feature stability in working with compressed audio data, laying a foundation for informed voice data use in evolving technological landscapes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    由于对社会和经济的重要性,财务困境识别仍然是科学文献中的重要主题。信息技术的进步和存储数据量的不断增加导致财务困境的出现,超越了财务报表及其指标(比率)的范围。特征空间可以通过纳入宏观经济学等特征数据类别的新观点来扩展,部门,社会,董事会,管理,司法事件,等。然而,维度的增加导致数据稀疏和模型过度拟合。本研究通过结合降维和机器学习技术,提出了一种有效的财务困境分类评估的新方法。拟议的框架旨在确定导致描述企业财务困境的损失函数最小化的特征子集。在研究期间,比较了15种具有不同特征数量的降维技术和17种机器学习模型。总的来说,使用2015年至2022年期间的立陶宛企业数据进行了1,432次实验。结果表明,使用随机森林均值递减Gini(RF_MDG)特征选择技术识别的具有30个排名特征的人工神经网络(ANN)模型提供了最高的AUC得分。此外,这项研究引入了一种新的特征提取方法,这可以改进财务困境分类模型。
    Financial distress identification remains an essential topic in the scientific literature due to its importance for society and the economy. The advancements in information technology and the escalating volume of stored data have led to the emergence of financial distress that transcends the realm of financial statements and its\' indicators (ratios). The feature space could be expanded by incorporating new perspectives on feature data categories such as macroeconomics, sectors, social, board, management, judicial incident, etc. However, the increased dimensionality results in sparse data and overfitted models. This study proposes a new approach for efficient financial distress classification assessment by combining dimensionality reduction and machine learning techniques. The proposed framework aims to identify a subset of features leading to the minimization of the loss function describing the financial distress in an enterprise. During the study, 15 dimensionality reduction techniques with different numbers of features and 17 machine-learning models were compared. Overall, 1,432 experiments were performed using Lithuanian enterprise data covering the period from 2015 to 2022. Results revealed that the artificial neural network (ANN) model with 30 ranked features identified using the Random Forest mean decreasing Gini (RF_MDG) feature selection technique provided the highest AUC score. Moreover, this study has introduced a novel approach for feature extraction, which could improve financial distress classification models.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在动态健康监测设置中自动和早期发现患有多发性神经病的糖尿病患者可能会减少糖尿病患者的主要危险因素。与疼痛和温度受损相关的增加和局部足底压力是多发性神经病患者中发展足部溃疡的组合。尽管在这一领域已经报道了许多有趣的研究工作,它们中的大多数强调信号采集过程和足底区域的足底压力分布。在这项工作中,使用足底压力和温度信号开发了一种机器学习辅助的低复杂性技术,该技术将对糖尿病性多发性神经病和健康受试者进行分类。分别使用主成分分析(PCA)和最大相关最小冗余(mRMR)方法进行特征提取和选择,然后使用k-NN分类器进行二元分类。使用来自43名受试者的100分钟公开可用的注释数据对所提出的技术进行了评估,并提供了盲测准确性。灵敏度,精度,F1分数,曲线下面积(AUC)为99.58%,99.50%,99.44%,分别为99.47%和99.56%。ARMv6控制器中的低资源硬件实现需要81.2kB的平均内存使用量和1.31s的延迟来处理从每个脚部区域的16个传感器通道收集的9s压力和温度数据。
    Automated and early detection of diabetics with polyneuropathy in an ambulatory health monitoring setup may reduce the major risk factors for diabetic patients. Increased and localized plantar pressure associated with impaired pain and temperature is a combination of developing foot ulcers in subjects with polyneuropathy. Although many interesting research works have been reported in this area, most of them emphasize on signal acquisition process and plantar pressure distribution in the foot region. In this work, a machine learning assisted low complexity technique was developed using plantar pressure and temperature signals which will classify between diabetic polyneuropathy and healthy subjects. Principal component analysis (PCA) and maximum relevance minimum redundancy (mRMR) methods were used for feature extraction and selection respectively followed by k-NN classifier for binary classification. The proposed technique was evaluated with 100 min of publicly available annotated data from 43 subjects and provides blind test accuracy, sensitivity, precision, F1-score, and area under curve (AUC) of 99.58%, 99.50%, 99.44%, 99.47% and 99.56% respectively. A low resource hardware implementation in ARM v6 controller required an average memory usage of 81.2 kB and latency of 1.31 s to process 9 s pressure and temperature data collected from 16 sensor channels for each of the foot region.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    脑-计算机接口(BCI)在人机协作中的新兴集成有望实现动态自适应交互。在辅助设备中使用脑电图(EEG)测量误差相关电位(ErrP)进行在线误差检测提供了一种提高此类设备可靠性的实用方法。然而,连续在线错误检测面临挑战,例如开发高效和轻量级的分类技术以进行快速预测,减少来自人工制品的错误警报,并处理脑电信号的非平稳性。进一步的研究对于解决在线会话中连续分类的复杂性至关重要。通过这项研究,我们展示了一种基于连续在线EEG的机器错误检测的综合方法,在第32届国际人工智能联席会议上成为比赛的获胜者。比赛包括两个阶段:使用预先录制的模型开发的离线阶段,标记的脑电图数据,线下阶段3个月后的在线阶段,这些模型在连续流的EEG数据上进行了实时测试,以实时检测矫形器运动中的错误。我们的方法结合了两个时间导数特征和基于效果大小的特征选择技术,用于模型训练,以及一种用于在线会话的轻量级噪声滤波方法,无需重新校准模型。在离线阶段训练的模型不仅导致所有参与者的平均交叉验证准确率高达89.9%,但在最初的数据收集后3个月的在线会议期间也表现出了显著的性能,而无需进一步校准,保持1.7%的低总体误报率和快速反应能力。我们的研究为该领域做出了两个重要贡献。首先,它证明了用基于效果大小的特征选择策略集成两个时间导数特征的可行性,特别是在基于在线脑电图的BCI中。其次,我们的工作介绍了一种创新的方法,设计用于连续在线误差预测,其中包括一个简单的噪声抑制技术,以减少误报。这项研究是对无缝错误检测方法的可行性研究,该方法有望在神经自适应技术和人机交互领域转变实际应用。
    The emerging integration of Brain-Computer Interfaces (BCIs) in human-robot collaboration holds promise for dynamic adaptive interaction. The use of electroencephalogram (EEG)-measured error-related potentials (ErrPs) for online error detection in assistive devices offers a practical method for improving the reliability of such devices. However, continuous online error detection faces challenges such as developing efficient and lightweight classification techniques for quick predictions, reducing false alarms from artifacts, and dealing with the non-stationarity of EEG signals. Further research is essential to address the complexities of continuous classification in online sessions. With this study, we demonstrated a comprehensive approach for continuous online EEG-based machine error detection, which emerged as the winner of a competition at the 32nd International Joint Conference on Artificial Intelligence. The competition consisted of two stages: an offline stage for model development using pre-recorded, labeled EEG data, and an online stage 3 months after the offline stage, where these models were tested live on continuously streamed EEG data to detect errors in orthosis movements in real time. Our approach incorporates two temporal-derivative features with an effect size-based feature selection technique for model training, together with a lightweight noise filtering method for online sessions without recalibration of the model. The model trained in the offline stage not only resulted in a high average cross-validation accuracy of 89.9% across all participants, but also demonstrated remarkable performance during the online session 3 months after the initial data collection without further calibration, maintaining a low overall false alarm rate of 1.7% and swift response capabilities. Our research makes two significant contributions to the field. Firstly, it demonstrates the feasibility of integrating two temporal derivative features with an effect size-based feature selection strategy, particularly in online EEG-based BCIs. Secondly, our work introduces an innovative approach designed for continuous online error prediction, which includes a straightforward noise rejection technique to reduce false alarms. This study serves as a feasibility investigation into a methodology for seamless error detection that promises to transform practical applications in the domain of neuroadaptive technology and human-robot interaction.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    深度学习深刻影响了各个领域,特别是医学图像分析。该领域的传统迁移学习方法依赖于在特定领域的医学数据集上预训练的模型,这限制了它们的通用性和可访问性。在这项研究中,我们提出了一个叫做真实世界特征迁移学习的新框架,它利用最初在大规模通用数据集如ImageNet上训练的骨干模型。与从头开始训练的模型相比,我们评估了这种方法的有效性和鲁棒性,专注于对X射线图像中的肺炎进行分类的任务。我们的实验,其中包括将灰度图像转换为RGB格式,证明了真实世界的特征迁移学习在各种性能指标上始终优于传统的训练方法。这一进步有可能通过利用从通用预训练模型学习的丰富特征表示来加速医学成像中的深度学习应用。所提出的方法克服了特定领域预训练模型的局限性,从而加速医疗诊断和医疗保健领域的创新。从数学的角度来看,我们形式化现实世界的特征迁移学习的概念,并提供了一个严格的数学公式的问题。我们的实验结果提供了支持这种方法有效性的经验证据,为进一步的理论分析和探索奠定基础。这项工作有助于更广泛地理解跨域的特征可转移性,并对开发准确有效的医学图像分析模型具有重要意义。即使在资源受限的环境中。
    Deep learning has profoundly influenced various domains, particularly medical image analysis. Traditional transfer learning approaches in this field rely on models pretrained on domain-specific medical datasets, which limits their generalizability and accessibility. In this study, we propose a novel framework called real-world feature transfer learning, which utilizes backbone models initially trained on large-scale general-purpose datasets such as ImageNet. We evaluate the effectiveness and robustness of this approach compared to models trained from scratch, focusing on the task of classifying pneumonia in X-ray images. Our experiments, which included converting grayscale images to RGB format, demonstrate that real-world-feature transfer learning consistently outperforms conventional training approaches across various performance metrics. This advancement has the potential to accelerate deep learning applications in medical imaging by leveraging the rich feature representations learned from general-purpose pretrained models. The proposed methodology overcomes the limitations of domain-specific pretrained models, thereby enabling accelerated innovation in medical diagnostics and healthcare. From a mathematical perspective, we formalize the concept of real-world feature transfer learning and provide a rigorous mathematical formulation of the problem. Our experimental results provide empirical evidence supporting the effectiveness of this approach, laying the foundation for further theoretical analysis and exploration. This work contributes to the broader understanding of feature transferability across domains and has significant implications for the development of accurate and efficient models for medical image analysis, even in resource-constrained settings.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    随着遥感技术的飞速发展,从高光谱遥感图像中获得的光谱信息越来越丰富,促进地球表面物体的详细光谱分析。然而,光谱信息的丰富对数据处理提出了一定的挑战,例如导致“休斯现象”的“维度诅咒”,“强相关性”由于高分辨率,和由变化的表面反射率引起的“非线性特性”。因此,高光谱数据的降维成为一项关键任务。本文首先阐述了基于流形理论和学习方法的高光谱图像降维的原理和过程,鉴于高光谱遥感数据中存在的非线性结构和特征,并制定了基于流形学习的降维过程。随后,这项研究探索了使用流形学习方法对高光谱图像进行特征提取和低维嵌入的能力,包括主成分分析(PCA),多维缩放(MDS),和线性方法的线性判别分析(LDA);和等距映射(Isomap),局部线性嵌入(LLE),拉普拉斯特征映射(LE),Hessian局部线性嵌入(HLLE),局部切线空间对齐(LTSA),和非线性方法的最大方差展开(MVU),基于印度松树高光谱数据集和帕维亚大学数据集。此外,本文研究了高光谱图像特征提取的最佳邻域计算时间和整体算法运行时间,不同流形学习方法的邻域k和固有维数d值的选择不同。基于特征提取的结果,本研究考察了各种流形学习方法的分类实验,比较和分析分类精度和Kappa系数随邻域k和固有维数d值的不同选择的变化。在这个基础上,LE方法中选择不同的高斯核带宽t和MVU方法选择不同的拉格朗日乘数λ对分类精度的影响,给定邻域k和固有维数d的不同选择,正在探索。通过这些实验,本文研究了不同流形学习方法在高光谱图像中特征提取和降维的能力和有效性,受邻域k和固有维数d值选择的影响,确定每种方法的最优邻域k和本征维数d值。分类精度的比较表明,与其他流形学习方法相比,LTSA方法可产生更出色的分类结果。该研究证明了流形学习方法在处理高光谱图像数据中的优势,流形学习方法为后续高光谱图像降维研究提供实验参考。
    With the rapid advancement of remote-sensing technology, the spectral information obtained from hyperspectral remote-sensing imagery has become increasingly rich, facilitating detailed spectral analysis of Earth\'s surface objects. However, the abundance of spectral information presents certain challenges for data processing, such as the \"curse of dimensionality\" leading to the \"Hughes phenomenon\", \"strong correlation\" due to high resolution, and \"nonlinear characteristics\" caused by varying surface reflectances. Consequently, dimensionality reduction of hyperspectral data emerges as a critical task. This paper begins by elucidating the principles and processes of hyperspectral image dimensionality reduction based on manifold theory and learning methods, in light of the nonlinear structures and features present in hyperspectral remote-sensing data, and formulates a dimensionality reduction process based on manifold learning. Subsequently, this study explores the capabilities of feature extraction and low-dimensional embedding for hyperspectral imagery using manifold learning approaches, including principal components analysis (PCA), multidimensional scaling (MDS), and linear discriminant analysis (LDA) for linear methods; and isometric mapping (Isomap), locally linear embedding (LLE), Laplacian eigenmaps (LE), Hessian locally linear embedding (HLLE), local tangent space alignment (LTSA), and maximum variance unfolding (MVU) for nonlinear methods, based on the Indian Pines hyperspectral dataset and Pavia University dataset. Furthermore, the paper investigates the optimal neighborhood computation time and overall algorithm runtime for feature extraction in hyperspectral imagery, varying by the choice of neighborhood k and intrinsic dimensionality d values across different manifold learning methods. Based on the outcomes of feature extraction, the study examines the classification experiments of various manifold learning methods, comparing and analyzing the variations in classification accuracy and Kappa coefficient with different selections of neighborhood k and intrinsic dimensionality d values. Building on this, the impact of selecting different bandwidths t for the Gaussian kernel in the LE method and different Lagrange multipliers λ for the MVU method on classification accuracy, given varying choices of neighborhood k and intrinsic dimensionality d, is explored. Through these experiments, the paper investigates the capability and effectiveness of different manifold learning methods in feature extraction and dimensionality reduction within hyperspectral imagery, as influenced by the selection of neighborhood k and intrinsic dimensionality d values, identifying the optimal neighborhood k and intrinsic dimensionality d value for each method. A comparison of classification accuracies reveals that the LTSA method yields superior classification results compared to other manifold learning approaches. The study demonstrates the advantages of manifold learning methods in processing hyperspectral image data, providing an experimental reference for subsequent research on hyperspectral image dimensionality reduction using manifold learning methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    近年来,自动睡眠分析的研究已经见证了显著的增长,反映了在理解睡眠模式及其对整体健康影响方面的进步。这篇综述综合了87篇论文的详尽分析结果,系统地从著名的数据库中检索,如谷歌学者,PubMed,IEEEXplore,和科学直接。选择标准优先研究采用的方法,利用的信号模态,和机器学习算法应用于自动睡眠分析。总体目标是批判性地评估拟议方法的优缺点,揭示了睡眠研究的当前景观和未来方向。对综述文献的深入探索揭示了自动化睡眠研究中采用的各种方法和机器学习方法。值得注意的是,K-最近邻居(KNN),合奏学习方法,支持向量机(SVM)作为多功能和有效的分类器出现,在各种应用中表现出高精度。然而,观察到性能可变性和计算需求等挑战,需要根据数据集的复杂性进行明智的分类器选择。此外,在睡眠相关研究中,传统特征提取方法与深层结构的整合以及不同深度神经网络的组合被认为是提高诊断准确性的有前景的策略.回顾的文献强调了对自适应分类器的需求,跨模态集成,合作努力推动该领域变得更加准确,健壮,和可访问的睡眠相关诊断解决方案。这项全面的审查为研究人员和从业人员奠定了坚实的基础,提供自动睡眠分析中知识的当前状态的有组织的综合。通过强调各种方法的优势和挑战,这篇综述旨在指导未来的研究朝着更有效和更细致的方法进行睡眠诊断。
    In recent years, research on automated sleep analysis has witnessed significant growth, reflecting advancements in understanding sleep patterns and their impact on overall health. This review synthesizes findings from an exhaustive analysis of 87 papers, systematically retrieved from prominent databases such as Google Scholar, PubMed, IEEE Xplore, and ScienceDirect. The selection criteria prioritized studies focusing on methods employed, signal modalities utilized, and machine learning algorithms applied in automated sleep analysis. The overarching goal was to critically evaluate the strengths and weaknesses of the proposed methods, shedding light on the current landscape and future directions in sleep research. An in-depth exploration of the reviewed literature revealed a diverse range of methodologies and machine learning approaches employed in automated sleep studies. Notably, K-Nearest Neighbors (KNN), Ensemble Learning Methods, and Support Vector Machine (SVM) emerged as versatile and potent classifiers, exhibiting high accuracies in various applications. However, challenges such as performance variability and computational demands were observed, necessitating judicious classifier selection based on dataset intricacies. In addition, the integration of traditional feature extraction methods with deep structures and the combination of different deep neural networks were identified as promising strategies to enhance diagnostic accuracy in sleep-related studies. The reviewed literature emphasized the need for adaptive classifiers, cross-modality integration, and collaborative efforts to drive the field toward more accurate, robust, and accessible sleep-related diagnostic solutions. This comprehensive review serves as a solid foundation for researchers and practitioners, providing an organized synthesis of the current state of knowledge in automated sleep analysis. By highlighting the strengths and challenges of various methodologies, this review aims to guide future research toward more effective and nuanced approaches to sleep diagnostics.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    白细胞分化的金标准是手工检查血涂片,这不仅是时间和劳动力密集型的,而且容易受到人为错误的影响。至于自动分类,仍然没有细胞分割的比较研究,特征提取,和细胞分类,将各种机器和深度学习模型与家庭开发的方法进行比较。在这项研究中,传统的K均值聚类的机器学习与U-Net的深度学习,使用U-Net+ResNet18和U-Net+ResNet34进行细胞分割,CellaVision数据集的分割精度为94.36%与99.17%,BCCD数据集的分割精度为93.20%与98.75%,证实深度学习在白细胞分类方面比传统机器学习产生更高的性能。此外,一系列深度学习方法,包括AlexNet,采用VGG16和ResNet18进行白细胞的特征提取和细胞分类,产生91.31%的分类准确率,97.83%,和100%的CellaVision以及81.18%,BCCD的91.64%和97.82%,证实了神经网络在白细胞分类中增加深度的能力。至于示威,本研究进一步对ALL-IDB2和PCB-HBC数据集进行了细胞类型分类,在所有文献中产生100%和98.49%的高精度,验证本研究中使用的深度学习模型。
    The gold standard of leukocyte differentiation is a manual examination of blood smears, which is not only time and labor intensive but also susceptible to human error. As to automatic classification, there is still no comparative study of cell segmentation, feature extraction, and cell classification, where a variety of machine and deep learning models are compared with home-developed approaches. In this study, both traditional machine learning of K-means clustering versus deep learning of U-Net, U-Net + ResNet18, and U-Net + ResNet34 were used for cell segmentation, producing segmentation accuracies of 94.36% versus 99.17% for the dataset of CellaVision and 93.20% versus 98.75% for the dataset of BCCD, confirming that deep learning produces higher performance than traditional machine learning in leukocyte classification. In addition, a series of deep-learning approaches, including AlexNet, VGG16, and ResNet18, was adopted to conduct feature extraction and cell classification of leukocytes, producing classification accuracies of 91.31%, 97.83%, and 100% of CellaVision as well as 81.18%, 91.64% and 97.82% of BCCD, confirming the capability of the increased deepness of neural networks in leukocyte classification. As to the demonstrations, this study further conducted cell-type classification of ALL-IDB2 and PCB-HBC datasets, producing high accuracies of 100% and 98.49% among all literature, validating the deep learning model used in this study.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号