GCN

GCN
  • 文章类型: Journal Article
    内存取证与深度学习相结合的恶意软件检测技术取得了一定进展,但是大多数现有方法将进程转储转换为图像进行分类,仍基于进程字节特征分类。恶意软件加载到内存后,原来的字节特征将改变。与字节特征相比,函数调用功能可以更有力地表示恶意软件的行为。因此,本文提出了ProcGCN模型,基于DGCNN(深图卷积神经网络)的深度学习模型,检测内存映像中的恶意进程。首先,从整个系统内存映像中提取进程转储;然后,提取过程的函数调用图(FCG),和基于词袋模型生成FCG中函数节点的特征向量;最后,FCG被输入到ProcGCN模型用于分类和检测。使用公共数据集进行实验,ProcGCN模型的准确率为98.44%,F1得分为0.9828.它显示了比现有的基于静态特征的深度学习方法更好的结果,它的检测速度更快,函数调用特征和图表示学习方法在记忆取证中的有效性。
    The combination of memory forensics and deep learning for malware detection has achieved certain progress, but most existing methods convert process dump to images for classification, which is still based on process byte feature classification. After the malware is loaded into memory, the original byte features will change. Compared with byte features, function call features can represent the behaviors of malware more robustly. Therefore, this article proposes the ProcGCN model, a deep learning model based on DGCNN (Deep Graph Convolutional Neural Network), to detect malicious processes in memory images. First, the process dump is extracted from the whole system memory image; then, the Function Call Graph (FCG) of the process is extracted, and feature vectors for the function node in the FCG are generated based on the word bag model; finally, the FCG is input to the ProcGCN model for classification and detection. Using a public dataset for experiments, the ProcGCN model achieved an accuracy of 98.44% and an F1 score of 0.9828. It shows a better result than the existing deep learning methods based on static features, and its detection speed is faster, which demonstrates the effectiveness of the method based on function call features and graph representation learning in memory forensics.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    卷积神经网络(CNN)在各个领域都表现出了出色的性能,比如人脸识别,物体检测,和图像分割。然而,CNN固有的缺乏透明度和有限的可解释性给医疗诊断等领域带来了挑战,自主驾驶,金融,和军事应用。一些研究探索了CNN的可解释性,并提出了各种事后可解释的方法。这些方法中的大多数是基于特征的,关注输入变量对输出的影响。很少有方法对CNN中的参数及其整体结构进行分析。为了探索CNN的结构并直观地理解其内部参数的作用,我们提出了一种基于归因图的CNN可解释方法(AGIC),该方法将CNN的整体结构建模为图形,并从全局和局部角度提供可解释性。CNN的运行时参数和每个图像样本的特征图被应用于构造归因图(At-GC),其中卷积内核表示为节点,内核输出之间的SHAP值被分配为边。然后使用这些At-GC来基于DeepGraphInfomax(DGI)预训练新设计的异构图编码器。为了全面研究CNN的整体结构,预训练的编码器用于两种类型的可解释任务:(1)一个分类器被附加到预训练的编码器,用于对At-GC进行分类,揭示了At-GC\的拓扑特征对图像样本类别的依赖性,(2)构建了一个评分聚合(SA)网络来评估At-GC中每个节点的重要性,从而反映出内核在CNN中的相对重要性。实验结果表明,At-GC的拓扑特征对其构造中使用的样本类别具有依赖性,这表明CNN中的内核显示出不同的组合激活模式,用于处理不同的图像类别,同时,从SA网络获得高分的内核对于特征提取至关重要,而低得分的内核可以在不影响模型性能的情况下被修剪,从而增强CNN的可解释性。
    Convolutional Neural Networks (CNNs) have demonstrated outstanding performance in various domains, such as face recognition, object detection, and image segmentation. However, the lack of transparency and limited interpretability inherent in CNNs pose challenges in fields such as medical diagnosis, autonomous driving, finance, and military applications. Several studies have explored the interpretability of CNNs and proposed various post-hoc interpretable methods. The majority of these methods are feature-based, focusing on the influence of input variables on outputs. Few methods undertake the analysis of parameters in CNNs and their overall structure. To explore the structure of CNNs and intuitively comprehend the role of their internal parameters, we propose an Attribution Graph-based Interpretable method for CNNs (AGIC) which models the overall structure of CNNs as graphs and provides interpretability from global and local perspectives. The runtime parameters of CNNs and feature maps of each image sample are applied to construct attribution graphs (At-GCs), where the convolutional kernels are represented as nodes and the SHAP values between kernel outputs are assigned as edges. These At-GCs are then employed to pretrain a newly designed heterogeneous graph encoder based on Deep Graph Infomax (DGI). To comprehensively delve into the overall structure of CNNs, the pretrained encoder is used for two types of interpretable tasks: (1) a classifier is attached to the pretrained encoder for the classification of At-GCs, revealing the dependency of At-GC\'s topological characteristics on the image sample categories, and (2) a scoring aggregation (SA) network is constructed to assess the importance of each node in At-GCs, thus reflecting the relative importance of kernels in CNNs. The experimental results indicate that the topological characteristics of At-GC exhibit a dependency on the sample category used in its construction, which reveals that kernels in CNNs show distinct combined activation patterns for processing different image categories, meanwhile, the kernels that receive high scores from SA network are crucial for feature extraction, whereas low-scoring kernels can be pruned without affecting model performance, thereby enhancing the interpretability of CNNs.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    非欧几里得数据,例如社交网络和文档之间的引用关系,具有节点和结构信息。图卷积网络(GCN)可以自动学习节点特征和节点之间的关联信息。图卷积网络的核心思想是利用边缘信息聚合节点信息,从而生成新的节点特征。在更新节点特征时,有两个核心影响因素。一个是中心节点的相邻节点的数量;另一个是相邻节点对中心节点的贡献。由于以前的GCN方法没有同时考虑相邻节点对中心节点的数量和不同贡献,我们设计了自适应注意力机制(AAM)。为了进一步增强模型的表示能力,我们利用多头图卷积(MHGC)。最后,我们采用交叉熵(CE)损失函数来描述节点类别的预测结果与地面实况(GT)之间的差异。结合反向传播,这最终实现了节点的准确分类。基于AAM,MHGC,CE,我们设计了新颖的图形自适应注意力网络(GAAN)。实验表明,分类精度在Cora上取得了突出的性能,Citeseer,和发布的数据集。
    Non-Euclidean data, such as social networks and citation relationships between documents, have node and structural information. The Graph Convolutional Network (GCN) can automatically learn node features and association information between nodes. The core ideology of the Graph Convolutional Network is to aggregate node information by using edge information, thereby generating a new node feature. In updating node features, there are two core influencing factors. One is the number of neighboring nodes of the central node; the other is the contribution of the neighboring nodes to the central node. Due to the previous GCN methods not simultaneously considering the numbers and different contributions of neighboring nodes to the central node, we design the adaptive attention mechanism (AAM). To further enhance the representational capability of the model, we utilize Multi-Head Graph Convolution (MHGC). Finally, we adopt the cross-entropy (CE) loss function to describe the difference between the predicted results of node categories and the ground truth (GT). Combined with backpropagation, this ultimately achieves accurate node classification. Based on the AAM, MHGC, and CE, we contrive the novel Graph Adaptive Attention Network (GAAN). The experiments show that classification accuracy achieves outstanding performances on Cora, Citeseer, and Pubmed datasets.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    用金属纳米颗粒突变的石墨氮化碳作为用于检测水中存在的有害离子的有效荧光传感器引起了极大的兴趣。在目前的工作中,以三聚氰胺为前体合成了bulk-gCN,并进一步通过原位直接还原沉积法制备了Au-gCN纳米复合材料。结构,形态学,组成,使用各种散射和光谱技术(如HRTEM)检查了块状gCN和Au-gCN纳米复合材料的稳定性和光学性能,XPS,XRD和SEM。由于其不均匀的表面形态,在对不同阳离子和阴离子的选择性研究中,合成的块状gCN散乱。然而,在Au-gCN中,金纳米颗粒均匀分布在gCN片上,这导致其相对于本体gCN的选择性增强。这导致了Fe3和Cr2离子的光学传感器的制造,检测极限为4.62和2.77μM,分别。Fe3+离子的传感对应于光诱导电子转移(PET)机制,而铬酸盐种类的检测与内部过滤效应(IFE)有关。还评估了该传感器对不同环境水样的实际适用性。稳定性高,灵敏度,Au-gCN纳米复合材料的特异性使其成为水样中Fe3+和Cr2离子的潜在荧光探针。
    Graphitic carbon nitride mutated with metal nanoparticles has captivated great interest as an effective fluorescent sensor for the detection of harmful ions present in water. In the present work, bulk-gCN was synthesized using melamine as precursor, and further Au-gCN nanocomposite were fabricated via in-situ direct reduction deposition method. The structural, morphological, compositional, stability and optical properties of bulk gCN and Au-gCN nanocomposite were examined using various scattering and spectroscopic techniques such as HRTEM, XPS, XRD and SEM. The synthesized bulk gCN straggles during selectivity studies with different cations and anions because of its uneven surface morphology, however in Au-gCN gold nanoparticles are uniformly distributed on the gCN sheets which results in its enhanced selectivity over bulk gCN. This leads to the fabrication of an optical sensor for Fe3+ and Cr2O72- ions with limit of detection of 4.62 and 2.77 μM, respectively. The sensing of Fe3+ ions corresponds to the photoinduced electron transfer (PET) mechanism, while the detection of chromate species is associated with an inner filter effect (IFE). The practical applicability of the sensor was also evaluated for different environmental water samples. The high stability, sensitivity, and specificity of Au-gCN nanocomposite make it a potential fluorescent probe for Fe3+ and Cr2O72- ions in water samples.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    破译植物中microRNAs(miRNAs)的靶标对于理解它们的功能和它们引起的表型变异至关重要。由于miRNA调控的高度细胞特异性,最近的计算方法通常利用表达数据来识别最生理相关的目标。虽然这些方法是有效的,它们通常需要大样本量和高深度测序来检测潜在的miRNA-靶标对,从而限制了它们在改进植物育种中的适用性。在这项研究中,我们提出了一种名为kmerPMTF(基于k-mer的植物miRNA-target预测框架)的新的miRNA-target预测框架。我们的框架通过利用k-mer分裂和深度自监督神经网络有效地提取序列的潜在语义嵌入。我们基于k-mer嵌入构建多个相似性网络,并采用图卷积网络来导出miRNA和靶标的深度表示,并计算潜在关联的概率。我们在四个典型的植物数据集上评估了kmerPMTF的性能:拟南芥,水稻,番茄红素,还有Prunuspersica.结果表明,它能够达到84.9%的AUPRC值,91.0%,80.1%,在5倍交叉验证中占82.1%,分别。与现有的几种最先进的方法相比,我们的框架在独立于阈值的评估指标上实现了更好的性能。总的来说,我们的研究为识别植物miRNA-靶关联提供了一种有效和简化的方法,这将有助于更深入地理解植物中的miRNA调控机制。
    Deciphering the targets of microRNAs (miRNAs) in plants is crucial for comprehending their function and the variation in phenotype that they cause. As the highly cell-specific nature of miRNA regulation, recent computational approaches usually utilize expression data to identify the most physiologically relevant targets. Although these methods are effective, they typically require a large sample size and high-depth sequencing to detect potential miRNA-target pairs, thereby limiting their applicability in improving plant breeding. In this study, we propose a novel miRNA-target prediction framework named kmerPMTF (k-mer-based prediction framework for plant miRNA-target). Our framework effectively extracts the latent semantic embeddings of sequences by utilizing k-mer splitting and a deep self-supervised neural network. We construct multiple similarity networks based on k-mer embeddings and employ graph convolutional networks to derive deep representations of miRNAs and targets and calculate the probabilities of potential associations. We evaluated the performance of kmerPMTF on four typical plant datasets: Arabidopsis thaliana, Oryza sativa, Solanum lycopersicum, and Prunus persica. The results demonstrate its ability to achieve AUPRC values of 84.9%, 91.0%, 80.1%, and 82.1% in 5-fold cross-validation, respectively. Compared with several state-of-the-art existing methods, our framework achieves better performance on threshold-independent evaluation metrics. Overall, our study provides an efficient and simplified methodology for identifying plant miRNA-target associations, which will contribute to a deeper comprehension of miRNA regulatory mechanisms in plants.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在利用图卷积网络(GCN)优化储能发电厂时,解决了模型性能欠佳,参数和操作过多的挑战,本文介绍了一种新的方法——分组交换图卷积网络。最初,建立了GCN极限学习机。从这个坚实的基础中汲取灵感,我们创新地制作了一个群交换图卷积模块。该模块利用组图卷积技术来合并独特的节点特征信息,根据各种分组为各种拓扑图矩阵量身定制。这种创新方法确保信息在不同的群体之间自由有效地流动。此外,我们设计了一个前沿的定时深度分离卷积模块,包括两个创新的组成部分。第一个组件引入时序深度分离卷积,彻底改变了原来的定时卷积模块。第二部分,分组交换图卷积网络,彻底改变了时间序列深度分离卷积过程。它通过在不同特征融合包之间采用1×1卷积层来实现这一点,实现不同数据包之间的无缝信息交换。实验结果证明了该模型的有效性,单步预测的均方根误差(RMSE)指标和均方根误差(MAE)指标在60分钟时达到46.08和26.22,分别。在多步骤测试中,与基准模型相比,所提出的模型在15分钟范围内的RMSE误差减少了14.71%,在60分钟范围内的RMSE误差减少了9.29%。这种性能改进提高了储能设备的运行效率和可靠性,特别是在时间序列的动态变化下。
    Addressing the challenges of suboptimal model performance and excessive parameters and operations in the optimization of energy storage power plants utilizing Graph Convolutional Network (GCN), this paper introduces a novel approach - the packet-switched graph convolutional network. Initially, a GCN extreme learning machine is established. Drawing inspiration from this solid foundation, we have innovatively crafted a group exchange graph convolution module. This module leverages group graph convolution techniques to amalgamate unique node feature information, tailored to diverse topology graph matrices based on various groupings. This innovative approach ensures that information flows freely and effectively among distinct groupings. Furthermore, we have designed a cutting-edge timing depth separation convolution module, comprising two innovative components. The first component introduces timing depth separation convolution, revolutionizing the original timing convolution module. The second component, the packet-switching graph convolutional network, revolutionizes the time sequence depth separation convolution process. It achieves this by employing 1 × 1 convolutional layers between different feature fusion packets, enabling seamless information exchange between distinct packets. Experimental results demonstrate the efficacy of the proposed model, with root mean square error (RMSE) metrics and root mean square error (MAE) metrics for single-step prediction reaching 46.08 and 26.22 at 60 min, respectively. In multi-step testing, the proposed model exhibits a 14.71 % reduction in RMSE error at the 15-min scale and a 9.29 % reduction at the 60-min scale compared to the benchmark model. This performance improvement enhances the operational efficiency and reliability of the energy storage plant, particularly under dynamic changes in the time series.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在人机交互(HRI)中,准确的3D手部姿势和网格估计具有至关重要的意义。然而,在严重的自遮挡和高自相似性中推断合理和准确的姿势仍然是一个固有的挑战。为了缓解HRI期间由不可见和相似关节引起的模糊性,我们提出了一种新的拓扑感知变压器网络,名为HandGCNFormer,以深度图像为输入,在对远程上下文信息进行建模的同时,将手运动学拓扑的先验知识纳入网络。具体来说,我们提出了一种新颖的图形形成器解码器,该解码器具有附加的节点偏移图形卷积层(NoffGConv)。Graphformer解码器优化了Transformer和GCN之间的协同作用,捕获关节之间的远程依赖关系和局部拓扑连接。最重要的是,我们用新颖的拓扑感知头替换标准MLP预测头,以更好地利用局部拓扑约束来实现更合理和准确的姿势。我们的方法在四个具有挑战性的数据集上实现了最先进的3D手部姿势估计性能,包括Hands2017,NYU,ICVL,MSRA。为了进一步证明我们提出的Graphformer解码器和拓扑感知头的有效性和可扩展性,我们将我们的框架扩展到HandGCNFormer-Mesh,用于3D手网格估计任务。扩展框架有效地将形状回归量与原始的Graphformer解码器和拓扑感知头集成在一起,生产马诺参数。HO-3D数据集上的结果,其中包含各种具有挑战性的遮挡,表明,与以前最先进的3D手网格估计方法相比,我们的HandGCNFormer-Mesh取得了有竞争力的结果。
    In Human-Robot Interaction (HRI), accurate 3D hand pose and mesh estimation hold critical importance. However, inferring reasonable and accurate poses in severe self-occlusion and high self-similarity remains an inherent challenge. In order to alleviate the ambiguity caused by invisible and similar joints during HRI, we propose a new Topology-aware Transformer network named HandGCNFormer with depth image as input, incorporating prior knowledge of hand kinematic topology into the network while modeling long-range contextual information. Specifically, we propose a novel Graphformer decoder with an additional Node-offset Graph Convolutional layer (NoffGConv). The Graphformer decoder optimizes the synergy between the Transformer and GCN, capturing long-range dependencies and local topological connections between joints. On top of that, we replace the standard MLP prediction head with a novel Topology-aware head to better exploit local topological constraints for more reasonable and accurate poses. Our method achieves state-of-the-art 3D hand pose estimation performance on four challenging datasets, including Hands2017, NYU, ICVL, and MSRA. To further demonstrate the effectiveness and scalability of our proposed Graphformer Decoder and Topology aware head, we extend our framework to HandGCNFormer-Mesh for the 3D hand mesh estimation task. The extended framework efficiently integrates a shape regressor with the original Graphformer Decoder and Topology aware head, producing Mano parameters. The results on the HO-3D dataset, which contains various and challenging occlusions, show that our HandGCNFormer-Mesh achieves competitive results compared to previous state-of-the-art 3D hand mesh estimation methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    空间转录组学(ST),含有细粒度的基因表达(即,不同的窗口)组织样本内的空间位置,在开发创新疗法方面已经变得至关重要。传统的ST技术,然而,依靠昂贵的专业商业设备。解决这个问题,我们的文章旨在创建一个具有成本效益的,使用标准组织图像进行基因表达预测的虚拟ST方法,消除了对昂贵设备的需求。该领域的常规方法通常忽略不同样本窗口之间的长距离空间依赖性或需要先前的基因表达数据。为了克服这些限制,我们提出了边缘-关系窗口-注意网络(ErwaNet),通过从组织图像中捕获局部相互作用和全局结构信息来增强基因预测,没有先前的基因表达数据。ErwaNet创新地构造异构图以对局部窗口交互进行建模,并结合了用于全局信息分析的注意力机制。这种双重框架不仅为基因表达预测提供了一种经济有效的解决方案,而且消除了先验知识基因表达信息的必要性。在癌症研究领域的一个显著优势,它使一个更有效和可访问的分析范式。ErwaNet是一种无先验且易于实现的图形卷积网络(GCN)方法,用于从组织图像中预测基因表达。对两个公共乳腺癌数据集的评估表明,ErwaNet,没有额外的信息,优于最先进的(SOTA)方法。代码可在https://github.com/biyecc/ErwaNet上获得。
    Spatial transcriptomics (ST), containing gene expression with fine-grained (i.e., different windows) spatial location within tissue samples, has become vital in developing innovative treatments. Traditional ST technology, however, rely on costly specialized commercial equipment. Addressing this, our article aims to creates a cost-effective, virtual ST approach using standard tissue images for gene expression prediction, eliminating the need for expensive equipment. Conventional approaches in this field often overlook the long-distance spatial dependencies between different sample windows or need prior gene expression data. To overcome these limitations, we propose the Edge-Relational Window-Attentional Network (ErwaNet), enhancing gene prediction by capturing both local interactions and global structural information from tissue images, without prior gene expression data. ErwaNet innovatively constructs heterogeneous graphs to model local window interactions and incorporates an attention mechanism for global information analysis. This dual framework not only provides a cost-effective solution for gene expression predictions but also obviates the necessity of prior knowledge gene expression information, a significant advantage in the field of cancer research where it enables a more efficient and accessible analytical paradigm. ErwaNet stands out as a prior-free and easy-to-implement Graph Convolution Network (GCN) method for predicting gene expression from tissue images. Evaluation of the two public breast cancer datasets shows that ErwaNet, without additional information, outperforms the state-of-the-art (SOTA) methods. Code is available at https://github.com/biyecc/ErwaNet.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    肝细胞癌(HCC)是世界上最常见的癌症之一,在癌症死亡中排名第四。原发性病理性坏死是肝细胞癌的有效预后指标。我们提出了一种基于GCN的方法,该方法模仿病理学家对坏死组织分布的整体评估的观点,以分析患者的生存率。具体来说,我们引入了一个图卷积神经网络来构建一个以坏死组织和肿瘤组织为图节点的空间图,旨在挖掘病理切片中坏死组织之间的上下文信息。我们使用来自浙江大学附属第一医院的303名患者的1381张幻灯片来训练模型,并使用TCGA-LIHC进行外部验证。我们方法的C指数比基线高出约4.45%,证明了GCN学习到的坏死空间分布信息对指导患者预后有一定意义。
    Hepatocellular carcinoma (HCC) is one of the most common cancers in the world which ranks fourth in cancer deaths. Primary pathological necrosis is an effective prognostic indicator for hepatocellular carcinoma. We propose a GCN-based approach that mimics the pathologist\'s perspective for global assessment of necrosis tissue distribution to analyze patient survival. Specifically, we introduced a graph convolutional neural network to construct a spatial map with necrotic tissue and tumor tissue as graph nodes, aiming to mine the contextual information between necrotic tissue in pathological sections. We used 1381 slides from 303 patients from the First Affiliated Hospital of Zhejiang University School to train the model and used TCGA-LIHC for external validation. The C-index of our method outperforms the baseline by about 4.45%, which proves that the information about the spatial distribution of necrosis learned by GCN is meaningful for guiding patient prognosis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    背景:准确、及时地评估儿童的发育状况对于早期诊断和干预至关重要。由于缺乏训练有素的医疗保健提供者和不精确的父母报告,更准确和自动化的发展评估至关重要。在发展的各个领域,众所周知,幼儿的粗大运动发育可以预测随后的童年发展。
    目的:这项研究的目的是开发一种模型来评估粗大运动行为,并将结果整合以确定幼儿的整体粗大运动状态。这项研究还旨在确定在评估总体总体运动技能方面很重要的行为,并检测关键时刻和重要的身体部位,以评估每种行为。
    方法:我们使用了18-35个月幼儿的行为视频。为了评估电机总体发展,我们选择了4种行为(爬楼梯,走下楼梯,扔球,并站在1英尺上),已通过韩国婴儿和儿童发育筛查测试进行了验证。在儿童行为视频中,我们将每个孩子的位置估计为边界框,并在框内提取人类关键点。在第一阶段,使用基于图形卷积网络(GCN)的算法分别评估具有每种行为的提取的人类关键点的视频。在第一阶段模型中获得的每个标签的概率值用作第二阶段模型的输入,极端梯度提升(XGBoost)算法,预测总体运动状态。为了可解释性,我们使用梯度加权类激活映射(Grad-CAM)来识别运动过程中的重要时刻和相关身体部位。Shapley加性解释方法用于评估变量重要性,以确定对整体发展评估贡献最大的运动。
    结果:从147名儿童中收集了4种粗大运动技能的行为视频,共产生2395个视频。评估每种行为的阶段1GCN模型的接受者工作特征曲线下面积(AUROC)为0.79至0.90。关键点映射Grad-CAM可视化识别了每个行为中的重要时刻以及重要身体部位的差异。评估总体粗大运动状态的阶段2XGBoost模型的AUROC为0.90。在这四种行为中,“下楼梯”对整体发展评估的贡献最大。
    结论:使用18-35个月幼儿的运动视频,我们开发了客观和自动化的模型来评估每个行为和评估每个孩子的整体粗大运动表现。我们确定了评估总体电机性能的重要行为,并开发了在评估总体电机性能时识别重要力矩和身体部位的方法。
    BACKGROUND: Accurate and timely assessment of children\'s developmental status is crucial for early diagnosis and intervention. More accurate and automated developmental assessments are essential due to the lack of trained health care providers and imprecise parental reporting. In various areas of development, gross motor development in toddlers is known to be predictive of subsequent childhood developments.
    OBJECTIVE: The purpose of this study was to develop a model to assess gross motor behavior and integrate the results to determine the overall gross motor status of toddlers. This study also aimed to identify behaviors that are important in the assessment of overall gross motor skills and detect critical moments and important body parts for the assessment of each behavior.
    METHODS: We used behavioral videos of toddlers aged 18-35 months. To assess gross motor development, we selected 4 behaviors (climb up the stairs, go down the stairs, throw the ball, and stand on 1 foot) that have been validated with the Korean Developmental Screening Test for Infants and Children. In the child behavior videos, we estimated each child\'s position as a bounding box and extracted human keypoints within the box. In the first stage, the videos with the extracted human keypoints of each behavior were evaluated separately using a graph convolutional networks (GCN)-based algorithm. The probability values obtained for each label in the first-stage model were used as input for the second-stage model, the extreme gradient boosting (XGBoost) algorithm, to predict the overall gross motor status. For interpretability, we used gradient-weighted class activation mapping (Grad-CAM) to identify important moments and relevant body parts during the movements. The Shapley additive explanations method was used for the assessment of variable importance, to determine the movements that contributed the most to the overall developmental assessment.
    RESULTS: Behavioral videos of 4 gross motor skills were collected from 147 children, resulting in a total of 2395 videos. The stage-1 GCN model to evaluate each behavior had an area under the receiver operating characteristic curve (AUROC) of 0.79 to 0.90. Keypoint-mapping Grad-CAM visualization identified important moments in each behavior and differences in important body parts. The stage-2 XGBoost model to assess the overall gross motor status had an AUROC of 0.90. Among the 4 behaviors, \"go down the stairs\" contributed the most to the overall developmental assessment.
    CONCLUSIONS: Using movement videos of toddlers aged 18-35 months, we developed objective and automated models to evaluate each behavior and assess each child\'s overall gross motor performance. We identified the important behaviors for assessing gross motor performance and developed methods to recognize important moments and body parts while evaluating gross motor performance.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号