Emotion recognition

情感识别
  • 文章类型: Journal Article
    儿童和青少年的冷酷无情(CU)特征与严重和持续的反社会行为有关。根据过去的实证研究,几个理论模型表明,CU特征可能部分解释为难以正确识别他人的情绪状态以及他们对他人眼睛的注意力减少,这对因果理论和治疗都很重要。本研究测试了CU性状之间的关系,在52名男孩的样本中,面部表情和视觉行为的情感识别被转诊为行为问题的诊所(Mage=10.29岁;SD=2.06)。我们通过儿童问题特征清单(CPTI)对CU特征进行了多方法和多信息评估,冷酷无情的清单(ICU),和亲社会情绪的临床评估-1.1版(CAPE)。该研究的主要目标是比较这些方法在形成情绪处理能力不同的青年亚组方面的效用。情绪识别任务评估每种情绪的识别准确性(错误百分比)和眼睛或嘴巴区域的绝对停留时间。重复测量的结果ANOVA显示,低CU组和高CU组在情绪识别准确性上没有差异,与CU性状的评估方法无关。然而,高CU组显示对恐惧和悲伤的面部表情(使用CPTI)或对所有情绪(使用CAPE)的关注减少。高CU组也显示对口腔区域的注意力普遍增加,但只有在CAPE评估时。这些发现提供了证据来支持CU特征升高的异常如何处理情绪刺激,尤其是通过临床访谈评估时,这可以指导对这一青年群体的适当评估和更成功的干预措施。
    Callous-unemotional (CU) traits in children and adolescents are linked to severe and persistent antisocial behavior. Based on past empirical research, several theoretical models have suggested that CU traits may be partly explained by difficulties in correctly identifying others\' emotional states as well as their reduced attention to others\' eyes, which could be important for both causal theory and treatment. This study tested the relationships among CU traits, emotion recognition of facial expressions and visual behavior in a sample of 52 boys referred to a clinic for conduct problems (Mage = 10.29 years; SD = 2.06). We conducted a multi-method and multi-informant assessment of CU traits through the Child Problematic Traits Inventory (CPTI), the Inventory of Callous-Unemotional (ICU), and the Clinical Assessment of Prosocial Emotions-Version 1.1 (CAPE). The primary goal of the study was to compare the utility of these methods for forming subgroups of youth that differ in their emotional processing abilities. An emotion recognition task assessed recognition accuracy (percentage of mistakes) and absolute dwell time on the eyes or mouth region for each emotion. Results from repeated measures ANOVAs revealed that low and high CU groups did not differ in emotion recognition accuracy, irrespective of the method of assessing CU traits. However, the high CU group showed reduced attention to the eyes of fearful and sad facial expressions (using the CPTI) or to all emotions (using the CAPE). The high CU group also showed a general increase in attention to the mouth area, but only when assessed by the CAPE. These findings provide evidence to support abnormalities in how those elevated on CU traits process emotional stimuli, especially when assessed by a clinical interview, which could guide appropriate assessment and more successful interventions for this group of youth.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    脑电图(EEG)信号为大脑中情绪生成的复杂性提供了宝贵的见解。然而,不同个体的EEG信号的可变性为经验实现提供了巨大的障碍。我们的研究创新地解决了这些挑战,关注不同受试者脑电图数据中的共性。我们介绍了一种名为对比学习图卷积网络(CLGCN)的新方法。这种方法捕获了与个人情绪状态相关的独特特征和关键通道节点。具体来说,CLGCN融合了对比学习的同步多主题数据学习和图卷积网络在破译大脑连接矩阵方面的熟练程度的双重好处。由于CLGCN在数据集的学习过程中生成标准化的大脑网络学习矩阵,因此可以理解多方面的大脑功能及其信息交换过程。我们的模型大大简化了新科目的再培训过程,只需要5%的初始样本量进行微调,以达到显著的92.8%的准确率。此外,我们的模型在DEAP和SEED数据集上进行了广泛的测试,证明了我们模型的有效性。 .
    Electroencephalogram (EEG) signals offer invaluable insights into the complexities of emotion generation within the brain. Yet, the variability in EEG signals across individuals presents a formidable obstacle for empirical implementations. Our research addresses these challenges innovatively, focusing on the commonalities within distinct subjects\' EEG data.We introduce a novel approach named Contrastive Learning Graph Convolutional Network (CLGCN). This method captures the distinctive features and crucial channel nodes related to individuals\' emotional states. Specifically, CLGCN merges the dual benefits of Contrastive Learning\'s synchronous multisubject data learning and the Graph Convolutional Network\'s proficiency in deciphering brain connectivity matrices.Understanding multifaceted brain functions and their information interchange processes is realized as CLGCN generates a standardized brain network learning matrix during a dataset\'s learning process. Our model significantly streamlines the retraining process for new subjects, requiring only 5% of the initial sample size for fine-tuning to attain a remarkable 92.8% accuracy rate. Additionally, our model has undergone extensive testing on the DEAP and SEED datasets, demonstrating the effectiveness of our model. .
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    针对现有的情感识别方法在时间上不能充分利用信息的问题,频率,和EEG信号中的空间域,这导致脑电情绪分类的准确性较低,本文提出了一种多特征,基于多频带的跨尺度注意卷积模型(CATM)。该模型主要由跨尺度的注意力模块,频率空间注意模块,一个功能转换模块,时间特征提取模块,和深度分类模块。首先,跨尺度注意卷积模块对预处理后的脑电信号提取不同尺度的空间特征;然后,频率空间注意模块为重要通道和空间位置分配更高的权重;接下来,时间特征提取模块提取脑电信号的时间特征;最后,深度分类模块将EEG信号分类为情绪。我们在DEAP数据集上评估了所提出的方法,在效价和唤醒二元分类实验中的准确率分别为99.70%和99.74%,效价-唤醒四分类实验的准确率分别为97.27%。此外,考虑到较少渠道的应用,我们还进行了5通道实验,效价和唤醒的二元分类准确率分别为97.96%和98.11%,分别。效价-唤醒四分类准确率为92.86%。实验结果表明,与其他最新方法相比,本文提出的方法具有更好的效果,并且在少数通道实验中也取得了更好的结果。
    Aiming at the problem that existing emotion recognition methods fail to make full use of the information in the time, frequency, and spatial domains in the EEG signals, which leads to the low accuracy of EEG emotion classification, this paper proposes a multi-feature, multi-frequency band-based cross-scale attention convolutional model (CATM). The model is mainly composed of a cross-scale attention module, a frequency-space attention module, a feature transition module, a temporal feature extraction module, and a depth classification module. First, the cross-scale attentional convolution module extracts spatial features at different scales for the preprocessed EEG signals; then, the frequency-space attention module assigns higher weights to important channels and spatial locations; next, the temporal feature extraction module extracts temporal features of the EEG signals; and, finally, the depth classification module categorizes the EEG signals into emotions. We evaluated the proposed method on the DEAP dataset with accuracies of 99.70% and 99.74% in the valence and arousal binary classification experiments, respectively; the accuracy in the valence-arousal four-classification experiment was 97.27%. In addition, considering the application of fewer channels, we also conducted 5-channel experiments, and the binary classification accuracies of valence and arousal were 97.96% and 98.11%, respectively. The valence-arousal four-classification accuracy was 92.86%. The experimental results show that the method proposed in this paper exhibits better results compared to other recent methods, and also achieves better results in few-channel experiments.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本文的目的是通过对面部肌电图(EMG)信号进行分类来识别用户的情绪。生物医学信号放大器,配备八个根据面部动作编码系统定位的有源电极,被用来记录肌电图信号。这些信号是在用户表现出各种情绪的过程中记录的:喜悦,悲伤,惊喜,厌恶,愤怒,恐惧,中立。为16个用户进行了录音。EMG信号的平均功率形成特征集。我们利用这些特征来训练和评估各种分类器。在主题相关模型中,KNN的平均分类准确率为96.3%,具有线性核的SVM为94.9%,具有立方核的SVM为94.6%,LDA为93.8%。在独立于主题的模型中,分类结果因测试用户而异,KNN分类器的范围从91.4%到48.6%,平均准确率为67.5%。具有立方内核的SVM表现稍差,平均准确率为59.1%,其次是线性内核为53.9%的SVM,和LDA分类器为41.2%。此外,该研究确定了区分情感对的最有效电极。
    The objective of the article is to recognize users\' emotions by classifying facial electromyographic (EMG) signals. A biomedical signal amplifier, equipped with eight active electrodes positioned in accordance with the Facial Action Coding System, was used to record the EMG signals. These signals were registered during a procedure where users acted out various emotions: joy, sadness, surprise, disgust, anger, fear, and neutral. Recordings were made for 16 users. The mean power of the EMG signals formed the feature set. We utilized these features to train and evaluate various classifiers. In the subject-dependent model, the average classification accuracies were 96.3% for KNN, 94.9% for SVM with a linear kernel, 94.6% for SVM with a cubic kernel, and 93.8% for LDA. In the subject-independent model, the classification results varied depending on the tested user, ranging from 91.4% to 48.6% for the KNN classifier, with an average accuracy of 67.5%. The SVM with a cubic kernel performed slightly worse, achieving an average accuracy of 59.1%, followed by the SVM with a linear kernel at 53.9%, and the LDA classifier at 41.2%. Additionally, the study identified the most effective electrodes for distinguishing between pairs of emotions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    从脑电图(EEG)信号中识别情绪是一项具有挑战性的任务,非线性,和大脑活动的非平稳特征。传统的方法往往无法捕捉到这些微妙的动态,而深度学习方法缺乏可解释性。在这项研究中,我们介绍了一种新颖的集成流形嵌入的三阶段方法,多级异质性复发分析(MHRA),和集成学习来解决这些限制在基于脑电图的情绪识别。
    使用SJTU-SEEDIV数据库评估了所提出的方法。我们首先应用均匀流形近似和投影(UMAP)将62导联EEG信号的流形嵌入到低维空间中。然后,我们开发了MHRA来表征跨多个过渡水平的脑活动的复杂复发动力学。最后,我们采用基于树的集成学习方法对四种情绪进行分类(中性,悲伤,恐惧,快乐)基于提取的MHRA特征。
    我们的方法实现了高性能,准确度为0.7885,AUC为0.7552,优于同一数据集上的现有方法。此外,我们的方法提供了在不同情绪中最一致的识别性能.敏感性分析显示特定的MHRA指标与每种情绪密切相关,提供对潜在神经动力学的有价值的见解。
    这项研究提出了一种基于EEG的情感识别的新颖框架,该框架有效地捕获了复杂的非线性和非平稳的大脑活动动力学,同时保持了可解释性。所提出的方法为提高我们对情绪处理的理解和开发更可靠的情绪识别系统提供了巨大的潜力,并在医疗保健及其他领域具有广泛的应用。
    UNASSIGNED: Recognizing emotions from electroencephalography (EEG) signals is a challenging task due to the complex, nonlinear, and nonstationary characteristics of brain activity. Traditional methods often fail to capture these subtle dynamics, while deep learning approaches lack explainability. In this research, we introduce a novel three-phase methodology integrating manifold embedding, multilevel heterogeneous recurrence analysis (MHRA), and ensemble learning to address these limitations in EEG-based emotion recognition.
    UNASSIGNED: The proposed methodology was evaluated using the SJTU-SEED IV database. We first applied uniform manifold approximation and projection (UMAP) for manifold embedding of the 62-lead EEG signals into a lower-dimensional space. We then developed MHRA to characterize the complex recurrence dynamics of brain activity across multiple transition levels. Finally, we employed tree-based ensemble learning methods to classify four emotions (neutral, sad, fear, happy) based on the extracted MHRA features.
    UNASSIGNED: Our approach achieved high performance, with an accuracy of 0.7885 and an AUC of 0.7552, outperforming existing methods on the same dataset. Additionally, our methodology provided the most consistent recognition performance across different emotions. Sensitivity analysis revealed specific MHRA metrics that were strongly associated with each emotion, offering valuable insights into the underlying neural dynamics.
    UNASSIGNED: This study presents a novel framework for EEG-based emotion recognition that effectively captures the complex nonlinear and nonstationary dynamics of brain activity while maintaining explainability. The proposed methodology offers significant potential for advancing our understanding of emotional processing and developing more reliable emotion recognition systems with broad applications in healthcare and beyond.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    本研究旨在剖析人工智能(AI)系统中情绪识别和反应机制的现状,探索取得的进展,面临的挑战,以及将情绪智力整合到AI中的隐式操作。本研究采用了全面的综述方法来研究情绪智力(EI)与人工智能(AI)系统的整合,专注于情绪识别和反应机制。审查过程需要制定研究问题,系统地搜索学术数据库,如PubMed,Scopus,和WebofScience,批判性地评估相关文献,综合数据,并以全面的格式介绍调查结果。这项研究强调了情绪识别模型的进步,包括使用深度识字方式和多模态数据乳液。它讨论了情感识别中的挑战,类似于凡人表达的可变性和实时处理的需要。强调语境信息和个体特征的整合,以增强对凡人情感的理解。该研究还涉及道德企业,类似于训练数据中的隔离和冲动。将情绪智力整合到AI系统中,为修改人类与计算机的关系提供了机会。情绪识别和反应机制取得重大进展,但挑战依然存在。未出生的探索方向包括增强情感识别模型的鲁棒性和可解释性,探索跨文化和环境忧虑的情感理解,解决长期的情绪阴影和适应问题。通过进一步探索人工智能系统中的情绪智力,可以开发更多富有同情心和反应灵敏的机器,与人类建立更深层次的情感联系。
    This study aims to dissect the current state of emotion recognition and response mechanisms in artificial intelligence (AI) systems, exploring the progress made, challenges faced, and implicit operations of integrating emotional intelligence into AI. This study utilized a comprehensive review approach to investigate the integration of emotional intelligence (EI) into artificial intelligence (AI) systems, concentrating on emotion recognition and response mechanisms. The review process entailed formulating research questions, systematically searching academic databases such as PubMed, Scopus, and Web of Science, critically evaluating relevant literature, synthesizing the data, and presenting the findings in a comprehensive format. The study highlights the advancements in emotion recognition models, including the use of deep literacy ways and multimodal data emulsion. It discusses the challenges in emotion recognition, similar to variability in mortal expressions and the need for real-time processing. The integration of contextual information and individual traits is emphasized as enhancing the understanding of mortal feelings. The study also addresses ethical enterprises, similar as sequestration and impulses in training data. The integration of emotional intelligence into AI systems presents openings to revise mortal-computer relations. Emotion recognition and response mechanisms have made significant progress, but challenges remain. Unborn exploration directions include enhancing the robustness and interpretability of emotion recognition models, exploring cross-cultural and environment-apprehensive emotion understanding, and addressing long-term emotion shadowing and adaption. By further exploring emotional intelligence in AI systems, further compassionate and responsive machines can be developed, enabling deeper emotional connections with humans.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    脑电图(EEG)情感识别在情感计算中起着至关重要的作用。EEG情感识别任务的局限性在于,由于缺乏有效的特征组织形式,因此很少同时将多个域的特征包括在分析中。本文提出了一种视频级特征组织方法来有效地组织时间,频域和空间域特征。此外,深度神经网络,信道注意力卷积聚合网络,旨在从视频级功能中探索更深层次的情感信息。该网络使用信道注意力机制来自适应地捕获关键EEG频带。然后通过多层卷积获得每个时间点的帧级表示。最后,帧级特征通过NeXtVLAD聚合以学习与时间序列相关的特征。本文提出的方法在SEED和DEAP数据集上取得了最好的分类性能。SEED数据集的平均准确度和标准偏差分别为95.80%和2.04%。在DEAP数据集中,唤醒和效价标准偏差的平均准确度为98.97%±1.13%和98.98%±0.98%,分别。实验结果表明,基于视频级特征的方法对脑电情感识别任务是有效的。
    Electroencephalogram (EEG) emotion recognition plays a vital role in affective computing. A limitation of the EEG emotion recognition task is that the features of multiple domains are rarely included in the analysis simultaneously because of the lack of an effective feature organization form. This paper proposes a video-level feature organization method to effectively organize the temporal, frequency and spatial domain features. In addition, a deep neural network, Channel Attention Convolutional Aggregation Network, is designed to explore deeper emotional information from video-level features. The network uses a channel attention mechanism to adaptively captures critical EEG frequency bands. Then the frame-level representation of each time point is obtained by multi-layer convolution. Finally, the frame-level features are aggregated through NeXtVLAD to learn the time-sequence-related features. The method proposed in this paper achieves the best classification performance in SEED and DEAP datasets. The mean accuracy and standard deviation of the SEED dataset are 95.80% and 2.04%. In the DEAP dataset, the average accuracy with the standard deviation of arousal and valence are 98.97% ± 1.13% and 98.98% ± 0.98%, respectively. The experimental results show that our approach based on video-level features is effective for EEG emotion recognition tasks.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    基于EEG的情感识别在脑机接口(BCI)中变得至关重要。目前,大多数研究集中在提高准确性,而忽略了对模型可解释性的进一步研究,我们致力于基于图结构分析不同脑区和信号频段对情绪生成的影响。因此,提出了一种双注意机制图卷积神经网络(DAMGCN)方法。具体来说,我们利用图卷积神经网络将大脑网络建模为图,以提取具有代表性的空间特征。此外,我们采用Transformer模型的自我注意机制,该机制将更多的电极通道权重和信号频带权重分配给重要的大脑区域和频带。注意力机制的可视化清楚地证明了DAMGCN学习的权重分配。在我们对DEAP模型的绩效评估中,SEED,和SEED-IV数据集,我们在SEED数据集上取得了最好的结果,显示受试者相关实验的准确率为99.42%,受试者独立实验的准确率为73.21%。结果明显优于基于EEG的情感识别领域中大多数现有模型的准确性。
    EEG-based emotion recognition is becoming crucial in brain-computer interfaces (BCI). Currently, most researches focus on improving accuracy, while neglecting further research on the interpretability of models, we are committed to analyzing the impact of different brain regions and signal frequency bands on emotion generation based on graph structure. Therefore, this paper proposes a method named Dual Attention Mechanism Graph Convolutional Neural Network (DAMGCN). Specifically, we utilize graph convolutional neural networks to model the brain network as a graph to extract representative spatial features. Furthermore, we employ the self-attention mechanism of the Transformer model which allocates more electrode channel weights and signal frequency band weights to important brain regions and frequency bands. The visualization of attention mechanism clearly demonstrates the weight allocation learned by DAMGCN. During the performance evaluation of our model on the DEAP, SEED, and SEED-IV datasets, we achieved the best results on the SEED dataset, showing subject-dependent experiments\' accuracy of 99.42% and subject-independent experiments\' accuracy of 73.21%. The results are demonstrably superior to the accuracies of most existing models in the realm of EEG-based emotion recognition.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    从动态面部视频中识别表情可以发现人类更自然的影响状态,由于面部的姿势变化,它在现实世界的场景中变得更具挑战性的任务,情绪序列的部分遮挡和微妙的动态变化。现有的基于变压器的方法通常侧重于自我关注来建模空间特征或时间特征之间的全局关系,对于野生表达视频,它们不能很好地关注与表达相关的重要局部性结构。为此,我们将不同的图结构整合到变压器中,并提出了一种CDGT方法来构建不同的图变压器,以从野外视频中进行有效的情感识别。具体来说,我们的方法包含空间对偶图变换器和时间双曲图变换器。前者部署了双重图约束注意力,以捕获局部空间令牌中潜在的与情感相关的图几何结构,以实现有效的特征表示,特别是对于具有姿态变化和部分遮挡的视频帧。后者采用双曲图约束的自注意力,探索双曲空间下重要的时间图结构信息,以模拟动态情绪的更细微变化。在野外基于视频的面部表情数据库上的大量实验结果表明,我们提出的CDGT优于其他最先进的方法。
    Recognizing expressions from dynamic facial videos can find more natural affect states of humans, and it becomes a more challenging task in real-world scenes due to pose variations of face, partial occlusions and subtle dynamic changes of emotion sequences. Existing transformer-based methods often focus on self-attention to model the global relations among spatial features or temporal features, which cannot well focus on important expression-related locality structures from both spatial and temporal features for the in-the-wild expression videos. To this end, we incorporate diverse graph structures into transformers and propose a CDGT method to construct diverse graph transformers for efficient emotion recognition from in-the-wild videos. Specifically, our method contains a spatial dual-graphs transformer and a temporal hyperbolic-graph transformer. The former deploys a dual-graph constrained attention to capture latent emotion-related graph geometry structures among local spatial tokens for efficient feature representation, especially for the video frames with pose variations and partial occlusions. The latter adopts a hyperbolic-graph constrained self-attention that explores important temporal graph structure information under hyperbolic space to model more subtle changes of dynamic emotion. Extensive experimental results on in-the-wild video-based facial expression databases show that our proposed CDGT outperforms other state-of-the-art methods.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    面部处理依赖于由低空间频率(LSF)驱动的预测过程,其在由高空间频率传达的精细信息之前传达粗略信息。然而,自闭症患者可能有非典型的预测过程,导致面部处理困难。这在自闭症女性中可能更正常,他们往往比男性表现出更好的社会交际能力。我们假设与自闭症男性相比,自闭症女性对社会情绪刺激的处理方式更为典型。为了检验这个假设,我们询问成年参与者(44名自闭症患者,51非自闭症患者)在中性面孔中检测到可怕的面孔,按两个顺序过滤:从粗到细(CtF)和从细到粗(FtC)。结果显示,与非自闭症(NA)个体相比,自闭症患者的恐惧检测d值更低,反应时间更长,无论过滤顺序如何。与FtC相比,两组CtF后P100潜伏期较短,与CtF相比,FtC后N170的振幅更大。然而,自闭症参与者在梭形中CtF和FtC之间的源活性差异减少。与NA女性相比,自闭症女性的激活模式在空间上也更为分散。最后,女性有更快的P100和N170延迟,以及FtC序列比男性更大的枕骨激活,与集团无关。总的来说,尽管在恐惧检测方面存在行为差异,但该结果并未表明自闭症患者LSF的预测过程受损.然而,它们确实表明自闭症患者的空间频率降低了大脑调制。此外,研究结果强调了性别差异,需要在理解自闭症女性时加以考虑。
    Face processing relies on predictive processes driven by low spatial frequencies (LSF) that convey coarse information prior to fine information conveyed by high spatial frequencies. However, autistic individuals might have atypical predictive processes, contributing to facial processing difficulties. This may be more normalized in autistic females, who often exhibit better socio-communicational abilities than males. We hypothesized that autistic females would display a more typical coarse-to-fine processing for socio-emotional stimuli compared to autistic males. To test this hypothesis, we asked adult participants (44 autistic, 51 non-autistic) to detect fearful faces among neutral faces, filtered in two orders: from coarse-to-fine (CtF) and from fine-to-coarse (FtC). Results show lower d\' values and longer reaction times for fearful detection in autism compared to non-autistic (NA) individuals, regardless of the filtering order. Both groups presented shorter P100 latency after CtF compared to FtC, and larger amplitude for N170 after FtC compared to CtF. However, autistic participants presented a reduced difference in source activity between CtF and FtC in the fusiform. There was also a more spatially spread activation pattern in autistic females compared to NA females. Finally, females had faster P100 and N170 latencies, as well as larger occipital activation for FtC sequences than males, irrespective of the group. Overall, the results do not suggest impaired predictive processes from LSF in autism despite behavioral differences in fear detection. However, they do indicate reduced brain modulation by spatial frequency in autism. In addition, the findings highlight sex differences that warrant consideration in understanding autistic females.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号