sensory modalities

  • 文章类型: Journal Article
    早期关于选择性注意力的工作使用基于听觉的任务,比如二分法倾听,阐明容量限制和这些限制的个体差异。今天,人们对注意力能力的个体差异非常感兴趣,但是该领域已经转向视觉模态任务。此外,大多数基于冲突的注意力控制测试由于低信噪比和使用差异分数而缺乏可靠性。严重的,目前还不清楚注意力控制在多大程度上在感官模式中推广,如果没有可靠的基于听觉的测试,这个问题的答案仍然难以捉摸。为此,我们开发了三种基于听觉的注意力控制测试,这些测试使用自适应反应截止日期(DL)来解释速度-准确性权衡:听觉西蒙DL,听觉侧翼DL,和听觉StroopDL。在大样本(N=316)中,我们研究了三种听觉冲突任务的心理测量特性,测试注意力控制是更好地建模为单一因素还是模态特定因素,并估计了模态特定因素的独特差异对双耳听力和多任务处理性能的预测贡献的程度。我们的分析表明,听觉冲突任务具有很强的心理测量特性,并通过注意力控制的视觉测试证明了收敛的有效性。即使控制了感知处理速度(r=.75),听觉和视觉注意力控制因素也高度相关(r=.81)。模态特定的注意力控制因素占模态匹配标准度量的独特差异,但大部分解释的差异是模态一般。结果表明,模态一般注意力控制和模态特定处理之间存在相互作用。
    Early work on selective attention used auditory-based tasks, such as dichotic listening, to shed light on capacity limitations and individual differences in these limitations. Today, there is great interest in individual differences in attentional abilities, but the field has shifted towards visual-modality tasks. Furthermore, most conflict-based tests of attention control lack reliability due to low signal-to-noise ratios and the use of difference scores. Critically, it is unclear to what extent attention control generalizes across sensory modalities, and without reliable auditory-based tests, an answer to this question will remain elusive. To this end, we developed three auditory-based tests of attention control that use an adaptive response deadline (DL) to account for speed-accuracy trade-offs: Auditory Simon DL, Auditory Flanker DL, and Auditory Stroop DL. In a large sample (N = 316), we investigated the psychometric properties of the three auditory conflict tasks, tested whether attention control is better modeled as a unitary factor or modality-specific factors, and estimated the extent to which unique variance in modality-specific factors contributed incrementally to the prediction of dichotic listening and multitasking performance. Our analyses indicated that the auditory conflict tasks have strong psychometric properties and demonstrate convergent validity with visual tests of attention control. Auditory and visual attention control factors were highly correlated (r = .81)-even after controlling for perceptual processing speed (r = .75). Modality-specific attention control factors accounted for unique variance in modality-matched criterion measures, but the majority of the explained variance was modality-general. The results suggest an interplay between modality-general attention control and modality-specific processing.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    走神现象(MW),作为一个与内部定向认知相关的经历家族,严重影响警惕性进化。特别是,远程操作中的人类在必要时进行手动控制之前监视部分自动化的车队,可能会看到由于内部来源而引起的注意力漂移;因此,它可以在出现外循环(OOTL)情况和相关的性能问题中发挥重要作用。跟随,量化、缓解这种现象,脑电图(EEG)系统已经证明了可靠的结果。随着MW产生了一种注意力上的解耦,ERP和大脑振荡都会受到影响。然而,在复杂环境中影响这些标记的因素仍未完全了解。在本文中,我们特别讨论了逐渐出现的注意解耦的可能性以及用于传达目标的感觉方式所产生的差异。18名参与者被要求(1)监督执行避障任务(视觉任务)的自动无人机,以及(2)尽可能快地响应不频繁的蜂鸣声(听觉任务)。我们通过脑电图测量事件相关电位和α波。我们还添加了40Hz幅度调制的棕色噪声以引起稳态听觉响应(ASSR)。报告的MW发作分为与任务相关的发作和与任务无关的发作。我们发现,蜂鸣声引起的N1ERP分量在任务无关的MW期间具有较低的幅度,而P3分量在任务相关MW期间具有较高的振幅,与其他注意状态相比。专注于顶枕区,与任务无关的MW期间的α波活性高于其他。这些结果支持任务无关MW而不是任务相关MW的解耦假设,突出显示解耦“深度”的可能变化,具体取决于MW事件。最后,我们发现注意状态对ASSR振幅没有影响。我们讨论解释原因的可能原因。结果强调了EEG在模仿生态环境的实验室任务中跟踪和研究MW的能力,以及感知解耦对操作者行为的复杂影响,特别是,脑电图测量。
    The phenomenon of mind wandering (MW), as a family of experiences related to internally directed cognition, heavily influences vigilance evolution. In particular, humans in teleoperations monitoring partially automated fleet before assuming manual control whenever necessary may see their attention drift due to internal sources; as such, it could play an important role in the emergence of out-of-the-loop (OOTL) situations and associated performance problems. To follow, quantify, and mitigate this phenomenon, electroencephalogram (EEG) systems already demonstrated robust results. As MW creates an attentional decoupling, both ERPs and brain oscillations are impacted. However, the factors influencing these markers in complex environments are still not fully understood. In this paper, we specifically addressed the possibility of gradual emergence of attentional decoupling and the differences created by the sensory modality used to convey targets. Eighteen participants were asked to (1) supervise an automated drone performing an obstacle avoidance task (visual task) and (2) respond to infrequent beeps as fast as possible (auditory task). We measured event-related potentials and alpha waves through EEG. We also added a 40-Hz amplitude modulated brown noise to evoke steady-state auditory response (ASSR). Reported MW episodes were categorized between task-related and task-unrelated episodes. We found that N1 ERP component elicited by beeps had lower amplitude during task-unrelated MW, whereas P3 component had higher amplitude during task-related MW, compared with other attentional states. Focusing on parieto-occipital regions, alpha-wave activity was higher during task-unrelated MW compared with others. These results support the decoupling hypothesis for task-unrelated MW but not task-related MW, highlighting possible variations in the \"depth\" of decoupling depending on MW episodes. Finally, we found no influence of attentional states on ASSR amplitude. We discuss possible reasons explaining why. Results underline both the ability of EEG to track and study MW in laboratory tasks mimicking ecological environments, as well as the complex influence of perceptual decoupling on operators\' behavior and, in particular, EEG measures.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    尽管人类感知时间的准确性,许多因素可以调节时间的主观体验。例如,据广泛报道,情绪可以扩大或缩小我们对时间的感知,并且听觉刺激标记的时间间隔比视觉刺激更长。在本研究中,我们旨在调查情绪对时间感知的影响是否可以通过情绪刺激的呈现顺序和呈现它们的感觉形态来改变。参与者被要求完成时间分割任务,在该任务中,情感刺激以听觉或视觉方式呈现,以及在要估计的间隔之前或之后。我们观察到了模态的主要影响(与视觉模态相比,感知的持续时间更长,听觉的变异性更低)以及情感的主要影响(与中性相比,对负面刺激的时间高估)。重要的是,情态和情绪的影响与情绪刺激的呈现顺序相互作用。在视觉条件下,当情绪刺激出现在时间间隔之后,参与者高估了时间,但是当首先呈现情绪刺激时,没有观察到阴性和中性刺激之间的差异。在听觉条件下,没有发现情绪对感知持续时间的显著影响。结果表明,负面情绪会影响我们对决策阶段而不是起搏器阶段的持续时间的感知。情绪听觉刺激对时间知觉没有影响。
    Despite human accuracy in perceiving time, many factors can modulate the subjective experience of time. For example, it is widely reported that emotion can expand or shrink our perception of time and that temporal intervals are perceived as longer when marked by auditory stimuli than by visual stimuli. In the present study, we aimed at investigating whether the influence of emotion on time perception can be altered by the order in which emotional stimuli are presented and the sensory modality in which they are presented. Participants were asked to complete a time bisection task in which emotional stimuli were presented either acoustically or visually, and either before or after interval to be estimated. We observed a main effect of modality (longer perceived duration and lower variability in the auditory than in the visual modality) as well as a main effect of emotion (temporal overestimation for negative stimuli compared to neutral). Importantly, the effects of modality and emotion interacted with the order of presentation of the emotional stimuli. In the visual condition, when emotional stimuli were presented after the temporal intervals, participants overestimated time, but no differences between negative and neutral stimuli were observed when emotional stimuli were presented first. In the auditory condition, no significant effect of emotion on perceived duration was found. Results suggest that negative emotions affect our perception of durations acting at the decision-making stage rather than at the pacemaker one. No effect on time perception was observed for emotional auditory stimuli.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    多模态交互(MMI)正在被广泛实施,特别是在增强现实(AR)系统等新技术中,因为它被认为支持更自然的,高效,灵活的互动形式。然而,研究了MMI在AR中的正确应用。更具体地说,在AR中MMI期间组合不同输入和输出模式的效果仍未完全理解。因此,本研究旨在研究典型AR任务中不同输入和输出模式的独立和综合效应.20名年轻人参加了一项对照实验,在该实验中,他们被要求使用AR设备在不同的输入(语音,手势,多模态)和输出(VV-VA,VV-NA,NV-VA,NV-NA)条件。结果表明,输入和输出模式对任务绩效的影响存在差异,工作量,感知的适当性,和用户偏好。在这项研究中,输入和输出条件对绩效指标的交互作用也很明显。这表明,尽管多模态输入通常是用户首选的,它应该谨慎实施,因为它的有效性受到系统输出的处理代码的高度影响。这项研究,这是同类中的第一个,揭示了MMI在AR系统中应用的几个新含义。
    Multimodal interaction (MMI) is being widely implemented, especially in new technologies such as augmented reality (AR) systems since it is presumed to support a more natural, efficient, and flexible form of interaction. However, limited research has been done to investigate the proper application of MMI in AR. More specifically, the effects of combining different input and output modalities during MMI in AR are still not fully understood. Therefore, this study aims to examine the independent and combined effects of different input and output modalities during a typical AR task. 20 young adults participated in a controlled experiment in which they were asked to perform a simple identification task using an AR device in different input (speech, gesture, multimodal) and output (VV-VA, VV-NA, NV-VA, NV-NA) conditions. Results showed that there were differences in the influence of input and output modalities on task performance, workload, perceived appropriateness, and user preference. Interaction effects between the input and output conditions on the performance metrics were also evident in this study, suggesting that although multimodal input is generally preferred by the users, it should be implemented with caution since its effectiveness is highly influenced by the processing code of the system output. This study, which is the first of its kind, has revealed several new implications regarding the application of MMI in AR systems.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    通信是一个发射器向一个或多个接收器传递信息以引起接收器的响应(行为或生理)的过程。交流在各种生物学功能中起着重要作用,可能涉及来自不同感觉方式的信号和线索。传统上,动物交流的调查集中在单一的感官模态上,然而通信通常是多模式的。由于这些不同的过程可能相当复杂,因此很难解开,一种方法是首先分别研究每种感觉形态。有了这种对个体感官的精致理解,揭示它们如何相互作用成为可能,因为可以解释每种模态的特征和属性,使多模式方法可行。使用这个框架,研究人员进行了系统的,对殖民地pinned物种中母犬识别过程的实验研究,澳大利亚海狮Neophocacinerea。该研究首先通过回放实验评估了母亲和幼崽通过声音识别彼此的能力。第二,他们评估了母亲和幼崽是否使用视觉线索来区分它们与特定物种,和/或女性是否将孝顺幼崽的气味与非孝顺幼崽的气味区分开来。最后,了解不同感官模态传递的信息是否协同分析,或者感官模态之间是否存在层次结构,同时进行涉及不同感觉线索的实验。这些发现是关于每个感觉线索的活动空间进行讨论的,以及通过评估来自不同模式的信息可能产生的潜在增强。
    Communication is the process by which one emitter conveys information to one or several receivers to induce a response (behavioral or physiological) by the receiver. Communication plays a major role in various biological functions and may involve signals and cues from different sensory modalities. Traditionally, investigations of animal communication focused on a single sensory modality, yet communication is often multimodal. As these different processes may be quite complex and therefore difficult to disentangle, one approach is to first study each sensorial modality separately. With this refined understanding of individual senses, revealing how they interact becomes possible as the characteristics and properties of each modality can be accounted for, making a multimodal approach feasible. Using this framework, researchers undertook systematic, experimental investigations on mother-pup recognition processes in a colonial pinniped species, the Australian sea lion Neophoca cinerea. The research first assessed the abilities of mothers and pups to identify each other by their voice using playback experiments. Second, they assessed whether visual cues are used by both mothers and pups to distinguish them from conspecifics, and/or whether females discriminate the odor of their filial pup from those from non-filial pups. Finally, to understand if the information transmitted by different sensory modalities is analyzed synergistically or if there is a hierarchy among the sensory modalities, experiments were performed involving different sensory cues simultaneously. These findings are discussed with regards to the active space of each sensory cue, and of the potential enhancements that may arise by assessing information from different modalities.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    关于狗(犬类)在物体辨别和识别任务中整合通过不同感觉方式获得的信息的能力的研究很少。这样的过程将表明多感官心理表征的形成。在实验1中,我们测试了3只天赋单词学习者(GWL)狗可以快速学习玩具的语言标签的能力,和10只典型的(T)狗区分最近与奖励相关的对象,来自干扰物,在光明和黑暗的条件下。虽然两组之间的成功率和条件没有差异,一项详细的行为分析表明,所有的狗在黑暗中搜索的时间更长,嗅闻更多。这表明,如果可能,狗主要依靠视觉,转而只使用其他感官模式,包括嗅觉,在黑暗中寻找在实验2中,我们调查了是否,对于GWL狗(N=4),听到对象的言语标签会激活对多感官心理表征的记忆。我们通过测试它们在黑暗和光照条件下根据它们的名字识别物体的能力来做到这一点。他们的成功率在两个条件之间没有差异,而狗的搜索行为,表明灵活使用不同的感觉方式。关于GWL狗识别标记物体的能力所涉及的认知机制知之甚少。这些发现为GWL狗提供了第一个证据,口头标签唤起了对象的多感官心理表征。
    Little research has been conducted on dogs\' (Canis familiaris) ability to integrate information obtained through different sensory modalities during object discrimination and recognition tasks. Such a process would indicate the formation of multisensory mental representations. In Experiment 1, we tested the ability of 3 Gifted Word Learner (GWL) dogs that can rapidly learn the verbal labels of toys, and 10 Typical (T) dogs to discriminate an object recently associated with a reward, from distractor objects, under light and dark conditions. While the success rate did not differ between the two groups and conditions, a detailed behavioral analysis showed that all dogs searched for longer and sniffed more in the dark. This suggests that, when possible, dogs relied mostly on vision, and switched to using only other sensory modalities, including olfaction, when searching in the dark. In Experiment 2, we investigated whether, for the GWL dogs (N = 4), hearing the object verbal labels activates a memory of a multisensory mental representation. We did so by testing their ability to recognize objects based on their names under dark and light conditions. Their success rate did not differ between the two conditions, whereas the dogs\' search behavior did, indicating a flexible use of different sensory modalities. Little is known about the cognitive mechanisms involved in the ability of GWL dogs to recognize labeled objects. These findings supply the first evidence that for GWL dogs, verbal labels evoke a multisensory mental representation of the objects.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Social perception is a multimodal process involving vision and audition as central input sources for human social cognitive processes. However, it remains unclear how profoundly deaf people assess others in the context of mating and social interaction. The current study explored the relative importance of different sensory modalities (vision, smell, and touch) in assessments of opposite- and same-sex strangers. We focused on potential sensory compensation processes in mate selection (i.e., increased importance of the intact senses in forming impressions of an opposite-sex stranger as a potential partner). A total of 74 deaf individuals and 100 normally hearing controls were included in the study sample. We found diminished importance of vision and smell in deaf participants compared with controls for opposite- and same-sex strangers, and increased importance of touch for the assessment of same-sex strangers. The results suggested that deaf people rely less on visual and olfactory cues in mating and social assessments, highlighting a possible role of sign language in shaping interpersonal tactile experience in non-romantic relationships.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    在过去的十年中,多模态学习的研究进展在多个领域迅速发展,尤其是在计算机视觉中。多模式数据流和深度学习算法的不断增长的潜力促进了深度多模式学习的日益普及。这涉及能够统一处理和分析多模态信息的模型的开发。非结构化的现实世界数据可以固有地采取多种形式,也被称为模态,通常包括视觉和文本内容。从这类数据中提取相关模式仍然是深度学习研究人员的激励目标。在本文中,我们寻求通过探索如何生成考虑跨感官模式的异构视觉线索的集成和组合的深度模型来提高对计算机视觉社区深度多模式学习的关键概念和算法的理解。特别是,我们从当前关于深度多模态学习的文献中总结了六个观点,即:多模态数据表示,多模态融合(即,传统和基于深度学习的方案),多任务学习,多模态对齐,多模态迁移学习,和零射学习。我们还调查了当前的多模式应用程序,并提供了一组基准数据集,用于解决各个视觉领域的问题。最后,我们强调了深度多模态学习的局限性和挑战,并为未来的研究提供了见解和方向。
    The research progress in multimodal learning has grown rapidly over the last decade in several areas, especially in computer vision. The growing potential of multimodal data streams and deep learning algorithms has contributed to the increasing universality of deep multimodal learning. This involves the development of models capable of processing and analyzing the multimodal information uniformly. Unstructured real-world data can inherently take many forms, also known as modalities, often including visual and textual content. Extracting relevant patterns from this kind of data is still a motivating goal for researchers in deep learning. In this paper, we seek to improve the understanding of key concepts and algorithms of deep multimodal learning for the computer vision community by exploring how to generate deep models that consider the integration and combination of heterogeneous visual cues across sensory modalities. In particular, we summarize six perspectives from the current literature on deep multimodal learning, namely: multimodal data representation, multimodal fusion (i.e., both traditional and deep learning-based schemes), multitask learning, multimodal alignment, multimodal transfer learning, and zero-shot learning. We also survey current multimodal applications and present a collection of benchmark datasets for solving problems in various vision domains. Finally, we highlight the limitations and challenges of deep multimodal learning and provide insights and directions for future research.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    应对是生物体的一种生存机制。它不仅仅是反应性的,但也涉及通过将感官信息转化为在生物体认知背景下具有意义的感知来理解环境。听音乐,另一方面,是一项包含感官的复杂任务,生理,行为,和认知水平的处理。既是一个依赖我们的进化工具包来应对世界的性格过程,又是一个更精细的感官创造技能,通过在感觉输入和效应器反应之间引入高阶中间变量,它超越了原始的动作-反应耦合。从应对的角度考虑听音乐,将音乐视为声音环境,而将听音乐视为涉及探索该环境以及与声音的相互作用的过程。在这方面考虑了几个问题,例如音乐作为可能的压力源的概念,适应性倾听的作用,应对和奖励之间的关系,自我调节策略在音乐选择中的重要性,以及音乐的器乐意义,即它可以用来修改听众的内部和外部环境。
    Coping is a survival mechanism of living organisms. It is not merely reactive, but also involves making sense of the environment by rendering sensory information into percepts that have meaning in the context of an organism\'s cognitions. Music listening, on the other hand, is a complex task that embraces sensory, physiological, behavioral, and cognitive levels of processing. Being both a dispositional process that relies on our evolutionary toolkit for coping with the world and a more elaborated skill for sense-making, it goes beyond primitive action-reaction couplings by the introduction of higher-order intermediary variables between sensory input and effector reactions. Consideration of music-listening from the perspective of coping treats music as a sound environment and listening as a process that involves exploration of this environment as well as interactions with the sounds. Several issues are considered in this regard such as the conception of music as a possible stressor, the role of adaptive listening, the relation between coping and reward, the importance of self-regulation strategies in the selection of music, and the instrumental meaning of music in the sense that it can be used to modify the internal and external environment of the listener.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Sci-hub)

       PDF(Pubmed)

  • 文章类型: Journal Article
    在两个实验中,作者研究了在特定运动任务中双峰整合的效果.沙滩排球运动员被要求做出战术决定,口头或通过运动反应做出反应,在被视觉呈现后,听觉,或者沙滩排球场景中的两种刺激。在实验1中,在视觉和听觉信息一致的情况下,玩家在游戏情况下做出正确决定的频率要高于在他们仅经历其中一种模态或不一致信息的试验中。当电机、而不是口头的,给出了答案。实验2使用不同的刺激材料复制了这种一致性效果,并显示了视觉刺激对决策的影响随着较短的视觉刺激持续时间而降低。总之,这项研究表明,与单峰信息相比,双峰整合一致的视觉和听觉信息可以更准确地做出运动决策。
    In 2 experiments, the authors investigated the effects of bimodal integration in a sport-specific task. Beach volleyball players were required to make a tactical decision, responding either verbally or via a motor response, after being presented with visual, auditory, or both kinds of stimuli in a beach volleyball scenario. In Experiment 1, players made the correct decision in a game situation more often when visual and auditory information were congruent than in trials in which they experienced only one of the modalities or incongruent information. Decision-making accuracy was greater when motor, rather than verbal, responses were given. Experiment 2 replicated this congruence effect using different stimulus material and showed a decreasing effect of visual stimulation on decision making as a function of shorter visual stimulus durations. In conclusion, this study shows that bimodal integration of congruent visual and auditory information results in more accurate decision making in sport than unimodal information.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

公众号