auditory perception

听觉感知
  • 文章类型: Journal Article
    Partially autonomous vehicles can help minimize human errors. However, being free from some driving subtasks can result in a low vigilance state, which can affect the driver\'s attention towards the road. The present study first tested whether drivers of partially autonomous vehicles would benefit from the addition of auditory versions of the messages presented in variable message signs (VMS), particularly, when they find themselves in a monotonous driving situation. A second aim was to test whether the addition of auditory messages would also produce an indirect effect on the driver\'s vigilance, improving performance on other driving subtasks not related to the message processing. Forty-three volunteers participated in a driving simulator study. They completed two tasks: (a) a VMS task, where they had to regain manual control of the car if the VMS message was critical, and (b) a car-following task, where they had to pay attention to the preceding car to respond to occasional brake events. Behavioral and EEG data were registered. Overall, results indicated that the addition of audio messages helped drivers process VMS information more effectively and maintain a higher level of vigilance throughout the driving time. These findings would provide useful information for the development of partially automated vehicles, as their design must guarantee that the driver remains attentive enough to assume control when necessary.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    \'Jump scares\' are particularly robust when visuals are paired with coherent sound. A new study demonstrates that connectivity between the superior colliculus and parabigeminal nucleus generates multimodal enhancement of visually triggered defensiveness, revealing a novel multisensory threat augmentation mechanism.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Some species have evolved the ability to use the sense of hearing to modify existing vocalizations, or even create new ones, which enlarges their repertoires and results in complex communication systems.1 This ability corresponds to various forms of vocal production learning that are all possessed by humans and independently displayed by distantly related vertebrates.1,2,3,4,5,6,7 Among mammals, a few species, including the Egyptian fruit bat,8,9,10 would possess such vocal production learning abilities.7 Yet the necessity of an intact auditory system for the development of the Egyptian fruit bat typical vocal repertoire has not been tested. Furthermore, a systematic causal examination of learned and innate aspects of the entire repertoire has never been performed in any vocal learner. Here we addressed these gaps by eliminating pups\' sense of hearing at birth and assessing its effects on vocal production in adulthood. The deafening treatment enabled us to both causally test these bats\' vocal learning ability and discern learned from innate aspects of their vocalizations. Leveraging wireless individual audio recordings from freely interacting adults, we show that a subset of the Egyptian fruit bat vocal repertoire necessitates auditory feedback. Intriguingly, these affected vocalizations belong to different acoustic groups in the vocal repertoire of males and females. These findings open the possibilities for targeted studies of the mammalian neural circuits that enable sexually dimorphic forms of vocal learning.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Docongruencebetweenaudialandvisualmomentsauditexperience?Whilecross-modalcorreslationsbetweenvisionandhearingarewell-documented,先前的研究显示,关于视听对应是否影响主观审美体验的结果相互矛盾。这里,与肯特勒国际绘画空间(纽约市,美国),我们从以前的研究出发,在专业策划的音乐作为图像和隐喻展览中使用专门创作的音乐与视觉艺术配对。我们预先注册的在线实验包括4个条件:音频,视觉,视听意欲(艺术家意欲的艺术/音乐配对),和视听随机(随机洗牌)。参与者(N=201)被展示了16件,并且可以在他们喜欢的时候点击进入下一件。我们使用花费的时间作为审美兴趣的隐含指标。此外,在每一块之后,参与者被问及他们的主观体验(例如,感到感动)。我们发现参与者花在音频上的时间明显更多,其次是视听,其次是视觉碎片;然而,他们在视听(双模式)条件下感到最感动。与视听随机条件相比,视听预期的视听对应评分明显更高;有趣的是,虽然,在任何其他主观评定量表上,预期条件和随机条件之间没有显着差异,或花费的时间。总的来说,这些结果质疑跨模态对应与审美之间的关系。此外,结果使使用时间作为一种隐含的审美体验度量变得更加复杂。
    Does congruence between auditory and visual modalities affect aesthetic experience? While cross-modal correspondences between vision and hearing are well-documented, previous studies show conflicting results regarding whether audiovisual correspondence affects subjective aesthetic experience. Here, in collaboration with the Kentler International Drawing Space (NYC, USA), we depart from previous research by using music specifically composed to pair with visual art in the professionally-curated Music as Image and Metaphor exhibition. Our pre-registered online experiment consisted of 4 conditions: Audio, Visual, Audio-Visual-Intended (artist-intended pairing of art/music), and Audio-Visual-Random (random shuffling). Participants (N = 201) were presented with 16 pieces and could click to proceed to the next piece whenever they liked. We used time spent as an implicit index of aesthetic interest. Additionally, after each piece, participants were asked about their subjective experience (e.g., feeling moved). We found that participants spent significantly more time with Audio, followed by Audiovisual, followed by Visual pieces; however, they felt most moved in the Audiovisual (bi-modal) conditions. Ratings of audiovisual correspondence were significantly higher for the Audiovisual-Intended compared to Audiovisual-Random condition; interestingly, though, there were no significant differences between intended and random conditions on any other subjective rating scale, or for time spent. Collectively, these results call into question the relationship between cross-modal correspondence and aesthetic appreciation. Additionally, the results complicate the use of time spent as an implicit measure of aesthetic experience.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    声音识别对人类来说是毫不费力的,但对人工听力系统构成了重大挑战。深度神经网络(DNN)尤其是卷积神经网络(CNN),最近在声音分类方面超过了传统的机器学习。然而,当前的DNN使用二进制分类变量将声音映射到标签,忽略标签之间的语义关系。认知神经科学研究表明,除了声学线索之外,人类听众还利用了此类语义信息。因此,我们的假设是,融合语义信息可以提高DNN的声音识别性能,模仿人类行为。在我们的方法中,声音识别是一个回归问题,被训练为将频谱图映射到来自NLP模型的连续语义表示(Word2Vec,BERT,和CLAP文本编码器)。训练了两种DNN类型:具有连续嵌入的semDNN和具有分类标签的catDNN,两者都具有从388211个声音集合中提取的数据集,并丰富了语义描述。跨四个外部数据集的评估,证实了semDNN与catDNN相比语义标记的优越性,保持更高层次的关系。重要的是,对自然声音的人类相似性评级的分析,表明semDNN比catDNN更接近人类听众的行为,其他DNN,NLP模型。我们的工作有助于理解语义在声音识别中的作用,弥合人工系统和人类听觉感知之间的差距。
    Sound recognition is effortless for humans but poses a significant challenge for artificial hearing systems. Deep neural networks (DNNs), especially convolutional neural networks (CNNs), have recently surpassed traditional machine learning in sound classification. However, current DNNs map sounds to labels using binary categorical variables, neglecting the semantic relations between labels. Cognitive neuroscience research suggests that human listeners exploit such semantic information besides acoustic cues. Hence, our hypothesis is that incorporating semantic information improves DNN\'s sound recognition performance, emulating human behaviour. In our approach, sound recognition is framed as a regression problem, with CNNs trained to map spectrograms to continuous semantic representations from NLP models (Word2Vec, BERT, and CLAP text encoder). Two DNN types were trained: semDNN with continuous embeddings and catDNN with categorical labels, both with a dataset extracted from a collection of 388,211 sounds enriched with semantic descriptions. Evaluations across four external datasets, confirmed the superiority of semantic labeling from semDNN compared to catDNN, preserving higher-level relations. Importantly, an analysis of human similarity ratings for natural sounds, showed that semDNN approximated human listener behaviour better than catDNN, other DNNs, and NLP models. Our work contributes to understanding the role of semantics in sound recognition, bridging the gap between artificial systems and human auditory perception.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    一个社会个体需要有效地管理他或她的环境中相对于他或她自己的目的的复杂信息的数量以获取相关信息。本文提出了一种神经架构,旨在重现在机器人视听任务期间对人类有效的注意力机制(警报/定向/选择)。我们根据其识别发出元音的受试者面部相关信息来源的能力来评估该系统。我们提出了一种视听注意力(MAVA)的发展模型,该模型结合了Hebbian学习和基于视觉运动和音频能量的显著性图之间的竞争。MAVA有效地结合了自下而上和自上而下的信息,以将系统定向到相关领域。该系统具有几个优点,包括在线和自主学习能力,低计算时间和对环境噪声的鲁棒性。MAVA优于其他人工模型,可在各种噪声条件下检测语音源。
    A social individual needs to effectively manage the amount of complex information in his or her environment relative to his or her own purpose to obtain relevant information. This paper presents a neural architecture aiming to reproduce attention mechanisms (alerting/orienting/selecting) that are efficient in humans during audiovisual tasks in robots. We evaluated the system based on its ability to identify relevant sources of information on faces of subjects emitting vowels. We propose a developmental model of audio-visual attention (MAVA) combining Hebbian learning and a competition between saliency maps based on visual movement and audio energy. MAVA effectively combines bottom-up and top-down information to orient the system toward pertinent areas. The system has several advantages, including online and autonomous learning abilities, low computation time and robustness to environmental noise. MAVA outperforms other artificial models for detecting speech sources under various noise conditions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    当在相似的感觉水平(SL)下呈现给听众时,可听的甚高频声音(VHFS)和超声波(US)被认为比较低频率声音更令人不愉快。在这项研究中,17名参与者将感官不愉快评为14-,16-,和18kHz音调和1kHz参考音调。每个人的音调都以相等的主观响度水平呈现,对应于在1kHz处测得的10、20和30dBSL的电平。根据他们归因于暴露于VHFS/US的自我报告的先前症状,参与者被归类为“有症状”或“无症状”。在这两组中,VHFS/US的主观响度随声压级的增加比1-kHz参考音调的增加更快,这与较高频率处的减小的动态范围一致。对于响度匹配的音调,参与者将VHFS/US评为比1kHz参考更令人不愉快。这些结果表明,在设计或部署发射VHFS/US的设备时,应考虑在高频下增加的感觉不愉快和减小的动态范围。
    Audible very-high frequency sound (VHFS) and ultrasound (US) have been rated more unpleasant than lower frequency sounds when presented to listeners at similar sensation levels (SLs). In this study, 17 participants rated the sensory unpleasantness of 14-, 16-, and 18-kHz tones and a 1-kHz reference tone. Tones were presented at equal subjective loudness levels for each individual, corresponding to levels of 10, 20, and 30 dB SL measured at 1 kHz. Participants were categorized as either \"symptomatic\" or \"asymptomatic\" based on self-reported previous symptoms that they attributed to exposure to VHFS/US. In both groups, subjective loudness increased more rapidly with sound pressure level for VHFS/US than for the 1-kHz reference tone, which is consistent with a reduced dynamic range at the higher frequencies. For loudness-matched tones, participants rated VHFS/US as more unpleasant than that for the 1-kHz reference. These results suggest that increased sensory unpleasantness and reduced dynamic range at high frequencies should be considered when designing or deploying equipment which emits VHFS/US that could be audible to exposed people.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    近年来,在声景评估中对感知情感品质(PAQs)的研究有所增加,方法从原位到实验室。通过技术进步,虚拟现实(VR)促进了对同一实验中多个位置的评估。在本文中,在大曼彻斯特的三个地点测试的在线和实验室环境中展示了不同城市地点的VR复制品(“公园”,\'广场\',和行人\'街道\')在两个人口密度(空的和繁忙的)使用ISO/TS12913-2(2018)声景PAQ。研究区域为360视频和双耳音频VR复制品准备了音频和视频记录。目的是观察方法中位置内的人口密度效应(Wilcoxon检验)和位置之间的变化(Mann-WhitneyU检验)。人口密度和不同地点之间的比较表明,对大多数PAQ有重大影响。结果还表明,大城市可以呈现均匀的声音,组成一个混合的城市声景,独立于功能。这些发现可以以低成本的方式支持城市设计,城市规划者可以测试不同的场景和干预措施。
    The study of the perceived affective qualities (PAQs) in soundscape assessments have increased in recent years, with methods varying from in-situ to laboratory. Through technological advances, virtual reality (VR) has facilitated evaluations of multiple locations in the same experiment. In this paper, VR reproductions of different urban sites were presented in an online and laboratory environment testing three locations in Greater Manchester (\'Park\', \'Plaza\', and pedestrian \'Street\') in two population densities (empty and busy) using ISO/TS 12913-2 (2018) soundscape PAQs. The studied areas had audio and video recordings prepared for 360 video and binaural audio VR reproductions. The aims were to observe population density effects within locations (Wilcoxon test) and variations between locations (Mann-Whitney U test) within methods. Population density and comparisons among locations demonstrated a significant effect on most PAQs. Results also suggested that big cities can present homogenous sounds, composing a \'blended\' urban soundscape, independently of functionality. These findings can support urban design in a low-cost approach, where urban planners can test different scenarios and interventions.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    听众对高频声音(ITDENV)的包络中携带的耳间时差敏感,但是这个线索的显著性取决于包络的某些属性,特别是,关于包络中振幅调制(AM)的存在/深度。这项研究检验了以下假设:患有感觉神经性听力损失的人,在某些条件下对AM表现出增强的敏感性,在这些条件下也会表现出优异的ITDENV灵敏度。第二个假设是,个体之间ITDENV敏感性的变化可能与对AM的敏感性变化有关。要启用直接比较,使用了标准的自适应AM检测任务以及用于测量ITDENV灵敏度的改进版本。刺激是以32、64或128Hz的速率调制的4kHz音调,并以30dB的感觉水平呈现。这两项任务均由16名听力正常的听众和16名听力损失的听众尝试。与假设一致,AM和ITDENV阈值相关,并且在听力损失的听众中趋于更好。一项对照实验强调,在解释群体效应时,绝对水平可能是一个考虑因素。
    Listeners are sensitive to interaural time differences carried in the envelope of high-frequency sounds (ITDENV), but the salience of this cue depends on certain properties of the envelope and, in particular, on the presence/depth of amplitude modulation (AM) in the envelope. This study tested the hypothesis that individuals with sensorineural hearing loss, who show enhanced sensitivity to AM under certain conditions, would also show superior ITDENV sensitivity under those conditions. The second hypothesis was that variations in ITDENV sensitivity across individuals can be related to variations in sensitivity to AM. To enable a direct comparison, a standard adaptive AM detection task was used along with a modified version of it designed to measure ITDENV sensitivity. The stimulus was a 4-kHz tone modulated at rates of 32, 64, or 128 Hz and presented at a 30 dB sensation level. Both tasks were attempted by 16 listeners with normal hearing and 16 listeners with hearing loss. Consistent with the hypotheses, AM and ITDENV thresholds were correlated and tended to be better in listeners with hearing loss. A control experiment emphasized that absolute level may be a consideration when interpreting the group effects.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    这项研究研究了用助听器聆听时刺激特性对声音外部化的影响。通常情况下,听众会从扬声器中获得宽带“令牌”(环境声音和语音),并使用连续量表对外部化进行评级。在单独的块中,他们在没有帮助的情况下或在佩戴耳后助听器时听,该助听器具有封闭的圆顶和低增益(线性或压缩)。代币对评级有重大影响,即使是独立倾听,助听器的效果取决于令牌。声学分析表明,助听器更有可能破坏低频强调的峰值声音的外部化。
    This study examined the influence of stimulus properties on sound externalization when listening with hearing aids. Normally hearing listeners were presented with broadband \"tokens\" (environmental sounds and speech) from loudspeakers, and rated externalization using a continuous scale. In separate blocks, they listened unaided or while wearing behind-the-ear hearing aids with closed domes and low gain (linear or compressive). There was a significant influence of token on ratings, even for unaided listening, and the effect of hearing aids depended on token. An acoustic analysis indicated that hearing aids were more likely to disrupt externalization for peakier sounds with a low-frequency emphasis.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号