sounds

声音
  • 文章类型: Journal Article
    背景:数字时代见证了对新闻和信息的数字平台的日益依赖,再加上“deepfake”技术的出现。Deepfakes,利用语音记录和图像的大量数据集的深度学习模型,对媒体真实性构成重大威胁,可能导致不道德的滥用,如冒充和传播虚假信息。
    目标:为了应对这一挑战,这项研究旨在引入先天生物过程的概念,以区分真实的人类声音和克隆的声音。我们建议存在或不存在某些感知特征,比如讲话中的停顿,可以有效区分克隆和真实的音频。
    方法:共招募了49名具有不同种族背景和口音的成年参与者。每个参与者贡献语音样本,用于训练多达3个不同的语音克隆文本到语音模型和3个控制段落。随后,克隆模型生成了控制段落的合成版本,产生由每个参与者多达9个克隆音频样本和3个对照样本组成的数据集。我们分析了呼吸等生物行为引起的语音停顿,吞咽,和认知过程。计算了对应于语音暂停简档的五个音频特征。评估了这些特征的真实音频和克隆音频之间的差异,和5个经典的机器学习算法实现了使用这些特征来创建预测模型。通过对看不见的数据进行测试,评估了最优模型的泛化能力,结合了一个朴素的生成器,一个模型天真的段落,和幼稚的参与者。
    结果:克隆音频显示暂停之间的时间显着增加(P<.001),语音段长度的变化减少(P=0.003),发言时间的总比例增加(P=.04),语音中的micro和macropauses比率降低(P=0.01)。使用这些功能实现了五个机器学习模型,AdaBoost模型展示了最高的性能,实现5倍交叉验证平衡精度为0.81(SD0.05)。其他模型包括支持向量机(平衡精度0.79,SD0.03),随机森林(平衡精度0.78,SD0.04),逻辑回归,和决策树(平衡精度0.76,SD0.10和0.72,SD0.06)。在评估最优AdaBoost模型时,在预测未知数据时,它实现了0.79的总体测试准确性。
    结论:引入感知,机器学习模型中的生物特征在区分真实的人类声音和克隆音频方面显示出有希望的结果。
    BACKGROUND: The digital era has witnessed an escalating dependence on digital platforms for news and information, coupled with the advent of \"deepfake\" technology. Deepfakes, leveraging deep learning models on extensive data sets of voice recordings and images, pose substantial threats to media authenticity, potentially leading to unethical misuse such as impersonation and the dissemination of false information.
    OBJECTIVE: To counteract this challenge, this study aims to introduce the concept of innate biological processes to discern between authentic human voices and cloned voices. We propose that the presence or absence of certain perceptual features, such as pauses in speech, can effectively distinguish between cloned and authentic audio.
    METHODS: A total of 49 adult participants representing diverse ethnic backgrounds and accents were recruited. Each participant contributed voice samples for the training of up to 3 distinct voice cloning text-to-speech models and 3 control paragraphs. Subsequently, the cloning models generated synthetic versions of the control paragraphs, resulting in a data set consisting of up to 9 cloned audio samples and 3 control samples per participant. We analyzed the speech pauses caused by biological actions such as respiration, swallowing, and cognitive processes. Five audio features corresponding to speech pause profiles were calculated. Differences between authentic and cloned audio for these features were assessed, and 5 classical machine learning algorithms were implemented using these features to create a prediction model. The generalization capability of the optimal model was evaluated through testing on unseen data, incorporating a model-naive generator, a model-naive paragraph, and model-naive participants.
    RESULTS: Cloned audio exhibited significantly increased time between pauses (P<.001), decreased variation in speech segment length (P=.003), increased overall proportion of time speaking (P=.04), and decreased rates of micro- and macropauses in speech (both P=.01). Five machine learning models were implemented using these features, with the AdaBoost model demonstrating the highest performance, achieving a 5-fold cross-validation balanced accuracy of 0.81 (SD 0.05). Other models included support vector machine (balanced accuracy 0.79, SD 0.03), random forest (balanced accuracy 0.78, SD 0.04), logistic regression, and decision tree (balanced accuracies 0.76, SD 0.10 and 0.72, SD 0.06). When evaluating the optimal AdaBoost model, it achieved an overall test accuracy of 0.79 when predicting unseen data.
    CONCLUSIONS: The incorporation of perceptual, biological features into machine learning models demonstrates promising results in distinguishing between authentic human voices and cloned audio.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:声乐生物标志物,从声音特征的声学分析中得出,提供非侵入性的医疗筛查途径,诊断,和监测。先前的研究证明了通过智能手机记录语音的声学分析来预测2型糖尿病的可行性。在这项工作的基础上,这项研究探讨了音频数据压缩对声学声乐生物标志物开发的影响,这对于在医疗保健中更广泛的适用性至关重要。
    目的:本研究的目的是分析常见的音频压缩算法(MP3,M4A,和WMA)由3种不同的转换工具以2种比特率应用,影响对声音生物标志物检测至关重要的特征。
    方法:使用转换为MP3,M4A的未压缩语音样本,研究了音频数据压缩对声学声乐生物标志物开发的影响。和WMA格式在2比特率(320和128kbps)与MediaHuman(MH)音频转换器,WonderShare(WS)UniConverter,和快进运动图像专家组(FFmpeg)。数据集包括来自505名参与者的记录,总共17298个音频文件,使用智能手机收集。参与者每天记录一个固定的英语句子,最多6次,最长14天。特征提取,包括音高,抖动,强度,和梅尔频率倒谱系数(MFCC),是使用Python和Parselmouth进行的。使用Wilcoxon符号秩检验和Bonferroni校正进行多重比较用于统计分析。
    结果:在这项研究中,最初从505名参与者那里录制了36,970个音频文件,筛选后,有17298张录音符合固定的句子标准。音频转换软件之间的差异,MH,WS,和FFmpeg,值得注意的是,影响压缩结果,如恒定或可变比特率。分析包括不同的数据压缩格式和广泛的语音特征和MFCC。Wilcoxon符号秩检验得出P值,低于Bonferroni校正的显著性水平的那些表明由于压缩引起的显著改变。结果表明了跨格式和比特率的压缩的特定特征影响。与WS转换的文件相比,MH转换的文件表现出更大的弹性。比特率也影响了功能稳定性,38例唯一受单一比特率影响。值得注意的是,语音特征在各种转换方法中显示出比MFCC更高的稳定性。
    结论:发现压缩效果具有特定特征,MH和FFmpeg表现出更大的弹性。某些功能一直受到影响,强调理解特征弹性对诊断应用的重要性。考虑到声乐生物标志物在医疗保健中的实施,为数据存储或传输目的找到通过压缩保持一致的功能是很有价值的。专注于特定的功能和格式,未来的研究可以拓宽范围,包括不同的特征,实时压缩算法,和各种记录方法。这项研究增强了我们对音频压缩对语音特征和MFCC的影响的理解,为跨领域开发应用程序提供见解。该研究强调了特征稳定性在处理压缩音频数据中的重要性,为在不断发展的技术环境中使用明智的语音数据奠定基础。
    BACKGROUND: Vocal biomarkers, derived from acoustic analysis of vocal characteristics, offer noninvasive avenues for medical screening, diagnostics, and monitoring. Previous research demonstrated the feasibility of predicting type 2 diabetes mellitus through acoustic analysis of smartphone-recorded speech. Building upon this work, this study explores the impact of audio data compression on acoustic vocal biomarker development, which is critical for broader applicability in health care.
    OBJECTIVE: The objective of this research is to analyze how common audio compression algorithms (MP3, M4A, and WMA) applied by 3 different conversion tools at 2 bitrates affect features crucial for vocal biomarker detection.
    METHODS: The impact of audio data compression on acoustic vocal biomarker development was investigated using uncompressed voice samples converted into MP3, M4A, and WMA formats at 2 bitrates (320 and 128 kbps) with MediaHuman (MH) Audio Converter, WonderShare (WS) UniConverter, and Fast Forward Moving Picture Experts Group (FFmpeg). The data set comprised recordings from 505 participants, totaling 17,298 audio files, collected using a smartphone. Participants recorded a fixed English sentence up to 6 times daily for up to 14 days. Feature extraction, including pitch, jitter, intensity, and Mel-frequency cepstral coefficients (MFCCs), was conducted using Python and Parselmouth. The Wilcoxon signed rank test and the Bonferroni correction for multiple comparisons were used for statistical analysis.
    RESULTS: In this study, 36,970 audio files were initially recorded from 505 participants, with 17,298 recordings meeting the fixed sentence criteria after screening. Differences between the audio conversion software, MH, WS, and FFmpeg, were notable, impacting compression outcomes such as constant or variable bitrates. Analysis encompassed diverse data compression formats and a wide array of voice features and MFCCs. Wilcoxon signed rank tests yielded P values, with those below the Bonferroni-corrected significance level indicating significant alterations due to compression. The results indicated feature-specific impacts of compression across formats and bitrates. MH-converted files exhibited greater resilience compared to WS-converted files. Bitrate also influenced feature stability, with 38 cases affected uniquely by a single bitrate. Notably, voice features showed greater stability than MFCCs across conversion methods.
    CONCLUSIONS: Compression effects were found to be feature specific, with MH and FFmpeg showing greater resilience. Some features were consistently affected, emphasizing the importance of understanding feature resilience for diagnostic applications. Considering the implementation of vocal biomarkers in health care, finding features that remain consistent through compression for data storage or transmission purposes is valuable. Focused on specific features and formats, future research could broaden the scope to include diverse features, real-time compression algorithms, and various recording methods. This study enhances our understanding of audio compression\'s influence on voice features and MFCCs, providing insights for developing applications across fields. The research underscores the significance of feature stability in working with compressed audio data, laying a foundation for informed voice data use in evolving technological landscapes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:高频听力损失是人口老龄化和有暴露于大声噪音史的人群中最常见的问题之一。这种类型的听力损失可能令人沮丧和残疾,使得难以理解言语交流和与世界有效互动。
    目的:这项研究旨在研究代表高频音素的空间独特触觉振动对日常情况下自我感知的理解对话能力的影响。
    方法:为了解决高频听力损失,开发了一种多电机腕带,使用机器学习来收听特定的高频音素。腕带在空间上唯一的位置振动以实时表示哪个音素存在。总共招募了16名高频听力损失的参与者,并要求他们佩戴腕带6周。每周使用助听器受益的缩写简介(APHAB)测量与听力损失相关的残疾程度。
    结果:在为期6周的研究结束时,所有参与者的平均APHAB福利得分达到12.39分,从基线40.32到最终评分27.93(SD13.11;N=16;P=0.002,双尾依赖t检验)。没有助听器的人在6周时平均APHAB获益评分比使用助听器的人提高10.78分(t14=2.14;P=.10,2尾独立t检验)。所有参与者的平均获益分数为15.44(SD13.88;N=16;P<.001,双尾依赖t检验)。所有参与者对背景噪声的平均获益分数为10.88(SD17.54;N=16;P=0.03,双尾依赖t检验)。所有参与者对混响的平均获益评分为10.84(SD16.95;N=16;P=.02,2尾依赖t检验)。
    结论:这些发现表明,腕带提供的振动触觉感觉替代,产生与高频音素相对应的空间可区分的振动,有助于高频听力损失的个体提高他们对言语交流的感知理解。无论一个人是否佩戴助听器,振动触觉反馈都能提供好处,尽管方式略有不同。最后,理解语音难度最大的人从振动触觉反馈中获得了最大的感知收益。
    BACKGROUND: High-frequency hearing loss is one of the most common problems in the aging population and with those who have a history of exposure to loud noises. This type of hearing loss can be frustrating and disabling, making it difficult to understand speech communication and interact effectively with the world.
    OBJECTIVE: This study aimed to examine the impact of spatially unique haptic vibrations representing high-frequency phonemes on the self-perceived ability to understand conversations in everyday situations.
    METHODS: To address high-frequency hearing loss, a multi-motor wristband was developed that uses machine learning to listen for specific high-frequency phonemes. The wristband vibrates in spatially unique locations to represent which phoneme was present in real time. A total of 16 participants with high-frequency hearing loss were recruited and asked to wear the wristband for 6 weeks. The degree of disability associated with hearing loss was measured weekly using the Abbreviated Profile of Hearing Aid Benefit (APHAB).
    RESULTS: By the end of the 6-week study, the average APHAB benefit score across all participants reached 12.39 points, from a baseline of 40.32 to a final score of 27.93 (SD 13.11; N=16; P=.002, 2-tailed dependent t test). Those without hearing aids showed a 10.78-point larger improvement in average APHAB benefit score at 6 weeks than those with hearing aids (t14=2.14; P=.10, 2-tailed independent t test). The average benefit score across all participants for ease of communication was 15.44 (SD 13.88; N=16; P<.001, 2-tailed dependent t test). The average benefit score across all participants for background noise was 10.88 (SD 17.54; N=16; P=.03, 2-tailed dependent t test). The average benefit score across all participants for reverberation was 10.84 (SD 16.95; N=16; P=.02, 2-tailed dependent t test).
    CONCLUSIONS: These findings show that vibrotactile sensory substitution delivered by a wristband that produces spatially distinguishable vibrations in correspondence with high-frequency phonemes helps individuals with high-frequency hearing loss improve their perceived understanding of verbal communication. Vibrotactile feedback provides benefits whether or not a person wears hearing aids, albeit in slightly different ways. Finally, individuals with the greatest perceived difficulty understanding speech experienced the greatest amount of perceived benefit from vibrotactile feedback.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    使用智能手机内置麦克风进行听诊是使用听诊器的可行替代方法,当医生申请时。
    这项横断面研究旨在评估该技术在父母-真正的预期最终用户使用时的可行性。
    医生在三级医院儿科就诊期间招募了46名儿童(男性:n=33,72%;年龄:平均11.3,SD3.1y;哮喘儿童:n=24,52%)。在4个位置(气管,右前胸,以及右肺和左肺基底),首先由医生(记录:n=297),然后由父母(记录:n=344)。所有录音(N=641)均由3个注释者对质量和不定声音的存在进行分类。家长填写了一份问卷,以提供有关该应用程序的反馈,使用李克特量表,范围从1(“完全不同意”)到5(“完全同意”)。
    大多数录音都有质量(医生录音:253/297,85.2%;父母录音:266/346,76.9%)。具有不定声音的医生录音(34/253,13.4%)和父母录音(31/266,11.7%)的比例相似。父母发现该应用程序易于使用(问卷:中位数5,IQR5-5),并愿意使用它(问卷:中位数5,IQR5-5)。
    我们的结果表明,由父母在临床背景下进行智能手机听诊是可行的,但是需要进一步的调查来测试它在现实生活中的可行性。
    UNASSIGNED: The use of a smartphone built-in microphone for auscultation is a feasible alternative to the use of a stethoscope, when applied by physicians.
    UNASSIGNED: This cross-sectional study aims to assess the feasibility of this technology when used by parents-the real intended end users.
    UNASSIGNED: Physicians recruited 46 children (male: n=33, 72%; age: mean 11.3, SD 3.1 y; children with asthma: n=24, 52%) during medical visits in a pediatric department of a tertiary hospital. Smartphone auscultation using an app was performed at 4 locations (trachea, right anterior chest, and right and left lung bases), first by a physician (recordings: n=297) and later by a parent (recordings: n=344). All recordings (N=641) were classified by 3 annotators for quality and the presence of adventitious sounds. Parents completed a questionnaire to provide feedback on the app, using a Likert scale ranging from 1 (\"totally disagree\") to 5 (\"totally agree\").
    UNASSIGNED: Most recordings had quality (physicians\' recordings: 253/297, 85.2%; parents\' recordings: 266/346, 76.9%). The proportions of physicians\' recordings (34/253, 13.4%) and parents\' recordings (31/266, 11.7%) with adventitious sounds were similar. Parents found the app easy to use (questionnaire: median 5, IQR 5-5) and were willing to use it (questionnaire: median 5, IQR 5-5).
    UNASSIGNED: Our results show that smartphone auscultation is feasible when performed by parents in the clinical context, but further investigation is needed to test its feasibility in real life.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    暴露于某些环境声音会对健康产生不利影响,并可能影响认知任务的表现。在这项研究中,我们分析了婴儿哭闹和狗叫声的邻里噪声对自主反应和认知功能的影响。
    二十名参与者被曝光,在单独的会议中,白噪声,一个婴儿在哭,一只小狗吠叫,一只大狗叫。在每次会议期间,心率,皮肤电导,反应时间,空间记忆,并在整个时间内采取数学处理措施。
    与暴露于白噪声相比,婴儿哭闹和狗叫声的声音导致心率和皮肤电导水平明显升高。结果与暴露于吠叫并不像婴儿那样一致。暴露于婴儿的哭声和狗叫声导致更快的反应时间,可能是由于自主系统激活的促进。没有发现对空间记忆的显着影响。相反,参与者表现更差和更慢的数学任务时,暴露于狗和婴儿的声音,而不是暴露在控制噪音下。
    暴露于哭闹的婴儿和狗吠叫的声音会导致交感神经反应增加和认知能力下降,与接触控制声音相比。应特别注意减轻对这些类型噪声的暴露。
    UNASSIGNED: The exposure to some environmental sounds has detrimental effects on health and might affect the performance in cognitive tasks. In this study, we analyze the effect of the neighborhood noises of a baby crying and dogs barking on the autonomic response and cognitive function.
    UNASSIGNED: Twenty participants were exposed, in separate sessions, to white noise, a baby crying, a small dog barking, and a large dog barking. During each session, heart rate, skin conductance, reaction times, spatial memory, and mathematical processing measures were taken throughout time.
    UNASSIGNED: The sounds of a baby crying and dogs barking led to significantly higher heart rates and skin conductance levels as opposed to exposure to white noise. Results were not as consistent with exposure to barking as they were to the baby. Exposure to the baby crying and dogs barking led to faster reaction times, possibly due to a facilitation by the autonomic system activation. No significant effects on spatial memory were found. Conversely, participants performed worse and slower in a mathematical task when exposed to the dog and baby sounds, than when exposed to control noise.
    UNASSIGNED: Exposure to the sound of crying babies and dogs barking leads to increased sympathetic response and decreased cognitive ability, as compared to exposure to control sounds. Special attention should be paid to the mitigation of exposure to these types of noises.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    高频听力损失是人口老龄化以及有暴露于大声噪音史的人群中最常见的问题之一。这种类型的听力损失可能令人沮丧和残疾,使得难以理解言语交流和与世界有效互动。
    这项研究旨在研究代表高频音素的空间独特触觉振动对日常情况下自我感知的能力的影响。
    为了解决高频听力损失,开发了一种多电机腕带,使用机器学习来收听特定的高频音素。腕带在空间上唯一的位置振动以实时表示哪个音素存在。总共招募了16名高频听力损失的参与者,并要求他们佩戴腕带6周。每周使用助听器受益的缩写简介(APHAB)测量与听力损失相关的残疾程度。
    在为期6周的研究结束时,所有参与者的平均APHAB福利得分达到12.39分,从基线40.32到最终评分27.93(SD13.11;N=16;P=0.002,双尾依赖t检验)。没有助听器的人在6周时平均APHAB获益评分比使用助听器的人提高10.78分(t14=2.14;P=.10,2尾独立t检验)。所有参与者的平均获益分数为15.44(SD13.88;N=16;P<.001,双尾依赖t检验)。所有参与者对背景噪声的平均获益分数为10.88(SD17.54;N=16;P=0.03,双尾依赖t检验)。所有参与者对混响的平均获益评分为10.84(SD16.95;N=16;P=.02,2尾依赖t检验)。
    这些发现表明,腕带产生与高频音素相对应的空间上可区分的振动所提供的振动触觉感觉替代,有助于高频听力损失的个人提高他们对言语交流的感知理解。无论一个人是否佩戴助听器,振动触觉反馈都能提供好处,尽管方式略有不同。最后,理解语音难度最大的人从振动触觉反馈中获得了最大的感知收益。
    High-frequency hearing loss is one of the most common problems in the aging population and with those who have a history of exposure to loud noises. This type of hearing loss can be frustrating and disabling, making it difficult to understand speech communication and interact effectively with the world.
    This study aimed to examine the impact of spatially unique haptic vibrations representing high-frequency phonemes on the self-perceived ability to understand conversations in everyday situations.
    To address high-frequency hearing loss, a multi-motor wristband was developed that uses machine learning to listen for specific high-frequency phonemes. The wristband vibrates in spatially unique locations to represent which phoneme was present in real time. A total of 16 participants with high-frequency hearing loss were recruited and asked to wear the wristband for 6 weeks. The degree of disability associated with hearing loss was measured weekly using the Abbreviated Profile of Hearing Aid Benefit (APHAB).
    By the end of the 6-week study, the average APHAB benefit score across all participants reached 12.39 points, from a baseline of 40.32 to a final score of 27.93 (SD 13.11; N=16; P=.002, 2-tailed dependent t test). Those without hearing aids showed a 10.78-point larger improvement in average APHAB benefit score at 6 weeks than those with hearing aids (t14=2.14; P=.10, 2-tailed independent t test). The average benefit score across all participants for ease of communication was 15.44 (SD 13.88; N=16; P<.001, 2-tailed dependent t test). The average benefit score across all participants for background noise was 10.88 (SD 17.54; N=16; P=.03, 2-tailed dependent t test). The average benefit score across all participants for reverberation was 10.84 (SD 16.95; N=16; P=.02, 2-tailed dependent t test).
    These findings show that vibrotactile sensory substitution delivered by a wristband that produces spatially distinguishable vibrations in correspondence with high-frequency phonemes helps individuals with high-frequency hearing loss improve their perceived understanding of verbal communication. Vibrotactile feedback provides benefits whether or not a person wears hearing aids, albeit in slightly different ways. Finally, individuals with the greatest perceived difficulty understanding speech experienced the greatest amount of perceived benefit from vibrotactile feedback.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Preprint
    睡眠对于将最近的经历巩固为长期记忆至关重要。作为一个关键的潜在神经元机制,睡眠期间发生的海马锐波波纹(SWRs)定义了最近经历的海马再激活时期,并且与记忆巩固有因果关系。睡眠期间海马SWR依赖性记忆巩固通常被称为发生在“离线”状态,致力于处理内部产生的神经活动模式,而不是外部刺激。然而,在睡眠期间,大脑并没有完全脱离环境。特别是,在睡眠中听到的声音由高度活跃的听觉系统处理,该系统投射到颞叶内侧的大脑区域,反映了海马活动声音调制的解剖学途径。虽然在睡眠过程中神经处理显著的声音,比如捕食者或后代,具有进化适应性,睡眠期间对环境声音的持续处理是否会干扰SWR依赖性记忆巩固仍不得而知.为了解决这个问题,我们使用闭环系统在睡眠大鼠的SWR期间或之后提供非清醒声音刺激。我们发现,在睡眠期间暴露于声音会抑制纹波功率并降低SWR的速率。此外,在SWR(开-SWR)期间传递的声音比在SWR(关-SWR)之后2秒传递的声音明显更抑制纹波功率。接下来,我们测试了睡眠期间声音表现对记忆巩固的影响。为此,在学习有条件的位置偏好范例后,在睡眠会话期间应用SWR触发的声音,其中老鼠学习了一个地方奖励协会。我们发现,学习后睡眠期间的On-SWR声音配对会在学习后24小时内完全消除记忆保留,同时在睡眠后立即保留记忆。相比之下,非SWR配对在学习后24小时以及学习后立即削弱了记忆。值得注意的是,与Off-SWR配对相比,On-SWR配对在学习后24小时引起的记忆明显更大的损害。一起,这些发现表明,在睡眠期间听到的声音会抑制SWRs和记忆巩固,并且这些影响的大小取决于声音SWR时序。这些结果表明,在睡眠期间暴露于环境声音可能会对记忆巩固过程构成风险。
    Sleep is critical for the consolidation of recent experiences into long-term memories. As a key underlying neuronal mechanism, hippocampal sharp-wave ripples (SWRs) occurring during sleep define periods of hippocampal reactivation of recent experiences and have been causally linked with memory consolidation. Hippocampal SWR-dependent memory consolidation during sleep is often referred to as occurring during an \"offline\" state, dedicated to processing internally generated neural activity patterns rather than external stimuli. However, the brain is not fully disconnected from the environment during sleep. In particular, sounds heard during sleep are processed by a highly active auditory system which projects to brain regions in the medial temporal lobe, reflecting an anatomical pathway for sound modulation of hippocampal activity. While neural processing of salient sounds during sleep, such as those of a predator or an offspring, is evolutionarily adaptive, whether ongoing processing of environmental sounds during sleep interferes with SWR-dependent memory consolidation remains unknown. To address this question, we used a closed-loop system to deliver non-waking sound stimuli during or following SWRs in sleeping rats. We found that exposure to sounds during sleep suppressed the ripple power and reduced the rate of SWRs. Furthermore, sounds delivered during SWRs (On-SWR) suppressed ripple power significantly more than sounds delivered 2 seconds after SWRs (Off-SWR). Next, we tested the influence of sound presentation during sleep on memory consolidation. To this end, SWR-triggered sounds were applied during sleep sessions following learning of a conditioned place preference paradigm, in which rats learned a place-reward association. We found that On-SWR sound pairing during post-learning sleep induced a complete abolishment of memory retention 24 h following learning, while leaving memory retention immediately following sleep intact. In contrast, Off-SWR pairing weakened memory 24 h following learning as well as immediately following learning. Notably, On-SWR pairing induced a significantly larger impairment in memory 24 h after learning as compared to Off-SWR pairing. Together, these findings suggest that sounds heard during sleep suppress SWRs and memory consolidation, and that the magnitude of these effects are dependent on sound-SWR timing. These results suggest that exposure to environmental sounds during sleep may pose a risk for memory consolidation processes.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Randomized Controlled Trial
    目标:心理健康问题正以惊人的速度增加,呼吁需要更具成本效益和更容易获得的干预措施。已经发现描绘自然的视觉图像和声音对个人的情绪和幸福有积极的影响;然而,图像和声音的综合效果几乎没有得到研究。因此,这项研究旨在比较观看与自然相关的舒缓图像与聆听舒缓声音以及两者的组合的情绪效果。
    方法:在本研究中,149名年龄在18-83岁之间的参与者(M=35.88,SD=15.63;72.5%为女性,男性26.8%,.7%的变性人)被随机分为三种干预条件:仅图像,声音仅或组合(图像和声音)。基线抑郁和焦虑症状被索引,和四个结果变量(正影响,负面影响,干预前后测量宁静情绪和抑郁情绪状态)。
    结果:研究结果表明,所有参与者,不分群体,报告负面影响减少,积极的情绪和抑郁情绪以及宁静情绪的增加(包括舒缓情绪)。然而,没有组间差异。探索性分析发现,基线时抑郁和焦虑症状水平较高的个体在负面影响和抑郁情绪状态方面经历了更大的减少。以及宁静影响的更大增加。
    结论:因此,这些发现提供了初步证据,经过进一步的研究和发展,描绘自然的图像和声音可能被开发用作改善情绪和福祉的有效工具。
    OBJECTIVE: Mental health problems are increasing at an alarming rate, calling for the need for more cost-effective and easily accessible interventions. Visual images and sounds depicting nature have been found to have positive effects on individuals\' mood and well-being; however, the combined effects of images and sounds have been scarcely investigated. This study therefore aimed to compare the mood effects of viewing nature-related soothing images versus listening to soothing sounds versus a combination of both.
    METHODS: In this study, 149 participants aged 18-83 years old (M = 35.88, SD = 15.63; 72.5% female, male 26.8%, .7% transgender) were randomised into three intervention conditions: images only, sounds only or combined (images and sounds). Baseline depressive and anxiety symptoms were indexed, and four outcome variables (positive affect, negative affect, serenity affect and depressive mood states) were measured pre- and post-intervention.
    RESULTS: Findings showed that all participants, regardless of group, reported a decrease in negative affect, positive affect and depressive mood as well as an increase in serenity affect (including feelings of soothe). However, there were no group differences. Exploratory analyses found that individuals with higher levels of depressive and anxiety symptoms at baseline experienced greater reduction in negative affect and depressive mood state, as well as a larger increase in serenity affect.
    CONCLUSIONS: These findings therefore provide preliminary evidence that, upon further research and development, images and sounds depicting nature can potentially be developed for use as an effective tool to improve mood and well-being.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Interpretation of breath sounds by auscultation has high inter-observer variability, even when performed by trained healthcare professionals. This can be mitigated by using Artificial Intelligence (AI) acoustic analysis. We aimed to develop and validate a novel breath sounds analysis system using AI-enabled algorithms to accurately interpret breath sounds in children. Subjects from the respiratory clinics and wards were auscultated by two independent respiratory paediatricians blinded to their clinical diagnosis. A novel device consisting of a stethoscope head connected to a smart phone recorded the breath sounds. The audio files were categorised into single label (normal, wheeze and crackles) or multi-label sounds. Together with commercially available breath sounds, an AI classifier was trained using machine learning. Unique features were identified to distinguish the breath sounds. Single label breath sound samples were used to validate the finalised Support Vector Machine classifier. Breath sound samples (73 single label, 20 multi-label) were collected from 93 children (mean age [SD] = 5.40 [4.07] years). Inter-rater concordance was observed in 81 (87.1%) samples. Performance of the classifier on the 73 single label breath sounds demonstrated 91% sensitivity and 95% specificity. The AI classifier developed could identify normal breath sounds, crackles and wheeze in children with high accuracy.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Journal Article
    Would you get close to a stinky perfume bottle or to a loudspeaker producing noise? In this paper, we present two procedures that allowed us to assess the ability of auditory and olfactory cues to elicit automatic approach/avoidance reactions toward their sources. The procedures resulted from an adaptation of the Visual Approach/Avoidance by the Self Task (VAAST; Rougier et al., 2018), a task having the peculiarity of simulating approach/avoidance reactions by using visual feedback coming from the whole-body movements. In the auditory VAAST (Experiment 1), participants were instructed to move forward or backward from a loudspeaker that produced spoken words differentiated by their level of distortion and thus by their hedonic value. In the olfactory VAAST (Experiment 2), participants were asked to move forward or backward from a perfume bottle that delivered pleasant and unpleasant odors. We expected, consistent with the approach/avoidance compatibility effect, shorter latencies for approaching positive stimuli and avoiding negative stimuli. In both experiments, we found an effect of the quality of the emotional stimulus on forward actions of participants, with undistorted words and pleasant odors inducing faster forward movements compared with that for distorted words and unpleasant odors. Notably, our results further suggest that the VAAST can successfully be used with implicit instructions, i.e., without requiring participants to explicitly process the valence of the emotional stimulus (in Experiment 1) or even the emotional stimulus itself (in Experiment 2). The sensitivity of our procedures is analyzed and its potential in cross-modal and (contextualized) consumer research discussed.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号