VOCAL

声乐
  • 文章类型: Journal Article
    为了评估一种称为SmartERA(智能子宫内膜容受性分析)的新型自动化技术的可重复性,用于子宫正常患者子宫内膜的自动分割和体积计算,,并比较SmartERA之间子宫内膜体积测量的一致性,半自动虚拟器官计算机辅助分析(VOCAL)技术和手动分割。这项回顾性研究评估了接受冻融胚胎移植(FET)的不育患者的子宫内膜体积测量。使用ResonaR9超声机进行经阴道三维超声扫描。数据收集自2021年至2022年的患者。包括子宫正常和最佳超声图像的患者。使用SmartERA测量子宫内膜体积,VOCAL在15°旋转,和手动分割。使用组内相关系数(ICC)和Bland-Altman分析评估了观察者内部的可重复性和技术之间的一致性。总共评估了407名女性患者(平均年龄33.2±4.7岁)。SmartERA的可重复性显示ICC为0.983(95%CI0.984-0.991)。SmartERA和手动方法之间的协议,智能时代和语音,和VOCAL和手动方法,根据国际商会的评估,为0.986(95%CI0.977-0.990),0.943(95%CI0.934-0.963),和0.951(95%CI0.918-0.969),分别。SmartERA技术需要大约3s来计算子宫内膜体积,而VOCAL大约需要5分钟,手动分割方法大约需要50分钟。Smart-ERA软件,它采用了一种新颖的三维分割算法,在子宫正常的女性中,具有出色的观察者内部可重复性,并且与VOCAL和手动分割的子宫内膜体积测量高度吻合。然而,这些发现应该谨慎解释,因为该算法的性能可能无法推广到具有不同子宫特征的人群。此外,与VOCAL和手动分段相比,智能ERA所需的时间明显减少。
    To evaluate the repeatability of a novel automated technique called Smart ERA (Smart Endometrial Receptivity Analysis) for the automated segmentation and volume calculation of the endometrium in patients with normal uteri,, and to compare the agreement of endometrial volume measurements between Smart ERA, the semi-automated Virtual Organ Computer-aided Analysis (VOCAL) technique and manual segmentation. This retrospective study evaluated endometrial volume measurement in infertile patients who underwent frozen-thawed embryo transfer (FET). Transvaginal three-dimensional ultrasound scans were performed using a Resona R9 ultrasound machine. Data was collected from patients between 2021 and 2022. Patients with normal uteri and optimal ultrasound images were included. Endometrial volumes were measured using Smart ERA, VOCAL at 15° rotation, and manual segmentation. Intra-observer repeatability and agreement between techniques were assessed using the intraclass correlation coefficient (ICC) and Bland-Altman analysis. A total of 407 female patients were evaluated (mean age 33.2 ± 4.7 years). The repeatability of Smart ERA showed an ICC of 0.983 (95% CI 0.984-0.991). The agreement between Smart ERA and the manual method, Smart ERA and VOCAL, and VOCAL and the manual method, as assessed by ICC, were 0.986 (95% CI 0.977-0.990), 0.943 (95% CI 0.934-0.963), and 0.951 (95% CI 0.918-0.969), respectively. The Smart ERA technique required approximately 3 s for endometrial volume calculation, while VOCAL took around 5 min and the manual segmentation method took approximately 50 min. The Smart-ERA software, which employs a novel three-dimensional segmentation algorithm, demonstrated excellent intra-observer repeatability and high agreement with both VOCAL and manual segmentation for endometrial volume measurement in women with normal uteri. However, these findings should be interpreted with caution, as the algorithm\'s performance may not be generalizable to populations with different uterine characteristic. Additionally, Smart ERA required significantly less time compared to VOCAL and manual segmentation.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:肥胖的全球患病率不断上升,需要探索新的诊断方法。最近的科学调查表明,与肥胖相关的语音特征可能发生变化,提示使用语音作为肥胖检测的非侵入性生物标志物的可行性。
    目的:本研究旨在通过对短录音的分析,使用深度神经网络来预测肥胖状态,研究声乐特征与肥胖的关系。
    方法:对696名参与者进行了一项初步研究,使用自我报告的BMI将个体分为肥胖和非肥胖组。参与者阅读简短脚本的录音被转换为频谱图,并使用改编的YOLOv8模型(Ultralytics)进行分析。使用准确性对模型性能进行了评估,召回,精度,和F1分数。
    结果:适应的YOLOv8模型显示出0.70的全局准确性和0.65的宏F1评分。在识别非肥胖(F1评分为0.77)方面比肥胖(F1评分为0.53)更有效。这种中等水平的准确性凸显了使用声乐生物标志物进行肥胖检测的潜力和挑战。
    结论:虽然该研究在基于语音的肥胖医学诊断领域显示出希望,它面临着一些限制,比如依赖自我报告的BMI数据,均匀的样本量。这些因素,再加上录音质量的可变性,需要使用更强大的方法和不同的样本进行进一步的研究,以增强这种新颖方法的有效性。这些发现为将来使用语音作为肥胖检测的非侵入性生物标志物的研究奠定了基础。
    BACKGROUND: The escalating global prevalence of obesity has necessitated the exploration of novel diagnostic approaches. Recent scientific inquiries have indicated potential alterations in voice characteristics associated with obesity, suggesting the feasibility of using voice as a noninvasive biomarker for obesity detection.
    OBJECTIVE: This study aims to use deep neural networks to predict obesity status through the analysis of short audio recordings, investigating the relationship between vocal characteristics and obesity.
    METHODS: A pilot study was conducted with 696 participants, using self-reported BMI to classify individuals into obesity and nonobesity groups. Audio recordings of participants reading a short script were transformed into spectrograms and analyzed using an adapted YOLOv8 model (Ultralytics). The model performance was evaluated using accuracy, recall, precision, and F1-scores.
    RESULTS: The adapted YOLOv8 model demonstrated a global accuracy of 0.70 and a macro F1-score of 0.65. It was more effective in identifying nonobesity (F1-score of 0.77) than obesity (F1-score of 0.53). This moderate level of accuracy highlights the potential and challenges in using vocal biomarkers for obesity detection.
    CONCLUSIONS: While the study shows promise in the field of voice-based medical diagnostics for obesity, it faces limitations such as reliance on self-reported BMI data and a small, homogenous sample size. These factors, coupled with variability in recording quality, necessitate further research with more robust methodologies and diverse samples to enhance the validity of this novel approach. The findings lay a foundational step for future investigations in using voice as a noninvasive biomarker for obesity detection.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:数字时代见证了对新闻和信息的数字平台的日益依赖,再加上“deepfake”技术的出现。Deepfakes,利用语音记录和图像的大量数据集的深度学习模型,对媒体真实性构成重大威胁,可能导致不道德的滥用,如冒充和传播虚假信息。
    目标:为了应对这一挑战,这项研究旨在引入先天生物过程的概念,以区分真实的人类声音和克隆的声音。我们建议存在或不存在某些感知特征,比如讲话中的停顿,可以有效区分克隆和真实的音频。
    方法:共招募了49名具有不同种族背景和口音的成年参与者。每个参与者贡献语音样本,用于训练多达3个不同的语音克隆文本到语音模型和3个控制段落。随后,克隆模型生成了控制段落的合成版本,产生由每个参与者多达9个克隆音频样本和3个对照样本组成的数据集。我们分析了呼吸等生物行为引起的语音停顿,吞咽,和认知过程。计算了对应于语音暂停简档的五个音频特征。评估了这些特征的真实音频和克隆音频之间的差异,和5个经典的机器学习算法实现了使用这些特征来创建预测模型。通过对看不见的数据进行测试,评估了最优模型的泛化能力,结合了一个朴素的生成器,一个模型天真的段落,和幼稚的参与者。
    结果:克隆音频显示暂停之间的时间显着增加(P<.001),语音段长度的变化减少(P=0.003),发言时间的总比例增加(P=.04),语音中的micro和macropauses比率降低(P=0.01)。使用这些功能实现了五个机器学习模型,AdaBoost模型展示了最高的性能,实现5倍交叉验证平衡精度为0.81(SD0.05)。其他模型包括支持向量机(平衡精度0.79,SD0.03),随机森林(平衡精度0.78,SD0.04),逻辑回归,和决策树(平衡精度0.76,SD0.10和0.72,SD0.06)。在评估最优AdaBoost模型时,在预测未知数据时,它实现了0.79的总体测试准确性。
    结论:引入感知,机器学习模型中的生物特征在区分真实的人类声音和克隆音频方面显示出有希望的结果。
    BACKGROUND: The digital era has witnessed an escalating dependence on digital platforms for news and information, coupled with the advent of \"deepfake\" technology. Deepfakes, leveraging deep learning models on extensive data sets of voice recordings and images, pose substantial threats to media authenticity, potentially leading to unethical misuse such as impersonation and the dissemination of false information.
    OBJECTIVE: To counteract this challenge, this study aims to introduce the concept of innate biological processes to discern between authentic human voices and cloned voices. We propose that the presence or absence of certain perceptual features, such as pauses in speech, can effectively distinguish between cloned and authentic audio.
    METHODS: A total of 49 adult participants representing diverse ethnic backgrounds and accents were recruited. Each participant contributed voice samples for the training of up to 3 distinct voice cloning text-to-speech models and 3 control paragraphs. Subsequently, the cloning models generated synthetic versions of the control paragraphs, resulting in a data set consisting of up to 9 cloned audio samples and 3 control samples per participant. We analyzed the speech pauses caused by biological actions such as respiration, swallowing, and cognitive processes. Five audio features corresponding to speech pause profiles were calculated. Differences between authentic and cloned audio for these features were assessed, and 5 classical machine learning algorithms were implemented using these features to create a prediction model. The generalization capability of the optimal model was evaluated through testing on unseen data, incorporating a model-naive generator, a model-naive paragraph, and model-naive participants.
    RESULTS: Cloned audio exhibited significantly increased time between pauses (P<.001), decreased variation in speech segment length (P=.003), increased overall proportion of time speaking (P=.04), and decreased rates of micro- and macropauses in speech (both P=.01). Five machine learning models were implemented using these features, with the AdaBoost model demonstrating the highest performance, achieving a 5-fold cross-validation balanced accuracy of 0.81 (SD 0.05). Other models included support vector machine (balanced accuracy 0.79, SD 0.03), random forest (balanced accuracy 0.78, SD 0.04), logistic regression, and decision tree (balanced accuracies 0.76, SD 0.10 and 0.72, SD 0.06). When evaluating the optimal AdaBoost model, it achieved an overall test accuracy of 0.79 when predicting unseen data.
    CONCLUSIONS: The incorporation of perceptual, biological features into machine learning models demonstrates promising results in distinguishing between authentic human voices and cloned audio.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    阴性分裂型性状可能可以使用客观的声音分析进行数字表型分析。先前的尝试在这方面显示出不同的成功,可能是因为声学分析依赖于小的,约束要素集。我们使用机器学习来(a)使用大型声学特征集优化和交叉验证自我报告的阴性分裂型的预测模型,(b)评估模型表现作为性别和说话任务的函数,(c)通过评估这些模型中的关键声学特征,了解潜在的负分裂型特征的潜在机制,和(d)检查模型性能与临床症状和认知功能的收敛性。准确性良好(>80%),并通过考虑说话任务和性别而提高。然而,被鉴定为最具阴性分裂型性状预测能力的特征通常不被认为对其概念定义至关重要.讨论了验证和实施数字表型以理解和量化阴性分裂型的含义。
    Negative schizotypal traits potentially can be digitally phenotyped using objective vocal analysis. Prior attempts have shown mixed success in this regard, potentially because acoustic analysis has relied on small, constrained feature sets. We employed machine learning to (a) optimize and cross-validate predictive models of self-reported negative schizotypy using a large acoustic feature set, (b) evaluate model performance as a function of sex and speaking task, (c) understand potential mechanisms underlying negative schizotypal traits by evaluating the key acoustic features within these models, and (d) examine model performance in its convergence with clinical symptoms and cognitive functioning. Accuracy was good (> 80%) and was improved by considering speaking task and sex. However, the features identified as most predictive of negative schizotypal traits were generally not considered critical to their conceptual definitions. Implications for validating and implementing digital phenotyping to understand and quantify negative schizotypy are discussed.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    暂无摘要。
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    这份手稿的目的是回顾适应症,功效,585nm脉冲染料激光(PDL)在非恶性喉部病变中的安全性。根据PRISMA声明的建议,三位独立作者搜索了发表在PubMed/MEDLINE上的文章,Cochrane图书馆,谷歌学者,Scielo,和WebofScience。根据NICE指导工具进行偏倚分析。从506个确定的出版物中,19项观察性研究符合纳入标准。与其他治疗方法相比,PDL客观和主观地改善了血管病变中的声带质量(p<0.005),并改善了异型增生/白细胞增生患者的声带质量,而不改变疾病的自然史。Reinke的水肿和肉芽肿平均需要1.5PDL疗程才能解决。复发性呼吸道乳头状瘤病的治疗需要多次治疗,在50-70%的患者中实现了完全缓解。不管病变如何,局部麻醉下的手术耐受性是例外的(84-97%),在回归和声音质量方面的结果是有希望的。并发症发生率很低,并且该程序不会干扰其他治疗方法。关于激光设置没有共识。在评估声乐结果时缺乏一致的使用,无论是客观的还是主观的,防止研究之间的可比性。对于非恶性喉病理患者,585nm脉冲染料激光似乎是一种有效且安全的治疗选择。需要进行未来的对照研究,以将585nm脉冲染料激光器与其他激光器或冷仪器程序进行比较。
    The objective of this manuscript was to review the indications, efficacy, and safety of a 585 nm pulsed dye laser (PDL) in non-malignant laryngeal lesions. Following the PRISMA statement recommendations, three independent authors searched for articles published in PubMed/MEDLINE, the Cochrane Library, Google Scholar, Scielo, and Web of Science. A bias analysis was performed following NICE guidance tools. From the 506 identified publications, 19 observational studies met the inclusion criteria. The PDL improves vocal quality objectively and subjectively in vascular lesions (p < 0.005) and improves vocal quality in patients with dysplasia/leukoplasia without changing the natural history of the disease compared to other treatments. Reinke\'s edema and granulomas require an average of 1.5 PDL sessions for resolution. Treatment of recurrent respiratory papillomatosis requires multiple sessions, with complete remission achieved in 50-70% of patients. Regardless of the lesion, the tolerance of the procedure under local anesthesia is exceptional (84-97%), and the results in terms of regression and vocal quality are promising. The complication rate is minimal, and the procedure does not interfere with other treatment alternatives. There is no consensus on laser settings. The lack of consistent use in evaluating vocal outcomes, whether objective or subjective, prevents the comparability between studies. The 585 nm pulsed dye laser appears to be an effective and safe therapeutic option in patients with non-malignant laryngeal pathology. Future controlled studies are needed to compare the 585 nm pulsed dye laser with other lasers or cold instrument procedures.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    背景:基于声乐生物标志物的机器学习方法在检测各种健康状况方面显示出有希望的结果,包括哮喘等呼吸道疾病。在这项研究中,我们旨在验证最初在哮喘和健康志愿者数据集上训练的呼吸反应性声带生物标志物(RRVB)平台的区分能力,没有修改,活跃的COVID-19感染与向美国和印度医院展示患者的健康志愿者。
    目的:本研究的目的是确定RRVB模型是否可以区分患有活动性COVID-19感染的患者。无症状健康志愿者通过评估其敏感性,特异性,和赔率比。另一个目的是评估RRVB模型输出是否与COVID-19的症状严重程度相关。
    方法:使用语音声学特征的加权和的逻辑回归模型先前在约1,700名确诊哮喘患者的数据集上进行了训练和验证类似数量的健康对照。相同的模型已显示出对慢性阻塞性肺疾病(COPD)患者的普遍性,间质性肺病(ILD),还有咳嗽.在本研究中,共有497名参与者(46%为男性,54%女性;94%<65岁,6%>=65岁;51%马拉地语,45%英语,5%的西班牙语使用者)在美国和印度的四个临床站点注册,并在其个人智能手机上提供语音样本和症状报告。参与者包括有症状的COVID-19阳性和阴性患者以及无症状的健康志愿者。通过与RT-PCR证实的COVID-19的临床诊断进行比较,评估了RRVB模型的性能。
    结果:RRVB模型区分呼吸系统疾病患者的能力与健康对照以前在哮喘的验证数据上得到了证明,COPD,ILD和咳嗽的比值比分别为4.3、9.1、3.1和3.9。本研究在COVID-19中进行的RRVB模型相同,灵敏度为73.2%,特异性为62.9%,比值比为4.64(p<0.0001)。出现呼吸道症状的患者比未出现呼吸道症状和完全无症状的患者更频繁地检测到(78.4%vs.67.4%与68.0%)。
    结论:RRVB模型在呼吸条件下显示出良好的泛化性,地理位置,和语言。COVID-19的结果表明,它有可能作为一种预筛查工具,用于结合温度和症状报告识别有COVID-19感染风险的受试者。虽然不是COVID-19测试,这些结果表明,RRVB模型可以鼓励有针对性的测试。此外,该模型在不同的语言和地理环境中检测呼吸道症状的通用性提示了开发和验证未来用于更广泛疾病监测和监测应用的基于语音的工具的潜在途径.
    背景:ClinicalTrials.gov(NCT04582331。
    Vocal biomarker-based machine learning approaches have shown promising results in the detection of various health conditions, including respiratory diseases, such as asthma.
    This study aimed to determine whether a respiratory-responsive vocal biomarker (RRVB) model platform initially trained on an asthma and healthy volunteer (HV) data set can differentiate patients with active COVID-19 infection from asymptomatic HVs by assessing its sensitivity, specificity, and odds ratio (OR).
    A logistic regression model using a weighted sum of voice acoustic features was previously trained and validated on a data set of approximately 1700 patients with a confirmed asthma diagnosis and a similar number of healthy controls. The same model has shown generalizability to patients with chronic obstructive pulmonary disease, interstitial lung disease, and cough. In this study, 497 participants (female: n=268, 53.9%; <65 years old: n=467, 94%; Marathi speakers: n=253, 50.9%; English speakers: n=223, 44.9%; Spanish speakers: n=25, 5%) were enrolled across 4 clinical sites in the United States and India and provided voice samples and symptom reports on their personal smartphones. The participants included patients who are symptomatic COVID-19 positive and negative as well as asymptomatic HVs. The RRVB model performance was assessed by comparing it with the clinical diagnosis of COVID-19 confirmed by reverse transcriptase-polymerase chain reaction.
    The ability of the RRVB model to differentiate patients with respiratory conditions from healthy controls was previously demonstrated on validation data in asthma, chronic obstructive pulmonary disease, interstitial lung disease, and cough, with ORs of 4.3, 9.1, 3.1, and 3.9, respectively. The same RRVB model in this study in COVID-19 performed with a sensitivity of 73.2%, specificity of 62.9%, and OR of 4.64 (P<.001). Patients who experienced respiratory symptoms were detected more frequently than those who did not experience respiratory symptoms and completely asymptomatic patients (sensitivity: 78.4% vs 67.4% vs 68%, respectively).
    The RRVB model has shown good generalizability across respiratory conditions, geographies, and languages. Results using data set of patients with COVID-19 demonstrate its meaningful potential to serve as a prescreening tool for identifying individuals at risk for COVID-19 infection in combination with temperature and symptom reports. Although not a COVID-19 test, these results suggest that the RRVB model can encourage targeted testing. Moreover, the generalizability of this model for detecting respiratory symptoms across different linguistic and geographic contexts suggests a potential path for the development and validation of voice-based tools for broader disease surveillance and monitoring applications in the future.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Case Reports
    鼻气管插管通常在口腔颌面手术中在全身麻醉下进行。为了手术的方便,鼻环-Adair-Elwyn(RAE)管主要使用。因为鼻RAE管弯曲成“L”形,插入深度有限。特别是,有必要准确确定儿童RAE管的适当深度。医疗市场上使用了几种类型的鼻RAE管,在材料和长度上有所不同。我们使用鼻RAE管进行了气管插管,用于双颌手术,但是即使袖带中的气压增加,空气泄漏仍然存在。用喉镜检查时,确认管子被推出了,袖口卡在声带上,导致空气泄漏。由于深入插入管子并不能解决问题,用鼻RAE管更换(Polar™,预成型气管导管,史密斯医疗,Inc.,美国)没有造成空气泄漏;因此,我们报告了这个病例。
    Nasotracheal intubation is commonly performed under general anesthesia in oral and maxillofacial surgery. For the convenience of surgery, nasal Ring-Adair-Elwyn (RAE) tubes are mainly used. Because the nasal RAE tubes were bent in an \"L\" shape, the insertion depth was limited. Particularly, it is necessary to accurately determine the appropriate depth of the RAE tubes in children. Several types of nasal RAE tubes are used in the medical market, which vary in material and length. We performed endotracheal intubation using a nasal RAE tube for double-jaw surgery, but air leakage persisted even when the air pressure in the cuff was increased. When checked with a laryngoscope, it was confirmed that the tube was pushed out, and the cuff was caught on the vocal cords, causing air leakage. Since inserting the tube deeply did not solve the problem, replacing it with a nasal RAE tube (Polar™, Preformed Tracheal Tube, Smith Medical, Inc., USA) did not cause air leakage; thus, we reported this case.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

  • 文章类型: Journal Article
    自闭症青年在情感识别方面表现出困难,然而,很少有研究研究声音情绪识别(VER)的行为和神经指标。当前的研究检查了自闭症和非自闭症青年中VER的行为和事件相关电位(N100,P200,晚期正电位[LPP])指数。参与者(N=164)完成了情绪识别任务,非语言准确性的诊断分析(DANVA-2),包括VER,在脑电图记录期间。响应于高强度VER,LPP振幅较大,社会认知预测了VER错误。言语智商,不是自闭症,与VER错误有关。VER强度与社交交流障碍之间的相互作用表明,这些障碍与低强度VER期间较大的LPP幅度有关。一起来看,VER的差异可能是由于高阶认知过程,不是基本的,早期感知(N100,P200),言语认知能力可能是行为的基础,然而闭塞神经,VER处理的差异。
    Autistic youth display difficulties in emotion recognition, yet little research has examined behavioral and neural indices of vocal emotion recognition (VER). The current study examines behavioral and event-related potential (N100, P200, Late Positive Potential [LPP]) indices of VER in autistic and non-autistic youth. Participants (N = 164) completed an emotion recognition task, the Diagnostic Analyses of Nonverbal Accuracy (DANVA-2) which included VER, during EEG recording. The LPP amplitude was larger in response to high intensity VER, and social cognition predicted VER errors. Verbal IQ, not autism, was related to VER errors. An interaction between VER intensity and social communication impairments revealed these impairments were related to larger LPP amplitudes during low intensity VER. Taken together, differences in VER may be due to higher order cognitive processes, not basic, early perception (N100, P200), and verbal cognitive abilities may underlie behavioral, yet occlude neural, differences in VER processing.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

    求助全文

  • 文章类型: Systematic Review
    声音交流在现存的脊椎动物中使用,在进化上是古老的,并被维护,在许多血统中。在这里,我回顾了支持代表性anuran中种内声学信号的神经电路架构,哺乳动物和鸟类以及两种无脊椎动物,果蝇和夏威夷板球。我专注于后脑运动控制图案及其与呼吸回路的联系,运动性腺类固醇受体的表达,感官,和边缘神经元以及引起声音反应的不同模式。声音交流的后脑和边缘参与者是高度保守的,虽然前脑参与者在无脑和哺乳动物之间有分歧,还有鸣鸟和啮齿动物。我讨论了自然选择和性选择在驱动物种形成中的作用,以及在呼吸中具有祖先作用的回路元件的暴露,用于产生声音和驾驶有节奏的人声特征。跨物种的全脑功能磁共振成像的最新技术进步将使声学信号伙伴的实时成像成为可能,将听觉感知与声乐制作联系起来。
    Vocal communication is used across extant vertebrates, is evolutionarily ancient, and been maintained, in many lineages. Here I review the neural circuit architectures that support intraspecific acoustic signaling in representative anuran, mammalian and avian species as well as two invertebrates, fruit flies and Hawaiian crickets. I focus on hindbrain motor control motifs and their ties to respiratory circuits, expression of receptors for gonadal steroids in motor, sensory, and limbic neurons as well as divergent modalities that evoke vocal responses. Hindbrain and limbic participants in acoustic communication are highly conserved, while forebrain participants have diverged between anurans and mammals, as well as songbirds and rodents. I discuss the roles of natural and sexual selection in driving speciation, as well as exaptation of circuit elements with ancestral roles in respiration, for producing sounds and driving rhythmic vocal features. Recent technical advances in whole brain fMRI across species will enable real time imaging of acoustic signaling partners, tying auditory perception to vocal production.
    导出

    更多引用

    收藏

    翻译标题摘要

    我要上传

       PDF(Pubmed)

公众号