Characterizing the recovery of spectral, temporal, and phonemic speech information from visual cues
表征从视觉线索中恢复频谱、时间和音位语音信息
基本信息
- 批准号:10563860
- 负责人:
- 金额:$ 55.04万
- 依托单位:
- 依托单位国家:美国
- 项目类别:
- 财政年份:2023
- 资助国家:美国
- 起止时间:2023-02-14 至 2028-01-31
- 项目状态:未结题
- 来源:
- 关键词:AcousticsAttentionAuditoryAuditory areaAuditory systemBiologicalBrainBrain InjuriesBrain NeoplasmsClassificationCochlear ImplantsCodeCompensationCrowdingCuesDataDevelopmentDevicesDimensionsDistributional ActivityElectrodesElectroencephalographyEmotionalFrequenciesFunctional Magnetic Resonance ImagingHealthHearingHumanIllusionsImpairmentIndividualLipreadingMapsMeasuresModalityModelingMovementNeuronsNoiseOralOral cavityParticipantPatientsPatternPerceptionPeriodicityPhysiological ProcessesPopulationPresbycusisProcessReaction TimeRecoveryRehabilitation therapyResearchResolutionResourcesRouteShapesSignal TransductionSocial InteractionSpeechSpeech PerceptionSpeech SoundStimulusStrokeSuperior temporal gyrusSystemTestingTitrationsTraining ProgramsTraumaVisionVisualVocationaudiovisual speechauditory stimulusdensityhealthy agingimprovedneural prosthesisneuromechanismprogramsresponserestorationsensory substitutionsocialspeech accuracyvisual informationvisual speech
项目摘要
Project Summary
Auditory speech perception is essential for social, vocational, and emotional health in hearing individuals.
However, the reliability of auditory signals varies widely in everyday settings (e.g., at a crowded party), requiring
supplemental processes to enable accurate speech perception. The principle mechanisms that support the
perception of degraded auditory speech signals are auditory-visual (crossmodal) interactions, which can
perceptually restore speech content using visual cues provided by lipreading, rhythmic articulatory movements,
and the natural correlations present between oral resonance and mouth shape. Moreover, receptive speech
processes can be limited through a variety of causes, including intrinsic brain tumor, stroke, cochlear implant
usage, and age-related hearing loss, making compensatory crossmodal mechanisms necessary for one to
continue working and maintaining healthy social interactions. However, the physiological processes that enable
vision to facilitate speech perception remain poorly understood and no integrative model exists for how these
multiple visual dimensions combine to enhance auditory speech perception. In the auditory domain, distributed
populations of neurons encode spectro-temporal information about acoustic cues that are then transcoded into
phonemes. We propose a dual-route perceptual model through which visual signals integrate with phoneme-
coded neurons. First, a direct path through which viseme-to-phoneme conversions generate semi-overlapping
distributions of activity in the superior temporal gyrus, leading to improved hearing through improved auditory
phoneme tuning functions. Second, an indirect path through which visual features restore spectral information
about speech frequencies and alter phoneme-response timing, resulting in improved auditory spectro-temporal
profiles (which in turn are transcoded into phonemes with greater precision). Finally, we will examine the
hypothesis that our perceptual system optimizes which of these visual dimensions is prioritized for recovery
based on what is missing from the auditory signal. These studies will provide a unified framework for how speech
perception benefits from different visual signals. By understanding biological approaches to crossmodally
restoring degraded auditory speech information, we can develop better targeted rehabilitation programs and
neural prostheses to maximize speech perception recovery after trauma or during healthy aging.
项目摘要
听觉言语感知对于听力个人的社会,职业和情感健康至关重要。
但是,听觉信号的可靠性在日常设置(例如,在拥挤的聚会上)有很大差异,需要
补充过程以实现准确的语音感知。支持的主要机制
降解的听觉语音信号的感知是听觉 - 视觉(Crossmodal)的交互,可以
感知恢复语音内容,使用唇读,有节奏的发音运动提供的视觉提示,
以及口腔共振和口形状之间存在的自然相关性。而且,接受演讲
过程可以通过多种原因限制,包括固有的脑肿瘤,中风,人工耳蜗
用法和与年龄相关的听力损失,使人需要的补偿性跨模式机制
继续工作并保持健康的社会互动。但是,可以实现的生理过程
促进言语感知的愿景仍然很众所周知,并且没有综合模型
多个视觉维度结合在一起,以增强听觉语音感知。在听觉域中,分布式
神经元的群体编码有关声明的光谱信息信息,然后将其转码
音素。我们提出了一个双路由感知模型,通过该模型,视觉信号与音素集成在一起
编码神经元。首先,通过该直接路径,通过该路径,Viseme-phoneme转换会生成半重叠的路径
高级颞回中活动的分布,从而通过改善听觉改善了听力
音素调整功能。其次,一个间接路径通过该路径恢复光谱信息
关于语音频率和更改音素响应时间,从而改善了听觉光谱时间
剖面(依次将其精确地转码为音素)。最后,我们将研究
假设我们的感知系统优化了这些视觉维度中的哪一个用于恢复
基于听觉信号所缺少的内容。这些研究将为语音如何提供一个统一的框架
感知受益于不同的视觉信号。通过了解跨模型的生物学方法
恢复降级的听觉语音信息,我们可以制定针对性更好的康复计划和
神经假体以最大程度地提高创伤后或健康衰老期间的语音感知恢复。
项目成果
期刊论文数量(0)
专著数量(0)
科研奖励数量(0)
会议论文数量(0)
专利数量(0)
数据更新时间:{{ journalArticles.updateTime }}
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
数据更新时间:{{ journalArticles.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ monograph.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ sciAawards.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ conferencePapers.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ patent.updateTime }}
David Brang其他文献
David Brang的其他文献
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
{{ truncateString('David Brang', 18)}}的其他基金
Networks Underlying Visual Modulation of Speech Perception
语音感知视觉调制的网络基础
- 批准号:
9337601 - 财政年份:2016
- 资助金额:
$ 55.04万 - 项目类别:
Networks Underlying Visual Modulation of Speech Perception
语音感知视觉调制的网络
- 批准号:
9353752 - 财政年份:2016
- 资助金额:
$ 55.04万 - 项目类别:
Networks underlying visual modulation of speech perception
语音感知视觉调制的网络
- 批准号:
8959922 - 财政年份:2014
- 资助金额:
$ 55.04万 - 项目类别:
相似国自然基金
人机共驾模式下驾驶人监管注意力弱化-恢复规律与调控机理
- 批准号:52302425
- 批准年份:2023
- 资助金额:30 万元
- 项目类别:青年科学基金项目
逆全球化下跨国企业动态能力形成的微观机理研究:高管注意力配置视角
- 批准号:72302220
- 批准年份:2023
- 资助金额:30 万元
- 项目类别:青年科学基金项目
注意力感知驱动的车载多模态传感器在线协同校正
- 批准号:42301468
- 批准年份:2023
- 资助金额:30 万元
- 项目类别:青年科学基金项目
基于两阶段注意力深度学习方法的系统性金融风险测度与预警研究
- 批准号:72301101
- 批准年份:2023
- 资助金额:30 万元
- 项目类别:青年科学基金项目
采用多种稀疏自注意力机制的Transformer隧道衬砌裂缝检测方法研究
- 批准号:62301339
- 批准年份:2023
- 资助金额:30.00 万元
- 项目类别:青年科学基金项目
相似海外基金
Optimizing bilateral and single-sided-deafness cochlear implants for functioning in complex auditory environments
优化双侧和单侧耳聋人工耳蜗植入物以在复杂的听觉环境中发挥作用
- 批准号:
10654316 - 财政年份:2023
- 资助金额:
$ 55.04万 - 项目类别:
Dynamic neural coding of spectro-temporal sound features during free movement
自由运动时谱时声音特征的动态神经编码
- 批准号:
10656110 - 财政年份:2023
- 资助金额:
$ 55.04万 - 项目类别:
Identifying acoustic-level and language-specific sensory processing mechanisms
识别声学级别和语言特定的感觉处理机制
- 批准号:
10711229 - 财政年份:2023
- 资助金额:
$ 55.04万 - 项目类别:
Characterizing the generative mechanisms underlying the cortical tracking of natural speech
表征自然语音皮质跟踪背后的生成机制
- 批准号:
10710717 - 财政年份:2023
- 资助金额:
$ 55.04万 - 项目类别:
Step 1 in Designing Appropriate Shams and Controls in Human TUS
在人类 TUS 中设计适当的假手术和对照的步骤 1
- 批准号:
10735292 - 财政年份:2023
- 资助金额:
$ 55.04万 - 项目类别: