High-resolution functional imaging of speech-induced sensory modulation

语音引起的感觉调制的高分辨率功能成像

基本信息

  • 批准号:
    10802563
  • 负责人:
  • 金额:
    $ 45.88万
  • 依托单位:
  • 依托单位国家:
    美国
  • 项目类别:
  • 财政年份:
    2023
  • 资助国家:
    美国
  • 起止时间:
    2023-09-18 至 2028-07-31
  • 项目状态:
    未结题

项目摘要

PROJECT SUMMARY / ABSTRACT The overall goal of this project is to test and refine a neurobiological systems model of speech-induced sensory modulation (SISM). Previous studies used magnetoencephalography (MEG) and electroencephalography (EEG) to reveal that auditory cortical responses evoked when speaking are reduced compared to passive listening, a phenomenon known as speaking-induced suppression (SIS). This effect has been proposed to reflect the action of learned forward models that predict sensory consequences of speech movements. While EEG/MEG studies of SIS have provided important insights, these methods have limited anatomical precision. For this reason, the auditory regions modulated by forward model predictions in speech, and the circuits that mediate them, remain relatively poorly understood. To close this knowledge gap, this project uses 7T fMRI to measure brainwide activity at exquisite resolution while participants complete a series of related auditory-motor tasks across multiple sessions. Guided by contemporary models of speech production, we hypothesize that three functional circuits link frontal speech planning areas with superior temporal auditory areas via cortical and trans-cerebellar pathways, providing sensory predictions at different levels of representation. In Aim 1, one EEG and three fMRI experiments will be used to map the sensory cortical areas modulated by self-generated speech. These studies will define subject-specific regions-of-interest (ROIs) that differentially respond to overt speech vs. passive listening to oneself or another speaker producing the same syllables or sentences. Multivariate encoding models test hypotheses about the speech features represented in these ROIs. In the same speakers, we will determine the extent and specificity of modulations in these ROIs during movement planning, prior to motor output. Aim 2 tests how these ROIs participate in error monitoring. Two fMRI experiments are designed to test hypotheses about responses to auditory errors that are either (1) “low-level” acoustic-phonetic errors induced by unexpected shifts in auditory feedback, or (2) “high-level” sound substitution errors elicited during a phonological error priming task. Because Aim 2 is carried out in the same speakers as Aim 1, within-subject comparisons will enable high- resolution individual-specific models and interpretations across observed effects. In Aim 3, we will determine the role of the cerebellum (CB) in implementing learned forward models that drive SISM. Individuals with cerebellar lesions and controls will be recruited to test the hypothesis that the CB is critical in learning predictive models for speech. In these participants, and those tested in Aims 1 and 2, 7T fMRI will measure neural activity changes during a speech motor adaptation task and a non-vocal auditory-motor learning task. Individuals with lesions in Lobule VI are predicted to show reduced learning, reduced SISM, and reduced associations between CB and auditory cortical activity. Together these studies will accelerate understanding of sensory-motor interactions in speech through in-depth analysis of individual brains, enable the refinement of detailed neurocomputational models, and set the stage for precision approaches to diagnosis and intervention in speech disorders.
项目摘要/摘要

项目成果

期刊论文数量(0)
专著数量(0)
科研奖励数量(0)
会议论文数量(0)
专利数量(0)

数据更新时间:{{ journalArticles.updateTime }}

{{ item.title }}
{{ item.translation_title }}
  • DOI:
    {{ item.doi }}
  • 发表时间:
    {{ item.publish_year }}
  • 期刊:
  • 影响因子:
    {{ item.factor }}
  • 作者:
    {{ item.authors }}
  • 通讯作者:
    {{ item.author }}

数据更新时间:{{ journalArticles.updateTime }}

{{ item.title }}
  • 作者:
    {{ item.author }}

数据更新时间:{{ monograph.updateTime }}

{{ item.title }}
  • 作者:
    {{ item.author }}

数据更新时间:{{ sciAawards.updateTime }}

{{ item.title }}
  • 作者:
    {{ item.author }}

数据更新时间:{{ conferencePapers.updateTime }}

{{ item.title }}
  • 作者:
    {{ item.author }}

数据更新时间:{{ patent.updateTime }}

Jason W Bohland其他文献

Jason W Bohland的其他文献

{{ item.title }}
{{ item.translation_title }}
  • DOI:
    {{ item.doi }}
  • 发表时间:
    {{ item.publish_year }}
  • 期刊:
  • 影响因子:
    {{ item.factor }}
  • 作者:
    {{ item.authors }}
  • 通讯作者:
    {{ item.author }}

{{ truncateString('Jason W Bohland', 18)}}的其他基金

An open, online course in neuronal data analysis for the practicing neuroscientist
面向执业神经科学家的神经元数据分析开放在线课程
  • 批准号:
    9043612
  • 财政年份:
    2015
  • 资助金额:
    $ 45.88万
  • 项目类别:

相似海外基金

Nonlinear Acoustics for the conditioning monitoring of Aerospace structures (NACMAS)
用于航空航天结构调节监测的非线性声学 (NACMAS)
  • 批准号:
    10078324
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
    BEIS-Funded Programmes
ORCC: Marine predator and prey response to climate change: Synthesis of Acoustics, Physiology, Prey, and Habitat In a Rapidly changing Environment (SAPPHIRE)
ORCC:海洋捕食者和猎物对气候变化的反应:快速变化环境中声学、生理学、猎物和栖息地的综合(蓝宝石)
  • 批准号:
    2308300
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Continuing Grant
University of Salford (The) and KP Acoustics Group Limited KTP 22_23 R1
索尔福德大学 (The) 和 KP Acoustics Group Limited KTP 22_23 R1
  • 批准号:
    10033989
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Knowledge Transfer Partnership
User-controllable and Physics-informed Neural Acoustics Fields for Multichannel Audio Rendering and Analysis in Mixed Reality Application
用于混合现实应用中多通道音频渲染和分析的用户可控且基于物理的神经声学场
  • 批准号:
    23K16913
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Grant-in-Aid for Early-Career Scientists
Combined radiation acoustics and ultrasound imaging for real-time guidance in radiotherapy
结合辐射声学和超声成像,用于放射治疗的实时指导
  • 批准号:
    10582051
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
Comprehensive assessment of speech physiology and acoustics in Parkinson's disease progression
帕金森病进展中言语生理学和声学的综合评估
  • 批准号:
    10602958
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
The acoustics of climate change - long-term observations in the arctic oceans
气候变化的声学——北冰洋的长期观测
  • 批准号:
    2889921
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Studentship
Collaborative Research: Estimating Articulatory Constriction Place and Timing from Speech Acoustics
合作研究:从语音声学估计发音收缩位置和时间
  • 批准号:
    2343847
  • 财政年份:
    2023
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Standard Grant
Flow Physics and Vortex-Induced Acoustics in Bio-Inspired Collective Locomotion
仿生集体运动中的流动物理学和涡激声学
  • 批准号:
    DGECR-2022-00019
  • 财政年份:
    2022
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Discovery Launch Supplement
Collaborative Research: Estimating Articulatory Constriction Place and Timing from Speech Acoustics
合作研究:从语音声学估计发音收缩位置和时间
  • 批准号:
    2141275
  • 财政年份:
    2022
  • 资助金额:
    $ 45.88万
  • 项目类别:
    Standard Grant
{{ showInfoDetail.title }}

作者:{{ showInfoDetail.author }}

知道了