Establishing a ground truth for focus placement in naturally-occurring speech
为自然发生的语音中的焦点放置建立基本事实
基本信息
- 批准号:1737846
- 负责人:
- 金额:$ 10.59万
- 依托单位:
- 依托单位国家:美国
- 项目类别:Standard Grant
- 财政年份:2017
- 资助国家:美国
- 起止时间:2017-07-01 至 2022-06-30
- 项目状态:已结题
- 来源:
- 关键词:
项目摘要
By emphasizing words acoustically, people can convey the information about which concepts they wish to contrast. This feature of speech, known as focus, is pervasive in English, yet is inadequately modeled in state-of-the-art speech technologies. The challenge, which this Early Grant for Exploratory Research addresses, is that it is often difficult to identify phonetic emphasis independently of semantic contrast: words whose meanings are focused are usually realized with increased acoustic prominence, but not all cases of increased acoustic prominence are due to focus. The project is innovative in its use both of speech that has been recorded in a laboratory under controlled conditions, and also of speech that occurs naturally, such as in podcasts and videos. Judgments of focus location in laboratory speech and in naturally-occurring speech are collected from ordinary, non-expert listeners using online crowd-sourcing. Using the comparative construction (for example, "He liked it better than I did" or "I like it better now than I did") in which focus can be independently verified, computational procedures are developed to mimic the judgment of subjects who read but do not listen to the utterance being investigated. The findings will inform research in speech synthesis and in automatic speech recognition. Commercial applications may include aids for the deaf and hearing impaired, robot assistants for the elderly, language instruction and speech therapy.In a previous proof-of-concept study, the researcher collected utterances of "than I did" in laboratory experiments and from transcribed podcasts available on the web. Machine learning classifiers (using linear discriminant analysis and support vector machines) were trained to detect focus from acoustic features alone, including measures of fundamental frequency, duration and intensity. Location of focus can be determined independently from prosody in the comparative construction by observing the presence or absence of co-reference between subjects in the main and comparative clauses. This research generalizes that study to variations of the comparative with different pronouns and auxiliaries and also introduces updated methods of acoustic extraction and classification. Then, a verification dataset is created in order to reject annotations from participants who annotate non-focal prominence or who mark focus location incorrectly. Finally, classifiers are trained to detect focus on pronouns and auxiliaries in contexts other than the comparative, using the crowd-sourced annotation data to infer correct location of focus independently from prosody.
通过在听觉上强调单词,人们可以传达他们希望对比的概念的信息。 这种语音特征,被称为焦点,在英语中很普遍,但在最先进的语音技术中没有充分建模。 这项探索性研究早期资助所面临的挑战是,通常很难独立于语义对比来识别语音强调:其含义被聚焦的单词通常会随着声学突出度的增加而实现,但并非所有声学突出度增加的情况都是由于聚焦。 该项目创新地使用了在受控条件下在实验室中记录的语音,以及自然发生的语音,如播客和视频。 实验室语音和自然发生的语音中的焦点位置的判断收集从普通的,非专家的听众使用在线众包。 使用比较结构(例如,“他比我更喜欢它”或“我现在比我更喜欢它”),其中焦点可以独立验证,计算程序被开发来模仿阅读但不听被调查话语的受试者的判断。 这些发现将为语音合成和自动语音识别的研究提供信息。 商业应用可能包括聋人和听力受损者的辅助设备,老年人的机器人助手,语言教学和言语治疗。在之前的一项概念验证研究中,研究人员在实验室实验中收集了“比我做的”的话语,并从网上转录的播客中获得。 机器学习分类器(使用线性判别分析和支持向量机)被训练成仅从声学特征(包括基频、持续时间和强度的测量)检测焦点。 在比较结构中,焦点的位置可以通过观察主句和比较分句中主语之间是否存在共指来独立于韵律来确定。 本研究将比较级的研究归纳到不同代词和助词的比较级变体中,并介绍了声学提取和分类的最新方法。 然后,创建验证数据集,以便拒绝来自注释非焦点突出或不正确地标记焦点位置的参与者的注释。 最后,训练分类器以检测在比较级之外的上下文中对代词和助动词的关注,使用众包注释数据来独立于韵律推断正确的焦点位置。
项目成果
期刊论文数量(0)
专著数量(0)
科研奖励数量(0)
会议论文数量(0)
专利数量(0)
数据更新时间:{{ journalArticles.updateTime }}
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
数据更新时间:{{ journalArticles.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ monograph.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ sciAawards.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ conferencePapers.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ patent.updateTime }}
Jonathan Howell其他文献
Phase 1 study of quercetin, a natural antioxidant for children and young adults with Fanconi anemia
槲皮素(一种天然抗氧化剂)用于儿童和年轻成人范可尼贫血的1期研究
- DOI:
10.1182/bloodadvances.2024015053 - 发表时间:
2025-04-22 - 期刊:
- 影响因子:7.100
- 作者:
Parinda A. Mehta;Adam Nelson;Sara Loveless;Adam Lane;Tsuyoshi Fukuda;Ashley Teusink-Cross;Deborah Elder;Denise Lagory;Erica Miller;Jose A. Cancelas;Jonathan Howell;Junfang Zhao;Kana Mizuno;Kasiani C. Myers;Kelly Lake;Kelly McIntosh;Kenneth D. R. Setchell;Nathan Luebbering;Stephanie Edwards;Tafadzwa Chihanga;Stella M. Davies - 通讯作者:
Stella M. Davies
Meaning And Prosody: On The Web, In The Lab And From The Theorist'S Armchair
意义和韵律:在网络上、在实验室里、在理论家的扶手椅上
- DOI:
- 发表时间:
2012 - 期刊:
- 影响因子:0
- 作者:
Jonathan Howell - 通讯作者:
Jonathan Howell
Acoustic Classification of Focus: On the Web and in the Lab
声学焦点分类:网络上和实验室中
- DOI:
10.5334/labphon.8 - 发表时间:
2017 - 期刊:
- 影响因子:0
- 作者:
Jonathan Howell;Mats Rooth;M. Wagner - 通讯作者:
M. Wagner
Harvesting speech datasets for linguistic research on the web
收集语音数据集以进行网络语言研究
- DOI:
- 发表时间:
2013 - 期刊:
- 影响因子:0
- 作者:
Mats Rooth;Jonathan Howell;M. Wagner - 通讯作者:
M. Wagner
Second Occurrence Focus and theAcoustics of Prominence
第二次发生焦点和突出的声学
- DOI:
10.1121/1.4787465 - 发表时间:
2006 - 期刊:
- 影响因子:0
- 作者:
Jonathan Howell - 通讯作者:
Jonathan Howell
Jonathan Howell的其他文献
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
{{ truncateString('Jonathan Howell', 18)}}的其他基金
Prosodic Event Annotation and Detection in Three Varieties of English
三种英语韵律事件标注与检测
- 批准号:
2316030 - 财政年份:2023
- 资助金额:
$ 10.59万 - 项目类别:
Standard Grant
相似国自然基金
Simulation and certification of the ground state of many-body systems on quantum simulators
- 批准号:
- 批准年份:2020
- 资助金额:40 万元
- 项目类别:
微生物灌浆:地基基础加固的新探索
- 批准号:51078202
- 批准年份:2010
- 资助金额:41.0 万元
- 项目类别:面上项目
变分与拓扑方法和Schrodinger方程中的Open 问题
- 批准号:10871109
- 批准年份:2008
- 资助金额:23.0 万元
- 项目类别:面上项目
相似海外基金
STEPPING OUT - Understanding the ground truth and mechanisms of falls and balance instability in community dwelling older people.
走出去——了解社区老年人跌倒和平衡不稳定的基本事实和机制。
- 批准号:
EP/Y029143/1 - 财政年份:2024
- 资助金额:
$ 10.59万 - 项目类别:
Research Grant
Ground Truth Validation of Crop Growth Cycle Using High Resolution Proximal and Remote Sensing
使用高分辨率近端和遥感对作物生长周期进行地面实况验证
- 批准号:
549723-2019 - 财政年份:2022
- 资助金额:
$ 10.59万 - 项目类别:
Alliance Grants
Ground Truth Validation of Crop Growth Cycle Using High Resolution Proximal and Remote Sensing
使用高分辨率近端和遥感对作物生长周期进行地面实况验证
- 批准号:
549723-2019 - 财政年份:2021
- 资助金额:
$ 10.59万 - 项目类别:
Alliance Grants
Ground-Truth Labeling of Images for AI Development in Computational Pathology
用于计算病理学人工智能开发的图像真实标签
- 批准号:
562132-2021 - 财政年份:2021
- 资助金额:
$ 10.59万 - 项目类别:
University Undergraduate Student Research Awards
Ground Truth Validation of Crop Growth Cycle Using High Resolution Proximal and Remote Sensing
使用高分辨率近端和遥感对作物生长周期进行地面实况验证
- 批准号:
549723-2019 - 财政年份:2020
- 资助金额:
$ 10.59万 - 项目类别:
Alliance Grants
Social Media, Violence, and Social Isolation Among At-Risk Adolescents: Exploring Ground Truth
高危青少年中的社交媒体、暴力和社会孤立:探索真相
- 批准号:
10265597 - 财政年份:2020
- 资助金额:
$ 10.59万 - 项目类别:
Social Media, Violence, and Social Isolation Among At-Risk Adolescents: Exploring Ground Truth
高危青少年中的社交媒体、暴力和社会孤立:探索真相
- 批准号:
10461047 - 财政年份:2020
- 资助金额:
$ 10.59万 - 项目类别:
Evaluation of Interpretable Deep Learning Approaches for Neuroimaging in Dementia using Stroke Classification as a Ground Truth for Brain Pathology
使用中风分类作为脑病理学的基本事实来评估痴呆症神经影像的可解释深度学习方法
- 批准号:
2407028 - 财政年份:2020
- 资助金额:
$ 10.59万 - 项目类别:
Studentship
MRI: Acquisition of Omnipercipient Chamber for Gathering Ground Truth and Enabling Research on Smart and Connected Things
MRI:获取全知室以收集地面真相并支持智能和互联事物的研究
- 批准号:
2018966 - 财政年份:2020
- 资助金额:
$ 10.59万 - 项目类别:
Standard Grant
Social Media, Violence, and Social Isolation Among At-Risk Adolescents: Exploring Ground Truth
高危青少年中的社交媒体、暴力和社会孤立:探索真相
- 批准号:
10701687 - 财政年份:2020
- 资助金额:
$ 10.59万 - 项目类别: