Unifying Pre-training and Multilingual Semantic Representation Learning for Low-resource Neural Machine Translation
统一预训练和多语言语义表示学习以实现低资源神经机器翻译
基本信息
- 批准号:22KJ1843
- 负责人:
- 金额:$ 1.09万
- 依托单位:
- 依托单位国家:日本
- 项目类别:Grant-in-Aid for JSPS Fellows
- 财政年份:2023
- 资助国家:日本
- 起止时间:2023-03-08 至 2024-03-31
- 项目状态:已结题
- 来源:
- 关键词:
项目摘要
In the past year, we focused on improving the efficiency of multilingual sentence representation learning and exploring novel methods for improving multilingual machine translation. Both research promotes the research for multilingual / low-resource neural machine translation.(1) We proposed an efficient and effective method for training and presented the work in 言語処理学会 2023. On the other hand, we proposed knowledge distillation for compressing a large model, and it has been accepted to EACL 2023 main conference, which leads to efficient model inference. With the above achievements, the process of collecting parallel sentences for training translation systems will be accelerated. Specifically, the model training phase can be accelerated by 4 - 16 times, and the model inference phase can achieve 2.5 - 5 times speedup with further faster speed on downstream tasks.(2) We explored novel ways to improve the multilingual translation system with a word-level contrastive learning technique and obtained better translation quality for low-resource language pairs, which was accepted by NAACL 2022 findings. We also explained the improvements by showing the relationship between BLEU scores and sentence retrieval performance of the NMT encoder, which motivates that future work can focus on further improving the encoder’s retrieval performance in many-to-many NMT and contrastive objective’s feasibility in a massively multilingual scenario.
In the past year, we focused on improving the efficiency of multilingual sentence representation learning and exploring novel methods for improving multilingual machine translation. Both research promotes the research for multilingual / low-resource neural machine translation. (1)We proposed an efficient and effective method for training and presented the work in Speech Processing Society 2023. On the other hand, we proposed knowledge distillation for compressing a large model, and it has been accepted to EACL 2023 main conference, which leads to efficient model inference. With the above achievements, the process of collecting parallel sentences for training translation systems will be accelerated. Specifically, the model training phase can be accelerated by 4 - 16 times, and the model inference phase can achieve 2.5 - 5 times speedup with further faster speed on downstream tasks. (2) We explored novel ways to improve the multilingual translation system with a word-level contrastive learning technique and obtained better translation quality for low-resource language pairs, which was accepted by NAACL 2022 findings. We also explained the improvements by showing the relationship between BLEU scores and sentence retrieval performance of the NMT encoder, which motivates that future work can focus on further improving the encoder’s retrieval performance in many-to-many NMT and contrastive objective’s feasibility in a massively multilingual scenario.
项目成果
期刊论文数量(11)
专著数量(0)
科研奖励数量(0)
会议论文数量(0)
专利数量(0)
When do Contrastive Word Alignments Improve Many-to-many Neural Machine Translation?
对比词对齐何时可以改善多对多神经机器翻译?
- DOI:
- 发表时间:2022
- 期刊:
- 影响因子:0
- 作者:Zhuoyuan Mao;Chenhui Chu;Raj Dabre;Haiyue Song;Zhen Wan and Sadao Kurohashi
- 通讯作者:Zhen Wan and Sadao Kurohashi
Textual Enhanced Contrastive Learning for Solving Math Word Problems
- DOI:10.48550/arxiv.2211.16022
- 发表时间:2022-11
- 期刊:
- 影响因子:3.9
- 作者:Yibin Shen;Qianying Liu;Zhuoyuan Mao;Fei Cheng;S. Kurohashi
- 通讯作者:Yibin Shen;Qianying Liu;Zhuoyuan Mao;Fei Cheng;S. Kurohashi
Rescue Implicit and Long-tail Cases: Nearest Neighbor Relation Extraction
- DOI:10.48550/arxiv.2210.11800
- 发表时间:2022-10
- 期刊:
- 影响因子:0
- 作者:Zhen Wan;Qianying Liu;Zhuoyuan Mao;Fei Cheng;S. Kurohashi;Jiwei Li
- 通讯作者:Zhen Wan;Qianying Liu;Zhuoyuan Mao;Fei Cheng;S. Kurohashi;Jiwei Li
Seeking Diverse Reasoning Logic: Controlled Equation Expression Generation for Solving Math Word Problems
- DOI:10.48550/arxiv.2209.10310
- 发表时间:2022-09
- 期刊:
- 影响因子:0
- 作者:Yibin Shen;Qianying Liu;Zhuoyuan Mao;Zhen Wan;Fei Cheng;S. Kurohashi
- 通讯作者:Yibin Shen;Qianying Liu;Zhuoyuan Mao;Zhen Wan;Fei Cheng;S. Kurohashi
Linguistically Driven Multi-Task Pre-Training for Low-Resource Neural Machine Translation
- DOI:10.1145/3491065
- 发表时间:2022-01
- 期刊:
- 影响因子:0
- 作者:Zhuoyuan Mao;Chenhui Chu;S. Kurohashi
- 通讯作者:Zhuoyuan Mao;Chenhui Chu;S. Kurohashi
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
数据更新时间:{{ journalArticles.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ monograph.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ sciAawards.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ conferencePapers.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ patent.updateTime }}
毛 卓遠其他文献
毛 卓遠的其他文献
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}