Deep transformers for integrating protein sequence, structure and interaction data to predict function
用于整合蛋白质序列、结构和相互作用数据以预测功能的深度转换器
基本信息
- 批准号:2308699
- 负责人:
- 金额:$ 63.79万
- 依托单位:
- 依托单位国家:美国
- 项目类别:Continuing Grant
- 财政年份:2023
- 资助国家:美国
- 起止时间:2023-06-01 至 2026-05-31
- 项目状态:未结题
- 来源:
- 关键词:
项目摘要
Proteins are fundamental macromolecules in the living systems. The knowledge about the function of proteins is important for biological research and technology development. However, the function of most proteins is still unknown. To fill the gap, this project aims to develop deep learning methods, one of the most powerful artificial intelligence (AI) technologies, to integrate multiple sources of protein data such as protein sequences, structures, and interaction to accurately predict protein function. The methods will advance the state of the art of protein function prediction and can be broadly applied in many domains such as life science research, biotechnology development, agriculture, and healthcare. The project will provide unique interdisciplinary research opportunities to train students at multiple levels including under-represented minority students with diverse backgrounds to apply AI to address fundamental scientific and technological problems. The project will develop deep transformer models based on self-attention to integrate protein sequence, structure, and interaction data to significantly advance the prediction of both protein-level function and amino acid-level function. Specifically, it aims to achieve three objectives: (1) develop 1D and 3D transformers to predict protein function from multiple sequence alignments and structures; (2) develop 2D graph transformers to predict protein function from protein-protein interactions and integrate them with sequences and structures; and (3) implement transformers as user-friendly, accurate, robust open-source protein function prediction tools for the community. Cutting-edge deep transformer models based on the self-attention mechanism will be developed to integrate protein sequence, structure, and interaction data to predict protein function for the first time. 1D sequence-based transformer, 2D graph transformer, and 3D-equivariant graph transformer can extract amino acid conservation and long-range co-evolutionary signals in multiple sequence alignments, long-range interactions in protein-protein networks, and rotation- and translation-invariant/equivariant properties of protein structures better than the existing deep learning methods based on traditional convolutional and recurrent mechanisms. Predicting both overall protein-level function terms and residue-level function sites via multi-task learning and novel deep learning architectures can leverage the compliment of the two prediction tasks to provide more accurate, more complete, and more interpretable function prediction. The project will deliver user-friendly open-source tools for the community to accurately predict function from sequence, structure, and interaction data, which will help reduce the vast knowledge gap between protein sequence and function. The open-source deep learning tools can be used to predict and study protein function in many domains. The methods and tools will be leveraged to train students at multiple levels and increase the diversity in scientific research and education. The results of the project can be found at https://calla.rnet.missouri.edu/cheng/nsf_protein_function.htmlThis award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
Proteins are fundamental macromolecules in the living systems. The knowledge about the function of proteins is important for biological research and technology development. However, the function of most proteins is still unknown. To fill the gap, this project aims to develop deep learning methods, one of the most powerful artificial intelligence (AI) technologies, to integrate multiple sources of protein data such as protein sequences, structures, and interaction to accurately predict protein function. The methods will advance the state of the art of protein function prediction and can be broadly applied in many domains such as life science research, biotechnology development, agriculture, and healthcare. The project will provide unique interdisciplinary research opportunities to train students at multiple levels including under-represented minority students with diverse backgrounds to apply AI to address fundamental scientific and technological problems. The project will develop deep transformer models based on self-attention to integrate protein sequence, structure, and interaction data to significantly advance the prediction of both protein-level function and amino acid-level function. Specifically, it aims to achieve three objectives: (1) develop 1D and 3D transformers to predict protein function from multiple sequence alignments and structures; (2) develop 2D graph transformers to predict protein function from protein-protein interactions and integrate them with sequences and structures; and (3) implement transformers as user-friendly, accurate, robust open-source protein function prediction tools for the community. Cutting-edge deep transformer models based on the self-attention mechanism will be developed to integrate protein sequence, structure, and interaction data to predict protein function for the first time. 1D sequence-based transformer, 2D graph transformer, and 3D-equivariant graph transformer can extract amino acid conservation and long-range co-evolutionary signals in multiple sequence alignments, long-range interactions in protein-protein networks, and rotation- and translation-invariant/equivariant properties of protein structures better than the existing deep learning methods based on traditional convolutional and recurrent mechanisms. Predicting both overall protein-level function terms and residue-level function sites via multi-task learning and novel deep learning architectures can leverage the compliment of the two prediction tasks to provide more accurate, more complete, and more interpretable function prediction. The project will deliver user-friendly open-source tools for the community to accurately predict function from sequence, structure, and interaction data, which will help reduce the vast knowledge gap between protein sequence and function. The open-source deep learning tools can be used to predict and study protein function in many domains. The methods and tools will be leveraged to train students at multiple levels and increase the diversity in scientific research and education. The results of the project can be found at https://calla.rnet.missouri.edu/cheng/nsf_protein_function.htmlThis award reflects NSF's statutory mission and has been deemed worthy of support through evaluation using the Foundation's intellectual merit and broader impacts review criteria.
项目成果
期刊论文数量(5)
专著数量(0)
科研奖励数量(0)
会议论文数量(0)
专利数量(0)
Combining protein sequences and structures with transformers and equivariant graph neural networks to predict protein function.
- DOI:10.1093/bioinformatics/btad208
- 发表时间:2023-06-30
- 期刊:
- 影响因子:0
- 作者:
- 通讯作者:
Geometry-complete perceptron networks for 3D molecular graphs
- DOI:10.1093/bioinformatics/btae087
- 发表时间:2022-11
- 期刊:
- 影响因子:5.8
- 作者:Alex Morehead;Jianlin Cheng
- 通讯作者:Alex Morehead;Jianlin Cheng
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
数据更新时间:{{ journalArticles.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ monograph.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ sciAawards.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ conferencePapers.updateTime }}
{{ item.title }}
- 作者:
{{ item.author }}
数据更新时间:{{ patent.updateTime }}
Jianlin Cheng其他文献
A Deep Learning Network Approach to ab initio Protein Secondary Structure Prediction.
- DOI:
10.1109/tcbb.2014.2343960 - 发表时间:
2015-01 - 期刊:
- 影响因子:0
- 作者:
Spencer M;Eickholt J;Jianlin Cheng - 通讯作者:
Jianlin Cheng
Curation of the Deep Green list of unannotated green lineage proteins to enable structural and functional characterization
整理未注释的绿色谱系蛋白的 Deep Green 列表,以实现结构和功能表征
- DOI:
10.1101/2022.09.30.510186 - 发表时间:
2022 - 期刊:
- 影响因子:0
- 作者:
E. Knoshaug;Peipei Sun;A. Nag;Huong Nguyen;Erin M. Mattoon;Ningning Zhang;Jian Liu;Chen Chen;Jianlin Cheng;Ru Zhang;Peter C. St. John;J. Umen - 通讯作者:
J. Umen
Predicting interchain contacts for homodimeric and homomultimeric protein complexes using multiple sequence alignments of monomers and deep learning
使用单体的多重序列比对和深度学习预测同二聚体和同多聚体蛋白质复合物的链间接触
- DOI:
10.1101/2020.11.09.373878 - 发表时间:
2020 - 期刊:
- 影响因子:0
- 作者:
Farhan Quadir;Rajashree Roy;Randal Halfmann;Jianlin Cheng - 通讯作者:
Jianlin Cheng
Machine Learning Algorithms for Protein Structure Prediction
- DOI:
- 发表时间:
2006 - 期刊:
- 影响因子:0
- 作者:
Jianlin Cheng - 通讯作者:
Jianlin Cheng
Protein Structure Refinement by Iterative Fragment Exchange
通过迭代片段交换优化蛋白质结构
- DOI:
10.1145/2506583.2506601 - 发表时间:
2013 - 期刊:
- 影响因子:0
- 作者:
Debswapna Bhattacharya;Jianlin Cheng - 通讯作者:
Jianlin Cheng
Jianlin Cheng的其他文献
{{
item.title }}
{{ item.translation_title }}
- DOI:
{{ item.doi }} - 发表时间:
{{ item.publish_year }} - 期刊:
- 影响因子:{{ item.factor }}
- 作者:
{{ item.authors }} - 通讯作者:
{{ item.author }}
{{ truncateString('Jianlin Cheng', 18)}}的其他基金
III: Medium: Collaborative Research: Guiding Exploration of Protein Structure Spaces with Deep Learning
III:媒介:协作研究:用深度学习指导蛋白质结构空间探索
- 批准号:
1763246 - 财政年份:2018
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
ABI Innovation: Deep learning methods for protein bioinformatics
ABI Innovation:蛋白质生物信息学的深度学习方法
- 批准号:
1759934 - 财政年份:2018
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
CAREER: Analysis, Construction and Visualization of 3D Genome Structures
职业:3D 基因组结构的分析、构建和可视化
- 批准号:
1149224 - 财政年份:2012
- 资助金额:
$ 63.79万 - 项目类别:
Continuing Grant
相似海外基金
University of Nottingham and Transformers & Rectifiers Limited KTP 22_23 R4
诺丁汉大学和变形金刚
- 批准号:
10056201 - 财政年份:2024
- 资助金额:
$ 63.79万 - 项目类别:
Knowledge Transfer Partnership
Development of magnetic coated rectangular wire for reducing AC copper loss and reducing size and weight of transformers and inductors
开发磁性涂层矩形线,以减少交流铜损并减小变压器和电感器的尺寸和重量
- 批准号:
23H01397 - 财政年份:2023
- 资助金额:
$ 63.79万 - 项目类别:
Grant-in-Aid for Scientific Research (B)
CAREER: Multi-level Bridge Tapped Resonant (MBTR) Solid-State Transformers (SSTs)
职业:多级桥抽头谐振 (MBTR) 固态变压器 (SST)
- 批准号:
2238472 - 财政年份:2023
- 资助金额:
$ 63.79万 - 项目类别:
Continuing Grant
Collaborative Research: Digital Twin Predictive Reliability Modeling of Solid-State Transformers
合作研究:固态变压器的数字孪生预测可靠性建模
- 批准号:
2228873 - 财政年份:2023
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
SHF: Small: Improving Efficiency of Vision Transformers via Software-Hardware Co-Design and Acceleration
SHF:小型:通过软硬件协同设计和加速提高视觉变压器的效率
- 批准号:
2233893 - 财政年份:2023
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
Advanced Electromagnetic Analysis and High-frequency Impedance Design for Magnetic Ferrite Inductors and Transformers
适用于磁性铁氧体电感器和变压器的先进电磁分析和高频阻抗设计
- 批准号:
2322529 - 财政年份:2023
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
Collaborative Research: Digital Twin Predictive Reliability Modeling of Solid-State Transformers
合作研究:固态变压器的数字孪生预测可靠性建模
- 批准号:
2228872 - 财政年份:2023
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
Inherently-Safer Hybrid Power Electronics Transformers (INSPIRE)
本质安全的混合电力电子变压器 (INSPIRE)
- 批准号:
10045685 - 财政年份:2022
- 资助金额:
$ 63.79万 - 项目类别:
Grant for R&D
ERI: Multiphysics cosimulation approach for optimal design of microgrid high frequency transformers
ERI:微电网高频变压器优化设计的多物理场协同仿真方法
- 批准号:
2138408 - 财政年份:2022
- 资助金额:
$ 63.79万 - 项目类别:
Standard Grant
Efficient Object Segmentation with Transformers
使用 Transformer 进行高效对象分割
- 批准号:
572477-2022 - 财政年份:2022
- 资助金额:
$ 63.79万 - 项目类别:
University Undergraduate Student Research Awards