4.7 Article

TALE: Transformer-based protein function Annotation with joint sequence-Label Embedding

期刊

BIOINFORMATICS
卷 37, 期 18, 页码 2825-2833

出版社

OXFORD UNIV PRESS
DOI: 10.1093/bioinformatics/btab198

关键词

-

资金

  1. National Institute of General Medical Sciences [R35GM124952]

向作者/读者索取更多资源

The study introduces a Transformer-based protein function annotation model TALE, which achieves significant breakthroughs in applicability and generalizability by using only sequence information. Compared to other methods, TALE outperforms when only sequence input is available and demonstrates superior generalizability to new species, low similarity proteins, and rare functions.
Motivation: Facing the increasing gap between high-throughput sequence data and limited functional insights, computational protein function annotation provides a high-throughput alternative to experimental approaches. However, current methods can have limited applicability while relying on protein data besides sequences, or lack generalizability to novel sequences, species and functions. Results: To overcome aforementioned barriers in applicability and generalizability, we propose a novel deep learning model using only sequence information for proteins, named Transformer-based protein function Annotation through joint sequence-Label Embedding (TALE). For generalizability to novel sequences we use self-attention-based transformers to capture global patterns in sequences. For generalizability to unseen or rarely seen functions (tail labels), we embed protein function labels (hierarchical GO terms on directed graphs) together with inputs/features (1D sequences) in a joint latent space. Combining TALE and a sequence similarity-based method, TALE+ outperformed competing methods when only sequence input is available. It even outperformed a state-of-the-art method using network information besides sequence, in two of the three gene ontologies. Furthermore, TALE and TALE+ showed superior generalizability to proteins of low similarity, new species, or rarely annotated functions compared to training data, revealing deep insights into the protein sequence-function relationship. Ablation studies elucidated contributions of algorithmic components toward the accuracy and the generalizability; and a GO term-centric analysis was also provided. Availability and implementation: The data, source codes and models are available at https://github.com/Shen-Lab/TALE.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据