Fig. 1: Architecture and embedding visualization of PTM-Mamba. | Nature Methods

Fig. 1: Architecture and embedding visualization of PTM-Mamba.

From: PTM-Mamba: a PTM-aware protein language model with bidirectional gated Mamba blocks

Fig. 1

a, Primitives of PTM-Mamba. Bottom left, given a sequence, with 80% probability, we perform standard 15% token masking, and, with 20% probability, we mask all the PTM tokens and randomly mask 15% of wild-type tokens. The bidirectional Mamba block in PTM-Mamba is built on top of the Mamba block (MB), which processes the sequences in both the forward (forward Mamba block) and backward (backward Mamba block) orientation. The gated embedding fusion module inputs ESM-2 and PTM embeddings and fuses them in a gated manner via a sigmoid-activated linear layer. SSM, state-space model. b, t-SNE visualization of PTM-Mamba embeddings of select wild-type and corresponding PTM protein sequences. Orange lines connect the corresponding embeddings. c, t-SNE visualization of labeled token embeddings. Conv, local 1D convolutional layer.

Back to article page