MarkTechPost@AI 10月23日 15:06
Anthrogen发布Odyssey蛋白质语言模型
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Anthrogen推出Odyssey,一个多模态蛋白质语言模型系列,旨在革新蛋白质序列和结构生成、编辑及条件化设计。该模型系列参数规模从1.2B到102B不等,采用名为Consensus的创新传播规则替代传统的自注意力机制,以O(L)的计算复杂度有效处理长序列,并具备对学习率选择的鲁棒性。Odyssey通过有限标量量化(FSQ)将3D结构转化为紧凑的离散标记,实现序列与结构的联合建模。其训练目标为离散扩散模型,在各项评估中均优于掩码语言模型,尤其在数据效率方面表现突出,可显著减少训练所需数据量,为蛋白质设计领域带来更高效、更强大的解决方案。

💡 **Odyssey蛋白质语言模型系列**:Anthrogen推出的Odyssey是一系列强大的多模态蛋白质语言模型,其参数规模涵盖1.2B至102B,专为蛋白质序列与结构生成、编辑以及条件化设计而优化。该模型能够融合序列、结构和功能性信息,为蛋白质设计提供全面的解决方案。

🌐 **创新的Consensus机制**:Odyssey摒弃了计算复杂度高的自注意力机制,转而采用名为Consensus的迭代传播规则。该机制能够实现O(L)的计算效率,有效处理长蛋白质序列,并且在处理大规模模型时对学习率选择表现出更强的鲁棒性,减少了训练过程中的不稳定性。

📐 **FSQ的结构编码能力**:为了实现序列与结构的联合建模,Odyssey利用有限标量量化(FSQ)技术,将蛋白质的3D几何结构转化为紧凑的离散标记。这使得模型能够像处理序列标记一样轻松地“阅读”和理解蛋白质结构信息。

🚀 **离散扩散模型的训练优势**:Odyssey采用离散扩散模型进行训练,其反向时间去噪器能够有效地重建序列和坐标,使其协同工作。在与掩码语言模型进行的对比评估中,离散扩散模型在多项指标上均表现出更优越的性能,尤其是在数据效率方面,显著降低了对训练数据的需求,这在数据稀缺的蛋白质建模领域尤为重要。

Anthrogen has introduced Odyssey, a family of protein language models for sequence and structure generation, protein editing, and conditional design. The production models range from 1.2B to 102B parameters. The Anthrogen’s research team positions Odyssey as a frontier, multimodal model for real protein design workloads, and notes that an API is in early access.

https://www.biorxiv.org/content/10.1101/2025.10.15.682677v1.full.pdf

What problem does Odyssey target?

Protein design couples amino acid sequence with 3D structure and with functional context. Many prior models adopt self attention, which mixes information across the entire sequence at once. Proteins follow geometric constraints, so long range effects travel through local neighborhoods in 3D. Anthrogen frames this as a locality problem and proposes a new propagation rule, called Consensus, that better matches the domain.

https://www.biorxiv.org/content/10.1101/2025.10.15.682677v1.full.pdf

Input representation and tokenization

Odyssey is multimodal. It embeds sequence tokens, structure tokens, and lightweight functional cues, then fuses them into a shared representation. For structure, Odyssey uses a finite scalar quantizer, FSQ, to convert 3D geometry into compact tokens. Think of FSQ as an alphabet for shapes that lets the model read structure as easily as sequence. Functional cues can include domain tags, secondary structure hints, orthologous group labels, or short text descriptors. This joint view gives the model access to local sequence patterns and long range geometric relations in a single latent space.

https://www.biorxiv.org/content/10.1101/2025.10.15.682677v1.full.pdf

Backbone change, Consensus instead of self attention

Consensus replaces global self attention with iterative, locality aware updates on a sparse contact or sequence graph. Each layer encourages nearby neighborhoods to agree first, then spreads that agreement outward across the chain and contact graph. This change alters compute. Self attention scales as O(L²) with sequence length L. Anthrogen reports that Consensus scales as O(L), which keeps long sequences and multi domain constructs affordable. The company also reports improved robustness to learning rate choices at larger scales, which reduces brittle runs and restarts.

https://www.biorxiv.org/content/10.1101/2025.10.15.682677v1.full.pdf

Training objective and generation, discrete diffusion

Odyssey trains with discrete diffusion on sequence and structure tokens. The forward process applies masking noise that mimics mutation. The reverse time denoiser learns to reconstruct consistent sequence and coordinates that work together. At inference, the same reverse process supports conditional generation and editing. You can hold a scaffold, fix a motif, mask a loop, add a functional tag, and then let the model complete the rest while keeping sequence and structure in sync.

Anthrogen reports matched comparisons where diffusion outperforms masked language modeling during evaluation. The page notes lower training perplexities for diffusion versus complex masking, and lower or comparable training perplexities versus simple masking. In validation, diffusion models outperform their masked counterparts, while a 1.2B masked model tends to overfit to its own masking schedule. The company argues that diffusion models the joint distribution of the full protein, which aligns with sequence plus structure co design.

https://www.biorxiv.org/content/10.1101/2025.10.15.682677v1.full.pdf

Key takeaways

    Odyssey is a multimodal protein model family that fuses sequence, structure, and functional context, with production models at 1.2B, 8B, and 102B parameters. Consensus replaces self attention with locality aware propagation that scales as O(L) and shows robust learning rate behavior at larger scales. FSQ converts 3D coordinates into discrete structure tokens for joint sequence and structure modeling. Discrete diffusion trains a reverse time denoiser and, in matched comparisons, outperforms masked language modeling during evaluation. Anthrogen reports better performance with about 10x less data than competing models, which addresses data scarcity in protein modeling.

Editorial Comments

Odyssey is impressive model because it operationalizes joint sequence and structure modeling with FSQ, Consensus, and discrete diffusion, enabling conditional design and editing under practical constraints. Odyssey scales to 102B parameters with O(L) complexity for Consensus, which lowers cost for long proteins and improves learning-rate robustness. Anthrogen reports diffusion outperforming masked language modeling in matched evaluations, which aligns with co-design objectives. The system targets multi-objective design, including potency, specificity, stability, and manufacturability. The research team emphasizes data efficiency near 10x versus competing models, which is material in domains with scarce labeled data.


Check out the Paperand Technical details. Feel free to check out our GitHub Page for Tutorials, Codes and Notebooks. Also, feel free to follow us on Twitter and don’t forget to join our 100k+ ML SubReddit and Subscribe to our Newsletter. Wait! are you on telegram? now you can join us on telegram as well.

The post Anthrogen Introduces Odyssey: A 102B Parameter Protein Language Model that Replaces Attention with Consensus and Trains with Discrete Diffusion appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Odyssey 蛋白质语言模型 Anthrogen AI 机器学习 结构生物学 蛋白质设计 Consensus 离散扩散模型 FSQ Odyssey Protein Language Model Anthrogen AI Machine Learning Structural Biology Protein Design Consensus Discrete Diffusion FSQ
相关文章