热点
关于我们
xx
xx
"
Mamba架构
" 相关文章
Bi-Mamba: Towards Accurate 1-Bit State Space Models
cs.AI updates on arXiv.org
2025-10-24T04:54:08.000000Z
CALM: A Causal Analysis Language Model for Tabular Data in Complex Systems with Local Scores, Conditional Independence Tests, and Relation Attributes
cs.AI updates on arXiv.org
2025-10-14T04:15:31.000000Z
Speak, Edit, Repeat: High-Fidelity Voice Editing and Zero-Shot TTS with Cross-Attentive Mamba
cs.AI updates on arXiv.org
2025-10-07T04:17:30.000000Z
DA-Mamba: Dialogue-aware selective state-space model for multimodal engagement estimation
cs.AI updates on arXiv.org
2025-09-23T05:26:05.000000Z
The pitfalls of next-token prediction
cs.AI updates on arXiv.org
2025-07-30T04:11:50.000000Z
Exploring State-Space-Model based Language Model in Music Generation
cs.AI updates on arXiv.org
2025-07-10T04:05:49.000000Z
Do Reasoning Models Really Need Transformers?: Researchers from TogetherAI, Cornell, Geneva, and Princeton Introduce M1—A Hybrid Mamba-Based AI that Matches SOTA Performance at 3x Inference Speed
MarkTechPost@AI
2025-04-18T04:05:34.000000Z
腾讯押注非Transformer!推理模型混合Mamba也能搞,深度思考也能秒回
智源社区
2025-03-23T02:01:48.000000Z
Testing which LLM architectures can do hidden serial reasoning
少点错误
2024-12-16T13:52:20.000000Z
多亏Transformer,Mamba更强了!仅用1%计算量达新SOTA
智源社区
2024-08-23T11:07:34.000000Z
多亏Transformer,Mamba更强了,仅用1%计算量达新SOTA
36kr
2024-08-22T09:03:19.000000Z
换掉Transformer,7B开源模型立刻登顶,任意长序列都能处理
36kr
2024-08-13T10:33:20.000000Z
Falcon Mamba 7B 开源模型登顶:换掉 Transformer,任意长序列都能处理
IT之家
2024-08-13T05:37:30.000000Z
Mamba写代码真的超越Transformer,原始论文入选顶流新会议
36kr-科技
2024-07-17T09:19:32.000000Z
Mamba真比Transformer更优吗?Mamba原作者:两个都要!混合架构才是最优解
智源社区
2024-07-16T06:06:06.000000Z