热点
关于我们
xx
xx
"
RNN
" 相关文章
Language Modeling With Factorization Memory
cs.AI updates on arXiv.org
2025-11-05T05:22:00.000000Z
RWKV7-G0a3 13.3B 发布:世界最强纯 RNN 大语言模型
RWKV元始智能
2025-11-03T11:59:12.000000Z
RWKV正史,与学术投机者杨某某的发家史:请将DeltaNet的credit给Schmidhuber
RWKV元始智能
2025-11-02T18:20:30.000000Z
TempoPFN: Synthetic Pre-training of Linear RNNs for Zero-shot Time Series Forecasting
cs.AI updates on arXiv.org
2025-10-30T04:19:51.000000Z
Accuracy, Memory Efficiency and Generalization: A Comparative Study on Liquid Neural Networks and Recurrent Neural Networks
cs.AI updates on arXiv.org
2025-10-10T04:11:01.000000Z
Accuracy, Memory Efficiency and Generalization: A Comparative Study on Liquid Neural Networks and Recurrent Neural Networks
cs.AI updates on arXiv.org
2025-10-10T04:11:01.000000Z
RWKV-8替代attention新机制名为ROSA,RWKV-7 7B单5090解码10000+ tok/s演示
RWKV元始智能
2025-10-09T04:41:18.000000Z
理解LLM系列:预告RWKV-8,高效解决RNN长文本问题
RWKV元始智能
2025-10-09T03:51:35.000000Z
RWKV7-G1a 2.9B 推理模型开源发布,继续推进纯 RNN 模型的思考能力
RWKV元始智能
2025-09-30T12:01:16.000000Z
Short window attention enables long-term memorization
cs.AI updates on arXiv.org
2025-09-30T04:07:01.000000Z
StateX: Enhancing RNN Recall via Post-training State Expansion
cs.AI updates on arXiv.org
2025-09-29T04:16:42.000000Z
Conceptual comparison of how attention works in (1) Plain RNN, (2) Models incorporating CNN, and (3) Transformers
Recent Questions - Artificial Intelligence Stack Exchange
2025-09-29T04:01:20.000000Z
Predict Stock Prices Using RNN: Part 1
Lil'Log
2025-09-25T10:02:44.000000Z
WAIC首日 | RWKV-7s 新型高效大模型架构正式亮相
RWKV元始智能
2025-09-25T10:01:46.000000Z
RWKV 社区七月动态:RWKV7-G0 7.2B 模型发布,8 篇高质量论文
RWKV元始智能
2025-09-25T10:01:45.000000Z
理解LLM系列:升级RNN的常见路线
RWKV元始智能
2025-09-25T10:01:45.000000Z
RNN的正确用法:大bsz才是关键(多agent,高并发)
RWKV元始智能
2025-09-25T10:01:45.000000Z
一文讲清 nn.LayerNorm 层归一化
掘金 人工智能
2025-09-16T03:13:41.000000Z
一文讲清 nn.LayerNorm 层归一化
掘金 人工智能
2025-09-16T03:13:41.000000Z
理解LLM系列:升级RNN的常见路线
RWKV元始智能
2025-09-15T12:50:11.000000Z