热点
关于我们
xx
xx
"
混合专家架构
" 相关文章
MedAlign: A Synergistic Framework of Multimodal Preference Optimization and Federated Meta-Cognitive Reasoning
cs.AI updates on arXiv.org
2025-10-27T06:16:56.000000Z
MTmixAtt: Integrating Mixture-of-Experts with Multi-Mix Attention for Large-Scale Recommendation
cs.AI updates on arXiv.org
2025-10-20T04:13:12.000000Z
Mix- and MoE-DPO: A Variational Inference Approach to Direct Preference Optimization
cs.AI updates on arXiv.org
2025-10-10T04:17:31.000000Z
Character.AI Open Sources pipeling-sft: A Scalable Framework for Fine-Tuning MoE LLMs like DeepSeek V3
Character AI Blog
2025-09-28T15:41:11.000000Z
美团开源推理大模型 LongCat-Flash-Thinking
oschina.net
2025-09-22T03:45:41.000000Z
Dynamic Adaptive Parsing of Temporal and Cross-Variable Patterns for Network State Classification
cs.AI updates on arXiv.org
2025-09-16T05:40:15.000000Z
MiniMax-M1:闪电注意力重塑大模型推理效率,百万上下文时代来临,附技术报告英中对照版
我爱自然语言处理
2025-09-11T19:55:55.000000Z
被 OpenAI 点名后,智谱刚刚发布新开源模型:国产最全能,一句话造出搜索引擎
爱范儿
2025-07-28T23:48:12.000000Z
Mixture of Experts in Large Language Models
cs.AI updates on arXiv.org
2025-07-16T04:28:40.000000Z
苦等一年 Meta终于放大招 正式发布开源大模型Llama 4
快科技资讯
2025-04-06T00:28:21.000000Z
Meta 发布 Llama 4 系列 AI 模型,引入“混合专家架构”提升效率
IT之家
2025-04-05T21:40:55.000000Z
Meta releases Llama 4, a new crop of flagship AI models
TechCrunch News
2025-04-05T20:07:35.000000Z
华尔街深度研究:DeepSeek是AI末日吗?
Cnbeta
2025-01-29T13:37:40.000000Z
Hunyuan-Large and the MoE Revolution: How AI Models Are Growing Smarter and Faster
Unite.AI
2024-12-20T17:16:00.000000Z
全球首次!时序大模型突破十亿参数,华人团队发布Time-MoE,预训练数据达3000亿个时间点
智源社区
2024-10-24T15:54:03.000000Z
全球首次,时序大模型突破十亿参数,华人团队发布Time-MoE,预训练数据达3000亿个时间点
36kr-科技
2024-10-22T23:19:11.000000Z
微軟釋出3款Phi-3.5模型
AI & Big Data
2024-08-21T09:32:46.000000Z