热点
"Pretraining" 相关文章
Correlation Dimension of Auto-Regressive Large Language Models
cs.AI updates on arXiv.org 2025-10-27T06:24:45.000000Z
喝点VC|YC对谈Anthropic预训练负责人:预训练团队也要考虑推理问题,如何平衡预训练和后训练仍在早期探索阶段
Z Potentials 2025-10-16T09:58:40.000000Z
喝点VC|YC对谈Anthropic预训练负责人:预训练团队也要考虑推理问题,如何平衡预训练和后训练仍在早期探索阶段
Z Potentials 2025-10-16T09:58:40.000000Z
NVIDIA Researchers Propose Reinforcement Learning Pretraining (RLP): Reinforcement as a Pretraining Objective for Building Reasoning During Pretraining
MarkTechPost@AI 2025-10-14T09:56:50.000000Z
NVIDIA Researchers Propose Reinforcement Learning Pretraining (RLP): Reinforcement as a Pretraining Objective for Building Reasoning During Pretraining
MarkTechPost@AI 2025-10-14T09:56:50.000000Z
Current Language Models Struggle to Reason in Ciphered Language
少点错误 2025-10-14T09:26:37.000000Z
庞若鸣还有苹果论文?改善预训练高质量数据枯竭困境
机器之心 2025-09-23T14:38:31.000000Z
斯坦福:优化器「诸神之战」?AdamW 凭「稳定」胜出
机器之心 2025-09-07T06:48:49.000000Z
From Pretraining to Post-Training: Why Language Models Hallucinate and How Evaluation Methods Reinforce the Problem
MarkTechPost@AI 2025-09-07T05:01:40.000000Z