热点
关于我们
xx
xx
"
持续预训练
" 相关文章
How Can We Effectively Expand the Vocabulary of LLMs with 0.01GB of Target Language Text?
cs.AI updates on arXiv.org
2025-10-28T04:14:37.000000Z
IKnow: Instruction-Knowledge-Aware Continual Pretraining for Effective Domain Adaptation
cs.AI updates on arXiv.org
2025-10-24T04:18:02.000000Z
DACIP-RC: Domain Adaptive Continual Instruction Pre-Training via Reading Comprehension on Business Conversations
cs.AI updates on arXiv.org
2025-10-10T04:16:19.000000Z
DACP: Domain-Adaptive Continual Pre-Training of Large Language Models for Phone Conversation Summarization
cs.AI updates on arXiv.org
2025-10-08T04:14:36.000000Z
LLM Research Papers: The 2024 List
Ahead of AI
2025-09-25T10:01:35.000000Z
如何在大模型基础上训练一个特定领域的专业模型
Datawhale
2025-09-15T07:31:54.000000Z
如何在大模型基础上训练一个特定领域的专业模型
Datawhale
2025-09-15T03:46:29.000000Z
MachineLearningLM: Scaling Many-shot In-context Learning via Continued Pretraining
cs.AI updates on arXiv.org
2025-09-11T15:51:56.000000Z
PLaMo 2 Technical Report
cs.AI updates on arXiv.org
2025-09-08T04:51:30.000000Z
Revisiting Replay and Gradient Alignment for Continual Pre-Training of Large Language Models
cs.AI updates on arXiv.org
2025-08-05T11:10:06.000000Z
A Practice of Post-Training on Llama-3 70B with Optimal Selection of Additional Language Mixture Ratio
cs.AI updates on arXiv.org
2025-07-29T04:22:41.000000Z
Data Mixing Agent: Learning to Re-weight Domains for Continual Pre-training
cs.AI updates on arXiv.org
2025-07-22T04:44:31.000000Z
HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training
cs.AI updates on arXiv.org
2025-07-16T04:29:02.000000Z
单向VLM变双向!人大斯坦福等提出MoCa框架:双向多模态编码器
新智元
2025-07-10T12:06:34.000000Z
Noteworthy AI Research Papers of 2024 (Part One)
Ahead of AI
2024-12-31T12:28:58.000000Z
Aquila-Med LLM:开创性的全流程开源医疗语言模型
智源研究院
2024-10-24T17:00:57.000000Z