热点
"微调" 相关文章
Attention Saturation and Gradient Suppression at Inflection Layers: Diagnosing and Mitigating Bottlenecks in Transformer Adaptation
cs.AI updates on arXiv.org 2025-11-05T05:27:04.000000Z
FlashEVA: Accelerating LLM inference via Efficient Attention
cs.AI updates on arXiv.org 2025-11-05T05:25:27.000000Z
Loquetier: A Virtualized Multi-LoRA Framework for Unified LLM Fine-tuning and Serving
cs.AI updates on arXiv.org 2025-11-05T05:18:52.000000Z
综述238篇遥感微调!清华院士团队指出9大方向 | CVMJ
新智元 2025-11-02T18:15:44.000000Z
Defeating the Training-Inference Mismatch via FP16
cs.AI updates on arXiv.org 2025-10-31T04:09:50.000000Z
苹果提出新型反向传播:一台iPhone 15 Pro Max就能微调LLM
机器之心 2025-10-30T05:34:47.000000Z
苹果提出新型反向传播:一台iPhone 15 Pro Max就能微调LLM
36氪 - 科技频道 2025-10-30T04:27:16.000000Z
Don't Blind Your VLA: Aligning Visual Representations for OOD Generalization
cs.AI updates on arXiv.org 2025-10-30T04:20:32.000000Z
Preventing Catastrophic Forgetting: Behavior-Aware Sampling for Safer Language Model Fine-Tuning
cs.AI updates on arXiv.org 2025-10-28T04:12:11.000000Z
Memory-Efficient Backpropagation for Fine-Tuning LLMs on Resource-Constrained Mobile Devices
machinelearning apple 2025-10-27T20:25:30.000000Z
$\alpha$-LoRA: Effective Fine-Tuning via Base Model Rescaling
cs.AI updates on arXiv.org 2025-10-27T06:25:36.000000Z
PLAN: Proactive Low-Rank Allocation for Continual Learning
cs.AI updates on arXiv.org 2025-10-27T06:24:32.000000Z
Compress to Impress: Efficient LLM Adaptation Using a Single Gradient Step on 100 Samples
cs.AI updates on arXiv.org 2025-10-24T04:51:30.000000Z
QKCV Attention: Enhancing Time Series Forecasting with Static Categorical Embeddings for Both Lightweight and Pre-trained Foundation Models
cs.AI updates on arXiv.org 2025-10-24T04:24:16.000000Z
Train an LLM on an NVIDIA Blackwell Desktop with Unsloth—and Scale It
Nvidia Developer 2025-10-23T17:58:03.000000Z
LLM Training Data Optimization: Fine-Tuning, RLHF & Red Teaming
Cogito Tech 2025-10-23T05:35:13.000000Z
LLM Training Data Optimization: Fine-Tuning, RLHF & Red Teaming
Cogito Tech 2025-10-23T05:35:13.000000Z
Data Efficient Adaptation in Large Language Models via Continuous Low-Rank Fine-Tuning
cs.AI updates on arXiv.org 2025-10-23T04:24:15.000000Z
Data Efficient Adaptation in Large Language Models via Continuous Low-Rank Fine-Tuning
cs.AI updates on arXiv.org 2025-10-23T04:24:15.000000Z
Unraveling Emotions with Pre-Trained Models
cs.AI updates on arXiv.org 2025-10-23T04:22:06.000000Z