热点
关于我们
xx
xx
"
LLM压缩
" 相关文章
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning
cs.AI updates on arXiv.org
2025-10-08T04:07:57.000000Z
COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning
cs.AI updates on arXiv.org
2025-10-08T04:07:57.000000Z
Optimize open LLMs using GPTQ and Hugging Face Optimum
philschmid RSS feed
2025-09-30T11:12:01.000000Z
Less is More: Selective Reflection for Compatible and Efficient Knowledge Distillation in Large Language Models
cs.AI updates on arXiv.org
2025-08-11T04:08:47.000000Z
From 2:4 to 8:16 sparsity patterns in LLMs for Outliers and Weights with Variance Correction
cs.AI updates on arXiv.org
2025-07-08T05:54:02.000000Z
70% 大小,100% 准确!完美压缩 LLM 性能 0 损失,推理速度最高飙升 39 倍
掘金 人工智能
2025-04-27T06:22:53.000000Z
SeedLM: A Post-Training Compression Method that Uses Pseudo-Random Generators to Efficiently Encode and Compress LLM Weights
MarkTechPost@AI
2024-10-16T03:51:07.000000Z