热点
关于我们
xx
xx
"
模型压缩
" 相关文章
Do Students Debias Like Teachers? On the Distillability of Bias Mitigation Methods
cs.AI updates on arXiv.org
2025-10-31T04:05:33.000000Z
The Structural Scalpel: Automated Contiguous Layer Pruning for Large Language Models
cs.AI updates on arXiv.org
2025-10-29T04:22:15.000000Z
Rethinking Inference Placement for Deep Learning across Edge and Cloud Platforms: A Multi-Objective Optimization Perspective and Future Directions
cs.AI updates on arXiv.org
2025-10-28T04:14:34.000000Z
“压缩不减智”!EPIC让多模态大模型以更少 Token 跑得更快、更稳|NeurIPS 2025
我爱计算机视觉
2025-10-27T08:53:42.000000Z
Learning Grouped Lattice Vector Quantizers for Low-Bit LLM Compression
cs.AI updates on arXiv.org
2025-10-27T06:22:17.000000Z
自媒体误读了 DeepSeek-OCR:一图胜千言
MacTalk - 微信公众号
2025-10-24T09:23:22.000000Z
C-SWAP: Explainability-Aware Structured Pruning for Efficient Neural Networks Compression
cs.AI updates on arXiv.org
2025-10-22T04:24:24.000000Z
Karpathy盛赞DeepSeek-OCR“淘汰”tokenizer!实测如何用Claude Code 让新模型跑在N卡上
AI前线
2025-10-21T11:30:40.000000Z
Neuronal Group Communication for Efficient Neural representation
cs.AI updates on arXiv.org
2025-10-21T04:27:23.000000Z
Stratos: An End-to-End Distillation Pipeline for Customized LLMs under Distributed Cloud Environments
cs.AI updates on arXiv.org
2025-10-21T04:16:30.000000Z
Stratos: An End-to-End Distillation Pipeline for Customized LLMs under Distributed Cloud Environments
cs.AI updates on arXiv.org
2025-10-21T04:16:30.000000Z
NeurIPS 2025 | 仅用20B tokens蒸出SOTA,小模型的「低秩时刻」到了
PaperWeekly
2025-10-20T16:35:42.000000Z
DeepSeek又发新模型,小而美玩出新高度
虎嗅
2025-10-20T13:56:57.000000Z
微软BitDistill将LLM压缩到1.58比特:10倍内存节省、2.65倍CPU推理加速
机器之心
2025-10-20T13:33:30.000000Z
微软BitDistill将LLM压缩到1.58比特:10倍内存节省、2.65倍CPU推理加速
机器之心
2025-10-20T13:33:30.000000Z
Finetuning and Quantization of EEG-Based Foundational BioSignal Models on ECG and PPG Data for Blood Pressure Estimation
cs.AI updates on arXiv.org
2025-10-20T04:15:03.000000Z
Microsoft AI Proposes BitNet Distillation (BitDistill): A Lightweight Pipeline that Delivers up to 10x Memory Savings and about 2.65x CPU Speedup
MarkTechPost@AI
2025-10-19T05:39:34.000000Z
Microsoft AI Proposes BitNet Distillation (BitDistill): A Lightweight Pipeline that Delivers up to 10x Memory Savings and about 2.65x CPU Speedup
MarkTechPost@AI
2025-10-19T05:39:34.000000Z
Towards Reversible Model Merging For Low-rank Weights
cs.AI updates on arXiv.org
2025-10-17T04:15:30.000000Z
LLM-Oriented Token-Adaptive Knowledge Distillation
cs.AI updates on arXiv.org
2025-10-14T04:20:30.000000Z