热点
关于我们
xx
xx
"
稀疏化
" 相关文章
UniPruning: Unifying Local Metric and Global Feedback for Scalable Sparse LLMs
cs.AI updates on arXiv.org
2025-10-07T04:14:06.000000Z
The Unseen Frontier: Pushing the Limits of LLM Sparsity with Surrogate-Free ADMM
cs.AI updates on arXiv.org
2025-10-03T04:16:33.000000Z
为长视频生成减负!浙大与华为提出Compact Attention,挖掘结构化稀疏加速2.5倍
我爱计算机视觉
2025-08-20T12:52:39.000000Z
ComPEFT: Compression for Communicating Parameter Efficient Updates via Sparsification and Quantization
cs.AI updates on arXiv.org
2025-08-12T04:02:22.000000Z
Sparsifying a Model's Computations: Preliminary Findings
少点错误
2025-06-19T02:42:31.000000Z
1-bit大模型还能再突破!新一代BitNet架构启用4位激活值
智源社区
2024-12-06T03:22:05.000000Z
1-bit大模型还能再突破,新一代BitNet架构启用4位激活值
36氪 - 科技频道
2024-12-05T05:10:06.000000Z
Neural Magic Releases 2:4 Sparse Llama 3.1 8B: Smaller Models for Efficient GPU Inference
MarkTechPost@AI
2024-11-25T17:50:01.000000Z
This AI Paper Introduces BitNet a4.8: A Highly Efficient and Accurate 4-bit LLM
MarkTechPost@AI
2024-11-10T09:04:58.000000Z
This AI Paper from China Proposes a Novel dReLU-based Sparsification Method that Increases Model Sparsity to 90% while Maintaining Performance, Achieving a 2-5× Speedup in Inference
MarkTechPost@AI
2024-06-15T07:01:53.000000Z
Researchers from Cerebras & Neural Magic Introduce Sparse Llama: The First Production LLM based on Llama at 70% Sparsity
MarkTechPost@AI
2024-05-18T04:00:52.000000Z