-
Your project are really helpful!
I am also trying to reproduce SparseGPT:
https://arxiv.org/pdf/2301.00774.pdf
Does the SparseGPT you tried work?
-
可以一些对大模型高效的剪枝方法,如SparseGPT:https://arxiv.org/abs/2301.00774
除了训练一个全连接的小模型,对大模型进行剪枝后得到的稀疏神经网络说不定也是个可行的思路
-
@karpathy
SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot
https://arxiv.org/abs/2301.00774
What do you think?
-
As shown in paper, CUTLASS library is used for speedup. But I did not find codes rely on these settlement.How should I verify SparseGPT is faster than dense models when doing inference? Even with end-…
-
Is there any demos for ppl test of wikitext2?
-
Hi,
Does this project support Mistral?
Thanks!
-
Hello, I have tried lots of different version combinations to make the LLaMA script work, it produces very bad results which is
also what I observed with my own implementation and some other implemen…
-
When I try the SparseGPT, it raises the error:
┌───────────────────── Traceback (most recent call last) ─────────────────────┐
│ E:\pythonProject\pruning.py:77 in │…
-
## Journey Series for Generative AI Application Architecture - Fine-tune SLM with Microsoft Olive
Link: https://techcommunity.microsoft.com/t5/educator-developer-blog/journey-series-for-generative-ai…
-
Great work going on with GGML. Bravo to so many contributors. You are champions!
Maybe more performance (on CPU) can be had with bringing sparsity into the workflow. Here is one of the many efforts…