princeton-nlp / LLM-Shearing

[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
https://arxiv.org/abs/2310.06694
MIT License
539 stars 41 forks source link

Missing Reference #1

Closed wutaiqiang closed 10 months ago

wutaiqiang commented 10 months ago

Nice work!

Targeted structured pruning seems to be similar to WID (https://arxiv.org/abs/2305.09098).

Please consider discussing it in the related work section.

Thanks!

xiamengzhou commented 10 months ago

Thanks for the pointer -- The work you mentioned is very relevant and we will include it in our next revision!

wutaiqiang commented 7 months ago

One kind reminder to add our paper as a reference. Really appreciate your time. Thanks!

wutaiqiang commented 7 months ago

@xiamengzhou

xiamengzhou commented 7 months ago

Hi, we are still in the process of preparing for camera ready. Will update the arxiv version soon!

wutaiqiang commented 7 months ago

Really appreciate it, thanks~