2
1
0

A new method was able to delete 40% of LLM layers with no drop in accuracy.
This makes them mich cheaper and faster.
The method combines pruning, quantization and PEFT.
They tested this across various open source models.
Each family of models had a maximum amount of layers…
时政
(
twitter.com
)
由
Lior⚡
提交
Markdown支持
评论加载中...
您可能感兴趣的:
更多