Shortgpt layers in large language models are more redundant than you expect.
ShortGPT
Layers in large language models are more redundant than you expect.
As Large Language Models (LLMs) continue to advance in performance, their size has escalated significantly, with current LLMs containing billions or even trillions of parameters…
Join the discussion on this paper page.
Comments are closed.