Natural Language Processing (NLP) faces major challenges in addressing the limitations of decoder-only Transformers, which are the backbone of large language models (LLMs). These models contend with issues like representational collapse and over-squashing, which severely hinder their functionality. Representational collapse happens when different sequences produce nearly the same results, while over-squashing occurs when the model…
