Google AI researchers have developed a new Transformer network dubbed TransformerFAM, aimed to enhance performance in handling extremely long context tasks. Despite Transformers proving revolutionary in the domain of deep learning, they have limitations due to their quadratic attention complexity— an aspect that curtails their ability to process infinitely long inputs. Existing Transformers often forget…
