The existing language learning models (LLMs) are advancing yet have been struggling with incorporating new knowledge without forgetting the previous information, a situation termed as "catastrophic forgetting." The present methods, such as retrieval-augmented generation (RAG), are not very effective in tasks demanding integration of new knowledge from various passages due to encoding each passage in…
