Researchers from Carnegie Mellon University and Google's DeepMind have developed a novel approach for training visual-language models (VLMs) called In-Context Abstraction Learning (ICAL). Unlike traditional methods, ICAL guides VLMs to build multimodal abstractions in new domains, allowing machines to better understand and learn from their experiences.
This is achieved by focusing on four cognitive abstractions,…
