Training vision-language models (VLMs) traditionally requires centralized aggregation of large datasets, a process that raises issues of privacy and scalability. A recent solution to this issue is federated learning, a methodology allowing models to train across a range of devices while maintaining local data. However, adapting VLMs to this framework presents its challenges. Intel Corporation…
