The challenges associated with training large language models (LLMs) given their memory-intensive nature can be significant. Traditional methods for reducing memory consumption frequently involve compressing model weights, commonly leading to a decrease in model performance. A new approach being called Gradient Low-Rank Projection (GaLore) is now being proposed by researchers from various institutions, including the…
