trainingresearch

Motivation

  • LoRA tries to approximate the weights update matrices using 2 lower-ranked matrices
  • the choice of affects how well the approximation can be and is fixed

Idea

Dynamically change during training

  • Separately for each layer
  • Start with a reasonable , during training
    • Lower (remove columns) when are redundant columns
    • Increase (add columns) when approximation score is low for too long
      • how to evaluate approximation score ?question
      • how long is too long ?question
      • must be so that
  • Adding/removing columns must not change the approximation significantly. How ?question