Formulas for LoRA: W_{updated} = W + \Delta W, \Delta W \approx AB, W_{updated} = W + AB
LoRA involves freezing all LLM layers and adding trainable low-rank weight matrices (AB) to specific layers to adjust their outputs and fine-tuning overall model predictions. This amounts to the low-rank matrices approximating the full weight updates that would have occurred during fine-tuning.