Activation Function Tuning Boosts Efficient Fine‑Tuning in AI Models
NoRA updates only 0.4% of parameters (≈0.02 M) and lifts CIFAR‑10 accuracy by 0.17%; on LLaMA‑3‑8B tuning, MMLU improves up to 0.8%. This low‑rank update adds minimal compute cost. getnews.me/activation-function-tuni... #activationfunction #peft
0
0
0
0