Do LLMs need rationales for learning from mistakes? ๐ค
When LLMs learn from previous incorrect answers, they typically observe corrective feedback in the form of rationales explaining each mistake. In our new preprint, we find these rationales do not help, in fact they hurt performance!
๐งต
1 year ago
21
9
1
3