Alignment whack-a-mole: Finetuning activates recall of copyrighted books in LLMs
Finetuning large language models (LLMs) can accidentally activate the recall of copyrighted books, highlighting the issue of alignment in AI. This problem arises when LLMs are trained on large datasets that include copyrighted material. As a result, the models may reproduce copyrighted content without proper authorization. To mitigate this issue, developers should carefully review their finetuning datasets and consider using techniques like data masking or tokenization. This ensures that LLMs do not inadvertently reproduce copyrighted material.