[2311.02105] Making Harmful Behaviors Unlearnable for Large Language Models