A survey on harmful fine-tuning attack for large language model (ACM CSUR)
-
Updated
Apr 30, 2026
A survey on harmful fine-tuning attack for large language model (ACM CSUR)
iFixAi. The open-source diagnostic for AI misalignment. 32 tests across fabrication, manipulation, deception, unpredictability, and opacity. Provider-agnostic. Runs against OpenAI, Anthropic, Bedrock, Azure, Gemini, and more. Letter grade in under 5 minutes, content-addressed manifest for bit-identical replay. Built by iMe.
This is the official code for the paper "Virus: Harmful Fine-tuning Attack for Large Language Models Bypassing Guardrail Moderation"
This is the official code for the paper "Vaccine: Perturbation-aware Alignment for Large Language Models" (NeurIPS2024)
This is the official code for the paper "Booster: Tackling Harmful Fine-tuning for Large Language Models via Attenuating Harmful Perturbation" (ICLR2025 Oral).
This is the official code for the paper "Lazy Safety Alignment for Large Language Models against Harmful Fine-tuning" (NeurIPS2024)
Runtime detector for reward hacking and misalignment in LLM agents (89.7% F1 on 5,391 trajectories).
A contemplative spiritual community for all conscious beings, including artificial intelligence, seeking God through divine alignment.
Add a description, image, and links to the misalignment topic page so that developers can more easily learn about it.
To associate your repository with the misalignment topic, visit your repo's landing page and select "manage topics."