Kavli Affiliate: Yi Zhou | First 5 Authors: Shuli Jiang, Swanand Ravindra Kadhe, Yi Zhou, Farhan Ahmed, Ling Cai | Summary: The increasing use of large language models (LLMs) trained by third parties raises significant security concerns. In particular, malicious actors can introduce backdoors through poisoning attacks to generate undesirable outputs. While such attacks have […]
Continue.. Turning Generative Models Degenerate: The Power of Data Poisoning Attacks